Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2022 Apr 12:5:787421.
doi: 10.3389/fdata.2022.787421. eCollection 2022.

Applications and Techniques for Fast Machine Learning in Science

Affiliations
Review

Applications and Techniques for Fast Machine Learning in Science

Allison McCarn Deiana et al. Front Big Data. .

Erratum in

  • Corrigendum: Applications and techniques for fast machine learning in science.
    Deiana AM, Tran N, Agar J, Blott M, Di Guglielmo G, Duarte J, Harris P, Hauck S, Liu M, Neubauer MS, Ngadiuba J, Ogrenci-Memik S, Pierini M, Aarrestad T, Bähr S, Becker J, Berthold AS, Bonventre RJ, Müller Bravo TE, Diefenthaler M, Dong Z, Fritzsche N, Gholami A, Govorkova E, Guo D, Hazelwood KJ, Herwig C, Khan B, Kim S, Klijnsma T, Liu Y, Lo KH, Nguyen T, Pezzullo G, Rasoulinezhad S, Rivera RA, Scholberg K, Selig J, Sen S, Strukov D, Tang W, Thais S, Unger KL, Vilalta R, von Krosigk B, Wang S, Warburton TK. Deiana AM, et al. Front Big Data. 2023 Oct 16;6:1301942. doi: 10.3389/fdata.2023.1301942. eCollection 2023. Front Big Data. 2023. PMID: 37908454 Free PMC article.

Abstract

In this community review report, we discuss applications and techniques for fast machine learning (ML) in science-the concept of integrating powerful ML methods into the real-time experimental data processing loop to accelerate scientific discovery. The material for the report builds on two workshops held by the Fast ML for Science community and covers three main areas: applications for fast ML across a number of scientific domains; techniques for training and implementing performant and resource-efficient ML algorithms; and computing architectures, platforms, and technologies for deploying these algorithms. We also present overlapping challenges across the multiple scientific domains where common solutions can be found. This community report is intended to give plenty of examples and inspiration for scientific discovery through integrated and accelerated ML solutions. This is followed by a high-level overview and organization of technical advances, including an abundance of pointers to source material, which can enable these breakthroughs.

Keywords: big data; codesign; coprocessors; fast machine learning; heterogeneous computing; machine learning for science; particle physics.

PubMed Disclaimer

Conflict of interest statement

MB was employed by the company Xilinx Inc. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The handling editor EC is currently organizing a Research Topic with the authors JD, ML, and JN.

Figures

Figure 1
Figure 1
The concept behind this review paper is to find the confluence of domain-specific challenges, machine learning, and experiment and computer system architectures to accelerate science discovery.
Figure 2
Figure 2
High-level overview of the stages in a GNN-based tracking pipeline. Only a subset of the typical edge weights are shown for illustration purposes. (A) Graph construction, (B) edge classification, and (C) track construction.
Figure 3
Figure 3
Simulated type Ia supernova light-curve and classification. Top: calibrated flux evolution in different DES band-passes as a function of normalized time (the first photometric measurement is set to time equals zero). Bottom: Baseline RNN classification probability evolution with respect of time, no host-galaxy redshift information was provided. At each photometric measurement, classification probability is obtained. The maximum light of the simulated supernova is shown in a gray dashed line and the simulated redshift of the supernovae is shown on the top z = 0.466. We highlight that redshift is not used for this classification but can improve results. Our baseline RNN classifies this light-curve as type Ia SN with great accuracy before maximum light, it only requires a handful of photometric epochs. (Möller and de Boissiére, 2019).
Figure 4
Figure 4
A 6GeV/c electron event in the ProtoDUNE detector. The x-axis shows the wire number. The y-axis shows the time tick in the unit of 0.5μs. The color scale represents the charge deposition.
Figure 5
Figure 5
Experimental 4D-STEM measurement of a dichalcogenide 2D material. Atomic map is inferred from the data, each diffraction pattern represents an average of 7 × 7 experimental images, green STEM probes are labeled for regions of the sample with one layer, vacuum, and two layers (Ophus, 2019).
Figure 6
Figure 6
The illustration of hardware-aware quantization and pruning. A given NN model can be compressed by using low precision quantization instead of single precision. The extreme case is to use 0-bit quantization which is equivalent to removing/pruning the corresponding neurons. The goal of compression is to find the best bit-precision setting for quantization/pruning to reduce model footprint/latency on a target hardware with minimal generalization loss.
Figure 7
Figure 7
Taxonomy of compute architectures, differentiating CPUs, GPUs and DPUs.
Figure 8
Figure 8
DPU architectures: Matrix of Processing Engines (MPE) on the left, and spatial architecture on the right.
Figure 9
Figure 9
FINN compiler flow.
Figure 10
Figure 10
Analog vector-by-matrix multiplication (VMM) in a crossbar circuit with adjustable crosspoint devices. For clarity, the output signal is shown for just one column of the array, while sense amplifier circuitry is not shown. Note that other VMM designs, e.g., utilizing duration of applied voltage pulses, rather than their amplitudes, for encoding inputs/outputs, are now being actively explored see, e.g., their brief review in Bavandpour et al. (2018).

References

    1. Aad G. (2008). The ATLAS experiment at the CERN large hadron collider. JINST 3:S08003. 10.1088/1748-0221/3/08/S08003 - DOI
    1. Aad G., Abajyan T., Abbott B., Abdallah J., Abdel Khalek S., Abdelalim A. A., et al. . (2012). Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC. Phys. Lett. B 716, 1–29. 10.1016/j.physletb.2012.08.020 - DOI
    1. Aarrestad T., Loncar V., Ghielmetti N., Pierini M., Summers S., Ngadiuba J., et al. . (2021). Fast convolutional neural networks on FPGAs with hls4ml. Mach. Learn. Sci. Tech., 2, 045015. 10.1088/2632-2153/ac0ea1 - DOI
    1. Aartsen M., The IceCube, Fermi-LAT, MAGIC, AGILE, ASAS-SN, HAWC et al. . (2018). Multimessenger observations of a flaring blazar coincident with high-energy neutrino IceCube-170922A. Science 361, eaat1378. 10.1126/science.aat1378 - DOI - PubMed
    1. Abbott B., Abbott R., Abbott T., Abernathy M., Acernese F., Ackley K., et al. . (2016a). Properties of the binary black hole merger gw150914. Phys. Rev. Lett. 116, 241102. 10.1103/PhysRevLett.116.241102 - DOI - PubMed