Literature DB >> 34278296

Editorial: Heterogeneous Computing for AI and Big Data in High Energy Physics.

Daniele D'Agostino1, Daniele Cesini2.   

Abstract

Entities:  

Keywords:  artificial intelligence; big data; heterogeneous computing; high energy physics; high performance computing

Year:  2021        PMID: 34278296      PMCID: PMC8282366          DOI: 10.3389/fdata.2021.652881

Source DB:  PubMed          Journal:  Front Big Data        ISSN: 2624-909X


× No keyword cloud information.
Heterogeneous computing denotes a scenario where different computing platforms are exploited for specific applications (Danovaro et al., 2014). While the demand for computational resources continues to grow with increasing need for querying and analyzing the volumes and rates of Big Data, energy efficiency is limiting the traditional approach to improve the compute capabilities of a data center by adding thousands of state-of-the-art x86 machines to an existing infrastructure in favor of adopting energy savvy devices (Cesini et al., 2017; D’Agostino et al., 2019). The result is that the computing nodes in data centers have different execution models, ranging from the traditional x68 architecture to GPUs, FPGAs (Papadimitriou et al., 2020) and other processor types like ARMs or more specialized processors as TPUs (Albrecht et al., 2019; Cass, 2019). For example, GPUs are used in many scientific applications based on regular domains and are delivering performance that is orders of magnitude better than traditional cores. They are also widely used in deep learning, especially the machine learning training phase. The FPGAs, being integrated circuits that can be configured by the programmer to implement a certain function, tries to close the gap between hardware and software. Within this context the research topic collects five papers showing very interesting experiences in adopting heterogeneous architectures, for AI and Big Data applications in High Energy Physics. In GPU-accelerated machine learning inference as a service for computing in neutrino experiments (Wang et al.) the authors discuss the performance achieved by exploiting GPU resources as a service for the ProtoDUNE-SP reconstruction chain developed in the context of the Deep Underground Neutrino Experiment (DUNE). This contribution represents one of the first experiences in which machine learning is acceleratedis accelerated with GPUs in neutrino software frameworks. The most time-consuming task, the track and particle shower hit identification, has been accelerated by a factor of 17. In Heterogeneous reconstruction of tracks and primary vertices with the CMS pixel tracker (Bocci et al.) the authors describe an heterogeneous implementation of pixel tracks and vertices reconstruction chain on GPUs able to achieve high performance speedup values. The resulting framework has been developed for being integrated in the CMS particle detector reconstruction software, CMSSW (http://cms-sw.github.io), which is employed to detect particle and phenomena resulting from high-energy collisions in the LHC by the CMS experiment. In Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics (Iiyama et al.) the authors present a novel method to export graph neural networks from complex modern machine learning packages to an efficient FPGA implementation. The main characteristics is that this implementation is able to perform nontrivial physics tasks with a sub-microsecond latency, therefore it represents a suitable solution for low-latency applications, such as the Level-1 trigger stage of LHC. In CLUE: A Fast Parallel Clustering Algorithm for High Granularity Calorimeters in High-Energy Physics (Rovere et al.) the authors propose a novel density-based clustering algorithm exploiting a new, highly parallelizable way of assigning points to clusters. Experiments have been performed on a synthetic dataset that resembles high occupancy events in high granularity calorimeters operated at HL-LHC with speedup values up to 112 on a GPU. In Porting HEP Parameterized Calorimeter Simulation Code to GPUs (Leggett et al.) the authors describe an accelerated version of the ATLAS Calorimeter Fast-Simulation on GPUs. Several techniques have been incorporated the implementation to increase the GPU overall utilization, because the compute-to-I/O ratio is not very large, and the achieved results are interesting. A key topic of the work is represented by the introduction of performance portability through Kokkos, a C++ cross-device compatibility library. Authos discuss the different issues encountered during the porting process, the adopted solutions and present a comparison of the achievable performance with respect to previous approaches.
  1 in total

1.  Design of Online Music Education System Based on Artificial Intelligence and Multiuser Detection Algorithm.

Authors:  Hua Yan
Journal:  Comput Intell Neurosci       Date:  2022-03-24
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.