GPU-Accelerated Machine Learning Inference as a Service for Computing in Neutrino Experiments

Machine learning algorithms are becoming increasingly prevalent and performant in the reconstruction of events in accelerator-based neutrino experiments. These sophisticated algorithms can be computationally expensive. At the same time, the data volumes of such experiments are rapidly increasing. Th...

Full description

Saved in:
Bibliographic Details
Published inFrontiers in big data Vol. 3; p. 604083
Main Authors Wang, Michael, Yang, Tingjun, Flechas, Maria Acosta, Harris, Philip, Hawks, Benjamin, Holzman, Burt, Knoepfel, Kyle, Krupa, Jeffrey, Pedro, Kevin, Tran, Nhan
Format Journal Article
LanguageEnglish
Published Switzerland Frontiers 14.01.2021
Frontiers Media S.A
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Machine learning algorithms are becoming increasingly prevalent and performant in the reconstruction of events in accelerator-based neutrino experiments. These sophisticated algorithms can be computationally expensive. At the same time, the data volumes of such experiments are rapidly increasing. The demand to process billions of neutrino events with many machine learning algorithm inferences creates a computing challenge. We explore a computing model in which heterogeneous computing with GPU coprocessors is made available as a web service. The coprocessors can be efficiently and elastically deployed to provide the right amount of computing for a given processing task. With our approach, Services for Optimized Network Inference on Coprocessors (SONIC), we integrate GPU acceleration specifically for the ProtoDUNE-SP reconstruction chain without disrupting the native computing workflow. With our integrated framework, we accelerate the most time-consuming task, track and particle shower hit identification, by a factor of 17. This results in a factor of 2.7 reduction in the total processing time when compared with CPU-only production. For this particular task, only 1 GPU is required for every 68 CPU threads, providing a cost-effective solution.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
AC02-07CH11359
USDOE Office of Science (SC), High Energy Physics (HEP)
FERMILAB-PUB-20-428-ND-SCD; arXiv:2009.04509
Edited by: Daniele D’Agostino, National Research Council (CNR), Italy
This article was submitted to Big Data and AI in High Energy Physics, a section of the journal Frontiers in Big Data
Reviewed by: Alexander Radovic, Borealis AI, Canada
Anushree Ghosh, University of Padua, Italy
ISSN:2624-909X
2624-909X
DOI:10.3389/fdata.2020.604083