
NSF Org: |
OAC Office of Advanced Cyberinfrastructure (OAC) |
Recipient: |
|
Initial Amendment Date: | May 1, 2020 |
Latest Amendment Date: | May 1, 2020 |
Award Number: | 2004645 |
Award Instrument: | Standard Grant |
Program Manager: |
Daniel F. Massey
dmassey@nsf.gov (703)292-5147 OAC Office of Advanced Cyberinfrastructure (OAC) CSE Directorate for Computer and Information Science and Engineering |
Start Date: | July 1, 2020 |
End Date: | June 30, 2024 (Estimated) |
Total Intended Award Amount: | $310,000.00 |
Total Awarded Amount to Date: | $310,000.00 |
Funds Obligated to Date: |
|
History of Investigator: |
|
Recipient Sponsored Research Office: |
77 MASSACHUSETTS AVE CAMBRIDGE MA US 02139-4301 (617)253-1000 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
Route de Meyrin, 385 Meyrin SZ |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): |
OFFICE OF MULTIDISCIPLINARY AC, COMPUTATIONAL PHYSICS, Software Institutes |
Primary Program Source: |
|
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
The development of the Standard Model (SM) of particle physics is a major intellectual achievement. The validity of this model was further confirmed by the discovery of the Higgs boson at the Large Hadron Collider (LHC) at CERN. However, the Standard Model leaves open many questions, including why matter dominates over anti-matter in the Universe and the properties of dark matter. Most explanations require new phenomena, which we call Beyond the Standard Model Physics (BSM), and which the LHCb experiment at CERN has been designed to explore. The LHC is the premier High Energy Physics particle accelerator in the world and is currently operating at the CERN laboratory near Geneva Switzerland, one of the foremost facilities for addressing these BSM questions. The LHCb experiment is one of four large experiments at the LHC and is designed to study in detail the decays of hadrons containing b or c quarks. The goal is to identify the existence of new physics beyond the Standard Model by examining the properties of hadrons containing these quarks. The new physics, or new forces, can be manifest by particles, as yet to be discovered, whose presence would modify decay rates and CP violating asymmetries of hadrons containing the b and c quarks, allowing new phenomena to be observed indirectly - or via direct observation of new force-carrying particles. The data sets collected by the LHC experiments are some of the largest in the world. For example, the sensor arrays of the LHCb experiment, in which both PIs participate, produce about 100 TB/s and close to a zettabyte per year. Even after drastic data-reduction performed by custom-built read-out electronics, the data volume is still about 10 exabytes per year. Such large data sets cannot be stored indefinitely; therefore, all high energy physics (HEP) experiments employ a second data-reduction scheme executed in real time by a data-ingestion system - referred to as a trigger system in HEP - to decide whether each event is to be persisted for future analysis or permanently discarded. The primary goal of this project is developing and deploying software that will maximize the performance of the LHCb trigger system - running its first processing stage on GPUs - so that the full physics discovery potential of LHCb is realized.
The LHCb detector is being upgraded for Run 3 (which will start to record data in 2022), when the trigger system will need to process 25 exabytes per year. Currently, only 0.3 of the 10 exabytes per year processed by the trigger is analyzed using high-level computing algorithms; the rest is discarded prior to this stage using simple algorithms executed on FPGAs. To significantly extend its physics reach in Run 3, LHCb plans to process the entire 25 exabytes each year using high-level computing algorithms. The PIs propose running the entire first trigger-processing stage on GPUs, which has zero (likely negative) net cost, and frees up all of the CPU resources for the second processing stage. The LHCb trigger makes heavy use of machine learning (ML) algorithms, which will need to be reoptimized both for Run 3 conditions but also for usage on GPUs. The specific objectives of this proposal are developing: GPU-based versions of the primary trigger-selection algorithms, which make heavy usage of ML; GPU-based calorimeter-clustering and electron-identification algorithms, likely using ML; and the infrastructure required to deploy ML algorithms within the GPU-based trigger framework. These advances will make it possible to explore many potential explanations for dark matter, e.g., dark photon decays, and the matter/anti-matter asymmetry of our universe using data that would be otherwise inaccessible due to trigger-system limitations.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
The data sets collected by the LHC experiments are some of the largest in the world. For example, the sensor arrays of the LHCb experiment produce about 100 TB/s and close to a zettabyte per year. Even after drastic data-reduction performed by custom-built read-out electronics, the data volume is still about 10 exabytes per year. Such large data sets cannot be stored indefinitely; therefore, all high energy physics (HEP) experiments employ a second data-reduction scheme executed in real time by a data-ingestion system - referred to as a trigger system in HEP - to decide whether each event is to be persisted for future analysis or permanently discarded. Trigger system design is dictated by the rate at which the sensors can be read out, the computational power of the system, and the available storage space. The LHCb detector has been upgraded for latest LHC data-taking run, with its trigger system now needing to process 25 exabytes per year. In the previous LHC run, only 0.3 of the 10 exabytes per year processed by the LHCb trigger were analyzed using high-level computing algorithms; the rest was discarded prior to this stage using simple algorithms executed on FPGAs. To significantly extend its physics reach in the latest run, LHCb now processes the entire 25 exabytes each year using high-level computing algorithms.
This grant supported work that made it possible to run the entire first trigger-processing stage on GPUs. The LHCb trigger makes heavy use of machine learning (ML) algorithms, which all needed to be reoptimized both for the latest running conditions but also for usage on GPUs. The primary achievements of this project were developing: GPU-based versions of the primary trigger-selection algorithms, which make heavy usage of ML; GPU-based calorimeter-clustering and electron-identification algorithms, the latter of which uses ML; and the infrastructure required to deploy ML algorithms within the GPU-based trigger framework. These advances are making it possible to explore many potential explanations for dark matter (e.g., dark photon decays to electron-positron pairs) and the matter/anti-matter asymmetry of our universe using data that would be otherwise inaccessible due to trigger-system limitations.
Last Modified: 07/02/2024
Modified by: Mike Williams
Please report errors in award information by writing to: awardsearch@nsf.gov.