Award Abstract # 2107257
Collaborative Research: SHF: Medium: Collaborative Automatic Parallelization

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: THE TRUSTEES OF PRINCETON UNIVERSITY
Initial Amendment Date: July 8, 2021
Latest Amendment Date: September 8, 2023
Award Number: 2107257
Award Instrument: Continuing Grant
Program Manager: Anindya Banerjee
abanerje@nsf.gov
 (703)292-7885
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: October 1, 2021
End Date: September 30, 2026 (Estimated)
Total Intended Award Amount: $600,000.00
Total Awarded Amount to Date: $600,000.00
Funds Obligated to Date: FY 2021 = $297,498.00
FY 2022 = $150,607.00

FY 2023 = $151,895.00
History of Investigator:
  • David August (Principal Investigator)
Recipient Sponsored Research Office: Princeton University
1 NASSAU HALL
PRINCETON
NJ  US  08544-2001
(609)258-3090
Sponsor Congressional District: 12
Primary Place of Performance: Princeton University
87 Prospect Avenue
Princeton
NJ  US  08544-2020
Primary Place of Performance
Congressional District:
12
Unique Entity Identifier (UEI): NJ1YPQXQG7U5
Parent UEI:
NSF Program(s): Software & Hardware Foundation
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
01002223DB NSF RESEARCH & RELATED ACTIVIT

01002324DB NSF RESEARCH & RELATED ACTIVIT

01002425DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7924, 7943
Program Element Code(s): 779800
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

In the context of the end of Moore's law, the greatest value of multicore is ultimately in its potential to accelerate sequential codes. This potential can only be realized with the reliable extraction of sufficient multicore-appropriate thread-level parallelism (MATLP) from programs. Yet, despite many new tools, languages, and libraries designed for multicore, difficulties in MATLP extraction keep multicore grossly underutilized. The energy and performance impact of this is nearly universal. To address this problem, this project's novelties are in (i) redefining traditional abstractions used within compilers to enable constructive and tight collaborations that aim to coordinate the multiple code analyses and transformations required for MATLP extraction, (ii) producing RAPPORT, the first publicly available compiler with full collaboration support, a necessary element for robust automatic parallelization. This project's impact is in making computing faster and more efficient with reliable MATLP extraction.

In conventional compilers, optimizations perform well greedily and independently, enabling easy compiler modularity without much performance impact. However, in MATLP extraction, key parallelization techniques may succeed only if other transformations clear the path, sometimes by de-optimizing the code. Over the last decade, researchers have made steady progress toward the goal of robust and routine automatic MATLP with new MATLP parallelization patterns, stronger memory analyses, and more efficient speculation techniques. This team believes these MATLP technologies are sufficient but lack the coordination necessary to realize their full potential. This work produces the technology necessary for reliable MATLP extraction by redefining compiler abstractions to enable transformations and analyses to work together actively without loss of modularity. This new technology enables a globally beneficial behavior by centralizing, in a modular way, the decentralized and greedy decision-making found in conventional compilers. In this way, it makes the reliable and robust extraction of MATLP possible.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Chaturvedi, Ishita and Godala, Bhargav Reddy and Wu, Yucan and Xu, Ziyang and Iliakis, Konstantinos and Eleftherakis, Panagiotis-Eleftherios and Xydis, Sotirios and Soudris, Dimitrios and Sorensen, Tyler and Campanoni, Simone and Aamodt, Tor M and August, "GhOST: a GPU Out-of-Order Scheduling Technique for Stall Reduction" , 2024 https://doi.org/10.1109/ISCA59077.2024.00011 Citation Details
Godala, Bhargav Reddy and Ramesh, Sankara Prasad and Pokam, Gilles A and Stark, Jared and Seznec, Andre and Tullsen, Dean and August, David I "PDIP: Priority Directed Instruction Prefetching" , 2024 https://doi.org/10.1145/3620665.3640394 Citation Details
Matni, Angelo and Deiana, Enrico Armenio and Su, Yian and Gross, Lukas and Ghosh, Souradip and Apostolakis, Sotiris and Xu, Ziyang and Tan, Zujun and Chaturvedi, Ishita and Homerding, Brian and McMichen, Tommy and August, David I. and Campanoni, Simone "NOELLE Offers Empowering LLVM Extensions" 2022 IEEE/ACM International Symposium on Code Generation and Optimization (CGO) , 2022 https://doi.org/10.1109/CGO53902.2022.9741276 Citation Details
Nagendra, Nayana Prasad and Godala, Bhargav Reddy and Chaturvedi, Ishita and Patel, Atmn and Kanev, Svilen and Moseley, Tipp and Stark, Jared and Pokam, Gilles A. and Campanoni, Simone and August, David I. "EMISSARY: Enhanced Miss Awareness Replacement Policy for L2 Instruction Caching" Proceedings of the 50th International Symposium on Computer Architecture (ISCA) , 2023 https://doi.org/10.1145/3579371.3589097 Citation Details
Popescu, Natalie and Xu, Ziyang and Apostolakis, Sotiris and August, David I. and Levy, Amit "Safer at any speed: automatic context-aware safety enhancement for Rust" Proceedings of the ACM on Programming Languages , v.5 , 2021 https://doi.org/10.1145/3485480 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page