Skip to feedback

Award Abstract # 2046816
CAREER: Foundations of Resource Efficient Machine Learning

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: REGENTS OF THE UNIVERSITY OF CALIFORNIA AT RIVERSIDE
Initial Amendment Date: January 8, 2021
Latest Amendment Date: June 26, 2024
Award Number: 2046816
Award Instrument: Continuing Grant
Program Manager: Phillip Regalia
pregalia@nsf.gov
 (703)292-2981
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: February 1, 2021
End Date: January 31, 2026 (Estimated)
Total Intended Award Amount: $559,029.00
Total Awarded Amount to Date: $447,740.00
Funds Obligated to Date: FY 2021 = $219,515.00
FY 2022 = $119,243.00

FY 2024 = $108,982.00
History of Investigator:
  • Samet Oymak (Principal Investigator)
    oymak@umich.edu
Recipient Sponsored Research Office: University of California-Riverside
200 UNIVERSTY OFC BUILDING
RIVERSIDE
CA  US  92521-0001
(951)827-5535
Sponsor Congressional District: 39
Primary Place of Performance: The Regents of the University of California
245 University Office Building
Riverside
CA  US  92521-0001
Primary Place of Performance
Congressional District:
39
Unique Entity Identifier (UEI): MR5QC5FCAVH5
Parent UEI:
NSF Program(s): Comm & Information Foundations
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
01002223DB NSF RESEARCH & RELATED ACTIVIT

01002324DB NSF RESEARCH & RELATED ACTIVIT

01002425DB NSF RESEARCH & RELATED ACTIVIT

01002526DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7937
Program Element Code(s): 779700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Contemporary machine learning techniques tend to be resource-intensive, often requiring good quality datasets, expensive hardware, or significant computing power. In a wide array of application domains, ranging from healthcare to mobile computing, these critical resources are lacking. Novel methodologies that enable the optimal utilization of resources can help unlock the full potential of the data science revolution for these domains. Towards this aim, this project will develop theoretically-grounded algorithms to facilitate the design of machine learning models under application-specific resource constraints. The outcomes of the project will help enable machine learning methods to operate with less human-annotated data, less computing power, and on a wider range of hardware platforms. To demonstrate interdisciplinary impact, the resulting algorithms will be employed in the design of efficient hydrological models which aid in predicting and managing water resources. The research will also be strongly coupled with education through the mentoring of undergraduate students, new undergraduate and graduate course development, and live broadcasts of the lectures over publicly accessible online platforms.

This project aims to develop the foundational theories and algorithms to guide the efficient use of statistical and computational resources. The research on the statistical front focuses on the data and will uncover the fundamental tradeoffs between the data amount, label quality, and the model accuracy. Understanding these tradeoffs will lead to the design of improved loss functions and regularization techniques. On the computational front, theory-inspired model compression schemes will be developed by exploring the interplay between the model size and accuracy. Secondly, the model performance will be enhanced by identifying the optimal model architecture via computationally-efficient algorithms that co-design the architecture, compression scheme, and the loss function. These theoretical and algorithmic investigations will utilize tools from statistical learning, optimization, deep learning theory, and high-dimensional probability. The proposed research is expected to provide much-needed theoretical basis for poorly-understood heuristics in fields spanning semi-supervised learning, model compression, neural architecture search, and will guide the design of next-generation algorithms achieving the optimal resource tradeoffs.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 27)
Ataee Tarzanagh, Davoud and Li, Yingcong and Zhang, Xuechen and Oymak, Samet "Max-Margin Token Selection in Attention Mechanism" Advances in neural information processing systems , 2023 Citation Details
Ahmed, Sk Miraj and Raychaudhuri, Dripta S. and Paul, Sujoy and Oymak, Samet and Roy-Chowdhury, Amit K. "Unsupervised Multi-source Domain Adaptation Without Access to Source Data" 2021 Conference on Computer Vision and Pattern Recognition , 2021 https://doi.org/10.1109/CVPR46437.2021.00997 Citation Details
Chang, Xiangyu and Li, Yingcong and Oymak, Samet and Thrampoulidis, Christos "Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks" The Thirty-Fifth AAAI Conference on Artificial Intelligence , 2021 Citation Details
Du, Zhe and Sattar, Yahya and Tarzanagh, Davoud Ataee and Balzano, Laura and Ozay, Necmiye and Oymak, Samet "Data-driven control of markov jump systems: Sample complexity and regret bounds" 2022 American Control Conference (ACC) , 2022 https://doi.org/10.23919/ACC53348.2022.9867863 Citation Details
Elamvazhuthi, Karthik and Zhang, Xuechen and Jacobs, Matthew and Oymak, Samet and Pasqualetti, Fabio "A Score-Based Deterministic Diffusion Algorithm with Smooth Scores for General Distributions" , 2024 Citation Details
Ildiz, Muhammed E and Huang, Yixiao and Li, Yingcong and Rawat, Ankit S and Oymak, Samet "From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers" , 2024 Citation Details
Ildiz, Muhammed E and Zhao, Zhe and Oymak, Samet "Understanding Inverse Scaling and Emergence in Multitask Representation Learning" , 2024 Citation Details
Kini, Ganesh R and Paraskevas, Orestis and Oymak, Samet and Thrampoulidis, Christos "Label-Imbalanced and Group-Sensitive Classification under Overparameterization" 35th Conference on Neural Information Processing Systems , 2021 Citation Details
Li, Mingchen and Zhang, Xuechen and Thrampoulidis, Christos and Chen, Jiasi and Oymak, Samet "AutoBalance: Optimized Loss Functions for Imbalanced Data" 35th Conference on Neural Information Processing Systems , 2021 Citation Details
Li, Y. and Ildiz, M. E. and Papailiopoulos, D. and & Oymak, S. "Transformers as algorithms: Generalization and stability in in-context learning" International Conference on Machine Learning , 2023 Citation Details
Li, Y. and Sreenivasan, K. and Giannou, A. and Papailiopoulos, D. and Oymak, S. "Dissecting Chain-of-Thought: Compositionality through In-Context Filtering and Learning" , 2023 Citation Details
(Showing: 1 - 10 of 27)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page