Award Abstract # 2007688
CIF: Small: Alpha Loss: A New Framework for Understanding and Trading Off Computation, Accuracy, and Robustness in Machine Learning

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: ARIZONA STATE UNIVERSITY
Initial Amendment Date: August 25, 2020
Latest Amendment Date: March 29, 2023
Award Number: 2007688
Award Instrument: Standard Grant
Program Manager: Alfred Hero
ahero@nsf.gov
 (703)292-0000
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: October 1, 2020
End Date: September 30, 2025 (Estimated)
Total Intended Award Amount: $507,999.00
Total Awarded Amount to Date: $539,999.00
Funds Obligated to Date: FY 2020 = $507,999.00
FY 2021 = $16,000.00

FY 2023 = $16,000.00
History of Investigator:
  • Lalitha Sankar (Principal Investigator)
    lalithasankar@asu.edu
  • Gautam Dasarathy (Co-Principal Investigator)
Recipient Sponsored Research Office: Arizona State University
660 S MILL AVENUE STE 204
TEMPE
AZ  US  85281-3670
(480)965-5479
Sponsor Congressional District: 04
Primary Place of Performance: Arizona State University
PO Box 876011
Tempe
AZ  US  85287-6011
Primary Place of Performance
Congressional District:
04
Unique Entity Identifier (UEI): NTLHJXM55KZ6
Parent UEI:
NSF Program(s): Special Projects - CCF,
Comm & Information Foundations
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
01002021DB NSF RESEARCH & RELATED ACTIVIT

01002122DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 075Z, 079Z, 2878, 7797, 7923, 7936, 9178, 9251
Program Element Code(s): 287800, 779700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

At the heart of the machine learning (ML) and artificial intelligence (AI) revolution are models that are trained using vast amounts of data. Given the increasing use of such data-driven modeling, there is an urgent need to understand and leverage the tradeoffs between various performance characteristics such as accuracy (statistical efficiency), computational speed (computational efficiency), and robustness (say to noise, adversarial tampering, and imbalance or biased data). This project develops a unified and powerful framework for understanding and trading off these facets by introducing the family of alpha-loss functions -- often-used loss functions such as the 0-1 loss, the log-loss, and the exponential-loss appear as instantiations of the alpha-loss framework. Over the past few years, we have seen a steadily growing recognition amongst advocates, regulators, and scientists that data-driven inference and decision engines pose significant challenges for ensuring non-discrimination, and fair and inclusive representation. The alpha-loss framework, combined with several technological advances, will allow practitioners to incorporate fairness as an explicit knob to be tuned during the development of machine learning models. Broader impacts of this work also include developing ML modules for a week-long summer camp for high school students as well as providing research opportunities for such students.

This project: (i) develops theoretical results on the behavior of the loss landscape as a function of the tuning parameter alpha, thereby illuminating the value and limitation of the industry standard log-loss, (ii) establishes accuracy-speed tradeoffs and generalization bounds, and (iii) designs practical adaptive algorithms with guarantees for tuning the hyperparameter alpha to achieve various operating points along the tradeoff. This project establishes the robustness properties of alpha-loss via the theory of influence functions. By introducing much-needed models for noise and adversarial examples, this work develops a principled method to choose alpha slightly larger than 1 to design models more robust to noise and adversaries. Using both influence functions and constrained learning settings such as fair classification, this project studies the efficacy of tuning alpha below one in order to enhance sensitivity to limited samples in highly imbalanced training datasets. Finally, this project also develops alpha-Boost as a tunable boosting algorithm with guaranteed convergence, robustness to noise and, where needed, online adaptation. Research is enhanced at every stage of this project through rigorous testing of algorithms on both synthetic and publicly available real datasets.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 29)
Alghamdi, Wael and Asoodeh, Shahab and Calmon, Flavio P and Felipe_Gomez, Juan and Kosut, Oliver and Sankar, Lalitha "Optimal Multidimensional Differentially Private Mechanisms in the Large-Composition Regime" , 2023 https://doi.org/10.1109/ISIT54713.2023.10206658 Citation Details
Alghamdi, Wael and Asoodeh, Shahab and Calmon, Flavio P and Felipe_Gomez, Juan and Kosut, Oliver and Sankar, Lalitha "Schrödinger Mechanisms: Optimal Differential Privacy Mechanisms for Small Sensitivity" , 2023 https://doi.org/10.1109/ISIT54713.2023.10206616 Citation Details
Alghamdi, Wael and Asoodeh, Shahab and Calmon, Flavio P. and Kosut, Oliver and Sankar, Lalitha and Wei, Fei "Cactus Mechanisms: Optimal Differential Privacy Mechanisms in the Large-Composition Regime" IEEE International Symposium on Information Theory , 2022 https://doi.org/10.1109/ISIT50566.2022.9834438 Citation Details
Alghamdi, Wael and Gomez, Juan_Felipe and Asoodeh, Shahab and Calmon, Flavio and Kosut, Oliver and Sankar, Lalitha "The Saddle-Point Method in Differential Privacy" , 2023 Citation Details
Asoodeh, Shahab and Liao, Jiachun and Calmon, Flavio P. and Kosut, Oliver and Sankar, Lalitha "Three Variants of Differential Privacy: Lossless Conversion and Applications" IEEE Journal on Selected Areas in Information Theory , v.2 , 2021 https://doi.org/10.1109/JSAIT.2021.3054692 Citation Details
Berisha, Visar and Krantsevich, Chelsea and Hahn, P. Richard and Hahn, Shira and Dasarathy, Gautam and Turaga, Pavan and Liss, Julie "Digital medicine and the curse of dimensionality" npj Digital Medicine , v.4 , 2021 https://doi.org/10.1038/s41746-021-00521-5 Citation Details
Chang, Andersen and Zheng, Lili and Dasarathy, Gautam and Allen, Genevera I. "Nonparanormal graph quilting with applications to calcium imaging" Stat , v.12 , 2023 https://doi.org/10.1002/sta4.623 Citation Details
Diaz, Mario and Kairouz, Peter and Liao, Jiachun and Sankar, Lalitha "Neural Network-based Estimation of the MMSE" International Symposium on Information Theory , 2021 https://doi.org/10.1109/ISIT45174.2021.9518063 Citation Details
Gilani, Atefeh and Kurri, Gowtham R and Kosut, Oliver and Sankar, Lalitha "Unifying Privacy Measures via Maximal ( , )-Leakage (MbeL)" IEEE Transactions on Information Theory , v.70 , 2024 https://doi.org/10.1109/TIT.2024.3384922 Citation Details
Inman, Joshua and Khandait, Tanmay and Pedrielli, Giulia and Sankar, Lalitha "Parameter Optimization With Conscious Allocation (POCA)" , 2023 Citation Details
Kairouz, Peter and Liao, Jiachun and Huang, Chong and Vyas, Maunil and Welfert, Monica and Sankar, Lalitha "Generating Fair Universal Representations Using Adversarial Models" IEEE Transactions on Information Forensics and Security , v.17 , 2022 https://doi.org/10.1109/TIFS.2022.3170265 Citation Details
(Showing: 1 - 10 of 29)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page