Skip to feedback

Award Abstract # 2134148
Collaborative Research: SCALE MoDL: Advancing Theoretical Minimax Deep Learning: Optimization, Resilience, and Interpretability

NSF Org: DMS
Division Of Mathematical Sciences
Recipient: REGENTS OF THE UNIVERSITY OF MINNESOTA
Initial Amendment Date: August 23, 2021
Latest Amendment Date: July 3, 2023
Award Number: 2134148
Award Instrument: Continuing Grant
Program Manager: Jodi Mead
jmead@nsf.gov
 (703)292-7212
DMS
 Division Of Mathematical Sciences
MPS
 Directorate for Mathematical and Physical Sciences
Start Date: September 1, 2021
End Date: August 31, 2024 (Estimated)
Total Intended Award Amount: $273,887.00
Total Awarded Amount to Date: $273,887.00
Funds Obligated to Date: FY 2021 = $93,161.00
FY 2022 = $94,630.00

FY 2023 = $86,096.00
History of Investigator:
  • Jie Ding (Principal Investigator)
    dingj@umn.edu
Recipient Sponsored Research Office: University of Minnesota-Twin Cities
2221 UNIVERSITY AVE SE STE 100
MINNEAPOLIS
MN  US  55414-3074
(612)624-5599
Sponsor Congressional District: 05
Primary Place of Performance: Regents of the University of Minnesota
224 Church St SE, 313 Ford Hall
Minneapolis
MN  US  55455-2070
Primary Place of Performance
Congressional District:
05
Unique Entity Identifier (UEI): KABJZBBJ4B54
Parent UEI:
NSF Program(s): OFFICE OF MULTIDISCIPLINARY AC,
Special Projects - CCF,
IIS Special Projects,
CDS&E-MSS
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
01002223DB NSF RESEARCH & RELATED ACTIVIT

01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 075Z, 079Z
Program Element Code(s): 125300, 287800, 748400, 806900
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.041, 47.049

ABSTRACT

The past decade has witnessed the great success of deep learning in broad societal and commercial applications. However, conventional deep learning relies on fitting data with neural networks, which is known to produce models that lack resilience. For instance, models used in autonomous driving are vulnerable to malicious attacks, e.g., putting an art sticker on a stop sign can cause the model to classify it as a speed limit sign; models used in facial recognition are known to be biased toward people of a certain race or gender; models in healthcare can be hacked to reconstruct the identities of patients that are used in training those models. The next-generation deep learning paradigm needs to deliver resilient models that promote robustness to malicious attacks, fairness among users, and privacy preservation. This project aims to develop a comprehensive learning theory to enhance the model resilience of deep learning. The project will produce fast algorithms and new diagnostic tools for training, enhancing, visualizing, and interpreting model resilience, all of which can have broad research and societal significance. The research activities will also generate positive educational impacts on undergraduate and graduate students. The materials developed by this project will be integrated into courses on machine learning, statistics, and data visualization and will benefit interdisciplinary students majoring in electrical and computer engineering, statistics, mathematics, and computer science. The project will actively involve underrepresented students and integrate research with education for undergraduate and graduate students in STEM. It will also produce introductory materials for K-12 students to be used in engineering summer camps.

In this project, the investigators will collaboratively develop a comprehensive minimax learning theory that advances the fundamental understanding of minimax deep learning from the perspectives of optimization, resilience, and interpretability. These complementary theoretical developments, in turn, will guide the design of novel minimax learning algorithms with substantially improved computational efficiency, statistical guarantees, and interpretability. The research includes three major thrusts. First, the investigators will develop a principled non-convex minimax optimization theory that supports scalable, fast, and convergent gradient-descent-ascent algorithms for training complex minimax deep learning models. The theory will focus on analyzing the convergence rate and sample complexity of the developed algorithms. Second, the investigators will formulate a measure of vulnerability of deep learning models and study how minimaxity can enhance their resilience against data, model, and task deviations. This theory will focus on the statistical limits of deep learning. Lastly, the investigators will establish the mathematical foundations for a set of novel visual analytics techniques that increase the model interpretability of minimax learning. In particular, the theory will provide guidance on visualizing and interpreting model resilience.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Chen, C. and Zhang, J. and Ding, J. and Zhou, Y. "Assisted Unsupervised Domain Adaptation" Abstracts of papers IEEE International Symposium on Information Theory , 2023 Citation Details
Chen, C. and Zhou, J. and Ding, J. and Zhou, Y. "Assisted Learning for Organizations with Limited Imbalanced Data" Transactions on machine learning research , 2023 Citation Details
Diao, E. and Ding, J. and Tarokh, V. "SemiFL: Semi-Supervised Federated Learning for Unlabeled Clients with Alternate Training" Advances in neural information processing systems , 2022 Citation Details
Li, Gen and Wang, Ganghua and Ding, Jie "Provable Identifiability of Two-Layer ReLU Neural Networks via LASSO Regularization" IEEE Transactions on Information Theory , 2023 https://doi.org/10.1109/TIT.2023.3274152 Citation Details
Morchdi, Chedi and Zhou, Yi and Ding, Jie and Wang, Bei "Exploring Gradient Oscillation in Deep Neural Network Training" , 2023 Citation Details
Mushtaq, Erum and He, Chaoyang and Ding, Jie and Avestimehr, Salman "Distributed Architecture Search Over Heterogeneous Distributions" Transactions on machine learning research , 2023 Citation Details
Wang, Xinran and Zhang, Jiawei and Hong, Mingyi and Yang, Yuhong and Ding, Jie "Parallel Assisted Learning" IEEE Transactions on Signal Processing , v.70 , 2022 https://doi.org/10.1109/TSP.2022.3229637 Citation Details
Xian, Xun and Hong, Mingyi and Ding, Jie "Mismatched Supervised Learning" International Conference on Acoustics, Speech, & Signal Processing (ICASSP) , 2022 https://doi.org/10.1109/ICASSP43922.2022.9747362 Citation Details
Zhang, Jiawei and Yang, Yuhong and Ding, Jie "Information criteria for model selection" WIREs Computational Statistics , v.15 , 2023 https://doi.org/10.1002/wics.1607 Citation Details
Zhou, Youjia and Zhou, Yi and Ding, Jie and Wang, Bei "Visualizing and Analyzing the Topology of Neuron Activations in Deep Adversarial Training" , 2023 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Overall Outcome:
Over the past three years, our project has advanced the capabilities of machine learning systems, focusing on making them more reliable and understandable, especially in challenging environments. By developing new theories and tools, we've enabled these systems to better handle unexpected changes and become more reliable in the complex open world. Our work has centered on creating methods that help these systems make better decisions, even when they are faced with incomplete or misleading information.

Intellectual Merit:
The intellectual essence of our project lies in its development of a solid theoretical foundation for what we call minimax deep learning. This approach has broadened our understanding of how learning systems can operate optimally under worst-case scenarios—imagine a system smartly navigating through the toughest situations by anticipating potential challenges and adapting accordingly. Our research has led to innovative strategies that improve how these systems learn from data that is diverse, incomplete, or unstructured. For instance, we've crafted algorithms that tailor learning models to perform well across various settings without needing to compromise on privacy or efficiency.

Broader Impact:
The broader impacts of our project are multifaceted:

First, we significantly enriched the educational experience for several PhD students, equipping them with cutting-edge knowledge that they are now taking into their new roles in academia and industry. This ripple effect ensures that the benefits of our research extend beyond our immediate project.

Second, our team has actively disseminated research findings through conferences, workshops, and seminars, engaging with a wide audience from academic peers to industry practitioners. This has sparked further innovation and interest in the field of machine learning.

Third, the practical applications of our work are vast, ranging from improving the security measures of digital systems to enhancing the fairness and transparency of automated decision-making in sectors like healthcare and public services. By increasing the robustness and reliability of these systems, we contribute to safer and more equitable technology use in everyday life.

 


Last Modified: 10/07/2024
Modified by: Jie Ding

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page