
NSF Org: |
DMS Division Of Mathematical Sciences |
Recipient: |
|
Initial Amendment Date: | August 23, 2021 |
Latest Amendment Date: | July 3, 2023 |
Award Number: | 2134148 |
Award Instrument: | Continuing Grant |
Program Manager: |
Jodi Mead
jmead@nsf.gov (703)292-7212 DMS Division Of Mathematical Sciences MPS Directorate for Mathematical and Physical Sciences |
Start Date: | September 1, 2021 |
End Date: | August 31, 2024 (Estimated) |
Total Intended Award Amount: | $273,887.00 |
Total Awarded Amount to Date: | $273,887.00 |
Funds Obligated to Date: |
FY 2022 = $94,630.00 FY 2023 = $86,096.00 |
History of Investigator: |
|
Recipient Sponsored Research Office: |
2221 UNIVERSITY AVE SE STE 100 MINNEAPOLIS MN US 55414-3074 (612)624-5599 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
224 Church St SE, 313 Ford Hall Minneapolis MN US 55455-2070 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): |
OFFICE OF MULTIDISCIPLINARY AC, Special Projects - CCF, IIS Special Projects, CDS&E-MSS |
Primary Program Source: |
01002223DB NSF RESEARCH & RELATED ACTIVIT 01002324DB NSF RESEARCH & RELATED ACTIVIT |
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.041, 47.049 |
ABSTRACT
The past decade has witnessed the great success of deep learning in broad societal and commercial applications. However, conventional deep learning relies on fitting data with neural networks, which is known to produce models that lack resilience. For instance, models used in autonomous driving are vulnerable to malicious attacks, e.g., putting an art sticker on a stop sign can cause the model to classify it as a speed limit sign; models used in facial recognition are known to be biased toward people of a certain race or gender; models in healthcare can be hacked to reconstruct the identities of patients that are used in training those models. The next-generation deep learning paradigm needs to deliver resilient models that promote robustness to malicious attacks, fairness among users, and privacy preservation. This project aims to develop a comprehensive learning theory to enhance the model resilience of deep learning. The project will produce fast algorithms and new diagnostic tools for training, enhancing, visualizing, and interpreting model resilience, all of which can have broad research and societal significance. The research activities will also generate positive educational impacts on undergraduate and graduate students. The materials developed by this project will be integrated into courses on machine learning, statistics, and data visualization and will benefit interdisciplinary students majoring in electrical and computer engineering, statistics, mathematics, and computer science. The project will actively involve underrepresented students and integrate research with education for undergraduate and graduate students in STEM. It will also produce introductory materials for K-12 students to be used in engineering summer camps.
In this project, the investigators will collaboratively develop a comprehensive minimax learning theory that advances the fundamental understanding of minimax deep learning from the perspectives of optimization, resilience, and interpretability. These complementary theoretical developments, in turn, will guide the design of novel minimax learning algorithms with substantially improved computational efficiency, statistical guarantees, and interpretability. The research includes three major thrusts. First, the investigators will develop a principled non-convex minimax optimization theory that supports scalable, fast, and convergent gradient-descent-ascent algorithms for training complex minimax deep learning models. The theory will focus on analyzing the convergence rate and sample complexity of the developed algorithms. Second, the investigators will formulate a measure of vulnerability of deep learning models and study how minimaxity can enhance their resilience against data, model, and task deviations. This theory will focus on the statistical limits of deep learning. Lastly, the investigators will establish the mathematical foundations for a set of novel visual analytics techniques that increase the model interpretability of minimax learning. In particular, the theory will provide guidance on visualizing and interpreting model resilience.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
Overall Outcome:
Over the past three years, our project has advanced the capabilities of machine learning systems, focusing on making them more reliable and understandable, especially in challenging environments. By developing new theories and tools, we've enabled these systems to better handle unexpected changes and become more reliable in the complex open world. Our work has centered on creating methods that help these systems make better decisions, even when they are faced with incomplete or misleading information.
Intellectual Merit:
The intellectual essence of our project lies in its development of a solid theoretical foundation for what we call minimax deep learning. This approach has broadened our understanding of how learning systems can operate optimally under worst-case scenarios—imagine a system smartly navigating through the toughest situations by anticipating potential challenges and adapting accordingly. Our research has led to innovative strategies that improve how these systems learn from data that is diverse, incomplete, or unstructured. For instance, we've crafted algorithms that tailor learning models to perform well across various settings without needing to compromise on privacy or efficiency.
Broader Impact:
The broader impacts of our project are multifaceted:
First, we significantly enriched the educational experience for several PhD students, equipping them with cutting-edge knowledge that they are now taking into their new roles in academia and industry. This ripple effect ensures that the benefits of our research extend beyond our immediate project.
Second, our team has actively disseminated research findings through conferences, workshops, and seminars, engaging with a wide audience from academic peers to industry practitioners. This has sparked further innovation and interest in the field of machine learning.
Third, the practical applications of our work are vast, ranging from improving the security measures of digital systems to enhancing the fairness and transparency of automated decision-making in sectors like healthcare and public services. By increasing the robustness and reliability of these systems, we contribute to safer and more equitable technology use in everyday life.
Last Modified: 10/07/2024
Modified by: Jie Ding
Please report errors in award information by writing to: awardsearch@nsf.gov.