Skip to feedback

Award Abstract # 1652491
CAREER: Beyond Worst-Case Analysis: New Approaches in Approximation Algorithms and Machine Learning

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: NORTHWESTERN UNIVERSITY
Initial Amendment Date: March 10, 2017
Latest Amendment Date: March 30, 2021
Award Number: 1652491
Award Instrument: Continuing Grant
Program Manager: Peter Brass
pbrass@nsf.gov
 (703)292-2182
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: March 15, 2017
End Date: February 29, 2024 (Estimated)
Total Intended Award Amount: $505,251.00
Total Awarded Amount to Date: $505,251.00
Funds Obligated to Date: FY 2017 = $55,436.00
FY 2018 = $100,269.00

FY 2019 = $106,032.00

FY 2020 = $106,795.00

FY 2021 = $136,719.00
History of Investigator:
  • Aravindan Vijayaraghavan (Principal Investigator)
    aravindv@northwestern.edu
Recipient Sponsored Research Office: Northwestern University
633 CLARK ST
EVANSTON
IL  US  60208-0001
(312)503-7955
Sponsor Congressional District: 09
Primary Place of Performance: Northwestern University
2133 N Sheridan Rd, 3-207
Evanston
IL  US  60208-3109
Primary Place of Performance
Congressional District:
09
Unique Entity Identifier (UEI): EXZVPWZBLUE8
Parent UEI:
NSF Program(s): Algorithmic Foundations
Primary Program Source: 01001718DB NSF RESEARCH & RELATED ACTIVIT
01001819DB NSF RESEARCH & RELATED ACTIVIT

01001920DB NSF RESEARCH & RELATED ACTIVIT

01002021DB NSF RESEARCH & RELATED ACTIVIT

01002122DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7926
Program Element Code(s): 779600
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Combinatorial optimization problems such as clustering, and unsupervised learning of probabilistic models are important computational problems that arise in diverse areas including machine learning, computer vision, operations research and data analysis. However, there is a large disconnect between our theoretical and practical understanding of these problems -- while theory tells us that many interesting computational problems in combinatorial optimization and machine learning are intractable in the worst case, practitioners in areas such as machine learning and computer vision have made significant progress in solving such theoretically hard problems. This project focuses on bridging the fundamental gap between theory and practice by developing paradigms and machinery that will allow us to reason about the performance of algorithms on real-world instances. This research has the potential to have broad impact on both theory and practice of computational problems across different areas of computer science, machine learning and statistics. The project will involve students at all levels of research, will integrate aspects of average-case analysis in both graduate and undergraduate courses, and will include outreach activities in high schools in Evanston and the broader Chicago area.

The PI will study several problems in machine learning and combinatorial optimization by using realistic average-case models and smoothed analysis. Broad goals include designing new model-independent algorithms with provable guarantees for realistic average-case models of graph partitioning and clustering and challenging average-case settings where there is no unique or planted solution. These algorithms will also lead to new algorithmic techniques for learning probabilistic models such as mixtures of Gaussians and stochastic block models that are robust to various kinds of modeling errors and noise. Another focus of the project is on developing new efficient algorithms for learning latent variable models and for reasoning about the performance of algorithms using smoothed analysis.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Awasthi, Pranjal and Chen, Xue and Vijayaraghavan, Aravindan "Estimating Principal Components under Adversarial Perturbations" Proceedings of Thirty Third Conference on Learning Theory, PMLR 125:323-362 , 2020 https://doi.org/ Citation Details
Awasthi, Pranjal and Dutta, Abhratanu and Vijayaraghavan, Aravindan "On Robustness to Adversarial Examples and Polynomial Optimization" Advances in Neural Information Processing Systems 32 (NIPS 2019) , 2019 Citation Details
Awasthi, Pranjal and Jain, Himanshu and Rawat, Ankit Singh and Vijayaraghavan, Aravindan "Adversarial robustness via robust low rank representations" Advances in Neural Information Processing Systems 33 (NeurIPS 2020) , 2020 https://doi.org/ Citation Details
Awasthi, Pranjal and Vijayaraghavan, Aravindan "Clustering Semi-Random Mixtures of Gaussians" Proceedings of Machine Learning Research , 2017 Citation Details
Awasthi, Pranjal and Vijayaraghavan, Aravindan "Clustering Semi-Random Mixtures of Gaussians" Proceedings of Machine Learning Research , 2018 Citation Details
Awasthi, Pranjal and Vijayaraghavan, Aravindan "Towards Learning Sparsely Used Dictionaries with Arbitrary Supports" 2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS) , 2018 10.1109/FOCS.2018.00035 Citation Details
Bhaskara, Aditya and Chen, Aidao and Perreault, Aidan and Vijayaraghavan, Aravindan "Smoothed Analysis in Unsupervised Learning via Decoupling" 2019 IEEE 60th Annual Symposium on Foundations of Computer Science (FOCS) , 2019 10.1109/FOCS.2019.00043 Citation Details
Chen, Aidao and De, Anindya and Vijayaraghavan, Aravindan "Learning a mixture of two subspaces over finite fields" The 32nd International Conference on Algorithmic Learning Theory , 2021 https://doi.org/ Citation Details
Lang, Hunter and Sontag, David and Vijayaraghavan, Aravindan "Block Stability for MAP inference" Proceedings of Machine Learning Research , v.86 , 2019 Citation Details
Maiti, Biswaroop and Rajaraman, Rajmohan and Stalfa, David and Svitkina, Zoya and Vijayaraghavan, Aravindan "Scheduling Precedence-Constrained Jobs on Related Machines with Communication Delay" 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS) , 2020 https://doi.org/10.1109/FOCS46700.2020.00082 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

This project focused on developing new theories and techniques for reasoning about algorithmic problems that arise in machine learning, and optimization though a beyond worst-case analysis lens. Traditional paradigms in theoretical computer science have focused on worst-case analysis of algorithms which can be pessimistic. Real-world instances of computational problems are unlikely to be in degenerate or pathological configurations. On the research side, the project led to the development of new models of real-world instances, new techniques to prove guarantees, and efficient algorithms that were also robust to different modeling errors and corruptions. These research outcomes were complemented with new courses and educational material on related topics, and the training of several PhD students and undergraduate students on beyond worst-case paradigms.  

On the research front, one of the highlights was the development of new smoothed analysis techniques for many basic problems in machine learning and high-dimensional data analysis that are computationally intractable in the worst-case. In smoothed analysis, the algorithm is analyzed on a small random perturbation of an arbitrary input. Smoothed analysis guarantees show that the bad instances for the algorithm are isolated – qualitatively they represent the strongest possible algorithmic guarantees in the absence of worst-case guarantees. In a sequence of works that also involved PhD students (funded by the project) and undergraduate students at Northwestern, we developed new techniques that are broadly useful for proving smoothed analysis guarantees (Bhaskara et al. FOCS 2019, Math Programming 2021, and STOC 2024). We developed new probabilistic ideas for establishing lower bounds on the least singular value of random matrices with limited randomness. Least singular value bounds often involve showing strong anti-concentration inequalities that are intricate and much less understood compared to concentration (or large deviation) bounds. These techniques allowed us to resolve open questions in smoothed analysis for well-studied problems like learning hidden Markov models, tensor decompositions, subspace clustering, learning mixtures of non-spherical decompositions and several other problems. Moreover, we used these ideas to also develop new that gave the first polynomial time algorithmic guarantees (through smoothed analysis) for learning depth-2 neural networks with general ReLU activation (Awasthi et al., NeurIPS 2021). Finally, in a sequence of recent projects (Johnston, Lovitz and Vijayaraghavan QIP 2022, FOCS 2023), we studied a broad class of algorithmic problems in algebraic geometry that capture important problems in quantum information and on low-rank tensor decompositions. While problem is known to be NP-hard in the worst-case, we showed that we could design efficient algorithms with smoothed analysis guarantees.

Another highlight of this project was the development of more realistic average-case models called semi-random models for common machine learning tasks. Semi-random models provide a framework to reason about robustness of algorithms to modeling errors by incorporating both adversarial and random choices in instance generation. In two different works (Awasthi and Vijayaraghvan FOCS 2018, NeurIPS 2018) we studied new semi-random models for two important problems in unsupervised learning: dictionary learning, and clustering mixtures of Gaussians, and designed efficient algorithms with robustness guarantees for these problems. This focus on robustness also led to the development of machine learning algorithms with provable guarantees in the presence of adversarial attacks and perturbations.  Adversarial robustness of machine learning methods is important from a reliability and security standpoint. In a sequence of works (Awasthi et al. COLT 2020, NeurIPS 2020, COLT 2021) we designed algorithms for basic machine learning primitives like finding low-dimensional data representations that are robust to adversarial perturbations. We gave polynomial time algorithms with approximately optimal guarantees, and also gave similar guarantees for downstream tasks. This approach was also used to get better adversarial robustness for deep neural networks in image classification. 

The project involved overall 6 PhD students (3 of whom graduated) and 3 undergraduate students (all went on to do PhDs). There were several graduate courses on related topics that were developed, along with a new undergraduate course on mathematical foundations that covered linear algebra, probability and optimization foundations for computer science. Finally, the PI was also involved in outreach activities in the broader Chicago area that targeted K-12 students, particularly from underrepresented minorities.

 

The project also supported organized activities and workshops for the broader scientific community. One of these workshops was the Junior Theorists Workshop, which provided opportunities for some of the best junior researchers (PhD students and postdoctoral researchers, who are not in faculty positions) to showcase their research. The research that came out of the project has also been published in several conferences and journals for dissemination, and in publicly accessible repositories like ArXiv. The PI also authored a publicly-accessible survey on smoothed analysis for tensor decompositions that also became a chapter in the book on Beyond the Worst-Case Analysis of Algorithms.  Finally, videos of many of the seminar talks and workshop talks have been made available publicly.

 


Last Modified: 07/02/2024
Modified by: Aravindan Vijayaraghavan

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page