Award Abstract # 2044149
CAREER: Generalization and Safety Guarantees for Learning-Based Control of Robots

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: THE TRUSTEES OF PRINCETON UNIVERSITY
Initial Amendment Date: February 24, 2021
Latest Amendment Date: July 25, 2022
Award Number: 2044149
Award Instrument: Continuing Grant
Program Manager: Cang Ye
cye@nsf.gov
 (703)292-4702
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: August 1, 2021
End Date: July 31, 2026 (Estimated)
Total Intended Award Amount: $545,980.00
Total Awarded Amount to Date: $545,980.00
Funds Obligated to Date: FY 2021 = $400,000.00
FY 2022 = $145,980.00
History of Investigator:
  • Anirudha Majumdar (Principal Investigator)
Recipient Sponsored Research Office: Princeton University
1 NASSAU HALL
PRINCETON
NJ  US  08544-2001
(609)258-3090
Sponsor Congressional District: 12
Primary Place of Performance: Princeton University
NJ  US  08544-2020
Primary Place of Performance
Congressional District:
12
Unique Entity Identifier (UEI): NJ1YPQXQG7U5
Parent UEI:
NSF Program(s): FRR-Foundationl Rsrch Robotics
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
01002223DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 6840
Program Element Code(s): 144Y00
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

The ability of machine learning techniques to process rich sensory inputs such as vision makes them highly appealing for use in robotic systems (e.g., micro aerial vehicles and robotic manipulators). However, the increasing adoption of learning-based components in the robotics perception and control pipeline poses an important challenge: how can we guarantee the safety and performance of such systems? As an example, consider a micro aerial vehicle that learns to navigate using a thousand different obstacle environments or a robotic manipulator that learns to grasp using a million objects in a dataset. How likely are these systems to remain safe and perform well on a novel (i.e., previously unseen) environment or object? How can we learn control policies for robotic systems that provably generalize well to environments that our robot has not previously encountered? Unfortunately, existing approaches either do not provide such guarantees or do so only under very restrictive assumptions. This Faculty Early Career Development (CAREER) project seeks to establish a foundational framework for learning-based control of safety-critical robotic systems with guaranteed generalization and safety. The project will impact challenging application domains such as aerial inspection and manipulation (e.g., for infrastructure repair tasks) and includes activities for (i) engaging regulatory agencies and industry entities in discussions regarding the certification of learning-based robotic systems, (ii) partnering with teacher preparation programs and other educational programs to engage high-school and undergraduate students in robotics, and (iii) widely disseminating materials from a new robotics course which uses hands-on labs with drones.

Motivated by the need for guaranteeing the safety of learning-based robotic systems, this project is developing a principled theoretical and algorithmic framework for learning control policies for robotic systems with provable guarantees on generalization to novel environments (i.e., environments that the robot has not previously encountered). The key technical insight of this project is to leverage and extend powerful techniques from generalization theory in theoretical machine learning. The resulting framework provides bounds on the expected performance of learned policies (including ones based on neural networks) across novel environments. The project is developing algorithms (based on convex optimization, gradient-based methods, and black-box optimization) for learning policies that explicitly optimize these bounds. The project also seeks to guarantee the robustness of learned policies to shifts in the distribution of environments that the robot encounters. An important part of the effort is to thoroughly validate the technical approach on hardware platforms including micro aerial vehicles performing navigation, inspection, and aerial manipulation tasks motivated by infrastructure repair applications.

This project is supported by the cross-directorate Foundational Research in Robotics program, jointly managed and funded by the Directorates for Engineering (ENG) and Computer and Information Science and Engineering (CISE).

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 11)
Agarwal, Abhinav and Veer, Sushant and Ren, Allen Z. and Majumdar, Anirudha "Stronger Generalization Guarantees for Robot Learning by Combining Generative Models and Real-World Data" IEEE International Conference on Robotics and Automation , 2022 https://doi.org/10.1109/ICRA46639.2022.9811565 Citation Details
Booker, Meghan and Majumdar, Anirudha "Switching Attention in Time-Varying Environments via Bayesian Inference of Abstractions" Proceedings IEEE International Conference on Robotics and Automation , 2023 Citation Details
Farid, Alec and Majumdar, Anirudha "Generalization Bounds for Meta-Learning via PAC-Bayes and Uniform Stability" Advances in neural information processing systems , 2021 Citation Details
Farid, Alec and Snyder, David and Ren, Allen Z. and Majumdar, Anirudha "Failure Prediction with Statistical Guarantees for Vision-Based Robot Control" Robotics: Science and Systems (RSS) , 2022 https://doi.org/10.15607/RSS.2022.XVIII.042 Citation Details
Farid, Alec and Veer, Sushant and Majumdar, Anirudha "Task-Driven Out-of-Distribution Detection with Statistical Guarantees for Robot Learning" Conference on Robot Learning (CoRL) , 2021 Citation Details
Ho, Michelle and Farid, Alec and Majumdar, Anirudha "Towards a Framework for Comparing the Complexity of Robotic Tasks" Workshop on the Algorithmic Foundations of Robotics (WAFR) , 2022 Citation Details
Hsu, Kai-Chieh and Ren, Allen Z. and Nguyen, Duy P. and Majumdar, Anirudha and Fisac, Jaime F. "Sim-to-Lab-to-Real: Safe reinforcement learning with shielding and generalization guarantees" Artificial Intelligence , v.314 , 2023 https://doi.org/10.1016/j.artint.2022.103811 Citation Details
Majumdar, Anirudha "Fundamental Tradeoffs in Learning with Prior Information" Proceedings of the International Conference on Machine Learning , 2023 Citation Details
Majumdar, Anirudha and Pacelli, Vincent "Fundamental Performance Limits for Sensor-Based Robot Control and Policy Learning" Robotics: Science and Systems (RSS) , 2022 https://doi.org/10.15607/RSS.2022.XVIII.036 Citation Details
Pacelli, Vincent and Majumdar, Anirudha "Robust Control Under Uncertainty via Bounded Rationality and Differential Privacy" IEEE International Conference on Robotics and Automation , 2022 https://doi.org/10.1109/ICRA46639.2022.9811557 Citation Details
Ren, Allen Z. and Majumdar, Anirudha "Distributionally Robust Policy Learning via Adversarial Environment Generation" IEEE Robotics and Automation Letters , v.7 , 2022 https://doi.org/10.1109/LRA.2021.3139949 Citation Details
(Showing: 1 - 10 of 11)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page