Skip to feedback

Award Abstract # 2231257
CPS: SMALL: Formal Methods for Safe, Efficient, and Transferable Learning-enabled Autonomy

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: WASHINGTON UNIVERSITY, THE
Initial Amendment Date: March 23, 2023
Latest Amendment Date: March 23, 2023
Award Number: 2231257
Award Instrument: Standard Grant
Program Manager: Ralph Wachter
rwachter@nsf.gov
 (703)292-8950
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: April 1, 2023
End Date: March 31, 2026 (Estimated)
Total Intended Award Amount: $413,694.00
Total Awarded Amount to Date: $413,694.00
Funds Obligated to Date: FY 2023 = $413,694.00
History of Investigator:
  • Ioannis Kantaros (Principal Investigator)
    ioannisk@wustl.edu
Recipient Sponsored Research Office: Washington University
1 BROOKINGS DR
SAINT LOUIS
MO  US  63130-4862
(314)747-4134
Sponsor Congressional District: 01
Primary Place of Performance: Washington University
ONE BROOKINGS DR
SAINT LOUIS
MO  US  63130-4862
Primary Place of Performance
Congressional District:
01
Unique Entity Identifier (UEI): L6NFUM28LQM5
Parent UEI:
NSF Program(s): CPS-Cyber-Physical Systems
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7918, 7923
Program Element Code(s): 791800
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Deep Reinforcement Learning (RL) has emerged as prominent tool to control cyber-physical systems (CPS) with highly non-linear, stochastic, and unknown dynamics. Nevertheless, our current lack of understanding of when, how, and why RL works necessitates the need for new synthesis and analysis tools for safety-critical CPS driven by RL controllers; this is the main scope of this project. The primary focus of this research is on mobile robot systems. Such CPS are often driven by RL controllers due to their inherent complex - and possibly uncertain/unknown - dynamics, unknown exogenous disturbances, or the need for real-time decision making. Typically, RL-based control design methods are data inefficient, they cannot be safely transferred to new mission & safety requirements or new environments, while they often lack performance guarantees. This research aims to address these limitations resulting in a novel paradigm in safe autonomy for CPS with RL controllers. Wide availability of the developed autonomy methods can enable safety-critical applications for CPS with significant societal impact on, e.g., environmental monitoring, infrastructure inspection, autonomous driving, and healthcare. The broader impacts of this research include its educational agenda involving K-12, undergraduate and graduate level education.

To achieve the research goal of safe, efficient, and transferable RL, three tightly coupled research thrusts are pursued: (i) accelerated & safe reinforcement learning for temporal logic control objectives; (ii) safe transfer learning for temporal logic control objectives; (iii) compositional verification of temporal logic properties for CPS with NN controllers. The technical approach in these thrusts relies on tools drawn from formal methods, machine learning, and control theory and requires overcoming intellectual challenges related to integration of computation, control, and sensing. The developed autonomy methods will be validated and demonstrated on mobile aerial and ground robots in autonomous surveillance, delivery, and mobile manipulation tasks.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Wang, Jun and He, Guocheng and Kantaros, Yiannis "Probabilistically Correct Language-Based Multi-Robot Planning Using Conformal Prediction" IEEE Robotics and Automation Letters , v.10 , 2025 https://doi.org/10.1109/LRA.2024.3504233 Citation Details
Kantaros, Yiannis and Wang, Jun "Sample-Efficient Reinforcement Learning with Temporal Logic Objectives: Leveraging the Task Specification to Guide Exploration" IEEE Transactions on Automatic Control , 2024 https://doi.org/10.1109/TAC.2024.3484290 Citation Details
Mitta, Rohan and Hasanbeig, Hosein and Wang, Jun and Kroening, Daniel and Kantaros, Yiannis and Abate, Alessandro "Safeguarded Progress in Reinforcement Learning: Safe Bayesian Exploration for Control Policy Synthesis" , 2024 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page