
NSF Org: |
CNS Division Of Computer and Network Systems |
Recipient: |
|
Initial Amendment Date: | March 23, 2023 |
Latest Amendment Date: | March 23, 2023 |
Award Number: | 2231257 |
Award Instrument: | Standard Grant |
Program Manager: |
Ralph Wachter
rwachter@nsf.gov (703)292-8950 CNS Division Of Computer and Network Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | April 1, 2023 |
End Date: | March 31, 2026 (Estimated) |
Total Intended Award Amount: | $413,694.00 |
Total Awarded Amount to Date: | $413,694.00 |
Funds Obligated to Date: |
|
History of Investigator: |
|
Recipient Sponsored Research Office: |
1 BROOKINGS DR SAINT LOUIS MO US 63130-4862 (314)747-4134 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
ONE BROOKINGS DR SAINT LOUIS MO US 63130-4862 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | CPS-Cyber-Physical Systems |
Primary Program Source: |
|
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
Deep Reinforcement Learning (RL) has emerged as prominent tool to control cyber-physical systems (CPS) with highly non-linear, stochastic, and unknown dynamics. Nevertheless, our current lack of understanding of when, how, and why RL works necessitates the need for new synthesis and analysis tools for safety-critical CPS driven by RL controllers; this is the main scope of this project. The primary focus of this research is on mobile robot systems. Such CPS are often driven by RL controllers due to their inherent complex - and possibly uncertain/unknown - dynamics, unknown exogenous disturbances, or the need for real-time decision making. Typically, RL-based control design methods are data inefficient, they cannot be safely transferred to new mission & safety requirements or new environments, while they often lack performance guarantees. This research aims to address these limitations resulting in a novel paradigm in safe autonomy for CPS with RL controllers. Wide availability of the developed autonomy methods can enable safety-critical applications for CPS with significant societal impact on, e.g., environmental monitoring, infrastructure inspection, autonomous driving, and healthcare. The broader impacts of this research include its educational agenda involving K-12, undergraduate and graduate level education.
To achieve the research goal of safe, efficient, and transferable RL, three tightly coupled research thrusts are pursued: (i) accelerated & safe reinforcement learning for temporal logic control objectives; (ii) safe transfer learning for temporal logic control objectives; (iii) compositional verification of temporal logic properties for CPS with NN controllers. The technical approach in these thrusts relies on tools drawn from formal methods, machine learning, and control theory and requires overcoming intellectual challenges related to integration of computation, control, and sensing. The developed autonomy methods will be validated and demonstrated on mobile aerial and ground robots in autonomous surveillance, delivery, and mobile manipulation tasks.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
Please report errors in award information by writing to: awardsearch@nsf.gov.