Award Abstract # 2038666
Collaborative Research: CPS: Medium: Spatio-Temporal Logics for Analyzing and Querying Perception Systems

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: ARIZONA STATE UNIVERSITY
Initial Amendment Date: September 4, 2020
Latest Amendment Date: April 14, 2022
Award Number: 2038666
Award Instrument: Standard Grant
Program Manager: Ralph Wachter
rwachter@nsf.gov
 (703)292-8950
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: January 1, 2021
End Date: December 31, 2024 (Estimated)
Total Intended Award Amount: $799,936.00
Total Awarded Amount to Date: $799,936.00
Funds Obligated to Date: FY 2020 = $799,936.00
History of Investigator:
  • Yezhou Yang (Principal Investigator)
    yz.yang@asu.edu
  • Georgios Fainekos (Former Principal Investigator)
  • Yezhou Yang (Former Co-Principal Investigator)
Recipient Sponsored Research Office: Arizona State University
660 S MILL AVENUE STE 204
TEMPE
AZ  US  85281-3670
(480)965-5479
Sponsor Congressional District: 04
Primary Place of Performance: Arizona State University
PO Box 876011
Tempe
AZ  US  85281-6011
Primary Place of Performance
Congressional District:
04
Unique Entity Identifier (UEI): NTLHJXM55KZ6
Parent UEI:
NSF Program(s): CPS-Cyber-Physical Systems
Primary Program Source: 01002021DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7924, 7918
Program Element Code(s): 791800
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

The goals of Automated Driving Systems (ADS) and Advanced Driver Assistance Systems (ADAS) include reduction in accidental deaths, enhanced mobility for differently abled people, and an overall improvement in the quality of life for the general public. Such systems typically operate in open and highly uncertain environments for which robust perception systems are essential. However, despite the tremendous theoretical and experimental progress in computer vision, machine learning, and sensor fusion, the form and conditions under which guarantees should be provided for perception components is still unclear. The state-of-the-art is to perform scenario-based evaluation of data against ground truth values, but this has only limited impact. The lack of formal metrics to analyze the quality of perception systems has already led to several catastrophic incidents and a plateau in ADS/ADAS development. This project develops formal languages for specifying and evaluating the quality and robustness of perception sub-systems within ADS and ADAS applications. To enable broader dissemination of this technology, the project develops graduate and undergraduate curricula to train engineers in the use of such methods, and new educational modules to explain the challenges in developing safe and robust ADS for outreach and public engagement activities. To broaden participation in computing, the investigators target the inclusion of undergraduate women in research and development phases through summer internships.

The formal language developed in this project is based on a new spatio-temporal logic pioneered by the investigators. This logic allows one to simultaneously perform temporal reasoning about streaming perception data, and spatially reason about objects both within a single frame of the data and across frames. The project also develops quantitative semantics for this logic, which provides the user with quantifiable quality metrics for perception sub-systems. These semantics enable comparisons between different perception systems and architectures. Crucially, the formal language facilitates the process of abstracting away implementation details, which in turn allows system designers and regulators to specify assumptions and guarantees for system performance at a higher-level of abstraction. An interesting benefit of this formal language is that it enables querying of databases with perception data for specific driving scenarios without the need for the highly manual process of creating ground truth annotations. Such a formal language currently does not exist, and this is a huge impediment to building a thriving marketplace for perception components used in safety-critical systems. This framework sets the foundation for a requirements language between suppliers of perception components and automotive companies. The open source and publicly available software tools developed in this project will assist with testing of perception systems by engineers and governmental agencies.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Cheng, S and Kong, D and Xie, J and Lee, K and Wu, Y and Yang, Y "Latent Space Energy-based Neural ODEs" Transactions on machine learning research , 2025 Citation Details
Buddareddygari, Prasanth and Zhang, Travis and Yang, Yezhou and Ren, Yi "Targeted Attack on Deep RL-based Autonomous Driving with Learned Visual Patterns" 2022 International Conference on Robotics and Automation (ICRA) , 2022 https://doi.org/10.1109/ICRA46639.2022.9811574 Citation Details
Balakrishnan, Anand and Deshmukh, Jyotirmoy and Hoxha, Bardh and Yamaguchi, Tomoya and Fainekos, Georgios "PerceMon: Online Monitoring for Perception Systems" Runtime Verification , v.12974 , 2021 https://doi.org/10.1007/978-3-030-88494-9_18 Citation Details
Anderson, Jacob and Hekmatnejad, Mohammad and Fainekos, Georgios "PyFoReL: A Domain-Specific Language for Formal Requirements in Temporal Logic" 2022 IEEE 30th International Requirements Engineering Conference (RE) , 2022 https://doi.org/10.1109/RE54965.2022.00037 Citation Details
Gokhale, Tejas "Covariate Shift Detection via Domain Interpolation Sensitivity" First Workshop on Interpolation Regularizers and Beyond at NeurIPS 2022 , 2022 Citation Details
Feinglass, Joshua and Yang, Yezhou "TROPE: TRaining-Free Object-Part Enhancement for Seamlessly Improving Fine-Grained Zero-Shot Image Captioning" , 2024 https://doi.org/10.18653/v1/2024.findings-emnlp.207 Citation Details
Feinglass, Joshua and Yang, Yezhou "Towards Addressing the Misalignment of Object Proposal Evaluation for Vision-Language Tasks via Semantic Grounding" Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2024 , 2024 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Overview: This CPS project focused on enhancing the reliability and robustness of perception systems by employing formal languages to specify and evaluate their quality. By developing new logics and computational frameworks, the research aimed to improve perception accuracy in applications such as autonomous driving and robotic systems. Throughout the project, significant progress was made in formalizing and applying spatio-temporal logics to real-world scenarios.


Intellectual Merit: The research introduced Spatio-Temporal Perception Logic (STPL) as an advanced method for evaluating and monitoring perception systems. This effort resulted in several major contributions, including the development of STPL as an extension of Timed Quality Temporal Logic (TQTL) with spatial operators, allowing for a more precise assessment of perception quality. Algorithmic advancements were made with the design of efficient offline and online monitoring algorithms for STPL, enabling real-time data stream analysis. The project demonstrated the practical application of these methods in both simulated and real-world driving environments, leading to improvements in object detection under adverse conditions. Another key achievement was the introduction of Latent Space Energy-based Neural ODEs (ODE-LEBM), a novel continuous-time sequence modeling technique that integrates an expressive energy-based prior to enhance perception models, improving both interpolation and extrapolation capabilities. The outcomes of this research were disseminated through high-impact journal publications, including papers presented at IJRR, WACV, EMNLP, and TMLR, which highlighted advancements in perception quality evaluation, object detection, and neural modeling.


Broader Impacts: The research significantly contributed to the advancement of autonomous systems by improving perception robustness, thereby enhancing safety in dynamic environments. In the educational sphere, the project supported multiple PhD students, two of whom successfully completed their doctorates and secured influential roles in industry and academia. Principal Investigator Yezhou Yang actively engaged in educational outreach by conducting courses at Arizona State University and delivering public lectures, including an invited talk on "AI and the Joy of Living" at the Mirabella invited talk series. Additionally, a high school intern was involved in research efforts, focusing on the analysis of perception metrics for AI-driven applications. The project also contributed to open-source development through the release of the PerceMon Python toolbox, which enables researchers and developers to apply STPL-based monitoring techniques in their work. Collaboration with Toyota further enriched the research, as the company provided robotic platforms for testing and validation.


Key Findings and Achievements: A fundamental contribution of this project was the formalization of STPL, which provided a mathematically rigorous approach to evaluating perception systems. A trust-aware control framework was developed, integrating trust-based monitoring into self-driving systems. The research led to improvements in object detection through the development of IA-YOLO, an enhanced object detection model that increases accuracy in challenging environments. The introduction of latent space ODE models significantly advanced continuous-time perception modeling by offering a novel energy-based neural ODE framework. These advancements were successfully implemented and tested in both simulated environments, such as CARLA, and real-world robotic systems, demonstrating their practical effectiveness.

Future Directions: The research conducted in this project establishes a foundation for further advancements in the field. Future work will focus on extending STPL to multi-modal perception, combining vision and LiDAR to improve environmental understanding. Additional efforts will be directed toward increasing AI interpretability through formal logics that explain and verify deep learning-based perception models. The project also paves the way for enhancing perception in dynamic environments by applying STPL techniques to complex multi-agent systems beyond autonomous driving. These contributions will continue to influence AI-driven perception, with applications spanning autonomous navigation, robotics, and intelligent transportation systems.


Last Modified: 03/23/2025
Modified by: Yezhou Yang

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page