Award Abstract # 1925403
NRI: FND: Scalable and Customizable Intent Inference and Motion Planning for Socially-Adept Autonomous Vehicles

NSF Org: CMMI
Division of Civil, Mechanical, and Manufacturing Innovation
Recipient: ARIZONA STATE UNIVERSITY
Initial Amendment Date: August 20, 2019
Latest Amendment Date: May 19, 2021
Award Number: 1925403
Award Instrument: Standard Grant
Program Manager: Jordan Berg
jberg@nsf.gov
 (703)292-5365
CMMI
 Division of Civil, Mechanical, and Manufacturing Innovation
ENG
 Directorate for Engineering
Start Date: September 1, 2019
End Date: August 31, 2023 (Estimated)
Total Intended Award Amount: $750,000.00
Total Awarded Amount to Date: $782,000.00
Funds Obligated to Date: FY 2019 = $750,000.00
FY 2020 = $16,000.00

FY 2021 = $16,000.00
History of Investigator:
  • Wenlong Zhang (Principal Investigator)
    Wenlong.Zhang@asu.edu
  • Yi Ren (Co-Principal Investigator)
  • Yezhou Yang (Co-Principal Investigator)
Recipient Sponsored Research Office: Arizona State University
660 S MILL AVENUE STE 204
TEMPE
AZ  US  85281-3670
(480)965-5479
Sponsor Congressional District: 04
Primary Place of Performance: Arizona State University
AZ  US  85287-6011
Primary Place of Performance
Congressional District:
04
Unique Entity Identifier (UEI): NTLHJXM55KZ6
Parent UEI:
NSF Program(s): Special Initiatives,
NRI-National Robotics Initiati
Primary Program Source: 01001920DB NSF RESEARCH & RELATED ACTIVIT
01002021DB NSF RESEARCH & RELATED ACTIVIT

01002122DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 8086, 9251, 9178, 9231, 116E
Program Element Code(s): 164200, 801300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.041

ABSTRACT

This project will promote the progress of science, and advance the national prosperity and safety, by tackling an important and challenging problem for autonomous vehicles (AVs): interaction of AVs with human-driven vehicles. Currently, there is a lack of theory that allows an autonomous vehicle to interact with multiple surrounding vehicles in a safe and socially-adept manner. This National Robotics Initiative (NRI) project will address this critical need by developing a novel algorithm framework for an autonomous vehicle to be able to anticipate other vehicles, behavior and customize its motion according to the local driving culture. This project serves the national interests by advancing knowledge in the fields of control engineering, machine learning, and cognitive science. The project will also make an important step in making widely adopted autonomous vehicles a reality, which promises to increase transportation system efficiency and safety. Project results will be disseminated through a project website, open-source simulation software, and public datasets. The impacts of this project will be broadened through various educational activities, including a new class on collaborative autonomous driving, undergraduate research projects, and outreach to the local community through lab tours.

This project aims at answering two fundamental research questions: 1) what formalisms of intent inference and motion planning are capable of creating socially-adept motions, and 2) what embodiment of these formalisms can achieve scalability for multi-vehicle interactions and customizability for changing driving cultures? To answer these questions, the research team will pursue the following three objectives. First, the project team will build a Bayesian game model to represent vehicle interactions, and develop mechanistic intent inference and motion planning policies. Second, a message passing neural network will be developed to enable scalable intent inference and motion prediction of multiple surrounding vehicles. Third, a social attention mechanism will be developed that allows an autonomous vehicle to actively prioritize its various control considerations, e.g., safety and courtesy towards the surrounding vehicles. The developed algorithms will be validated with real-world driving scenarios such as intersections and highways in a driving simulator designed through collaboration with autonomous vehicle manufacturers and research institutes.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Amatya, Sunny and Ghimire, Mukesh and Ren, Yi and Xu, Zhe and Zhang, Wenlong "When Shall I Estimate Your Intent? Costs and Benefits of Intent Inference in Multi-Agent Interactions" 2022 American Control Conference (ACC) , 2022 https://doi.org/10.23919/ACC53348.2022.9867155 Citation Details
Buddareddygari, Prasanth and Zhang, Travis and Yang, Yezhou and Ren, Yi "Targeted Attack on Deep RL-based Autonomous Driving with Learned Visual Patterns" 2022 International Conference on Robotics and Automation (ICRA) , 2022 https://doi.org/10.1109/ICRA46639.2022.9811574 Citation Details
Chen, Yi and Zhang, Lei and Merry, Tanner and Amatya, Sunny and Zhang, Wenlong and Ren, Yi "When Shall I Be Empathetic? The Utility of Empathetic Parameter Estimation in Multi-Agent Interactions" 2021 IEEE International Conference on Robotics and Automation (ICRA) , 2021 https://doi.org/10.1109/ICRA48506.2021.9561079 Citation Details
Fang, Zhiyuan and Wang, Jianfeng and Hu, Xiaowei and Liang, Lin and Gan, Zhe and Wang, Lijuan and Yang, Yezhou and "Injecting Semantic Concepts Into End-to-End Image Captioning" Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 2022 Citation Details
Gunasekar, Kausic and Qiu, Qiang and Yang, Yezhou "Low to High Dimensional Modality Hallucination Using Aggregated Fields of View" IEEE Robotics and Automation Letters , v.5 , 2020 https://doi.org/10.1109/LRA.2020.2970679 Citation Details
Wang, Yiwei and Ren, Yi and Elliott, Steven and Zhang, Wenlong "Enabling Courteous Vehicle Interactions through Game-based and Dynamics-aware Intent Inference" IEEE Transactions on Intelligent Vehicles , v.5 , 2020 10.1109/TIV.2019.2955897 Citation Details
Zhang, Lei and Ghimire, Mukesh and Zhang, Wenlong and Xu, Zhe and Ren, Yi "Approximating Discontinuous Nash Equilibrial Values of Two-Player General-Sum Differential Games" 2023 IEEE International Conference on Robotics and Automation (ICRA) , 2023 https://doi.org/10.1109/ICRA48891.2023.10160219 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

This project studied the intent estimation, planning and control, and perception problems for an autonomous vehicle (AV) to interact with human-driven vehicles (HV) in a safe and socially adept manner. The research team ("we" thereafter) formulated the AV-HV interaction problem as an incomplete-information differential game, explored learning methods to solve incomplete-information differential game, and developed semantically driven and robust sensing techniques for vehicle sensing.

In an incomplete-information differential game setup, the AV and HV estimate the intent of each other simultaneously. We developed a novel intent inference scheme, called empathetic intent inference, which explicitly considers the mutual intent inference between the AV and HV. Simulation studies showed that the new intent inference algorithm allowed the AV to dynamically negotiate with the HV like another human driver, which improved the safety and efficiency of future transportation. We further studied the utility of the proposed empathetic intent inference algorithm, and results showed that empathy helped agents choose policies that led to higher social values when agents had common beliefs inconsistent with the ground truth. Considering the computational intensity for empathetic intent inference, we developed an intermittent intent inference algorithm triggered by reinforcement learning. Using an uncontrolled intersection case, we showed that the proposed new intent inference mechanism led to safe interactions between an AV and HV while significantly reducing the computation load.  To study corner cases such as unprotected left turns and uncontrolled intersections where traffic accidents often happen due to misunderstanding between drivers, we focused on extending the game-theoretic framework for robotics applications in two important directions.

First, we studied two-player zero-sum differential games with both incomplete information and state constraints. Our study was the first to prove value existence for such games. As a byproduct, the proof led to the subdynamic programming principles that governed the behavioral strategies of informed and uninformed players. In comparison with studies on imperfect-information games, our theoretical development revealed unique structures of values and strategies for games where the incomplete information is on player payoff types.

Second, we studied the open challenge in approximating values for complete-information differential games with Markov payoffs and state constraints. For games either without constraints or without continuous payoffs (e.g., reach-avoid), the curse of dimensionality can be alleviated through Monte Carlo methods such as physics-informed neural networks (PINN). When both payoffs and state constraints are present, the value of the game becomes discontinuous in which case standard PINN fails to converge. To this end, we developed a Lagrangian PINN that forced explicit learning of discontinuous co-states along with the value. We showed that our method led to significant improvement in value generalization and safety performance compared with existing PINNs for Lipschitz-continuous partial differential equations (PDEs).

For situation awareness sensing research thrust, in our project's first stage, we leveraged traffic monitoring cameras as integral tools for traffic management and autonomous driving systems. We developed CAROM, or "CARs On the Map," a framework for vehicle localization and traffic scene reconstruction. This framework processed traffic monitoring videos to generate anonymous data structures detailing vehicle type, 3D shape, position, and velocity, facilitating accurate scene reconstruction and replay. In the second stage, we aimed to enhance sensing models without relying on detectors. Introducing ViTCAP, a pure vision transformer-based model utilizing grid representations, we incorporated a Concept Token Network (CTN) to predict semantic concepts and improve end-to-end sensing performance. CTN, based on a vision transformer, effectively predicted concept tokens, enriching the sensing process with valuable semantic information. In the last stage, we addressed visual feature selection crucial for Visual Odometry (VO) and Simultaneous Localization and Mapping (SLAM). Our novel self-supervised VO method, VOCAL (Visual Odometry via ContrAstive Learning), intelligently selected and normalized features using contrastive learning based on associated 3D camera motions. VOCAL eliminated the need for intricate feature extraction and preprocessing steps, demonstrating robust and generalizable performance in experiments on the KITTI challenge dataset compared to baseline methods. VOCAL presents a promising alternative pathway for achieving accurate VO through self-supervised contrastive learning methodologies.

This project has generated positive societal impacts by improving the safety and intelligence of AVs when operating around human-driven vehicles. The research team have disseminated their research findings in various workshops and presentations in control, robotics, and computer vision conferences, and they have shared their vision on future AVs and transportation in outreach events (e.g., Phoenix Mobile Festival) to local communities. Results on vehicle perception from this project led to a new start-up company by co-PI Yang. All the publications resulting from this project have been made available in the NSF Public Access Repository.

Seven Ph.D. students and one M.S. student have been involved and at least partially supported by this project, and one Ph.D. graduate is now working in academia. Eight undergraduate students have participated in this project, including five as Research Experience for Undergraduates (REU) trainees.


Last Modified: 03/12/2024
Modified by: Wenlong Zhang

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page