Skip to feedback

Award Abstract # 2405103
NRI: Hierarchical Representation Learning for Robot Assistants

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: THE LELAND STANFORD JUNIOR UNIVERSITY
Initial Amendment Date: November 24, 2023
Latest Amendment Date: December 13, 2023
Award Number: 2405103
Award Instrument: Standard Grant
Program Manager: Jie Yang
jyang@nsf.gov
 (703)292-4768
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: October 1, 2023
End Date: January 31, 2027 (Estimated)
Total Intended Award Amount: $1,500,000.00
Total Awarded Amount to Date: $1,186,632.00
Funds Obligated to Date: FY 2021 = $1,175,197.00
FY 2022 = $11,435.00
History of Investigator:
  • Shuran Song (Principal Investigator)
    shuran@stanford.edu
Recipient Sponsored Research Office: Stanford University
450 JANE STANFORD WAY
STANFORD
CA  US  94305-2004
(650)723-2300
Sponsor Congressional District: 16
Primary Place of Performance: Stanford University
450 JANE STANFORD WAY
STANFORD
CA  US  94305-2004
Primary Place of Performance
Congressional District:
16
Unique Entity Identifier (UEI): HJD6G4D6TJY5
Parent UEI:
NSF Program(s): IIS Special Projects,
NRI-National Robotics Initiati
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
01002223DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 8086, 9251
Program Element Code(s): 748400, 801300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

More than eighteen million people in North America have a physical disability due to limited mobility, restricting their independence, lifestyle, and ability to perform daily activities. One in five older adults struggle with mobility, millions of people with limited mobility are veterans, and a significant number of people have limited mobility because of diseases and accidents. Due to recent advances in artificial intelligence, robots hold great promise to provide timely assistance to people with disabilities, and drive improvements to their quality of life, independence, and productivity. This project introduces an automated robot assistant that is able to recognize a person?s goal, and provide them the right object at the right time, thereby helping people perform complex activities, such as cooking, object repair, and housekeeping. From both sight and dialogue, the research products will be able to anticipate what objects a person will need in the near future, and deliver it at exactly the right moment. Furthermore, the project will generate new educational opportunities at the intersection of robotics, computer vision and natural language processing through a series of systematically designed curriculum and annual capstone projects for assistive robotics. Due to the tight integration of multiple disciplines and the large practical impact, these educational programs will serve as an excellent platform for training the next generation of roboticists and increasing the diversity in the field.

This research project introduces a novel hierarchical representation learning framework for assistive robots, which serves as a common interface to drive integration between robotics, computer vision, and natural language understanding. The project includes three thrusts. First, the team will develop hierarchical task representations. Second, human intention prediction and verification will developed. The final thrust will address intention-aware planning. Unlike established state representations in robotics, the new representation leverages non-Euclidean geometry, such as hyperbolic manifolds. Since hyperbolic space is a continuous analog of a tree, it provides new opportunities for learning task hierarchies from large-scale unlabeled instructional videos. This task representation is able to anticipate the activities of people, steer dialogue to reduce uncertainty, and provide dense rewards for long-horizon planning. This representation is learned from large-scale unlabeled instructional videos, making this approach flexible and adaptable to the many real-world applications of just-in-time object delivery for people with limited mobility.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Chi, C and Xu, Z and Pan, C and Cousineau, E and Burchfiel, B and Feng, S and Tedrake, R and Song, S "Universal Manipulation Interface:In-The-Wild Robot Teaching Without In-The-Wild Robots" , 2024 Citation Details
Chi, Cheng and Xu, Zhenjia and Pan, Chuer and Cousineau, Eric and Burchfiel, Benjamin and Feng, Siyuan and Tedrake, Russ and Song, Shuran "Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots" , 2024 Citation Details
Hayamizu, Yohei and Yu, Zhou and Zhang, Shiqi "Learning Joint Policies for Human-Robot Dialog and Co-Navigation" , 2023 https://doi.org/10.1109/IROS55552.2023.10341663 Citation Details
Liang, J and Liu, R and Ozguroglu, E and Sudhakar, S and Dave, A and Tokmakov, P and Song, S and Vondrick, C "Dreamitate:Real-World Visuomotor Policy Learning via Video Generation" , 2024 Citation Details
Liu, Z and Chi, C and Cousineau, E and Kuppuswamy, N and Burchfiel, B and Song, S "ManiWAV: Learning RobotManipulation from In-the-Wild Audio-Visual Data" , 2024 Citation Details
Mandi, Zhao and Jain, Shreeya and Song, Shuran "RoCo: Dialectic Multi-Robot Collaboration with Large Language Models" , 2024 https://doi.org/10.1109/ICRA57147.2024.10610855 Citation Details
Maximillian Chen, Zhou Yu "Pre-Finetuning for Few-Shot Emotional Speech Recognition" INTERSPEECH , 2023 Citation Details
Purva Tendulkar, Dídac Surís "FLEX: Full-Body Grasping Without Full-Body Grasps" CVPR , 2023 Citation Details
Wu, Jimmy and Antonova, Rika and Kan, Adam and Lepert, Marion and Zeng, Andy and Song, Shuran and Bohg, Jeannette and Rusinkiewicz, Szymon and Funkhouser, Thomas "TidyBot: personalized robot assistance with large language models" Autonomous Robots , v.47 , 2023 https://doi.org/10.1007/s10514-023-10139-z Citation Details
Zeyi Liu Arpit Bahety Shuran Song "REFLECT: Summarizing Robot Experiences for FaiLure Explanation and CorrecTion" Conference of Robot Learning , 2023 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page