Skip to feedback

Award Abstract # 2239301
CAREER: FLEXIBLE HIERARCHICAL ABSTRACTIONS FOR ACTIONABLE VISUAL PERCEPTION

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: TRUSTEES OF THE UNIVERSITY OF PENNSYLVANIA, THE
Initial Amendment Date: January 30, 2023
Latest Amendment Date: May 29, 2024
Award Number: 2239301
Award Instrument: Continuing Grant
Program Manager: Jie Yang
jyang@nsf.gov
 (703)292-4768
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: May 1, 2023
End Date: April 30, 2028 (Estimated)
Total Intended Award Amount: $600,000.00
Total Awarded Amount to Date: $246,505.00
Funds Obligated to Date: FY 2023 = $119,773.00
FY 2024 = $126,732.00
History of Investigator:
  • Dinesh Jayaraman (Principal Investigator)
    dineshj@seas.upenn.edu
Recipient Sponsored Research Office: University of Pennsylvania
3451 WALNUT ST STE 440A
PHILADELPHIA
PA  US  19104-6205
(215)898-7293
Sponsor Congressional District: 03
Primary Place of Performance: University of Pennsylvania
3330 Walnut Street
PHILADELPHIA
PA  US  19104-6228
Primary Place of Performance
Congressional District:
03
Unique Entity Identifier (UEI): GM1XX56LEP58
Parent UEI: GM1XX56LEP58
NSF Program(s): Robust Intelligence
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
01002425DB NSF RESEARCH & RELATED ACTIVIT

01002526DB NSF RESEARCH & RELATED ACTIVIT

01002627DB NSF RESEARCH & RELATED ACTIVIT

01002728DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7495
Program Element Code(s): 749500
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Humans excel at prioritizing and attending to different parts of the environment as the need arises. For example, a nurse feeding a patient might carefully consider individual morsels of food on a plate while picking them up, but afterwards devote his/her attention to the patient's mouth and facial expressions. By comparison, today's robots are rigid and inflexible in the ways that they observe, process, and interact with the world. They either focus on everything at the cost of becoming sluggish or instead cut corners indiscriminately and become prone to failures. This project will build innovative software technologies to allow general-purpose robots to flexibly adapt to task requirements, much like humans can, improving their agility and efficiency and making it easier for them to learn new tasks. This will enable applications of such robots not only in the household and hospital tasks which this project will use to develop the research, but also in many other socially relevant settings as farms, constructions sites, and small-scale manufacturing. Two graduate students and several undergraduate students will receive research training directly through this project. Further, this project will also draw from its research findings to improve graduate and undergraduate courses and summer outreach courses for high-school students.

This project explores the hypothesis that one key missing piece is agile and efficient perception-action loops that can evolve on-the-fly in response to task requirements. Consider a robot feeding a child food from a rice bowl. The locations and shapes of individual morsels of food, and the child's detailed pose and mouth configuration are all relevant over the course of the task and in each of the eating phases. Such time-varying task requirements are ubiquitous, yet they are poorly accounted for by the locked-in abstractions in today's standard computer vision algorithms and control loops. This project aims to advance visual recognition approaches in the context of embodied action, and to develop robot learning approaches in tandem that exploit these advances. It redesigns control loops to be flexible according to task demands, hierarchically factorized to permit novel compositions of the factor components, and self-learnable for scalability. To this end, it develops a staged but end-to-end differentiable control loop structure with interleaved task-specific, fine-tunable components and task-generic, reusable component stages. It further proposes novel active self-learning approaches that exploit the agent's own embodiment to teach it. These advances will enable improved sample efficiency for robot learning approaches as well as computational efficiency for task execution.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page