Award Abstract # 1749204
CAREER: Safe and Efficient Robot Learning from Demonstration in the Real World

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: UNIVERSITY OF TEXAS AT AUSTIN
Initial Amendment Date: March 23, 2018
Latest Amendment Date: June 1, 2020
Award Number: 1749204
Award Instrument: Continuing Grant
Program Manager: Juan Wachs
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: June 1, 2018
End Date: May 31, 2023 (Estimated)
Total Intended Award Amount: $524,605.00
Total Awarded Amount to Date: $524,605.00
Funds Obligated to Date: FY 2018 = $137,302.00
FY 2019 = $141,013.00

FY 2020 = $24,412.00
History of Investigator:
  • Scott Niekum (Principal Investigator)
    sniekum@cs.umass.edu
Recipient Sponsored Research Office: University of Texas at Austin
110 INNER CAMPUS DR
AUSTIN
TX  US  78712-1139
(512)471-6424
Sponsor Congressional District: 25
Primary Place of Performance: University of Texas at Austin
2317 Speedway
Austin
TX  US  78712-1757
Primary Place of Performance
Congressional District:
25
Unique Entity Identifier (UEI): V6AFQPN18437
Parent UEI:
NSF Program(s): Robust Intelligence
Primary Program Source: 01001819DB NSF RESEARCH & RELATED ACTIVIT
01001920DB NSF RESEARCH & RELATED ACTIVIT

01002021DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7495, 1045
Program Element Code(s): 749500
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

General purpose robots are poised to enter the home and workplace in unprecedented numbers in coming years, but face the significant challenge of customization - the ability to perform user-specified tasks in many different unstructured environments. In response to this need, robot learning from demonstration (LfD) has emerged as a paradigm that allows users to quickly and naturally program robots by simply showing them how to perform a task, rather than by writing code. This methodology aims to allow non-expert users to program robots, as well as communicate embodied knowledge that is difficult to translate into formal code. However, current state-of-the-art LfD algorithms are not yet ready for widespread deployment, as they are often unreliable, need too much data, and are designed to learn in a single session in a laboratory setting. This work addresses these issues to help enable future robots to perform important tasks ranging from in-home elderly care to reconfigurable manufacturing.

Specifically, this work identifies three significant technical improvements to current LfD algorithms that are needed before they can be deployed in the real world: the need for safety guarantees, the ability to learn from very limited amounts of data, and the ability to continually improve in an ongoing, life-long fashion. A formal theory of safe LfD is developed, along with practical algorithms that provide strong probabilistic lower bounds on agent performance. Algorithmic efficiency is addressed via a re-examining of common statistical assumptions (such as independent and identically distributed data) and through the use of multimodal side-information, such as natural language and gaze. Finally, active learning strategies and modeling of human beliefs are used to enable interactive, continual learning.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 33)
Brown, D and Goo, W and Niekum, S "Better-than-Demonstrator Imitation Learning via Automatically-Ranked Demonstrations" Conference on Robot Learning , 2019 Citation Details
Brown, D and Niekum, S and Petrik, M "Bayesian Robust Optimization for Imitation Learning" Neural Information Processing Systems , 2020 https://doi.org/ Citation Details
Brown, Daniel and Cui, Yuchen and Niekum, Scott "Risk-Aware Active Inverse Reinforcement Learning" Conference on Robot Learning , 2018 Citation Details
Brown, D and Coleman, R and Srinivasan, R and Niekum, S "Safe Imitation Learning via Fast Bayesian Reward Inference from Preferences" International Conference on Machine Learning , 2020 Citation Details
Brown, Daniel and Goo, Wonjoon and Nagarajan, Prabhat and Niekum, Scott "Extrapolating Beyond Suboptimal Demonstrations via Inverse Reinforcement Learning from Observations" International Conference on Machine Learning , 2019 Citation Details
Brown, Daniel and Niekum, Scott "Machine Teaching for Inverse Reinforcement Learning: Algorithms and Applications" Proceedings of the ... AAAI Conference on Artificial Intelligence , 2019 Citation Details
Brown, Daniel and Niekum, Scott "Machine Teaching for Inverse Reinforcement Learning:Algorithms and Applications" Proceedings of the ... AAAI Conference on Artificial Intelligence , 2019 Citation Details
Brown, Daniel S and Schneider, Jordan and Dragan, Anca and Niekum, Scott "Value Alignment Verification" International Conference on Machine Learning , 2021 Citation Details
Chandak, Y and Niekum, S and Castro da Silva, B and Learned-Miller, E and Brunskill, E and Thomas, P "Universal Off-Policy Evaluation" Neural Information Processing Systems , 2021 Citation Details
Cui, Y and Koppol, P and Admoni, H and Niekum, S and Simmons, R and Steinfeld, A and Fitzgerald, T "Understanding the Relationship between Interactions and Outcomes in Human-in-the-Loop Machine Learning" International Joint Conference on Artificial Intelligence , 2021 https://doi.org/10.24963/ijcai.2021/599 Citation Details
Cui, Y and Zhang, Q and Allievi, A and Stone, P and Niekum, S and Knox, W "The EMPATHIC Framework for Task Learning from Implicit Human Feedback" Conference on Robot Learning , 2020 https://doi.org/ Citation Details
(Showing: 1 - 10 of 33)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page