
NSF Org: |
IIS Division of Information & Intelligent Systems |
Recipient: |
|
Initial Amendment Date: | April 22, 2014 |
Latest Amendment Date: | April 30, 2018 |
Award Number: | 1350337 |
Award Instrument: | Continuing Grant |
Program Manager: |
Sylvia Spengler
sspengle@nsf.gov (703)292-7347 IIS Division of Information & Intelligent Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | May 1, 2014 |
End Date: | December 31, 2020 (Estimated) |
Total Intended Award Amount: | $549,863.00 |
Total Awarded Amount to Date: | $549,863.00 |
Funds Obligated to Date: |
FY 2015 = $110,508.00 FY 2016 = $114,926.00 FY 2017 = $94,091.00 FY 2018 = $97,855.00 |
History of Investigator: |
|
Recipient Sponsored Research Office: |
10 W 35TH ST CHICAGO IL US 60616-3717 (312)567-3035 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
10 W 31st St Chicago IL US 60616-3729 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | Info Integration & Informatics |
Primary Program Source: |
01001516DB NSF RESEARCH & RELATED ACTIVIT 01001617DB NSF RESEARCH & RELATED ACTIVIT 01001718DB NSF RESEARCH & RELATED ACTIVIT 01001819DB NSF RESEARCH & RELATED ACTIVIT |
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
Machine learning models are trained on data that are annotated (labeled) by humans. The accuracy of the trained models generally improves with the number of annotated data examples. Yet, annotating takes time, money, and effort. Active learning aims to minimize the costs by determining which exemples are most informative and directing the human labeler to them. Improvements in active learning will lower the costs associated with data annotation and lead to faster implementations of intelligent systems for a range of applications including robotics, speech technology, error and anomaly detection (for example in medicine, financial fraud, and condition-based maintenance of infrastructure), targeted advertising, human-computer interfaces, and bioinformatics.
In traditional active learning approaches, algorithms are limited in the types of information they can acquire, and they often do not provide any rationale to the user as to why a particular exemplar is chosen for annotation. This CAREER project develops a new paradigm dubbed "rich and transparent active learning." This new paradigm opens a communication channel between algorithms and users whereby they can exchange a rich set of queries, answers, and explanations. By using rich feedback from users the algorithms will be able to learn the target concept more economically, reducing the resources required to build an accurate predictive model. By explaining their reasoning, these algorithms will achieve transparency, build trust, and open themselves to scrutiny.
Towards that end, the project develops methods that allow algorithms to use a rich set of queries for resource-efficient model training, and generate explanations that are informative but not overwhelming for the users. The methods developed build on expected loss minimization, information theory, and principles from human-computer interaction. Approaches are evaluated using publicly available datasets and user studies carried out as part of the project. The project develops case studies on two high-impact real-world problems: detecting fraudulent health-care claims, and identifying patients at risk of disease.
The rich and transparent active learning paradigm provides unique educational opportunities. In contrast to standard machine learning algorithms, operated as black boxes, interactive and transparent machine learning is expected to raise students' interest and motivation for data science. Two PhD and several undergraduate and high school students are being trained under this award. A new graduate course on interactive machine learning is being developed. Finally the PI ensures effective outreach to under-represented groups by partnering with a Chicago public high school whose student population includes 90% minorities.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
Predictive models are utilized for a number of applications, including medical diagnosis, traffic light and sign recognition by self-driving cars, face detection and recognition, speech recognition, and more. Predictive models are typically built using training datasets that are annotated by us humans. In simple terms, the bigger and the higher quality the data, the more accurate the models are. However, curating high-quality and high-quantity datasets are costly; they require human time, effort, laboratory tests, experiments, and simulations.
This project developed algorithms that enabled a rich and transparent set of interactions between the predictive models and the human annotators. The predictive models have been equipped with the ability to ask a rich set of questions to the human annotators during the curation of a training dataset. With the help of the new rich set of queries, the models improved their learning efficiency significantly, reducing the time and costs for curation of high-quality datasets.
Moreover, the predictive models have been equipped with the ability to generate explanations regarding their rich set of questions to the human annotators. When the model explained its reasons for asking a question (for e.g., explaining why it is uncertain about a particular case), the human annotators were able to provide more targeted answers and rationales, further increasing learning efficiency.
The research resulted in numerous publications in several AI conferences and journals. Two PhD theses by two female students were fully funded by this research. Moreover, several undergraduate students received training through this award. The research by the PhD and undergraduate students resulted in several publications and open-source code and data that are available at http://www.cs.iit.edu/~ml/. A new course titled Interactive and Transparent Machine Learning has been developed and added to the curriculum.
Last Modified: 07/27/2021
Modified by: Mustafa Bilgic
Please report errors in award information by writing to: awardsearch@nsf.gov.