Award Abstract # 2147061
FAI: An Interpretable AI Framework for Care of Critically Ill Patients Involving Matching and Decision Trees

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: DUKE UNIVERSITY
Initial Amendment Date: March 7, 2022
Latest Amendment Date: March 7, 2022
Award Number: 2147061
Award Instrument: Standard Grant
Program Manager: Sylvia Spengler
sspengle@nsf.gov
 (703)292-7347
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: July 1, 2022
End Date: June 30, 2026 (Estimated)
Total Intended Award Amount: $625,000.00
Total Awarded Amount to Date: $625,000.00
Funds Obligated to Date: FY 2022 = $625,000.00
History of Investigator:
  • Cynthia Rudin (Principal Investigator)
    cynthia@cs.duke.edu
  • Sudeepa Roy (Co-Principal Investigator)
  • Alexander Volfovsky (Co-Principal Investigator)
Recipient Sponsored Research Office: Duke University
2200 W MAIN ST
DURHAM
NC  US  27705-4640
(919)684-3030
Sponsor Congressional District: 04
Primary Place of Performance: Duke University
2200 W. Main St, Suite 710
Durham
NC  US  27705-4010
Primary Place of Performance
Congressional District:
04
Unique Entity Identifier (UEI): TP7EK8DZV6N5
Parent UEI:
NSF Program(s): Fairness in Artificial Intelli,
Fairness in Artificial Intelli
Primary Program Source: 01002223DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 075Z
Program Element Code(s): 114y00, 114Y00
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

This project introduces a framework for interpretable, patient-centered causal inference and policy design for in-hospital patient care. This framework arose from a challenging problem, which is how to treat critically ill patients who are at risk for seizures (subclinical seizures) that can severely damage a patient's brain. In this high-stakes application of artificial intelligence, the data are complex, including noisy time-series, medical history, and demographic information. The goal is to produce interpretable causal estimates and policy decisions, allowing doctors to understand exactly how data were combined, permitting better troubleshooting, uncertainty quantification, and ultimately, trust. The core of the project's framework consists of novel and sophisticated matching techniques, which match each treated patient in the dataset with other (similar) patients who were not treated. Matching emulates a randomized controlled trial, allowing the effect of the treatment to be estimated for each patient, based on the outcomes from their matched group. A second important element of the framework involves interpretable policy design, where sparse decision trees will be used to identify interpretable subgroups of individuals who should receive similar treatments.


The matching techniques developed in this project will be within the new family of "almost-matching-exactly" (AME) techniques. AME techniques use machine learning on a training set to determine how to construct high-quality matched groups. In applying AME techniques to analyze seizure risk and treatment for critically ill patients, there are two major challenges that this project addresses: how to incorporate mechanistic models for drug absorption, and how to perform uncertainty quantification. Importantly, the project also addresses the release of AME code in several formats to be used by non-experts. The policy design aspect of the project involves the optimization of sparse decision trees. This project involves a close collaboration between experts in machine learning, causal inference, databases, and neurology, with the goal to improve patient care in high-stakes hospital settings where experiments cannot be conducted, and the only way to assess causal effects is through the analysis of observational data.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Donnelly, Jon and Katta, Srikar and Rudin, Cynthia and Browne, Edward P. "The Rashomon Importance Distribution: Getting RID of Unstable, Single Model-based Variable Importance" Advances in Neural Information Processing Systems , 2023 Citation Details
Ali Behrouz, Mathias Lecuyer "Fast Optimization of Weighted Sparse Decision Trees for use in Optimal Treatment Regimes and Optimal Policy Design" Workshop on Advances in Interpretable Machine Learning and Artificial Intelligence (AIMLAI) at CIKM, 2022. , 2022 Citation Details
, Harsh Parikh and , Cynthia Rudin and , Alexander Volfovsky "MALTS: Matching After Learning to Stretch" Journal of machine learning research , v.23 , 2022 Citation Details
Harsh Parikh, Quinn Lanners "Safe and Interpretable Estimation of Optimal Treatment Regimes" Proceedings of the International Conference on Artificial Intelligence and Statistics , 2024 Citation Details
Liu, Jiachang and Zhong, Chudi and Li, Boxuan and Seltzer, Margo and Rudin, Cynthia "FasterRisk: Fast and Accurate Interpretable Risk Scores" Advances in Neural Information Processing Systems , 2022 Citation Details
Parikh, Harsh and Hoffman, Kentaro and Sun, Haoqi and Zafar, Sahar F and Ge, Wendong and Jing, Jin and Liu, Lin and Sun, Jimeng and Struck, Aaron and Volfovsky, Alexander and Rudin, Cynthia and Westover, M Brandon "Effects of epileptiform activity on discharge outcome in critically ill patients in the USA: a retrospective cross-sectional study" The Lancet Digital Health , v.5 , 2023 https://doi.org/10.1016/S2589-7500(23)00088-2 Citation Details
Quinn Lanners, Harsh Parikh "From Feature Importance to Distance Metric: An Almost Exact Matching Approach for Causal Inference" Uncertainty in artificial intelligence , 2023 Citation Details
Seale-Carlisle, Travis and Jain, Saksham and Lee, Courtney and Levenson, Caroline and Ramprasad, Swathi and Garrett, Brandon and Roy, Sudeepa and Rudin, Cynthia and Volfovsky, Alexander "Evaluating Pre-trial Programs Using Interpretable Machine Learning Matching Algorithms for Causal Inference" Proceedings of the AAAI Conference on Artificial Intelligence , v.38 , 2024 https://doi.org/10.1609/aaai.v38i20.30239 Citation Details
Srikar Katta, Harsh Parikh "Interpretable Causal Inference for Analyzing Wearable, Sensor, and Distributional Data" Proceedings of the International Conference on Artificial Intelligence and Statistics , 2024 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page