Award Abstract # 2302968
Collaborative Research: NSF-CSIRO: HCC: Small: Understanding Bias in AI Models for the Prediction of Infectious Disease Spread

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: EMORY UNIVERSITY
Initial Amendment Date: February 10, 2023
Latest Amendment Date: February 10, 2023
Award Number: 2302968
Award Instrument: Standard Grant
Program Manager: Hector Munoz-Avila
hmunoz@nsf.gov
 (703)292-4481
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: April 1, 2023
End Date: March 31, 2026 (Estimated)
Total Intended Award Amount: $373,667.00
Total Awarded Amount to Date: $373,667.00
Funds Obligated to Date: FY 2023 = $373,667.00
History of Investigator:
  • Andreas Zuefle (Principal Investigator)
    azufle@emory.edu
  • Li Xiong (Co-Principal Investigator)
Recipient Sponsored Research Office: Emory University
201 DOWMAN DR NE
ATLANTA
GA  US  30322-1061
(404)727-2503
Sponsor Congressional District: 05
Primary Place of Performance: Emory University
201 DOWMAN DR
ATLANTA
GA  US  30322-1007
Primary Place of Performance
Congressional District:
05
Unique Entity Identifier (UEI): S352L5PJLMP8
Parent UEI:
NSF Program(s): Info Integration & Informatics
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7923, 7364
Program Element Code(s): 736400
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Artificial intelligence (AI) provides powerful techniques for understanding and prediction of complex systems such as modeling and predicting the spread of infectious diseases. Despite this, these predictive capabilities are rarely adopted by public health decision-makers to support policy interventions. One of the issues preventing their adoption is that AI methods are known to amplify the bias in the data they are trained on. This is especially problematic in infectious disease models which leverage available large and inherently biased spatiotemporal data. These biases may propagate through the modeling pipeline to decision-making, resulting in inequitable and ineffective policy interventions. This project investigates how the AI disease modeling pipeline can lead from biased data to biased predictions and to derive solutions that mitigate this bias in three aims: 1) creating an AI system to predict the spread of emerging infectious diseases in space and time, 2) simulating a population from which we will collect data often used as input for AI systems in a way that the bias is controlled, and 3) exploring links between bias in the collected data and the resulting bias in the AI model and deriving solutions for their mitigation. The project will enable AI-driven infectious disease models and predictions that will support fair and equitable decision-making and interventions. The project will enrich education and training related to ethical AI practices and will support professional development opportunities for early-career researchers, graduate, undergraduate, and high school students in the United States and Australia.

In Aim 1, the team of researchers will use a self-supervised contrastive learning approach that uses mobility prediction as a pre-text task to learn representations of spatial regions. These representations can be used for infectious disease spread prediction given only very little infectious disease ground truth data. The investigators hypothesize that such a model is susceptible to data bias. Thus, in Aim 2, the team of researchers will leverage a large-scale agent-based simulation that will serve as a sandbox world for which we have perfect knowledge of and from which we can collect data and inject various types of bias. For Aim 3, the team of researchers will investigate how different types of simulated data bias leads to biased AI predictions by leveraging different metrics of fairness in AI and studying how these fairness measures can be incorporated into the AI optimization procedure to mitigate bias. By understanding, measuring, and mitigating bias inherent to traditional AI solutions, the project will enable accurate, scalable, and rapid predictions to support fair and equitable decision-making for pandemic prevention.

This is a joint project between researchers in the United States and Australia funded by the Collaboration Opportunities in Responsible and Equitable AI under the U.S. NSF and the Australian Commonwealth Scientific and Industrial Research Organization (CSIRO).

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 13)
Amiri, Hossein and Kong, Ruochen and Züfle, Andreas "Urban Anomalies: A Simulated Human Mobility Dataset with Injected Anomalies" , 2024 https://doi.org/10.1145/3681765.3698459 Citation Details
Kato, Fumiyuki and Xiong, Li and Takagi, Shun and Cao, Yang and Yoshikawa, Masatoshi "Uldp-FL: Federated Learning with Across-Silo User-Level Differential Privacy" Proceedings of the VLDB Endowment , v.17 , 2024 https://doi.org/10.14778/3681954.3681966 Citation Details
Kong, Ruochen and Anderson, Taylor and Heslop, David and Zufle, Andreas "An Infectious Disease Spread Simulation to Control Data Bias" , 2024 https://doi.org/10.1145/3678717.3691293 Citation Details
Liu, Junxu and Lou, Jian and Xiong, Li and Liu, Jinfei and Meng, Xiaofeng "Cross-silo Federated Learning with Record-level Personalized Differential Privacy" , 2024 https://doi.org/10.1145/3658644.3670351 Citation Details
Liu, Ruixuan and Lee, Hong Kyu and Bhavani, Sivasubramanium V and Jiang, Xiaoqian and Ohno-Machado, Lucila and Xiong, Li "Patient-Centered and Practical Privacy to Support AI for Healthcare" , 2024 https://doi.org/10.1109/TPS-ISA62245.2024.00038 Citation Details
Liu, Ruixuan and Wang, Tianhao and Cao, Yang and Xiong, Li "PreCurious: How Innocent Pre-Trained Language Models Turn into Privacy Traps" , 2024 https://doi.org/10.1145/3658644.3690279 Citation Details
Liu, Yixuan and Liu, Yuhan and Xiong, Li and Gu, Yujie and Chen, Hong "Enhanced Privacy Bound for Shuffle Model with Personalized Privacy" , 2024 https://doi.org/10.1145/3627673.3679911 Citation Details
Razmi, Fereshteh and Lou, Jian and Hong, Yuan and Xiong, Li "Interpretation Attacks and Defenses on Predictive Models using Electronic Health Records" European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases , 2023 Citation Details
Takagi, Shun and Xiong, Li and Kato, Fumiyuki and Cao, Yang and Yoshikawa, Masatoshi "HRNet: Differentially Private Hierarchical and Multi-Resolution Network for Human Mobility Data Synthesization" Proceedings of the VLDB Endowment , v.17 , 2024 https://doi.org/10.14778/3681954.3681983 Citation Details
Wang, Wenjie and Tang, Pengfei and Lou, Jian and Shao, Yuanming and Waller, Lance and Ko, Yi-an and Xiong, Li "IGAMT: Privacy-Preserving Electronic Health Record Synthesization with Heterogeneity and Irregularity" The 38th Annual AAAI Conference on Artificial Intelligence , 2024 https://doi.org/10.1609/aaai.v38i14.29491 Citation Details
Xie, Han and Xiong, Li and Yang, Carl "Federated Node Classification over Distributed Ego-Networks with Secure Contrastive Embedding Sharing" , 2024 https://doi.org/10.1145/3627673.3679834 Citation Details
(Showing: 1 - 10 of 13)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page