Skip to feedback

Award Abstract # 2237880
CAREER: Human-Machine Supervision Cycle for Trustworthy Biometrics

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: UNIVERSITY OF NOTRE DAME DU LAC
Initial Amendment Date: March 10, 2023
Latest Amendment Date: June 26, 2024
Award Number: 2237880
Award Instrument: Continuing Grant
Program Manager: Anna Squicciarini
asquicci@nsf.gov
 (703)292-5177
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: June 1, 2023
End Date: May 31, 2028 (Estimated)
Total Intended Award Amount: $556,639.00
Total Awarded Amount to Date: $219,278.00
Funds Obligated to Date: FY 2023 = $107,988.00
FY 2024 = $111,290.00
History of Investigator:
  • Adam Czajka (Principal Investigator)
    aczajka@nd.edu
Recipient Sponsored Research Office: University of Notre Dame
940 GRACE HALL
NOTRE DAME
IN  US  46556-5708
(574)631-7432
Sponsor Congressional District: 02
Primary Place of Performance: University of Notre Dame
940 Grace Hall
NOTRE DAME
IN  US  46556-5708
Primary Place of Performance
Congressional District:
02
Unique Entity Identifier (UEI): FPU6XGFXMBE9
Parent UEI: FPU6XGFXMBE9
NSF Program(s): Secure &Trustworthy Cyberspace
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
01002425DB NSF RESEARCH & RELATED ACTIVIT

01002526DB NSF RESEARCH & RELATED ACTIVIT

01002627DB NSF RESEARCH & RELATED ACTIVIT

01002728DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 025Z, 1045
Program Element Code(s): 806000
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Dominant approaches of biometric attack detection make strong assumptions about the type of deviations from authentic information. This creates a critical gap between reliability observed in laboratory settings and the performance expected in the real world, where future attacks are unknown. This project fills this gap and builds an effective symbiosis between Artificial Intelligence (AI) and humans. The project novelties are new methods that (a) allow the AI to effectively learn from humans how to increase detectability of unknown attacks on biometric systems, and (b) support humans in their examination of fake biometric inputs. The project's broader significance and importance are: (a) trustworthy biometric systems that better recognize never-seen presentation attacks, and thus better protect consumer devices, bank accounts and strengthen the US border control processes; (b) a strong educational program that exposes K-12, undergraduate and graduate students to both the security- and ethics-related aspects of biometrics, and broadens their knowledge in a relevant topic of national concern; (c) publicly available lectures prepared by the investigator, which will broaden the awareness of responsible use of biometrics.

In this project, a holistic framework for human-machine supervision cycle will be established to enable (a) human-guided design of computer vision methods to make the biometric presentation attack detection mechanisms generalize better to unknown attack instruments and (b) creation of computer-aided methods of assisting human examiners in detecting of fake inputs. Fundamental technical contributions of this project include (1) broadening knowledge about mechanisms that govern human perception of fake visual signals, (2) discovering the most effective human-interpretable representations of information to support their decisions and speed up their learning of new types of attacks, (3) developing quantitative metrics of trust assessment and linking human and machine decisions into a trustworthy tandem that makes better judgements on the authenticity of biometric inputs, and (4) application of the framework to iris recognition of newborns, contributing to a better linkage with mothers and/or guardians, resulting in improved chances to benefit from healthcare systems, especially in developing countries.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Tinsley, Patrick and Purnapatra, Sandip and Mitcheff, Mahsa and Boyd, Aidan and Crum, Colton and Bowyer, Kevin and Flynn, Patrick and Schuckers, Stephanie and Czajka, Adam and Fang, Meiling and Damer, Naser and Liu, Xingyu and Wang, Caiyong and Sun, Xiany "Iris Liveness Detection Competition (LivDet-Iris) The 2023 Edition" , 2023 https://doi.org/10.1109/IJCB57857.2023.10448637 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page