Award Abstract # 2040638
NSF Convergence Accelerator Track D: Data & AI Methods for Modeling Facial Expressions in Language with Applications to Privacy for the Deaf, ASL Education & Linguistic Research

NSF Org: ITE
Innovation and Technology Ecosystems
Recipient: RUTGERS, THE STATE UNIVERSITY
Initial Amendment Date: September 4, 2020
Latest Amendment Date: October 14, 2020
Award Number: 2040638
Award Instrument: Standard Grant
Program Manager: Mike Pozmantier
ITE
 Innovation and Technology Ecosystems
TIP
 Directorate for Technology, Innovation, and Partnerships
Start Date: September 15, 2020
End Date: May 31, 2022 (Estimated)
Total Intended Award Amount: $960,000.00
Total Awarded Amount to Date: $960,000.00
Funds Obligated to Date: FY 2020 = $960,000.00
History of Investigator:
  • Dimitris Metaxas (Principal Investigator)
    dnm@cs.rutgers.edu
  • Carol Neidle (Co-Principal Investigator)
  • Matt Huenerfauth (Co-Principal Investigator)
  • Mariapaola D'Imperio (Co-Principal Investigator)
Recipient Sponsored Research Office: Rutgers University New Brunswick
3 RUTGERS PLZ
NEW BRUNSWICK
NJ  US  08901-8559
(848)932-0150
Sponsor Congressional District: 12
Primary Place of Performance: Rutgers University New Brunswick
33 Knightsbridge Road
Pisacataway
NJ  US  08854-8019
Primary Place of Performance
Congressional District:
06
Unique Entity Identifier (UEI): M1LVPE5GLSD9
Parent UEI:
NSF Program(s): Convergence Accelerator Resrch
Primary Program Source: 01002021DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s):
Program Element Code(s): 131Y00
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.084

ABSTRACT

The NSF Convergence Accelerator supports use-inspired, team-based, multidisciplinary efforts that address challenges of national importance and will produce deliverables of value to society in the near future. American Sign Language (ASL) is the third most studies ?foreign? language in the United States. This project is building 4-dimensional face-tracking algorithms that could be used to separate facial geometry from facial movement and expression. The work supports an application for teaching American Sign Language (ASL) to ASL-learners, an application for anonymizing the signer when privacy is a concern, and research into the role of facial expressions in both sign and spoken language. The privacy preserving application being developed by this project will enable ASL speakers to have private conversations about sensitive topics they would otherwise.

This team of linguists, computer scientists, deaf and hearing experts on ASL, and industry partners will address research and societal challenges through three types of deliverables targeted to diverse user and research communities: 1) Modifications and extension of AI methods and publicly shared ASL data and tools to encompass spoken language. Although facial expressions and head gestures, essential to the grammar of signed languages, also play an important role in speech, this is not well understood because resources of the kind developed by this project have not been available. New data and analyses will open the door to comparative study of the role of facial expressions across modalities, and the role of facial expressions in signed language vs. spoken language. Shared raw data, analyses, and visualizations will open up new avenues for linguistic and computer science research into the role of spatiotemporal synchronization of nonmanual expressions in conjunction with speech and signing. 2) An application to help ASL learners produce facial expressions and head gestures to convey grammatical information in signed languages; and 3) Development of a tool for real-time anonymization of ASL videos to preserve grammatical information expressed non-manually, while de-identifying the signer.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Dafnis, Konstantinos M. and Chroni, Evgenia and Neidle, Carol and Metaxas, Dimitris "Bidirectional Skeleton-Based Isolated Sign Recognition using Graph Convolution Networks" Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022), Marseille, 20-25 June 2022. , 2022 Citation Details
Dafnis, Konstantinos M. and Chroni, Evgenia and Neidle, Carol and Metaxas, Dimitris "Isolated Sign Recognition using ASL Datasets with Consistent Text-based Gloss Labeling and Curriculum Learning" Seventh International Workshop on Sign Language Translation and Avatar Technology: The Junction of the Visual and the Textual (SLTAT 2022). LREC, Marseille, France, June 2022 , 2022 Citation Details
Hassan, Saad and Lee, Sooyeon and Metaxas, Dimitris and Neidle, Carol and Huenerfauth, Matt "Understanding ASL Learners Preferences for a Sign Language Recording and Automatic Feedback System to Support Self-Study" Proceedings of the 24th International ACM SIGACCESS Conference on Computers and Accessibility , 2022 https://doi.org/10.1145/3517428.3550367 Citation Details
Lee, Sooyeon and Glasser, Abraham and Dingman, Becca and Xia, Zhaoyang and Metaxas, Dimitris and Neidle, Carol and Huenerfauth, Matt "American Sign Language Video Anonymization to Support Online Participation of Deaf and Hard of Hearing Users" Proceedings of ASSETS '21: The 23rd International ACM SIGACCESS Conference on Computers and Accessibility , 2021 https://doi.org/10.1145/3441852.3471200 Citation Details
Neidle, Carol and Opoku, Augustine and Ballard, Carey and Dafnis, Konstantinos M. and Chroni, Evgenia and Metaxas, Dimitris "Resources for Computer-Based Sign Recognition from Video, and the Criticality of Consistency of Gloss Labeling across Multiple Large ASL Video Corpora" Proceedings of the LREC2022 10th Workshop on the Representation and Processing of Sign Languages: Multilingual Sign Language Resources, Marseille, France, 25 June 2022 , 2022 Citation Details
Neidle, Carol and Opoku, Augustine and Metaxas, Dimitris "ASL Video Corpora & Sign Bank: Resources Available through the American Sign Language Linguistic Research Project (ASLLRP)" ArXivorg , v.2201 , 2022 Citation Details
Xia, Zhaoyang and Chen, Yuxiao and Zhangli, Qilong and Huenerfauth, Matt and Neidle, Carol and Metaxas, Dimitris "Sign Language Video Anonymization" Proceedings of the LREC2022 10th Workshop on the Representation and Processing of Sign Languages: Multilingual Sign Language Resources, Marseille, France, 25 June 2022 , 2022 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

We have been developing sustainable robust AI methods for facial analytics, potentially applicable across domains but targeted here to new applications that address important problems related to use of facial expressions and head gestures in natural language. In sign language, critical linguistic information of many kinds is conveyed exclusively by facial expressions and head gestures. The fact that the face carries essential linguistic information poses major challenges for Deaf signers and for students of ASL as a non-native language. 

Problem #1: The >500,000 US ASL signers have no way to communicate anonymously through videos in their native language, e.g., about sensitive topics (such as medical issues). This is perceived to be a significant problem by the Deaf community. It also means, for example, that signed submissions to scholarly journals cannot be reviewed anonymously. 

Problem #2: 2nd-language learners of ASL (the 3rd most studied "foreign" language, with US college enrollments >107,000 as of 2016) have difficulty learning to produce these essential expressions, in part because they do not see their own face when signing. 

    In spoken language, these expressions also play an important role, but they function differently.

Problem #3: The role of co-speech gestures, including facial expressions and head movements, is not well understood because of inadequacies in current analytic tools. This has held back applications that rely on such correlations, such as the development of realistic speaking avatars and robots; technology for the elderly and those with disabilities; and detection of, e.g., deception and intent. 

    To address these problems, we worked closely with the prospective users of these tools and conducted user studies to identify and respond to their needs and preferences, in order to create prototypes for applications (1) to enable ASL signers to share videos anonymously by disguising their face without loss of linguistic information; (2) to help ASL learners produce these expressions correctly; and (3) to help speech scientists study co-speech gestures. 

    Our AI approach to analysis of facial expressions and head gestures -- combining 3D modeling, Machine Learning (ML), and linguistic knowledge derived from our annotated video corpora -- overcomes limitations of prior research. It is distinctive in its ability to capture subtle facial expressions, even with significant head rotations. 

    Beyond the benefits that would accrue to the three target populations, this research has other potential applications, as well, e.g. for sanitizing other data involving video of human faces, medical applications, security, driving safety, and the arts. 

    In conjunction with the research conducted for this project, we have expanded and enhanced our collection of publicly shared, linguistically annotated ASL video data and refined the software we distribute through an MIT license for linguistic annotation of visual language data, SignStream(R). Data are available from https://dai.cs.rutgers.edu/dai/s/dai and https://dai.cs.rutgers.edu/dai/s/signbank. The software is available from http://www.bu.edu/asllrp/SignStream/3/. The data and software shared freely through our websites will also contribute to linguistic and computational research on sign language. This is especially valuable for students and has been used for many research projects in the US and abroad. 

 


Last Modified: 08/29/2022
Modified by: Carol J Neidle

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page