Award Abstract # 2235405
NSF Convergence Accelerator Track H: AI-based Tools to Enhance Access and Opportunities for the Deaf

NSF Org: ITE
Innovation and Technology Ecosystems
Recipient: RUTGERS, THE STATE UNIVERSITY
Initial Amendment Date: December 8, 2022
Latest Amendment Date: December 8, 2022
Award Number: 2235405
Award Instrument: Standard Grant
Program Manager: Alex Vadati
alvadati@nsf.gov
 (703)292-7068
ITE
 Innovation and Technology Ecosystems
TIP
 Directorate for Technology, Innovation, and Partnerships
Start Date: December 15, 2022
End Date: November 30, 2024 (Estimated)
Total Intended Award Amount: $750,000.00
Total Awarded Amount to Date: $750,000.00
Funds Obligated to Date: FY 2023 = $750,000.00
History of Investigator:
  • Dimitris Metaxas (Principal Investigator)
    dnm@cs.rutgers.edu
  • Carol Neidle (Co-Principal Investigator)
  • Matt Huenerfauth (Co-Principal Investigator)
Recipient Sponsored Research Office: Rutgers University New Brunswick
3 RUTGERS PLZ
NEW BRUNSWICK
NJ  US  08901-8559
(848)932-0150
Sponsor Congressional District: 12
Primary Place of Performance: Rutgers University New Brunswick
110 Frelinghuysen Road
Piscataway
NJ  US  08854-8072
Primary Place of Performance
Congressional District:
06
Unique Entity Identifier (UEI): M1LVPE5GLSD9
Parent UEI:
NSF Program(s): Convergence Accelerator Resrch
Primary Program Source: 01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s):
Program Element Code(s): 131Y00
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.084

ABSTRACT

We propose to develop sustainable, robust AI methods to overcome obstacles to digital communication and information access faced by Deaf and Hard-of-Hearing (DHH) individuals, empowering them personally and professionally. Users of American Sign Language (ASL), which has no standard written form, lack parity with hearing users in the digital arena. The proposed tools for privacy protection for ASL video communication and video search-by-example for access to multimedia digital resources build on prior NSF-funded AI research on linguistically-informed computer-based analysis and recognition of ASL from videos.
PROBLEM #1. ASL signers cannot communicate anonymously about sensitive topics through videos in their native language; this is perceived by the Deaf community to be a serious problem.
PROBLEM #2. There is no good way to look up a sign in a dictionary. Many ASL dictionaries enable sign look-up based on English translations, but what if the user does not understand the sign, or does not know its English translation? Others allow for search based on properties of ASL signs (e.g., handshape, location, movement type), but this is cumbersome, and a user must often look through hundreds of pictures of signs to find a target sign (if it is present at all in that dictionary).

The tools to be developed will enable signers to anonymize ASL videos while preserving essential linguistic information conveyed by hands, arms, facial expressions, and head movements; and enable searching for a sign based on ASL input from a webcam or a video clip.

Participants include DHH individuals, Deaf-owned companies, and members of other underrepresented minorities. The products will serve the >500,000 US signers and could be extended to other sign languages. The proposed application development brings together state-of-the-art research on: (1) video anonymization (using an asymmetric encoder-decoder structured image generator to generate high-resolution target frames driven by the original signing from the low-resolution source frames for anonymization, based on optical flow and confidence maps); (2) computer-based sign recognition from video (bidirectional skeleton-based isolated sign recognition using Graph Convolution Networks); and (3) HCI, including DHH user studies to assess desiderata for user interfaces for the proposed applications.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Bohacek, Matyas and Hassan, Saad "Sign Spotter: Design and Initial Evaluation of an Automatic Video-Based American Sign Language Dictionary System" , 2023 https://doi.org/10.1145/3597638.3614497 Citation Details
de_Lacerda_Pataca, Caluã and Hassan, Saad and Tinker, Nathan and Peiris, Roshan Lalintha and Huenerfauth, Matt "Caption Royale: Exploring the Design Space of Affective Captions from the Perspective of Deaf and Hard-of-Hearing Individuals" , 2024 https://doi.org/10.1145/3613904.3642258 Citation Details
Hassan, Saad and De_Lacerda_Pataca, Caluã and Nourian, Laleh and Tigwell, Garreth W and Davis, Briana and Silver_Wagman, Will Zhenya "Designing and Evaluating an Advanced Dance Video Comprehension Tool with In-situ Move Identification Capabilities" , 2024 https://doi.org/10.1145/3613904.3642710 Citation Details
Hassan, Saad and Ding, Yao and Kerure, Agneya Abhimanyu and Miller, Christi and Burnett, John and Biondo, Emily and Gilbert, Brenden "Exploring the Design Space of Automatically Generated Emotive Captions for Deaf or Hard of Hearing Users" Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems (CHI EA '23) , 2023 https://doi.org/10.1145/3544549.3585880 Citation Details
Xia, Zhaoyang and Neidle, Carol and Metaxas, Dimitris N. "DiffSLVA: Harnessing Diffusion Models for Sign Language Video Anonymization" arXiv , 2023 Citation Details
Neidle, Carol "Challenges for Linguistically-Driven Computer-Based Sign Recognition from Continuous Signing for American Sign Language" arXiv.org , 2023 Citation Details
Amin, Akhter Al and Hassan, Saad and Huenerfauth, Matt and Alm, Cecilia Ovesdotter "Modeling Word Importance in Conversational Transcripts: Toward improved live captioning for Deaf and hard of hearing viewers" Proceedings of the 20th International Web for All Conference , 2023 https://doi.org/10.1145/3587281.3587290 Citation Details
Amin, Akhter Al and Hassan, Saad and Lee, Sooyeon and Huenerfauth, Matt "Understanding How Deaf and Hard of Hearing Viewers Visually Explore Captioned Live TV News" Proceedings of the 20th International Web for All Conference , 2023 https://doi.org/10.1145/3587281.3587287 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page