Award Abstract # 1949649
SaTC: CORE: Medium: Collaborative: BaitBuster 2.0: Keeping Users Away From Clickbait

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: UNIVERSITY OF MARYLAND, COLLEGE PARK
Initial Amendment Date: June 17, 2020
Latest Amendment Date: June 23, 2021
Award Number: 1949649
Award Instrument: Standard Grant
Program Manager: Dan Cosley
dcosley@nsf.gov
 (703)292-8832
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: July 1, 2020
End Date: June 30, 2024 (Estimated)
Total Intended Award Amount: $228,401.00
Total Awarded Amount to Date: $244,401.00
Funds Obligated to Date: FY 2020 = $228,401.00
FY 2021 = $16,000.00
History of Investigator:
  • Naeemul Hassan (Principal Investigator)
    nhassan@umd.edu
Recipient Sponsored Research Office: University of Maryland, College Park
3112 LEE BUILDING
COLLEGE PARK
MD  US  20742-5100
(301)405-6269
Sponsor Congressional District: 04
Primary Place of Performance: University of Maryland College Park
College Park
MD  US  20742-5103
Primary Place of Performance
Congressional District:
04
Unique Entity Identifier (UEI): NPU8ULVAAS23
Parent UEI: NPU8ULVAAS23
NSF Program(s): Secure &Trustworthy Cyberspace
Primary Program Source: 01002021DB NSF RESEARCH & RELATED ACTIVIT
01002122DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7924, 9178, 9251, 025Z
Program Element Code(s): 806000
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Social media sites such as Facebook are popular platforms for spreading clickbait, links with misleading titles that do not deliver on their promises. Not only does clickbait waste users' time, it often directs users to phishing sites and sites containing spyware and malware. A large number of users fall victim to scams on social media, including those spread through clickbait, due to both a lack of awareness and a lack of appropriate warnings on social media platforms. These users are vulnerable to identity theft, online hacking, and the exposure of sensitive information to adversaries. Thus, it is critical to limit the impact of clickbait on users' security. This project is developing novel techniques to detect various forms of clickbait, especially video-based clickbait, and study user behavior on social media to design effective warning systems. The findings from this research are being incorporated into an open-source browser extension called Baitbuster 2.0, building on the original Baitbuster tool for detecting text-based clickbait. To enhance the impact of this tool, the researchers will design new training methods to raise security awareness and help users avoid clickbait in social media. The project also aims to engage underrepresented groups via outreach efforts and through developing videos to encourage women to consider cybersecurity as a career.

Detecting clickbait is a major challenge, particularly as video becomes a more prominent form of media online, undermining efforts to detect misleading text. To address this challenge, the research team will take an integrated approach examining the effects of techniques used to attract clicks from users, presentation and distribution of clickbait, personalization of clickbait through crawling users? personal information (i.e., targeted clickbait), automatic generation of face-swapping clickbait, and risk perceptions and security awareness of users. As a first step, the researchers are collecting and analyzing clickbait datasets to explore ways of identifying clickbait on social media. Using these datasets, they are developing novel applications of state-of-the-art machine learning techniques such as optical character recognition and video understanding to automatically identify video clickbait. In another thrust of this project, the researchers are studying users' clicking behavior and corresponding security mental models to better understand their vulnerability to clickbait and examine the effects of a wide range of social engineering techniques used to attract clicks from users. The findings are being used to design warning systems, which will be integrated into BaitBuster 2.0, to warn users intelligently and effectively to avoid clickbait. Finally, the usability and efficacy of the warning system and BaitBuster 2.0 are being evaluated through in-depth user studies.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Sung, Yoo and Boyd-Graber, Jordan and Hassan, Naeemul "Not all Fake News is Written: A Dataset and Analysis of Misleading Video Headlines" , 2023 https://doi.org/10.18653/v1/2023.emnlp-main.1010 Citation Details

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Our project focused on understanding and detecting misleading information in online videos, an increasingly important issue in today’s digital world. We explored how video titles can misrepresent content, leading viewers to misunderstand or be misled by what they watch. Below are outcomes of this project-

1. A Taxonomy of Misleading Information
We identified and categorized different strategies used in misleading video titles. For example, some titles cherry-pick specific details that distort the overall message, while others use exaggeration or undermining techniques to mislead viewers. This taxonomy helps researchers and platforms better understand how misinformation spreads through subtle manipulations in headlines.

2. Dataset Development Method for Subjective Data
Since "misleading" is often a subjective concept, we developed a crowdsourcing method to annotate videos reliably. This approach gathers multiple perspectives from diverse individuals, ensuring a balanced view on whether a video title is misleading in relation to its content.

3. Curated Dataset of Misleading Videos
Using our crowdsourced annotations, we created a carefully curated dataset of misleading videos. This resource is publicly available and can support future research on misinformation, content moderation, and media literacy efforts.

4. A Multi-Modal Detection Model
We also developed an advanced deep learning model capable of analyzing both video content and its title (text) to detect misleading information. By combining insights from different types of data (multi-modality), the model can more accurately identify when a video title doesn’t align with the content, paving the way for better tools to combat misinformation.

Broader Impacts
This project contributes to the fight against online misinformation, providing tools and knowledge that can help platforms, educators, and policymakers. By improving our understanding of how misleading content works and developing methods to detect it, we’re helping to create a safer, more informed digital environment.

 


Last Modified: 02/20/2025
Modified by: Naeemul Hassan

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page