
NSF Org: |
CNS Division Of Computer and Network Systems |
Recipient: |
|
Initial Amendment Date: | June 17, 2020 |
Latest Amendment Date: | June 23, 2021 |
Award Number: | 1949649 |
Award Instrument: | Standard Grant |
Program Manager: |
Dan Cosley
dcosley@nsf.gov (703)292-8832 CNS Division Of Computer and Network Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | July 1, 2020 |
End Date: | June 30, 2024 (Estimated) |
Total Intended Award Amount: | $228,401.00 |
Total Awarded Amount to Date: | $244,401.00 |
Funds Obligated to Date: |
FY 2021 = $16,000.00 |
History of Investigator: |
|
Recipient Sponsored Research Office: |
3112 LEE BUILDING COLLEGE PARK MD US 20742-5100 (301)405-6269 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
College Park MD US 20742-5103 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | Secure &Trustworthy Cyberspace |
Primary Program Source: |
01002122DB NSF RESEARCH & RELATED ACTIVIT |
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
Social media sites such as Facebook are popular platforms for spreading clickbait, links with misleading titles that do not deliver on their promises. Not only does clickbait waste users' time, it often directs users to phishing sites and sites containing spyware and malware. A large number of users fall victim to scams on social media, including those spread through clickbait, due to both a lack of awareness and a lack of appropriate warnings on social media platforms. These users are vulnerable to identity theft, online hacking, and the exposure of sensitive information to adversaries. Thus, it is critical to limit the impact of clickbait on users' security. This project is developing novel techniques to detect various forms of clickbait, especially video-based clickbait, and study user behavior on social media to design effective warning systems. The findings from this research are being incorporated into an open-source browser extension called Baitbuster 2.0, building on the original Baitbuster tool for detecting text-based clickbait. To enhance the impact of this tool, the researchers will design new training methods to raise security awareness and help users avoid clickbait in social media. The project also aims to engage underrepresented groups via outreach efforts and through developing videos to encourage women to consider cybersecurity as a career.
Detecting clickbait is a major challenge, particularly as video becomes a more prominent form of media online, undermining efforts to detect misleading text. To address this challenge, the research team will take an integrated approach examining the effects of techniques used to attract clicks from users, presentation and distribution of clickbait, personalization of clickbait through crawling users? personal information (i.e., targeted clickbait), automatic generation of face-swapping clickbait, and risk perceptions and security awareness of users. As a first step, the researchers are collecting and analyzing clickbait datasets to explore ways of identifying clickbait on social media. Using these datasets, they are developing novel applications of state-of-the-art machine learning techniques such as optical character recognition and video understanding to automatically identify video clickbait. In another thrust of this project, the researchers are studying users' clicking behavior and corresponding security mental models to better understand their vulnerability to clickbait and examine the effects of a wide range of social engineering techniques used to attract clicks from users. The findings are being used to design warning systems, which will be integrated into BaitBuster 2.0, to warn users intelligently and effectively to avoid clickbait. Finally, the usability and efficacy of the warning system and BaitBuster 2.0 are being evaluated through in-depth user studies.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
Our project focused on understanding and detecting misleading information in online videos, an increasingly important issue in today’s digital world. We explored how video titles can misrepresent content, leading viewers to misunderstand or be misled by what they watch. Below are outcomes of this project-
1. A Taxonomy of Misleading Information
We identified and categorized different strategies used in misleading video titles. For example, some titles cherry-pick specific details that distort the overall message, while others use exaggeration or undermining techniques to mislead viewers. This taxonomy helps researchers and platforms better understand how misinformation spreads through subtle manipulations in headlines.
2. Dataset Development Method for Subjective Data
Since "misleading" is often a subjective concept, we developed a crowdsourcing method to annotate videos reliably. This approach gathers multiple perspectives from diverse individuals, ensuring a balanced view on whether a video title is misleading in relation to its content.
3. Curated Dataset of Misleading Videos
Using our crowdsourced annotations, we created a carefully curated dataset of misleading videos. This resource is publicly available and can support future research on misinformation, content moderation, and media literacy efforts.
4. A Multi-Modal Detection Model
We also developed an advanced deep learning model capable of analyzing both video content and its title (text) to detect misleading information. By combining insights from different types of data (multi-modality), the model can more accurately identify when a video title doesn’t align with the content, paving the way for better tools to combat misinformation.
Broader Impacts
This project contributes to the fight against online misinformation, providing tools and knowledge that can help platforms, educators, and policymakers. By improving our understanding of how misleading content works and developing methods to detect it, we’re helping to create a safer, more informed digital environment.
Last Modified: 02/20/2025
Modified by: Naeemul Hassan
Please report errors in award information by writing to: awardsearch@nsf.gov.