@inproceedings{sung-etal-2023-fake,
title = "Not all Fake News is Written: A Dataset and Analysis of Misleading Video Headlines",
author = "Sung, Yoo Yeon and
Boyd-Graber, Jordan and
Hassan, Naeemul",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.emnlp-main.1010",
doi = "10.18653/v1/2023.emnlp-main.1010",
pages = "16241--16258",
abstract = "Polarization and the marketplace for impressions have conspired to make navigating information online difficult for users, and while there has been a significant effort to detect false or misleading text, multimodal datasets have received considerably less attention. To complement existing resources, we present multimodal Video Misleading Headline (VMH), a dataset that consists of videos and whether annotators believe the headline is representative of the video{'}s contents. After collecting and annotating this dataset, we analyze multimodal baselines for detecting misleading headlines. Our annotation process also focuses on why annotators view a video as misleading, allowing us to better understand the interplay of annotators{'} background and the content of the videos.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="sung-etal-2023-fake">
<titleInfo>
<title>Not all Fake News is Written: A Dataset and Analysis of Misleading Video Headlines</title>
</titleInfo>
<name type="personal">
<namePart type="given">Yoo</namePart>
<namePart type="given">Yeon</namePart>
<namePart type="family">Sung</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jordan</namePart>
<namePart type="family">Boyd-Graber</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Naeemul</namePart>
<namePart type="family">Hassan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-12</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</title>
</titleInfo>
<name type="personal">
<namePart type="given">Houda</namePart>
<namePart type="family">Bouamor</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Juan</namePart>
<namePart type="family">Pino</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Kalika</namePart>
<namePart type="family">Bali</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Singapore</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Polarization and the marketplace for impressions have conspired to make navigating information online difficult for users, and while there has been a significant effort to detect false or misleading text, multimodal datasets have received considerably less attention. To complement existing resources, we present multimodal Video Misleading Headline (VMH), a dataset that consists of videos and whether annotators believe the headline is representative of the video’s contents. After collecting and annotating this dataset, we analyze multimodal baselines for detecting misleading headlines. Our annotation process also focuses on why annotators view a video as misleading, allowing us to better understand the interplay of annotators’ background and the content of the videos.</abstract>
<identifier type="citekey">sung-etal-2023-fake</identifier>
<identifier type="doi">10.18653/v1/2023.emnlp-main.1010</identifier>
<location>
<url>https://aclanthology.org/2023.emnlp-main.1010</url>
</location>
<part>
<date>2023-12</date>
<extent unit="page">
<start>16241</start>
<end>16258</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Not all Fake News is Written: A Dataset and Analysis of Misleading Video Headlines
%A Sung, Yoo Yeon
%A Boyd-Graber, Jordan
%A Hassan, Naeemul
%Y Bouamor, Houda
%Y Pino, Juan
%Y Bali, Kalika
%S Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
%D 2023
%8 December
%I Association for Computational Linguistics
%C Singapore
%F sung-etal-2023-fake
%X Polarization and the marketplace for impressions have conspired to make navigating information online difficult for users, and while there has been a significant effort to detect false or misleading text, multimodal datasets have received considerably less attention. To complement existing resources, we present multimodal Video Misleading Headline (VMH), a dataset that consists of videos and whether annotators believe the headline is representative of the video’s contents. After collecting and annotating this dataset, we analyze multimodal baselines for detecting misleading headlines. Our annotation process also focuses on why annotators view a video as misleading, allowing us to better understand the interplay of annotators’ background and the content of the videos.
%R 10.18653/v1/2023.emnlp-main.1010
%U https://aclanthology.org/2023.emnlp-main.1010
%U https://doi.org/10.18653/v1/2023.emnlp-main.1010
%P 16241-16258
Markdown (Informal)
[Not all Fake News is Written: A Dataset and Analysis of Misleading Video Headlines](https://aclanthology.org/2023.emnlp-main.1010) (Sung et al., EMNLP 2023)
ACL