Proceedings of the 26th ACM International Conference on Multimedia 2018
DOI: 10.1145/3240508.3240707
|View full text |Cite
|
Sign up to set email alerts
|

Deep Multimodal Image-Repurposing Detection

Abstract: Nefarious actors on social media and other platforms often spread rumors and falsehoods through images whose metadata (e.g., captions) have been modified to provide visual substantiation of the rumor/falsehood. This type of modification is referred to as image repurposing, in which often an unmanipulated image is published along with incorrect or manipulated metadata to serve the actor's ulterior motives. We present the Multimodal Entity Image Repurposing (MEIR) dataset, a substantially challenging dataset ove… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
79
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 43 publications
(79 citation statements)
references
References 18 publications
0
79
0
Order By: Relevance
“…The reuse of unmanipulated images to spread misinformation about a possibly unrelated entity or event was introduced in [8] as the verification of semantic integrity of data with image assets. Image repurposing with manipulated textual and location data has been studied more specifically in [24]. Our work falls in this category and we propose a novel generalized framework for image repurposing detection.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…The reuse of unmanipulated images to spread misinformation about a possibly unrelated entity or event was introduced in [8] as the verification of semantic integrity of data with image assets. Image repurposing with manipulated textual and location data has been studied more specifically in [24]. Our work falls in this category and we propose a novel generalized framework for image repurposing detection.…”
Section: Related Workmentioning
confidence: 99%
“…Previous works in this domain [8,24] focused on the detection of metadata modalities that are continuous in nature, such as captions and GPS coordinates in the form of latent encodings. The method of [8] is not suitable for structured metadata because it relies on learning a joint representation of images and captions.…”
Section: Baseline and State-of-the-art Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…The first considers pairs of images and captions [4], and then trains a network to decide how consistent the caption is with the associated image. The second work [5] altered GPS coordinates, captions, and the actual image pixels for image re-purposing detection. In contrast to these methods, our approach operates only at the image pixel level and does not need any metadata, which may not be always available at hand.…”
Section: Related Workmentioning
confidence: 99%