2022
DOI: 10.1109/access.2022.3202976
|View full text |Cite
|
Sign up to set email alerts
|

CAMM: Cross-Attention Multimodal Classification of Disaster-Related Tweets

Abstract: During the past decade, social media platforms have been extensively used during a disaster for information dissemination by the affected community and humanitarian agencies. Although many studies have been done recently to classify the informative and non-informative messages from social media posts, most are unimodal, i.e., have independently used textual or visual data to build the deep learning models. In the present study, we integrate the complementary information provided by the text and image messages … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 50 publications
0
7
0
Order By: Relevance
“…Deep learning, a subset of ML, has also found extensive applications in handling Twitter data due to its ability to automatically learn composite patterns and representations from large-scale data. In the latest literature mainly focused on deep networks, the authors [5] proposed a new deep neural network called Cross-Attention Multi-Modal (CAMM) to classify disaster data that contains both text and images. The authors [6] proposed a novel method, called Stacking-based Ensemble [43], using Statistical features and Informative words to tackle the challenges of damage assessment in tweets.…”
Section: Deep Learning and Neural Network Approachesmentioning
confidence: 99%
See 3 more Smart Citations
“…Deep learning, a subset of ML, has also found extensive applications in handling Twitter data due to its ability to automatically learn composite patterns and representations from large-scale data. In the latest literature mainly focused on deep networks, the authors [5] proposed a new deep neural network called Cross-Attention Multi-Modal (CAMM) to classify disaster data that contains both text and images. The authors [6] proposed a novel method, called Stacking-based Ensemble [43], using Statistical features and Informative words to tackle the challenges of damage assessment in tweets.…”
Section: Deep Learning and Neural Network Approachesmentioning
confidence: 99%
“…Continuous Bag of Words (CBOW): Based on the context words around a target word, CBOW attempts to anticipate that term. To maximize the chance of correctly predicting the target word "w_t" given its context words "w_c," where "c" spans from "-C" to "C" (excluding 0 as the target word itself), one must start with a context window of size "C" (number of words on each side of the target word) shown in equation [5] π‘€π‘Žπ‘₯π‘–π‘šπ‘–π‘§π‘’ ( 1 𝑇 ) * Ξ£(Log P(w t |w c ))…”
Section: Word To Vector (Word2vec)mentioning
confidence: 99%
See 2 more Smart Citations
“…Multimodal learning is a general method for building artificial intelligence (AI) models that extract and correlate information from multimodal data (Baltrusaitis et al 2019). Multimodal learning has been used in several areas (Khattar & Quadri 2022), such as visual question answering, emotion recognition, machine translation, cross-modal retrieval, and speech recognition. With the development of large survey telescopes, a massive amount of multi-source heterogeneous astronomical data, such as spectral and photometric data of astronomical objects, have been generated.…”
Section: Introductionmentioning
confidence: 99%