Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.253
|View full text |Cite
|
Sign up to set email alerts
|

MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance Detection

Abstract: Much of natural language processing is focused on leveraging large capacity language models, typically trained over single messages with a task of predicting one or more tokens. However, modeling human language at higher-levels of context (i.e., sequences of messages) is underexplored. In stance detection and other social media tasks where the goal is to predict an attribute of a message, we have contextual data that is loosely semantically connected by authorship. Here, we introduce Message-Level Transformer … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 20 publications
(7 reference statements)
0
2
0
Order By: Relevance
“…HaRT. Recent works (Lynn et al, 2020;Matero et al, 2021b; have highlighted the importance of incorporating author context into the message representations through the use of history and multi-level modeling. We use the Human aware Recurrent Transformer model which is built on GPT2 (Radford et al, 2019), to produce message representations that encode the latent representation of the author as well.…”
Section: Task Amentioning
confidence: 99%
“…HaRT. Recent works (Lynn et al, 2020;Matero et al, 2021b; have highlighted the importance of incorporating author context into the message representations through the use of history and multi-level modeling. We use the Human aware Recurrent Transformer model which is built on GPT2 (Radford et al, 2019), to produce message representations that encode the latent representation of the author as well.…”
Section: Task Amentioning
confidence: 99%
“…HaRT. Recent works (Lynn et al, 2020;Matero et al, 2021b;Soni et al, 2022) have highlighted the importance of incorporating author context into the message representations through the use of history and multi-level modeling. We use the Human aware Recurrent Transformer model (Soni et al, 2022) which is built on GPT2 (Radford et al, 2019), to produce message representations that encode the latent representation of the author as well.…”
Section: Task Amentioning
confidence: 99%
“…Many of the works have focused on data from Twitter, incorporating conversational and interactional context [9,10] in order to better classify the stances of the users in a thread of tweets or simply by taking the tweets in an independent way [11][12][13][14]. The SemEval-2017 task 8 [15] proposes to use the interactional context of Twitter threads, focusing on rumouroriented stance classification, where the objective is to identify support towards a rumour and an entire statement, rather than individual target concepts.…”
Section: Introductionmentioning
confidence: 99%
“…Since then, the conversation has been integrated with graphical models that allow taking into account its dynamics [22][23][24] through the different successive speech turns of the participants. Neural networks [12,[25][26][27] fall into this type of model and can even be pre-trained for the conversation setting [10] to understand better the conversational context to analyse stances in Twitter threads.…”
Section: Introductionmentioning
confidence: 99%