2021 IEEE Spoken Language Technology Workshop (SLT) 2021
DOI: 10.1109/slt48900.2021.9383453
|View full text |Cite
|
Sign up to set email alerts
|

Go Beyond Plain Fine-Tuning: Improving Pretrained Models for Social Commonsense

Abstract: Pretrained language models have demonstrated outstanding performance in many NLP tasks recently. However, their social intelligence, which requires commonsense reasoning about the current situation and mental states of others, is still developing. Towards improving language models' social intelligence, in this study we focus on the Social IQA dataset, a task requiring social and emotional commonsense reasoning. Building on top of the pretrained RoBERTa and GPT2 models, we propose several architecture variation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 38 publications
0
2
0
Order By: Relevance
“…Improvements in these simulators contribute to overall enhancements in dialogue system performance and their ability to handle diverse user inputs and scenarios. Lastly, we explore end-to-end approaches that aim to directly generate dialogue without explicitly defining intermediate steps or modules [4,19,28,32,58]. End-to-end models offer the advantage of encapsulating the entire dialogue generation process within a single model, simplifying both training and inference procedures.…”
Section: Contentmentioning
confidence: 99%
“…Improvements in these simulators contribute to overall enhancements in dialogue system performance and their ability to handle diverse user inputs and scenarios. Lastly, we explore end-to-end approaches that aim to directly generate dialogue without explicitly defining intermediate steps or modules [4,19,28,32,58]. End-to-end models offer the advantage of encapsulating the entire dialogue generation process within a single model, simplifying both training and inference procedures.…”
Section: Contentmentioning
confidence: 99%
“…Another direction in the field of commonsense reasoning involves combining pretrained language models with commonsense-specific fine-tuning techniques. Chang et al (2021) propose several architectural variations, leverage external commonsense corpora, and employ commonsense-specific fine-tuning techniques for the Social IQA task (Sap et al, 2019). Through their work, they demonstrate that these optimizations can enhance the model's performance in tasks related to social intelligence.…”
Section: Physicalmentioning
confidence: 99%