2022
DOI: 10.1609/aaai.v36i11.21496
|View full text |Cite
|
Sign up to set email alerts
|

Commonsense Knowledge Reasoning and Generation with Pre-trained Language Models: A Survey

Abstract: While commonsense knowledge acquisition and reasoning has traditionally been a core research topic in the knowledge representation and reasoning community, recent years have seen a surge of interest in the natural language processing community in developing pre-trained models and testing their ability to address a variety of newly designed commonsense knowledge reasoning and generation tasks. This paper presents a survey of these tasks, discusses the strengths and weaknesses of state-of-the-art pre-trained mod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(9 citation statements)
references
References 48 publications
(61 reference statements)
0
2
0
Order By: Relevance
“…To represent words or sentences with vectors, a pre-trained model is welcomed due to its high-dimensional space and semantic representation. Such models have been widely accepted by both academic and industrial researchers in the past ten years [30,33,34]. They are pre-trained on an original task with a large corpus and used on a target task by tuning the corresponding parameters according to the characteristics of the target task.…”
Section: Word Embeddingmentioning
confidence: 99%
“…To represent words or sentences with vectors, a pre-trained model is welcomed due to its high-dimensional space and semantic representation. Such models have been widely accepted by both academic and industrial researchers in the past ten years [30,33,34]. They are pre-trained on an original task with a large corpus and used on a target task by tuning the corresponding parameters according to the characteristics of the target task.…”
Section: Word Embeddingmentioning
confidence: 99%
“…It is widely acknowledged that LLMs, trained on a huge amount of data, are able to obtain broad knowledge covering a wide range of domains (Rae et al 2021;Hoffmann et al 2022;Touvron et al 2023;Du et al 2022a;Guo et al 2023), including commonsense knowledge (West et al 2022;Bian et al 2023;Bang et al 2023). However, commonsense reasoning is still regarded as a major challenge for LLMs (Zhou et al 2020;Bhargava and Ng 2022). Studies disclose that LLMs fall short in performing adequate commonsense reasoning (Wei et al 2022).…”
Section: Introductionmentioning
confidence: 99%
“…by failing to account for the figurative interpretation of the IE (Balahur et al, 2010). This study focuses on injecting IE-related knowledge into small-frame PTLMs known for their wide use, such as BERT (Devlin et al, 2019) and BART (Lewis et al, 2020), considering their struggle to understand the figurative meanings of IEs (Bhargava and Ng, 2022;Zeng and Bhat, 2022). We discuss the corresponding capabilities of large PTLMs, such as GPT-3.5, in the limitation section.…”
Section: Introductionmentioning
confidence: 99%
“…We rely on psycholinguistic findings about the impact of IE-related aspects, such as mental states, emotions, and likely actions, on human IE comprehension (Rohani et al, 2012;Saban-Bezalel and Mashal, 2019), to explore the use of commonsense knowledge about IEs towards their comprehension. Specifically, we build on the findings that commonsense knowledge graphs (KGs), e.g., ATOMIC 20 20 (Hwang et al, 2021), organized as if-then relations for inferential knowledge enable linguistic and social reasoning abilities for PTLMs (Bhargava and Ng, 2022). Indeed, models relying on their applications have benefited figurative language processing, such as their interpretation (Chakrabarty et al, 2022a) and generation (Chakrabarty et al, 2021b).…”
Section: Introductionmentioning
confidence: 99%