Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021
DOI: 10.24963/ijcai.2021/612
|View full text |Cite
|
Sign up to set email alerts
|

Pretrained Language Model for Text Generation: A Survey

Abstract: Text generation has become one of the most important yet challenging tasks in natural language processing (NLP). The resurgence of deep learning has greatly advanced this field by neural generation models, especially the paradigm of pretrained language models (PLMs). In this paper, we present an overview of the major advances achieved in the topic of PLMs for text generation. As the preliminaries, we present the general task definition and briefly describe the mainstream architectures of PLMs for text gener… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
58
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 94 publications
(58 citation statements)
references
References 5 publications
(12 reference statements)
0
58
0
Order By: Relevance
“…In this work, we use pre-trained natural language models in order to extract the contextual semantic patterns in a collection of open-responses. Pre-trained natural language models [10] are commonly used for a wide range of tasks such as text generation [11], building dialogue systems [12], text classification [13], hate speech detection [14], sentiment analysis [15], named entity recognition [16], question answering [17], and text summarization [18,19].…”
Section: Methodsmentioning
confidence: 99%
“…In this work, we use pre-trained natural language models in order to extract the contextual semantic patterns in a collection of open-responses. Pre-trained natural language models [10] are commonly used for a wide range of tasks such as text generation [11], building dialogue systems [12], text classification [13], hate speech detection [14], sentiment analysis [15], named entity recognition [16], question answering [17], and text summarization [18,19].…”
Section: Methodsmentioning
confidence: 99%
“…Our solution is inspired by the excellent few-shot capabilities of pretrained language models (PLMs) on language understanding and generation tasks (Brown et al, 2020;Chen et al, 2020;Li et al, 2021a). Pretrained on the large-scale corpora, PLMs encode vast amounts of world knowledge into their parameters (Li et al, 2021b), which is potentially beneficial to understand and describe the KG facts in our task.…”
Section: Kg Descriptive Textmentioning
confidence: 99%
“…Recent years have witnessed prominent achievement of PLMs in NLP tasks (Devlin et al, 2019;Radford et al, 2019). Pretrained on massive corpora, pretrained models showcase unprecedented generalization ability to solve related downstream tasks (Li et al, 2021b). However, most of existing PLMs were conditioned on text data (Radford et al, 2019;Lewis et al, 2020), lacking consideration of structured data input.…”
Section: Related Workmentioning
confidence: 99%
“…Pre-trained language models (PLMs) (Peters et al, 2018;Devlin et al, 2019), are now used in almost all NLP applications, e.g., machine translation (Li et al, 2021), question answering (Zhang et al, 2020), dialogue systems (Ni et al, 2021), and sentiment analysis (Minaee et al, 2020). They have sometimes been referred to as "foundation models" (Bommasani et al, 2021) due to their significant impact on research and industry.…”
Section: Introductionmentioning
confidence: 99%