Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.18
|View full text |Cite
|
Sign up to set email alerts
|

Few-Shot NLG with Pre-Trained Language Model

Abstract: Neural-based end-to-end approaches to natural language generation (NLG) from structured data or knowledge are data-hungry, making their adoption for real-world applications difficult with limited data. In this work, we propose the new task of few-shot natural language generation. Motivated by how humans tend to summarize tabular data, we propose a simple yet effective approach and show that it not only demonstrates strong performance but also provides good generalization across domains. The design of the model… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
106
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 76 publications
(107 citation statements)
references
References 20 publications
(28 reference statements)
1
106
0
Order By: Relevance
“…Most recently, large-scale pre-trained models (Radford et al, 2019;Song et al, 2019;Raffel et al, 2019) have achieved new state-ofthe-arts on various generation tasks. Chen et al (2019b) demonstrate that a simple pre-training based method can achieve very reasonable performance on the WikiBio dataset (Lebret et al, 2016) under few-shot setting. More recent works begin to focus on fidelity preserving of the generation, such as (Dhingra et al, 2019;Tian et al, 2019).…”
Section: Related Workmentioning
confidence: 89%
See 2 more Smart Citations
“…Most recently, large-scale pre-trained models (Radford et al, 2019;Song et al, 2019;Raffel et al, 2019) have achieved new state-ofthe-arts on various generation tasks. Chen et al (2019b) demonstrate that a simple pre-training based method can achieve very reasonable performance on the WikiBio dataset (Lebret et al, 2016) under few-shot setting. More recent works begin to focus on fidelity preserving of the generation, such as (Dhingra et al, 2019;Tian et al, 2019).…”
Section: Related Workmentioning
confidence: 89%
“…Considering that acquiring a large amount of (logical form, description) pairs in real-world cases is expensive, we also include a few-shot learning task for our dataset, where the model is only provided with hundreds of paired examples. Previous works have shown that the pre-trained language models obtain strong NLG performance even with a handful of fine-tuning instances (Chen et al, 2019b). Therefore we still use the best-performing GPT-2 model for this study.…”
Section: Few-shot Settingmentioning
confidence: 99%
See 1 more Smart Citation
“…In similar studies, pretrained GPT models (Radford et al, 2019) were used by Chen et al (2020) and Peng et al (2020), who fine-tune them on a small set of in-domain data, but they did not distill these models into ones suitable for production. Interestingly, Wen et al (2016) demonstrated that the structure of arguments in existing dialogues can be used to guide data collection for low-resource domain adaptation, which is similar to the bucketing strategies explored here.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, advances in neural-network-based (conditional) language generation prompted a new direction in NLG research (Novikova et al, 2017;Budzianowski et al, 2018;Chen et al, 2020;Bal-akrishnan et al, 2019;Peng et al, 2020). The process is typically split into two steps: (1) serialization of input data into a flattened meaning representation (MR), and (2) using the neural generation model to generate a natural language response conditioned on the MR.…”
Section: Introductionmentioning
confidence: 99%