2020
DOI: 10.1162/tacl_a_00302
|View full text |Cite
|
Sign up to set email alerts
|

A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation

Abstract: Story generation, namely, generating a reasonable story from a leading context, is an important but challenging task. In spite of the success in modeling fluency and local coherence, existing neural language generation models (e.g., GPT-2) still suffer from repetition, logic conflicts, and lack of long-range coherence in generated stories. We conjecture that this is because of the difficulty of associating relevant commonsense knowledge, understanding the causal relationships, and planning entities and events … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
149
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 169 publications
(150 citation statements)
references
References 27 publications
1
149
0
Order By: Relevance
“…One future path lies in combining language models with knowledge bases: curated databases of declarative facts. In work presented at last year's Association for Computational Linguistics meeting 9 , researchers fine-tuned GPT-2 on sentences explicitly stating facts and inferences from a compendium of common sense (for instance, if someone cooks spaghetti, that person wants to eat). As a result, it wrote short stories that were more logical.…”
Section: Seeking Common Sensementioning
confidence: 99%
“…One future path lies in combining language models with knowledge bases: curated databases of declarative facts. In work presented at last year's Association for Computational Linguistics meeting 9 , researchers fine-tuned GPT-2 on sentences explicitly stating facts and inferences from a compendium of common sense (for instance, if someone cooks spaghetti, that person wants to eat). As a result, it wrote short stories that were more logical.…”
Section: Seeking Common Sensementioning
confidence: 99%
“…To address this defect, incorporating external commonsense knowledge to enhance models' reasoning ability has been widely explored (Lin et al, 2019;Ye et al, 2019;Lv et al, 2019). In language generation, previous work (Bhagavatula et al, 2020;Guan et al, 2020) transfers commonsense knowledge into pre-trained language models by utilizing triple information in commonsense knowledge bases such as ConceptNet (Speer and Havasi, 2012) and ATOMIC .…”
Section: Conceptnet Roc Storymentioning
confidence: 99%
“…First, recovering knowledge triples at the posttraining stage (Guan et al, 2020) hardly enables the model to utilize the encoded knowledge in fine-tuning generation tasks which requires reasoning over underlying commonsense knowledge. Second, it ignores the abundant structural relational relevance of the concepts in the knowledge graph (Guan et al, 2020;Bhagavatula et al, 2020) that may provide multiple plausible evidence for complex reasoning.…”
Section: Conceptnet Roc Storymentioning
confidence: 99%
See 2 more Smart Citations