2021
DOI: 10.48550/arxiv.2105.11174
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Retrieval Enhanced Model for Commonsense Generation

Abstract: Commonsense generation is a challenging task of generating a plausible sentence describing an everyday scenario using provided concepts. Its requirement of reasoning over commonsense knowledge and compositional generalization ability even puzzles strong pre-trained language generation models. We propose a novel framework using retrieval methods to enhance both the pre-training and fine-tuning for commonsense generation. We retrieve prototype sentence candidates by concept matching and use them as auxiliary inp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…Baselines (1) Concept2Sentence: We consider several recent submissions to the leaderboard of CommonGen that leverage auxiliary information for GCSR. KFCNet (Li et al, 2021), Re-T5 (Wang et al, 2021), and EKI-BART (Fan et al, 2020) are prototype-based models, which retrieve sentences containing as many input concepts as possible from external captions and NLI datasets, and then use these sentences as auxiliary inputs. VisCTG (Feng et al, 2021b) is an image-augmented model which retrieves images from Google by using concepts as a query, followed by an image captioning model that generates captions as auxiliary inputs.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Baselines (1) Concept2Sentence: We consider several recent submissions to the leaderboard of CommonGen that leverage auxiliary information for GCSR. KFCNet (Li et al, 2021), Re-T5 (Wang et al, 2021), and EKI-BART (Fan et al, 2020) are prototype-based models, which retrieve sentences containing as many input concepts as possible from external captions and NLI datasets, and then use these sentences as auxiliary inputs. VisCTG (Feng et al, 2021b) is an image-augmented model which retrieves images from Google by using concepts as a query, followed by an image captioning model that generates captions as auxiliary inputs.…”
Section: Methodsmentioning
confidence: 99%
“…This is because LMs have no intrinsic mechanism to reason over high-level relations between concepts . To close the knowledge gap, recent work augment LM input with knowledge graph triples (e.g., (dog, CapableOf, catch)) retrieved from ConceptNet Li et al, 2020), or prototype sentences that cover input concepts retrieved from external text corpora (Fan et al, 2020;Wang et al, 2021). However, despite the input augmentation, GCSR skills are implicitly learned based on the concept-text pairs in the training data, without explicit supervision.…”
Section: Learning To Verbalizementioning
confidence: 99%
See 1 more Smart Citation