Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.9
|View full text |Cite
|
Sign up to set email alerts
|

Mention Flags (MF): Constraining Transformer-based Text Generators

Abstract: This paper focuses on Seq2Seq (S2S) constrained text generation where the text generator is constrained to mention specific words, which are inputs to the encoder, in the generated outputs. Pre-trained S2S models such as T5 or a Copy Mechanism can be trained to copy the surface tokens from encoders to decoders, but they cannot guarantee constraint satisfaction. Constrained decoding algorithms always produce hypotheses satisfying all constraints. However, they are computationally expensive and can lower the gen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…In the future, we would explore extending KG-S2S to other Seq2Seq PLMs, such as BART (Lewis et al, 2020) and MASS (Song et al, 2019). In addition, it is interesting to combine KG-S2S with other knowledge-intensive NLP tasks, such as conversation recommendation (Li et al, 2018b) and commonsense generation (Wang et al, 2021b) in the Seq2Seq framework, and see if the KG knowledge could benefit these downstream tasks.…”
Section: Discussionmentioning
confidence: 99%
“…In the future, we would explore extending KG-S2S to other Seq2Seq PLMs, such as BART (Lewis et al, 2020) and MASS (Song et al, 2019). In addition, it is interesting to combine KG-S2S with other knowledge-intensive NLP tasks, such as conversation recommendation (Li et al, 2018b) and commonsense generation (Wang et al, 2021b) in the Seq2Seq framework, and see if the KG knowledge could benefit these downstream tasks.…”
Section: Discussionmentioning
confidence: 99%
“…where score(b) is the score of the current beam state, logGen(w) is the output logit of the generator, f ( * ) are functions that scores word w weighted by α i , and V suc is a predefined vocabulary. Similarily, Mention Flags (Wang et al, 2021) tries to identify the presence of tokens in the hypothesis given a set of flags. Both methods face the same problem since they operate on surface tokens.…”
Section: Related Workmentioning
confidence: 99%
“…These decomposition strategies showed high performance while introducing more detailed annotation to model training [5,7]. Inspired by the success of pretrained language models and the corresponding natural language generation-based paradigm for various NLP tasks [4,[21][22][23] tackle event extraction as controlled event generation. [6] is an end-to-end conditional generation method with manually designed discrete prompts for each event type, which needs more human effort to find the Fig.…”
Section: Related Workmentioning
confidence: 99%