Proceedings of the 2nd Workshop on Machine Reading for Question Answering 2019
DOI: 10.18653/v1/d19-5821
|View full text |Cite
|
Sign up to set email alerts
|

A Recurrent BERT-based Model for Question Generation

Abstract: In this study, we investigate the employment of the pre-trained BERT language model to tackle question generation tasks. We introduce three neural architectures built on top of BERT for question generation tasks. The first one is a straightforward BERT employment, which reveals the defects of directly using BERT for text generation. Accordingly, we propose another two models by restructuring our BERT employment into a sequential manner for taking information from previous decoded results. Our models are traine… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
51
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 122 publications
(72 citation statements)
references
References 17 publications
0
51
0
Order By: Relevance
“…Reinforcement learning is a popular approach to train the neural QG models, where the reward is defined as the evaluation metrics (Song et al, 2017;Kumar et al, 2018), or the QA accuracy/likelihood (Yuan et al, 2017;Hosking and Riedel, 2019;Zhang and Bansal, 2019). State-ofthe-art QG models Chan and Fan, 2019) use pre-trained language models. Question-Answer Pair Generation (QAG) from contexts, which is our main target, is a relatively less explored topic tackled by only a few recent works (Du and Cardie, 2018;.…”
Section: Related Workmentioning
confidence: 99%
“…Reinforcement learning is a popular approach to train the neural QG models, where the reward is defined as the evaluation metrics (Song et al, 2017;Kumar et al, 2018), or the QA accuracy/likelihood (Yuan et al, 2017;Hosking and Riedel, 2019;Zhang and Bansal, 2019). State-ofthe-art QG models Chan and Fan, 2019) use pre-trained language models. Question-Answer Pair Generation (QAG) from contexts, which is our main target, is a relatively less explored topic tackled by only a few recent works (Du and Cardie, 2018;.…”
Section: Related Workmentioning
confidence: 99%
“…However, only few attempts have been made so far to make use of these pre-trained models for conditional language modeling. Dong et al (2019) and Chan and Fan (2019) use a single BERT model for both encoding and decoding and achieve state-of-the-art results in QG. However, both of them use the [MASK] token as the input for predicting the word in place, which makes the training slower as it warranties recurrent generation (Chan and Fan, 2019) or generation with random masking (Dong et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…For evaluating our models, we report standard metrics of BLEU4, METEOR and ROUGE-L. As baselines, we take two of the non-BERT state-of-the-art models (Du and Cardie, 2018;Zhang and Bansal, Model BLEU4 METEOR ROUGE-L CorefNQG (Du and Cardie, 2018) 15.16 19.12 -SemdriftQG (Zhang and Bansal, 2019) 18.37 22.65 6.68 Recurrent-BERT (Chan and Fan, 2019) 20.33 23.88 48.23 UniLM (Dong et al, 2019) 22 Du et al (2017). BERT refers to BERT-Large(cased) model (Devlin et al, 2019) 2019) and the two BERT-based QG models (Dong et al, 2019;Chan and Fan, 2019). We experimented with 4 settings: one without using any copy mechanism (No Copy), one using normal copy (Normal Copy; §3.3.1), one using self-copy (Self-Copy; §3.3.2) and finally with two-hop selfcopy (Two-Hop Self-Copy; §3.3.3).…”
Section: Evaluation Metrics and Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…One of the benefits of our architecture is that the modules are not bounded by any specific model. For the current work, we employ the QG model proposed by Chan and Fan (2019) for the two assistants of the teacher and for the student, taking advantage of BERT .…”
Section: Question Generation Modulementioning
confidence: 99%