Proceedings of the Web Conference 2020 2020
DOI: 10.1145/3366423.3380270
|View full text |Cite
|
Sign up to set email alerts
|

Asking Questions the Human Way: Scalable Question-Answer Generation from Text Corpus

Abstract: The ability to ask questions is important in both human and machine intelligence. Learning to ask questions helps knowledge acquisition, improves question-answering and machine reading comprehension tasks, and helps a chatbot to keep the conversation flowing with a human. Existing question generation models are ineffective at generating a large amount of high-quality question-answer pairs from unstructured text, since given an answer and an input passage, question generation is inherently a one-to-many mapping… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 61 publications
(31 citation statements)
references
References 57 publications
(130 reference statements)
0
28
0
Order By: Relevance
“…As a dual task of question-answering, QG can be used to improve QA performance. Some works [7,11,26,46,58] take QG as a generator to harvest question-answer pairs from passages, and use this harvested data to pre-train QA models, which subsequently resulted in improved QA model effectiveness. QG is also widely used in IR tasks, such as improving search system effectiveness by generating clarifying questions [57], or generating questions from e-commercial customers reviews [55].…”
Section: Question Generationmentioning
confidence: 99%
“…As a dual task of question-answering, QG can be used to improve QA performance. Some works [7,11,26,46,58] take QG as a generator to harvest question-answer pairs from passages, and use this harvested data to pre-train QA models, which subsequently resulted in improved QA model effectiveness. QG is also widely used in IR tasks, such as improving search system effectiveness by generating clarifying questions [57], or generating questions from e-commercial customers reviews [55].…”
Section: Question Generationmentioning
confidence: 99%
“…We used heuristics to filter out low-quality generated QA pairs, dropping questions that are longer than 20 words or shorter than 5 words and answers that are longer than 10 words, keeping questions that have at least one interrogative word, and removing n-gram repetition in questions. While some existing works used the BERT QA model or an entailment model as a data filter Zhang and Bansal, 2019;Liu et al, 2020), our heuristics are enough to obtain improvement in the downstream QA task as shown in §4.6. Some samples in our datasets are given in Table 3, showing that the diverse QA pairs are generated.…”
Section: £ ¢ ¡mentioning
confidence: 99%
“…We also chose 100 samples from SQuAD Du test . In addition to the three items proposed by Liu et al (2020), we asked annotators if an given answer is important, i.e., it is worth being asked about. We showed the workers a triple (passage, question, answer) and asked them to answer the four questions shown in Table 4.…”
Section: Human Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…As with the real-world QA, most existing works use the encoder-decoder framework that takes as input a question and a sequence of words in a given context and then generates answer words [55]- [57]. For augmenting the question and answer pairs to improve QA performance, the methods of generating questions as well as answers have also been proposed [58], [59]. However, these works mainly focus on a single domain for answer generation.…”
Section: Natural Answer Generationmentioning
confidence: 99%