Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2016
DOI: 10.48550/arxiv.1611.09830
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

NewsQA: A Machine Comprehension Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
111
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 82 publications
(137 citation statements)
references
References 0 publications
2
111
0
1
Order By: Relevance
“…In principle, as a dual task of QA, any QA datasets can be used for QG [50]. SQuAD [58], MS-MARCO [4] and newsQA [73] are three famous datasets used for answer-extraction QG, collected from Wikipedia, Bing search logs, and CNN news respectively. Unlike the previous three datasets, Nar-rativeQA [35] does not restrict the answers to be the span of texts in the articles, therefore, it can be used as an answer-abstraction QG dataset.…”
Section: Related Work 21 Question Generationmentioning
confidence: 99%
“…In principle, as a dual task of QA, any QA datasets can be used for QG [50]. SQuAD [58], MS-MARCO [4] and newsQA [73] are three famous datasets used for answer-extraction QG, collected from Wikipedia, Bing search logs, and CNN news respectively. Unlike the previous three datasets, Nar-rativeQA [35] does not restrict the answers to be the span of texts in the articles, therefore, it can be used as an answer-abstraction QG dataset.…”
Section: Related Work 21 Question Generationmentioning
confidence: 99%
“…The famous SQuAD dataset [27,26] for the first time introduces human-generated free-form questions, which requires the machine to understand natural language to select the correct span in Wikipedia pages. Similar datasets follow this trend of using free-form questions and adopt reading documents from a variety of sources, such as news articles [37,17] and dialogues [18,28,4]. In addition to these datasets where the answers can be directly extracted from the document, another popular type of datasets, i.e., abstractive datasets, ask the reader to generate an answer that may not be found in the given context [23,12].…”
Section: Related Workmentioning
confidence: 99%
“…The original version of SQuAD [5] was published in 2016. There are also a variety of benchmarks related to question-answering -NewsQA [12], SearchQA [13], TriviaQA [14], HotpotQA [15], and Natural Questions [16]. Nonetheless, SQuAD is one of the most commonly used question-answering benchmarks.…”
Section: Related Workmentioning
confidence: 99%