Proceedings of the Workshop on Machine Reading for Question Answering 2018
DOI: 10.18653/v1/w18-2607
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic Classification of Knowledge, Reasoning, and Context within the ARC Dataset

Abstract: The recent work of introduces the AI2 Reasoning Challenge (ARC) and the associated ARC dataset that partitions open domain, complex science questions into an Easy Set and a Challenge Set. That paper includes an analysis of 100 questions with respect to the types of knowledge and reasoning required to answer them; however, it does not include clear definitions of these types, nor does it offer information about the quality of the labels. We propose a comprehensive set of definitions of knowledge and reasoning … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
28
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(29 citation statements)
references
References 34 publications
1
28
0
Order By: Relevance
“…Many systems answer science questions by composing multiple facts from semi-structured and unstructured knowl-edge sources (Khashabi et al 2016;Khot, Sabharwal, and Clark 2017;Jansen et al 2017;Khashabi et al 2018b). However, these often require careful manual tuning due to the large variety of reasoning techniques needed for these questions (Boratko et al 2018) and the large number of facts that often must be composed together (Jansen 2018;Jansen et al 2016). By limiting QASC to require exactly 2 hops (thereby avoiding semantic drift issues with longer paths (Fried et al 2015;Khashabi et al 2019)) and explicitly annotating these hops, we hope to constrain the problem enough so as to enable the development of supervised models for identifying and composing relevant knowledge.…”
Section: Comparison With Existing Datasetsmentioning
confidence: 99%
“…Many systems answer science questions by composing multiple facts from semi-structured and unstructured knowl-edge sources (Khashabi et al 2016;Khot, Sabharwal, and Clark 2017;Jansen et al 2017;Khashabi et al 2018b). However, these often require careful manual tuning due to the large variety of reasoning techniques needed for these questions (Boratko et al 2018) and the large number of facts that often must be composed together (Jansen 2018;Jansen et al 2016). By limiting QASC to require exactly 2 hops (thereby avoiding semantic drift issues with longer paths (Fried et al 2015;Khashabi et al 2019)) and explicitly annotating these hops, we hope to constrain the problem enough so as to enable the development of supervised models for identifying and composing relevant knowledge.…”
Section: Comparison With Existing Datasetsmentioning
confidence: 99%
“…LoBue and Yates (2011) and Sammons et al (2010) analyzed entailment phenomena using detailed classifications in RTE. For the ARC dataset, Boratko et al (2018) proposed knowledge and reasoning types.…”
Section: Related Workmentioning
confidence: 99%
“…Similarly, we leverage this dataset to build an essential term selector using a neural network-based algorithm. More recently, Boratko et al (2018) developed a labeling interface to obtain high quality labels for the ARC dataset. One finding is that human annotators tend to retrieve better evidence after they reformulate the search queries which are originally constructed by a simple concatenation of question and answer choice.…”
Section: Related Workmentioning
confidence: 99%