2013
DOI: 10.1007/978-3-642-40802-1_29
|View full text |Cite
|
Sign up to set email alerts
|

QA4MRE 2011-2013: Overview of Question Answering for Machine Reading Evaluation

Abstract: This paper describes the methodology for testing the performance of Machine Reading systems through Question Answering and Reading Comprehension Tests. This was the attempt of the QA4MRE challenge which was run as a Lab at CLEF 2011-2013. The traditional QA task was replaced by a new Machine Reading task, whose intention was to ask questions that required a deep knowledge of individual short texts and in which systems were required to choose one answer, by analysing the corresponding test document in conjuncti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
24
0
2

Year Published

2013
2013
2019
2019

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 39 publications
(26 citation statements)
references
References 8 publications
0
24
0
2
Order By: Relevance
“…Inherently, answers cannot be derived from external resources for this type of question. Although such questions may be reduced into textual entailment between a given text and choices Kasahara et al 2010;Peñas et al 2011], these cases were excluded from the present study.…”
Section: Reading Comprehensionmentioning
confidence: 99%
See 1 more Smart Citation
“…Inherently, answers cannot be derived from external resources for this type of question. Although such questions may be reduced into textual entailment between a given text and choices Kasahara et al 2010;Peñas et al 2011], these cases were excluded from the present study.…”
Section: Reading Comprehensionmentioning
confidence: 99%
“…Third, the performance of entailment recognition can be mapped to the test scores, which are comparable with human performance for the same task. Dagan et al [2006], Kasahara et al [2010], and Peñas et al [2011] investigated the concept of using reading comprehension tests for the evaluation of textual entailment recognition. They considered reading comprehension tests to be intended to test human ability to compute entailment relations, and they considered such tests to be applicable to the evaluation of automatic systems.…”
Section: Textual Entailment Recognitionmentioning
confidence: 99%
“…These clues may include Tang poem and Song iambic verse, domainspecific expressions, even some mixture of mod- ern Chinese and excerpt from ancient books and etc. The dependence of background knowledge makes the models that are designed for reading comprehension such as (Peñas et al, 2013;Richardson et al, 2013) fail. Thirdly, the diversity of candidates' granularity, i.e.…”
Section: Introductionmentioning
confidence: 99%
“…Оценка качества работы таких систем осуществляется на тестовых данных, разрабо-танных обществом экспертов и представленных в виде фрагмента текста, вопроса по этому тексту и предполагаемого ответа. В случае, если система по запросу выдает результат, не соответствующий пред-полагаемому ответу, результат не засчитывается.В 2012 году проводились соревнования по качеству работы вопросно-ответных систем в области биомедицины, наилучший результат (с точностью 55 %) показала система [2]. Отечественных систем, способных составить конкуренцию зарубежным системам по качеству работы, на данный момент не су-ществует.…”
unclassified
“…В 2012 году проводились соревнования по качеству работы вопросно-ответных систем в области биомедицины, наилучший результат (с точностью 55 %) показала система [2]. Отечественных систем, способных составить конкуренцию зарубежным системам по качеству работы, на данный момент не су-ществует.…”
unclassified