2020
DOI: 10.48550/arxiv.2010.04898
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Open-Domain Question Answering Goes Conversational via Question Rewriting

Abstract: We introduce a new dataset for Question Rewriting in Conversational Context (QReCC), which contains 14K conversations with 81K question-answer pairs. The task in QReCC is to find answers to conversational questions within a collection of 10M web pages (split into 54M passages). Answers to questions in the same conversation may be distributed across several web pages. QReCC provides annotations that allow us to train and evaluate individual subtasks of question rewriting, passage retrieval and reading comprehen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(17 citation statements)
references
References 10 publications
0
11
0
Order By: Relevance
“…Following previous work, we perform both intrinsic and extrinsic evaluation [2,8]. In intrinsic evaluation, we compare rewrites produced by QR methods with manual rewrites produced by human annotators using ROUGE-1 Precision (P), Recall (R) and F-measure (F) [2].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Following previous work, we perform both intrinsic and extrinsic evaluation [2,8]. In intrinsic evaluation, we compare rewrites produced by QR methods with manual rewrites produced by human annotators using ROUGE-1 Precision (P), Recall (R) and F-measure (F) [2].…”
Section: Methodsmentioning
confidence: 99%
“…Following previous work, we perform both intrinsic and extrinsic evaluation [2,8]. In intrinsic evaluation, we compare rewrites produced by QR methods with manual rewrites produced by human annotators using ROUGE-1 Precision (P), Recall (R) and F-measure (F) [2]. 4 In extrinsic evaluation, we measure PR performance when using different QR methods using standard ranking metrics: NDCG@3, MRR and Recall@1000.…”
Section: Methodsmentioning
confidence: 99%
“…Both open-domain QA and ad hoc search datasets are frequently used as additional (weak) supervision data in conversational search. Recently, the QReCC dataset [Anantha et al, 2020] is released as yet another conversational search benchmark. It extends CAsT and QuAC with additional conversations constructed from Natural Questions and additional passages sampled from Common Crawl.…”
Section: Other Related Resourcesmentioning
confidence: 99%
“…The approach has a main shortcoming, namely, it introduces a great amount of noise, since not everything in the previous turns is relevant. An alternative approach is Question Rewriting (QR), in which the question is rewritten in a self-contained form based on the previous conversational information (Vakulenko et al, 2021a;Anantha et al, 2020). QR selects only the relevant information in previous turns, thus improving over concatenation.…”
Section: Introductionmentioning
confidence: 99%