Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.589
|View full text |Cite
|
Sign up to set email alerts
|

Neural Conversational QA: Learning to Reason vs Exploiting Patterns

Abstract: Neural Conversational QA tasks like ShARC require systems to answer questions based on the contents of a given passage. On studying recent state-of-the-art models on the ShARC QA task, we found indications that the models learn spurious clues/patterns in the dataset. Furthermore, we show that a heuristic-based program designed to exploit these patterns can have performance comparable to that of the neural models. In this paper we share our findings about four types of patterns found in the ShARC corpus and des… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(16 citation statements)
references
References 7 publications
(6 reference statements)
0
14
0
Order By: Relevance
“…The major differences lie in two sides: 1) machines are required to formulate follow-up questions for clarification before confident enough to make the decision, 2) machines have to make a question-related conclusion by interpreting a set of complex decision rules, instead of simply extracting the answer from the text. Existing works (Zhong and Zettlemoyer, 2019;Lawrence et al, 2019;Verma et al, 2020;Gao et al, 2020a,b) have made progress in improving the reasoning ability by modeling the interactions between the rule document and other elements. As a widely-used manner, the existing models commonly extracted the rule documents into individual rule items, and track the rule fulfillment for the dialogue states.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The major differences lie in two sides: 1) machines are required to formulate follow-up questions for clarification before confident enough to make the decision, 2) machines have to make a question-related conclusion by interpreting a set of complex decision rules, instead of simply extracting the answer from the text. Existing works (Zhong and Zettlemoyer, 2019;Lawrence et al, 2019;Verma et al, 2020;Gao et al, 2020a,b) have made progress in improving the reasoning ability by modeling the interactions between the rule document and other elements. As a widely-used manner, the existing models commonly extracted the rule documents into individual rule items, and track the rule fulfillment for the dialogue states.…”
Section: Related Workmentioning
confidence: 99%
“…The corresponding rule document and the question are marked in the same color in the figure . The major challenges for the conversational machine reading include the rule document interpretation, and reasoning with the background knowledge, e.g., the provided rule document, user scenario and the input question. Existing works (Zhong and Zettlemoyer, 2019;Lawrence et al, 2019;Verma et al, 2020;Gao et al, 2020a,b) have made progress in improving the reasoning ability by modeling the interactions among rule document, user scenario and the other elements implicitly. As for rule document interpretation, most existing approaches simply split the rule document into several rule conditions to be satisfied.…”
Section: Introductionmentioning
confidence: 99%
“…The left part introduces the retrieval and tagging process for rule documents, which is then fed into the encoder together with other necessary information. Lawrence et al, 2019;Verma et al, 2020;Gao et al, 2020a,b;Ouyang et al, 2021) have made progress in modeling the matching relationships between the rule document and other elements such as user scenarios and questions. These studies are based on the hypothesis that the supporting information for answering the question is provided, which does not meet the real-world applications.…”
Section: Decision Makingmentioning
confidence: 99%
“…Existing studies deal with decision making and question generation independently (Zhong and Zettlemoyer, 2019;Lawrence et al, 2019;Verma et al, 2020;Gao et al, 2020a,b), and use hard-label decisions to activate question generation. These methods inevitably suffer from error propagation if the model makes the wrong decisions.…”
Section: Double-channel Decodermentioning
confidence: 99%
See 1 more Smart Citation