Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.583
|View full text |Cite
|
Sign up to set email alerts
|

Will this Question be Answered? Question Filtering via Answer Model Distillation for Efficient Question Answering

Abstract: In this paper we propose a novel approach towards improving the efficiency of Question Answering (QA) systems by filtering out questions that will not be answered by them. This is based on an interesting new finding: the answer confidence scores of state-of-the-art QA systems can be approximated well by models solely using the input question text. This enables preemptive filtering of questions that are not answered by the system due to their answer confidence scores being lower than the system threshold. Speci… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
3
1

Relationship

2
8

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 55 publications
(52 reference statements)
0
9
0
Order By: Relevance
“…The softmax probability assigned to class '1' by this calibrator is used as the confidence estimator for selective prediction. We refer to this approach as Calib C. We also train a transformer-based model for calibration (Calib T) that leverages the entire input text for this classifi-cation task instead of the syntactic features (Garg and Moschitti, 2021).…”
Section: Calibrationmentioning
confidence: 99%
“…The softmax probability assigned to class '1' by this calibrator is used as the confidence estimator for selective prediction. We refer to this approach as Calib C. We also train a transformer-based model for calibration (Calib T) that leverages the entire input text for this classifi-cation task instead of the syntactic features (Garg and Moschitti, 2021).…”
Section: Calibrationmentioning
confidence: 99%
“…These confidence scores can be used in a "selective QA" setting (Kamath et al, 2020), where the model can abstain on a certain fraction of questions where it assigns low confidence to its answers. We use the area under coverage-accuracy curve(AUC) to evaluate how well a model is calibrated as in past literature (Kamath et al, 2020;Garg and Moschitti, 2021;Ye and Durrett, 2022). The curve plots the average accuracy with varying fractions (coverage) of questions being answered (examples in Figure 5).…”
Section: Calibrating Advhotpotmentioning
confidence: 99%
“…The performance of AS2 systems in practical applications is typically (Garg and Moschitti, 2021) measured using the Accuracy in providing correct answers for the questions (the percentage of correct responses provided by the system), also called the Precision-at-1 (P@1). In addition to P@1, we use Mean Average Precision (MAP) and Mean Reciprocal Recall (MRR) to evaluate the ranking produced of the set of candidates by the model.…”
Section: B3 Metricsmentioning
confidence: 99%