Proceedings of the 2nd Workshop on Machine Reading for Question Answering 2019
DOI: 10.18653/v1/d19-5827
|View full text |Cite
|
Sign up to set email alerts
|

Generalizing Question Answering System with Pre-trained Language Model Fine-tuning

Abstract: With a large number of datasets being released and new techniques being proposed, Question answering (QA) systems have witnessed great breakthroughs in reading comprehension (RC) tasks. However, most existing methods focus on improving in-domain performance, leaving open the research question of how these models and techniques can generalize to out-ofdomain and unseen RC tasks. To enhance the generalization ability, we propose a multi-task learning framework that learns the shared representation across differe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
60
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 47 publications
(62 citation statements)
references
References 26 publications
0
60
0
Order By: Relevance
“…Harbin Institute of Technology HLTC (Su et al, 2019) Hong Kong University of Science & Technology BERT-cased-whole-word Aristo @ AI2 CLER (Takahashi et al, 2019) Fuji Xerox Co., Ltd. Adv. Train (Lee et al, 2019) 42Maru and Samsung Research BERT-Multi-Finetune Beijing Language and Culture University PAL IN DOMAIN University of California Irvine HierAtt (Osama et al, 2019) Alexandria University (Longpre et al, 2019) 82.3 68.5 66.9 74.6 70.8 FT XLNet 82.9 68.0 66.7 74.4 70.5 HLTC (Su et al, 2019) 81.0 65.9 65.0 72.9 69.0 BERT-cased-whole-word 79.4 61.1 61.4 71.2 66.3 CLER (Takahashi et al, 2019) 80.2 62.7 62.5 69.7 66.1 Adv. Train (Lee et al, 2019) 76.8 57.1 57.9 66.5 62.…”
Section: Ft Xlnetmentioning
confidence: 99%
See 2 more Smart Citations
“…Harbin Institute of Technology HLTC (Su et al, 2019) Hong Kong University of Science & Technology BERT-cased-whole-word Aristo @ AI2 CLER (Takahashi et al, 2019) Fuji Xerox Co., Ltd. Adv. Train (Lee et al, 2019) 42Maru and Samsung Research BERT-Multi-Finetune Beijing Language and Culture University PAL IN DOMAIN University of California Irvine HierAtt (Osama et al, 2019) Alexandria University (Longpre et al, 2019) 82.3 68.5 66.9 74.6 70.8 FT XLNet 82.9 68.0 66.7 74.4 70.5 HLTC (Su et al, 2019) 81.0 65.9 65.0 72.9 69.0 BERT-cased-whole-word 79.4 61.1 61.4 71.2 66.3 CLER (Takahashi et al, 2019) 80.2 62.7 62.5 69.7 66.1 Adv. Train (Lee et al, 2019) 76.8 57.1 57.9 66.5 62.…”
Section: Ft Xlnetmentioning
confidence: 99%
“…Within these restrictions, we encouraged participants to explore how to best utilize the provided data. Inspired by Talmor and Berant (2019), two submissions (Su et al, 2019;Longpre et al, 2019) analyzed similarities between datasets. Unsurprisingly, the performance improved significantly when fine-tuned on the training dataset most similar to the evaluation dataset of interest.…”
Section: Summary Of Findingsmentioning
confidence: 99%
See 1 more Smart Citation
“…We combine all the bias weights and use them to adapt the distillation loss. Su et al (2019) achieve considerable improvements by simply fine-tuning XLNet instead of BERT, and Longpre et al (2019) achieve further improvements by augmenting the training data with additional unanswerable questions.…”
Section: Related Workmentioning
confidence: 99%
“…Existing approaches to improve generalization in QA either are only applicable when there exist multiple training domains (Talmor and Berant, 2019;Takahashi et al, 2019; or rely on models and ensembles with larger capacity (Longpre et al, 2019;Su et al, 2019;. In contrast, our novel debiasing approach can be applied to both single and multi-domain scenarios, and it improves the model generalization without requiring larger pre-trained language models.…”
Section: Introductionmentioning
confidence: 99%