Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1226
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Pre-Trained Language Representations with Rich Knowledge for Machine Reading Comprehension

Abstract: Machine reading comprehension (MRC) is a crucial and challenging task in NLP. Recently, pre-trained language models (LMs), especially BERT, have achieved remarkable success, presenting new state-of-the-art results in MRC. In this work, we investigate the potential of leveraging external knowledge bases (KBs) to further improve BERT for MRC. We introduce KT-NET, which employs an attention mechanism to adaptively select desired knowledge from KBs, and then fuses selected knowledge with BERT to enable context-and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
64
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 122 publications
(67 citation statements)
references
References 27 publications
0
64
0
Order By: Relevance
“…Table 4 shows EM (%) and F 1 (%) of human performance, the PSH-SJTU system as well as baselines on the development and test sets of task 2. Compared with the best baseline, KT-NET (Yang et al, 2019a), PSH-SJTU achieves significantly better scores. On the hidden test set, they improve EM by 10.08%, and F 1 by 8.98%.…”
Section: Participantsmentioning
confidence: 94%
See 1 more Smart Citation
“…Table 4 shows EM (%) and F 1 (%) of human performance, the PSH-SJTU system as well as baselines on the development and test sets of task 2. Compared with the best baseline, KT-NET (Yang et al, 2019a), PSH-SJTU achieves significantly better scores. On the hidden test set, they improve EM by 10.08%, and F 1 by 8.98%.…”
Section: Participantsmentioning
confidence: 94%
“…KT-NET (Yang et al, 2019a) employs an attention mechanism to adaptively select desired knowledge from knowledge bases, and then fuses selected knowledge with BERT to enable contextand knowledge-aware predictions for machine reading comprehension. (Seo et al, 2016) and self-attention, both of which are widely used in MC models.…”
Section: Task 2 Baselinesmentioning
confidence: 99%
“…The first one, mutual attention, is aimed at fusing the question representations into the passage so as to obtain the question-aware passage representations; the second one, self-attention, is aimed at fusing the question-aware passage representations into themselves so as to obtain the final passage representations. Yang et al [35] proposed KT-NET, which employs an attention mechanism to adaptively select desired knowledge from KBs, and then fuses selected knowledge with BERT to enable context-and knowledge-aware predictions. However, most previous methods retrieve the relevant knowledge base before encoding it into the MRC, which can only refer to KB locally.…”
Section: Knowledge Fusionmentioning
confidence: 99%
“…This has led many to leverage knowledge graphs (KGs) (Mihaylov and Frank, 2018;Lin et al, 2019;Yang et al, 2019). KGs represent relational knowledge between entities with multi-relational edges for models to acquire.…”
Section: Introductionmentioning
confidence: 99%