2018
DOI: 10.48550/arxiv.1809.03275
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multilingual Extractive Reading Comprehension by Runtime Machine Translation

Abstract: Despite recent work in Reading Comprehension (RC), progress has been mostly limited to English due to the lack of large-scale datasets in other languages. In this work, we introduce the first RC system for languages without RC training data. Given a target language without RC training data and a pivot language with RC training data (e.g. English), our method leverages existing RC resources in the pivot language by combining a competitive RC model in the pivot language with an attentive Neural Machine Translati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
35
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 36 publications
(36 citation statements)
references
References 30 publications
1
35
0
Order By: Relevance
“…we still often see improvement over the baseline, but the improvement is always less than when using XLDA only over the context channel. This is in keeping with the findings of Asai et al (2018), which show that the ability to correctly translate questions is crucial for question answering. In other words, SQuAD is extremely sensitive to the translation quality of the question, and it is not surprising that machine translations of the questions are less effective than translating the context, which is less sensitive.…”
Section: Xlda For Squadsupporting
confidence: 89%
“…we still often see improvement over the baseline, but the improvement is always less than when using XLDA only over the context channel. This is in keeping with the findings of Asai et al (2018), which show that the ability to correctly translate questions is crucial for question answering. In other words, SQuAD is extremely sensitive to the translation quality of the question, and it is not surprising that machine translations of the questions are less effective than translating the context, which is less sensitive.…”
Section: Xlda For Squadsupporting
confidence: 89%
“…However, all this work focuses on English attacks. Further, although many multilingual QA datasets exist (He et al, 2017;Asai et al, 2018;Mozannar et al, 2019;Artetxe et al, 2020;Lewis et al, 2020), no prior work has explored adversarial evaluation and exposed vulnerabilities over large pre-trained multi-lingual language models. .…”
Section: Related Workmentioning
confidence: 99%
“…Unlike NBC and Fox, CBS does not have a Spanish-language outlet of its own that could broadcast the game (though per league policy, a separate Spanish play-by-play call was carried on CBS's second audio program channel for over-the-air viewers). […] 3 . We here evaluate the performance of the fine-tuned multilingual BERT on them and compare the results to a baseline [3].…”
Section: Fine-tuning Bert On English Squadmentioning
confidence: 99%