2021
DOI: 10.48550/arxiv.2106.03193
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The FLORES-101 Evaluation Benchmark for Low-Resource and Multilingual Machine Translation

Abstract: One of the biggest challenges hindering progress in low-resource and multilingual machine translation is the lack of good evaluation benchmarks. Current evaluation benchmarks either lack good coverage of low-resource languages, consider only restricted domains, or are low quality because they are constructed using semi-automatic procedures. In this work, we introduce the FLORES-101 evaluation benchmark, consisting of 3001 sentences extracted from English Wikipedia and covering a variety of different topics and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 26 publications
0
5
0
Order By: Relevance
“…For source transcripts, we reuse the transcripts produced by human translators from [20]. We maintain the English translated transcripts, which are useful for tasks such as multi-modal speech translation evaluations.…”
Section: Textual Datamentioning
confidence: 99%
See 1 more Smart Citation
“…For source transcripts, we reuse the transcripts produced by human translators from [20]. We maintain the English translated transcripts, which are useful for tasks such as multi-modal speech translation evaluations.…”
Section: Textual Datamentioning
confidence: 99%
“…In machine translation, the release of new benchmarks like FLoRes-101 [20] has enabled advances in publicly available massively multilingual machine translation systems [21]. With FLEURS, we hope to provide a resource that could catalyze research towards building massively multilingual speech and text representations and their evaluation on a variety of tasks.…”
Section: Introductionmentioning
confidence: 99%
“…At first, we translate the ChAII dataset from Hindi and Tamil to English and then to Bengali, Marathi, Malayalam, and Telugu. In the FLORES devset benchmark (Goyal et al, 2021), the BLEU scores of IndicTrans for translating Hindi and Tamil to English are 37.9 and 28.6, respectively. The scores (Radford et al, 2021) for translating English to Bengali, Marathi, Malayalam,and Telugu are 20.3,16.1,16.3,and 22.0, respectively.…”
Section: Translation and Transliteration Detailsmentioning
confidence: 99%
“…Accompanied by the increase of publicly released parallel corpus such as FLORES-101 (Goyal et al, 2021) and AI hub, the importance of evaluating and improving the quality of the parallel corpus becomes higher. Especially for the data construction process, assessing the quality of the corpus is regarded as an essential process.…”
Section: Parallel Corpus Quality Assessmentmentioning
confidence: 99%