Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.147
|View full text |Cite
|
Sign up to set email alerts
|

SimAlign: High Quality Word Alignments Without Parallel Training Data Using Static and Contextualized Embeddings

Abstract: Word alignments are useful for tasks like statistical and neural machine translation (NMT) and cross-lingual annotation projection. Statistical word aligners perform well, as do methods that extract alignments jointly with translations in NMT. However, most approaches require parallel training data, and quality decreases as less training data is available. We propose word alignment methods that require no parallel data. The key idea is to leverage multilingual word embeddings -both static and contextualized -f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
39
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 45 publications
(41 citation statements)
references
References 50 publications
2
39
0
Order By: Relevance
“…It can be observed that XLM-ALIGN consistently improves the results over XLM-R base across these layers. Moreover, it shows a parabolic trend across the layers of XLM-R base , which is consistent with the results in (Jalili Sabet et al, 2020). In contrast to XLM-R base , XLM-ALIGN alleviates this trend and greatly reduces AER in the last few layers.…”
Section: Word Alignmentsupporting
confidence: 89%
“…It can be observed that XLM-ALIGN consistently improves the results over XLM-R base across these layers. Moreover, it shows a parabolic trend across the layers of XLM-R base , which is consistent with the results in (Jalili Sabet et al, 2020). In contrast to XLM-R base , XLM-ALIGN alleviates this trend and greatly reduces AER in the last few layers.…”
Section: Word Alignmentsupporting
confidence: 89%
“…We would like to point out that parallel to the present work, Sabet et al (2020) also introduced the first two of the four methods. Since they aim to extract an explicit alignment between source and target they do not construct a score for a sentence pair and do not consider the use in a data filtering task.…”
Section: Source ↔ Target Embedding Similaritymentioning
confidence: 99%
“…For a long time, IBM-model-based frameworks like GIZA++ (Och and Ney, 2003) or fastalign (Dyer et al, 2013) produced the best word alignments. However, recently Sabet et al (2020) report equally good results by using a word similarity matrix calculated from cross-lingual word embeddings.…”
Section: Related Workmentioning
confidence: 99%
“…Statistical models such as IBM models (Brown et al, 1993), Giza++ (Och andNey, 2003), fast-align (Dyer et al, 2013) and Eflomal (Östling and Tiedemann, 2016b) are widely used. Recently, neural models were proposed, such as SimAlign (Jalili Sabet et al, 2020), Awesome-align (Dou and Neubig, 2021), and methods that are based on neural machine translation (Garg et al, 2019;Zenkel et al, 2020). We use Eflomal and SimAlign for generating alignments.…”
Section: Related Workmentioning
confidence: 99%