Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1084
|View full text |Cite
|
Sign up to set email alerts
|

A Discriminative Neural Model for Cross-Lingual Word Alignment

Abstract: We introduce a novel discriminative word alignment model, which we integrate into a Transformer-based machine translation model. In experiments based on a small number of labeled examples (∼1.7K-5K sentences) we evaluate its performance intrinsically on both English-Chinese and English-Arabic alignment, where we achieve major improvements over unsupervised baselines (11-27 F1). We evaluate the model extrinsically on data projection for Chinese NER, showing that our alignments lead to higher performance when us… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
22
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 19 publications
(22 citation statements)
references
References 28 publications
0
22
0
Order By: Relevance
“…Li et al (2019) propose two methods to extract alignments from NMT models, however they do not outperform fast-align. Stengel-Eskin et al (2019) compute similarity matrices of encoder-decoder representations that are leveraged for word alignments, together with supervised learning, which requires manually annotated alignment. We find our proposed methods to be competitive with these approaches.…”
Section: Part-of-speech Analysismentioning
confidence: 99%
“…Li et al (2019) propose two methods to extract alignments from NMT models, however they do not outperform fast-align. Stengel-Eskin et al (2019) compute similarity matrices of encoder-decoder representations that are leveraged for word alignments, together with supervised learning, which requires manually annotated alignment. We find our proposed methods to be competitive with these approaches.…”
Section: Part-of-speech Analysismentioning
confidence: 99%
“…It formalizes word alignment as a collection of SQuAD-style span prediction problems (Rajpurkar et al, 2016) and solves them with multilingual BERT (Devlin et al, 2019). We experimentally show that our proposed model significantly outperformed both (Garg et al, 2019) and (Stengel-Eskin et al, 2019).…”
Section: Introductionmentioning
confidence: 90%
“…Most previous works that use them for word alignment (Yang et al, 2013;Tamura et al, 2014;Legrand et al, 2016) achieved accuracies that are basically comparable to GIZA++. However, the accuracy of recent works (Garg et al, 2019;Stengel-Eskin et al, 2019;Zenkel et al, 2020) based on the Transformer (Vaswani et al, 2017), which is the state-of-the art neural machine translation model, have started to outperform GIZA++. Garg et al (2019) made the attention of the Transformer more closely resembled the word alignment, and achieved better accuracy than GIZA++ when they used alignments obtained from it for supervision.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations