Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.307
|View full text |Cite
|
Sign up to set email alerts
|

Point, Disambiguate and Copy: Incorporating Bilingual Dictionaries for Neural Machine Translation

Abstract: This paper proposes a sophisticated neural architecture to incorporate bilingual dictionaries into Neural Machine Translation (NMT) models. By introducing three novel components: Pointer, Disambiguator, and Copier, our method PDC achieves the following merits inherently compared with previous efforts:(1) Pointer leverages the semantic information from bilingual dictionaries, for the first time, to better locate source words whose translation in dictionaries can potentially be used;(2) Disambiguator synthesizes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 18 publications
0
4
0
Order By: Relevance
“…The initial embedding is based on BERT and the hyperparameter is the same as their paper. GatedGCN exploits gate diversity and the overall contextual importance scores of the words on graph convolution neural networks 4 . The dependency tree is built on Stanford CoreNLP toolkit.…”
Section: Appendix D Comparision Methods Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…The initial embedding is based on BERT and the hyperparameter is the same as their paper. GatedGCN exploits gate diversity and the overall contextual importance scores of the words on graph convolution neural networks 4 . The dependency tree is built on Stanford CoreNLP toolkit.…”
Section: Appendix D Comparision Methods Detailsmentioning
confidence: 99%
“…MOGANED code is available in https://github.com/ll0ruc/MOGANED3 DMBERT code is available in https://github.com/Bakser/DMBERT4 GatedGCN code is available in https://github.com/laiviet/ed-gated-gcn 5 EE-GCN code is available in https://github.com/cuishiyao96/eegcned6 MLBiNet code is available in https://github.com/zjunlp/DocED…”
mentioning
confidence: 99%
“…Low-Frequency Word Translation is a persisting challenge for NMT due to the token imbalance phenomenon. Conventional researches range from introducing finegrained translation units (Luong and Manning 2016;Lee, Cho, and Hofmann 2017), seeking optimal vocabulary (Wu et al 2016;Sennrich, Haddow, and Birch 2016;Gowda and May 2020;Liu et al 2021), to incorporating external lexical knowledge (Luong et al 2015;Arthur, Neubig, and Nakamura 2016;Zhang et al 2021). Recently, some approaches alleviate this problem by well-designed loss function with adaptive weights, in light of the token frequency (Gu et al 2020) or bilingual mutual information (Xu et al 2021b).…”
Section: Related Workmentioning
confidence: 99%
“…Implementation Details We examine our model based on the advanced Transformer architecture and base setting (Vaswani et al 2017). All the baseline systems and our models are implemented on top of THUMT toolkit (Zhang et al 2017). During training, the dropout rate and label smoothing are set to 0.1.…”
Section: Experimental Settingsmentioning
confidence: 99%