2020
DOI: 10.14712/00326585.005
|View full text |Cite
|
Sign up to set email alerts
|

Every Layer Counts: Multi-Layer Multi-Head Attention for Neural Machine Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…The suggested RPE MHA NMT model presents a strategy for analyzing the alignment of terms in the resulting translation with terms in the source sentence. This strategy is illustrated in Figure 9 by visualizing the annotation weights [29,30]. Every row of the matrix displays the weights associated with annotations, with the x-axis representing the AD sentence and the y-axis representing the resulting sentence in the MSA.…”
Section: Attention Analysismentioning
confidence: 99%
“…The suggested RPE MHA NMT model presents a strategy for analyzing the alignment of terms in the resulting translation with terms in the source sentence. This strategy is illustrated in Figure 9 by visualizing the annotation weights [29,30]. Every row of the matrix displays the weights associated with annotations, with the x-axis representing the AD sentence and the y-axis representing the resulting sentence in the MSA.…”
Section: Attention Analysismentioning
confidence: 99%
“…For some small languages and dialects, there is a shortage of relevant talents. With the help of English machine translation, the translation quality can meet the basic task requirements to make up for the lack of good and bad translators [13][14][15]. When the number of translations is small, the difference between the cost of manual translation and English machine translation is not particularly obvious.…”
Section: Introductionmentioning
confidence: 99%