Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing 2022
DOI: 10.18653/v1/2022.emnlp-main.754
|View full text |Cite
|
Sign up to set email alerts
|

Sampling-Based Approximations to Minimum Bayes Risk Decoding for Neural Machine Translation

Abstract: In NMT we search for the mode of the model distribution to form predictions. The mode and other high-probability translations found by beam search have been shown to often be inadequate in a number of ways. This prevents improving translation quality through better search, as these idiosyncratic translations end up selected by the decoding algorithm, a problem known as the beam search curse. Recently, an approximation to minimum Bayes risk (MBR) decoding has been proposed as an alternative decision rule that w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(11 citation statements)
references
References 43 publications
(76 reference statements)
0
6
0
Order By: Relevance
“…For MBR, we generate 1024 samples per segment using epsilon sampling and re-use the same samples as references. While this approach does not guarantee that the estimation of the expected utility is unbiased (Eikema and Aziz, 2022), it has empirically been found to work well (Freitag et al, 2023).…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…For MBR, we generate 1024 samples per segment using epsilon sampling and re-use the same samples as references. While this approach does not guarantee that the estimation of the expected utility is unbiased (Eikema and Aziz, 2022), it has empirically been found to work well (Freitag et al, 2023).…”
Section: Methodsmentioning
confidence: 99%
“…A line of research has focused on improving the efficiency of sampling-based MBR. Eikema and Aziz (2022) propose coarse-to-fine MBR, which prunes the hypotheses based on a cheaper metric, and N-by-S MBR, which uses fewer references than hypotheses. Cheng and Vlachos (2023) propose confidence-based pruning, where the number of hypotheses is iteratively reduced based on an increasing number of references.…”
Section: Background and Related Workmentioning
confidence: 99%
See 3 more Smart Citations