Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing - EMNLP '06 2006
DOI: 10.3115/1610075.1610084
|View full text |Cite
|
Sign up to set email alerts
|

Phrasetable smoothing for statistical machine translation

Abstract: We discuss different strategies for smoothing the phrasetable in Statistical MT, and give results over a range of translation settings. We show that any type of smoothing is a better idea than the relativefrequency estimates that are often used. The best smoothing techniques yield consistent gains of approximately 1% (absolute) according to the BLEU metric.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
42
0

Year Published

2007
2007
2016
2016

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 50 publications
(48 citation statements)
references
References 18 publications
(20 reference statements)
2
42
0
Order By: Relevance
“…log probability), phrase translation and lexical translation scores, word and phrase penalties, and a linear distortion score. The phrase translation probabilities are smoothed with Good-Turing smoothing (Foster et al, 2006). We used the hierarchical lexicalized reordering model (Galley and Manning, 2008) with 4 possible orientations (monotone, swap, discontinuous left and discontinuous right) in both left-to-right and right-to-left direction.…”
Section: Baseline Featuresmentioning
confidence: 99%
“…log probability), phrase translation and lexical translation scores, word and phrase penalties, and a linear distortion score. The phrase translation probabilities are smoothed with Good-Turing smoothing (Foster et al, 2006). We used the hierarchical lexicalized reordering model (Galley and Manning, 2008) with 4 possible orientations (monotone, swap, discontinuous left and discontinuous right) in both left-to-right and right-to-left direction.…”
Section: Baseline Featuresmentioning
confidence: 99%
“…The core features of our model are a 5-gram LM score, phrase translation and lexical translation scores, word and phrase penalties, and a linear distortion score. The phrase translation probabilities are smoothed with GoodTuring smoothing (Foster et al, 2006). We used the hierarchical lexicalized reordering model (Galley and Manning, 2008) with 4 possible orientations (monotone, swap, discontinuous left and discontinuous right) in both left-to-right and rightto-left direction.…”
Section: Baseline Featuresmentioning
confidence: 99%
“…Smoothing is still an open problem in SMT (Foster et al, 2006) and SFSTs are not an exception: they are not completely smoothed just by smoothing the corresponding SFSAs. Future research on SFST smoothing should be based on specific criteria for SFSTs as well.…”
Section: Smoothingmentioning
confidence: 99%