Proceedings of the Fourth Workshop on Statistical Machine Translation - StatMT '09 2009
DOI: 10.3115/1626431.1626461
|View full text |Cite
|
Sign up to set email alerts
|

The University of Maryland statistical machine translation system for the Fourth Workshop on Machine Translation

Abstract: This paper describes the techniques we explored to improve the translation of news text in the German-English and Hungarian-English tracks of the WMT09 shared translation task. Beginning with a convention hierarchical phrase-based system, we found benefits for using word segmentation lattices as input, explicit generation of beginning and end of sentence markers, minimum Bayes risk decoding, and incorporation of a feature scoring the alignment of function words in the hypothesized translation. We also explored… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(15 citation statements)
references
References 20 publications
0
15
0
Order By: Relevance
“…In previous shared tasks of WMT, there have been submissions that use other metrics for tuning (e.g. Dyer et al 2009) in order to achieve higher correlation with human judgment. In our experiments, however, tuning on Meteor or Meteor-SCP can be better than tuning on Bleu even if our aim is to obtain a higher Bleu score (cf.…”
Section: Results and Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…In previous shared tasks of WMT, there have been submissions that use other metrics for tuning (e.g. Dyer et al 2009) in order to achieve higher correlation with human judgment. In our experiments, however, tuning on Meteor or Meteor-SCP can be better than tuning on Bleu even if our aim is to obtain a higher Bleu score (cf.…”
Section: Results and Analysismentioning
confidence: 99%
“…He, A. Way also been some efforts to tune against other criteria, such as Bleu鈭扵er 2 (Dyer et al 2009) or the IQ MT metric (Lambert et al 2006).…”
Section: Introductionmentioning
confidence: 99%
“…Although sparse features are useful, training of sparse features is an extremely difficult optimization problem, and at this point there is still no method that has been widely demonstrated as being able to robustly estimate the parameters of millions of features. Because of this, a third approach of first training the parameters of sparse features, then condensing the sparse features into dense features and performing one more optimization pass (potentially with a different algorithm), has been widely used in a large number of research papers and systems (Dyer et al 2009;He and Deng 2012;Flanigan, Dyer, and Carbonell 2013;Setiawan and Zhou 2013). A dense feature created from a large group of sparse features and their weights is generally called a summary feature, and can be expressed as follows h sum ( f , e, d) = w sparse h sparse ( f , e, d)…”
Section: Summary Featuresmentioning
confidence: 99%
“…Finally, there is also some work on optimizing multiple evaluation metrics at one time. The easiest way to do so is to simply use the linear interpolation of two or more metrics as the error function (Dyer et al 2009; He and Way 2009; Servan and Schwenk 2011):…”
Section: Evaluation Measures and Optimizationmentioning
confidence: 99%
“…We begin by adapting the lattice technique of Dyer et al (2009) to Finnish. We train a standard phrasebased machine translation model on a new corpus: on the source side we concatenate the original data with its one-best segmentation according to a Morfessor (Creutz and Lagus, 2007) model trained on the original data, and on the target side we simply concatenate it with itself.…”
Section: Finnish Tokenization Using Morfessor and Word-latticesmentioning
confidence: 99%