Proceedings of the 22nd International Conference on Computational Linguistics - COLING '08 2008
DOI: 10.3115/1599081.1599225
|View full text |Cite
|
Sign up to set email alerts
|

A systematic comparison of phrase-based, hierarchical and syntax-augmented statistical MT

Abstract: Probabilistic synchronous context-free grammar (PSCFG) translation models define weighted transduction rules that represent translation and reordering operations via nonterminal symbols. In this work, we investigate the source of the improvements in translation quality reported when using two PSCFG translation models (hierarchical and syntax-augmented), when extending a state-of-the-art phrasebased baseline that serves as the lexical support for both PSCFG models. We isolate the impact on translation quality f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
26
1

Year Published

2009
2009
2012
2012

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(28 citation statements)
references
References 14 publications
1
26
1
Order By: Relevance
“…A significant improvement in English-Danish machine translation has been achieved [8]. The phrase level representation of source language, converting the source language phrases into target language phrases and reordering the target language phrases generate fluent target output [9]. Phrase based representation and translation develops better Chinese-English, Arabic-English and Urdu-English machine translation systems [9].…”
Section: Knowledge Representation Of Urdu Text and Predicate Logicmentioning
confidence: 99%
“…A significant improvement in English-Danish machine translation has been achieved [8]. The phrase level representation of source language, converting the source language phrases into target language phrases and reordering the target language phrases generate fluent target output [9]. Phrase based representation and translation develops better Chinese-English, Arabic-English and Urdu-English machine translation systems [9].…”
Section: Knowledge Representation Of Urdu Text and Predicate Logicmentioning
confidence: 99%
“…These translation models comprise translationally equivalent sequences of words-so-called phrase pairsthat are extracted from aligned sentence pairs using heuristics over a statistical word alignment. While phrase-based models have achieved state-of-the-art translation quality, evidence suggests there is a limit as to what can be accomplished using only simple phrases, for example, satisfactory capturing of context-sensitive reordering phenomena between language pairs (Zollmann et al 2008). This assertion has been acknowledged within the field as illustrated by the recent shift in focus towards more linguistically motivated models.…”
Section: Introductionmentioning
confidence: 96%
“…If the rule set can be reduced without reducing translation quality, both memory efficiency and translation speed can be increased. Previously published approaches to reducing the rule set include: enforcing a minimum span of two words per non-terminal (Lopez, 2008), which would reduce our set to 115M rules; or a minimum count (mincount) threshold (Zollmann et al, 2008), which would reduce our set to 78M (mincount=2) or 57M (mincount=3) rules. Shen et al (2008) describe the result of filtering rules by insisting that target-side rules are well-formed dependency trees.…”
Section: Rule Filtering By Patternmentioning
confidence: 99%
“…As another reference point, Chiang (2007) reports Chinese-to-English translation experiments based on 5.5M rules. Zollmann et al (2008) report that filtering rules en masse leads to degradation in translation performance. Rather than apply a coarse filtering, such as a mincount for all rules, we follow a more syntactic approach and further classify our rules according to their pattern and apply different filters to each pattern depending on its value in translation.…”
Section: Rule Filtering By Patternmentioning
confidence: 99%
See 1 more Smart Citation