Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2014
DOI: 10.3115/v1/d14-1210
|View full text |Cite
|
Sign up to set email alerts
|

Latent-Variable Synchronous CFGs for Hierarchical Translation

Abstract: Data-driven refinement of non-terminal categories has been demonstrated to be a reliable technique for improving monolingual parsing with PCFGs. In this paper, we extend these techniques to learn latent refinements of single-category synchronous grammars, so as to improve translation performance. We compare two estimators for this latent-variable model: one based on EM and the other is a spectral algorithm based on the method of moments. We evaluate their performance on a Chinese-English translation task. The … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2015
2015
2017
2017

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 21 publications
0
8
0
Order By: Relevance
“…• Our model is completely monolingual and unlexicalized (does not condition its reordering on the translation) in contrast with the Latent SCFG used in (Saluja et al, 2014), • Our Latent PCFG label splits are defined as refinements of prime permutations, i.e., specifically designed for learning reordering, whereas (Saluja et al, 2014) aims at learning label splitting that helps predicting NDTs from source sentences, • Our model exploits all PETs and all derivations, both during training (latent treebank) and during inferences. In (Saluja et al, 2014) only left branching NDT derivations are used for learning the model.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…• Our model is completely monolingual and unlexicalized (does not condition its reordering on the translation) in contrast with the Latent SCFG used in (Saluja et al, 2014), • Our Latent PCFG label splits are defined as refinements of prime permutations, i.e., specifically designed for learning reordering, whereas (Saluja et al, 2014) aims at learning label splitting that helps predicting NDTs from source sentences, • Our model exploits all PETs and all derivations, both during training (latent treebank) and during inferences. In (Saluja et al, 2014) only left branching NDT derivations are used for learning the model.…”
Section: Related Workmentioning
confidence: 99%
“…If we are to choose a single PET per training instance, then learning RG from only left-branching PETs (the one usually chosen in other work, e.g. (Saluja et al, 2014)) performs slightly worse than the right-branching PET. This is possibly because English is mostly rightbranching.…”
Section: Extrinsic Evaluation In Mtmentioning
confidence: 99%
See 2 more Smart Citations
“…Learning latent-variable SCFGs for hierarchical translation is explored by Saluja et al (2014). This work uses spectral learning or the EM-algorithm to learn tensors that capture the latent variable information of rules.…”
Section: Learning Labelsmentioning
confidence: 99%