Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2015
DOI: 10.3115/v1/n15-1095
|View full text |Cite
|
Sign up to set email alerts
|

Joint Generation of Transliterations from Multiple Representations

Abstract: Machine transliteration is often referred to as phonetic translation.We show that transliterations incorporate information from both spelling and pronunciation, and propose an effective model for joint transliteration generation from both representations. We further generalize this model to include transliterations from other languages, and enhance it with reranking and lexicon features. We demonstrate significant improvements in transliteration accuracy on several datasets.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2018
2018

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 14 publications
(14 reference statements)
0
6
0
Order By: Relevance
“…The data and the dictionaries can be accessed via https://www.hucompute.org/ressourcen/corpora so that our findings may be used as a starting point for related research. In future work, we intend to investigate more sophisticated combination techniques for combining outputs of several spell checkers, e.g., on the character-level, as done in Cortes et al (2014); Eger (2015d,c); Yao and Kondrak (2015). We also intend to evaluate neural-network based techniques in the present scenario (Sutskever et al, 2014;Yao and Zweig, 2015).…”
Section: Resultsmentioning
confidence: 99%
“…The data and the dictionaries can be accessed via https://www.hucompute.org/ressourcen/corpora so that our findings may be used as a starting point for related research. In future work, we intend to investigate more sophisticated combination techniques for combining outputs of several spell checkers, e.g., on the character-level, as done in Cortes et al (2014); Eger (2015d,c); Yao and Kondrak (2015). We also intend to evaluate neural-network based techniques in the present scenario (Sutskever et al, 2014;Yao and Zweig, 2015).…”
Section: Resultsmentioning
confidence: 99%
“…The JOINT system was trained on a held-out set composed of the outputs of the base systems generated for each source word. The second set of experiments followed the original design of Yao and Kondrak (2015), in which the supplemental data consists of transliterations of a source word in other languages. We extracted the supplemental transliterations from the NEWS 2015 Shared Task training and development sets for which English was the source language.…”
Section: Jointmentioning
confidence: 99%
“…Yao and Kondrak (2015) propose a JOINT generation approach that can incorporate multiple transliterations as input, and show that it outperforms the reranking approach of Bhargava and Kondrak (2012). The JOINT system is a modified version of DIRECTL+ that utilizes aligned supplemental transliterations to learn additional features.…”
Section: Jointmentioning
confidence: 99%
“…The work by Oh and Choi actually widened a new direction of research on transliteration started by Al-Onaizan and Knight [24] and later Bilac and Tanaka [81], of combining several sources and/or including supplemental transliterations. This has led to several recent works such as Kumaran et al [77], Bhargava and Kondrak (2011Kondrak ( , 2012 and Yao and Kondrak [82].…”
Section: Hybrid Approachesmentioning
confidence: 99%