Proceedings of the 2009 Named Entities Workshop: Shared Task on Transliteration - NEWS '09 2009
DOI: 10.3115/1699705.1699712
|View full text |Cite
|
Sign up to set email alerts
|

DirecTL

Abstract: We present DIRECTL: an online discriminative sequence prediction model that employs a many-to-many alignment between target and source. Our system incorporates input segmentation, target character prediction, and sequence modeling in a unified dynamic programming framework. Experimental results suggest that DIRECTL is able to independently discover many of the language-specific regularities in the training data.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2011
2011
2018
2018

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…Here, g represents a given grapheme sequence, where ϕ is the most likely pronunciation of the grapheme sequence g. Soon after, Jiampojamarn et al represented the joint n-grams model for G2P conversion as an online discriminative sequence-prediction model, which used a many-tomany alignment between grapheme and phoneme sequences and a feature vector consisting of n-grams context features, HMM-like transition features, and linear-chain features [6]. For each training iteration, the feature weight vector was updated using the MIRA algorithm; this system is called DirecTL.…”
Section: Mira-based Methods For G2p Conversion (Directl+)mentioning
confidence: 99%
See 1 more Smart Citation
“…Here, g represents a given grapheme sequence, where ϕ is the most likely pronunciation of the grapheme sequence g. Soon after, Jiampojamarn et al represented the joint n-grams model for G2P conversion as an online discriminative sequence-prediction model, which used a many-tomany alignment between grapheme and phoneme sequences and a feature vector consisting of n-grams context features, HMM-like transition features, and linear-chain features [6]. For each training iteration, the feature weight vector was updated using the MIRA algorithm; this system is called DirecTL.…”
Section: Mira-based Methods For G2p Conversion (Directl+)mentioning
confidence: 99%
“…Thus, researchers have proposed various data-driven meth-ods using many-to-many mapping techniques between graphemes and phonemes. Methods have been proposed based on hidden Markov models (HMMs) [2], [3], artificial neural network (ANNs) [4], joint-sequences [5], margininfused relaxed algorithms (MIRAs) [6], [7], a weighted finite-state transducer (WFST) [8], an adaptive regularization of weight vectors (AROW) [9], a narrow adaptive regularization of weight vectors (NAROW) [10], and structured soft-margin confidence weighted learning (SSMCW) [11]. Most of these methods, and especially SSMCW-based G2P conversion that is implemented in the Slearp toolkit * , have demonstrated significantly accurate results.…”
Section: Introductionmentioning
confidence: 99%
“…Once these mappings are learned, a common approach involves using a generative model that attempts to generate all possible transliterations of a source word, given the character mappings between two languages, and restricting the output to words in the target language (El-Kahki et al, 2011;Noeman and Madkour, 2010). Other approaches include the use of locality sensitive hashing (Udupa and Kumar, 2010) and classification (Jiampojamarn et al, 2010). Another dramatically different approaches involves the unsupervised learning of transliteration mappings from a large parallel corpus instead of transliteration pairs (Sajjad et al, 2012).…”
Section: Discussionmentioning
confidence: 99%
“…The typical first step in training a G2P system involves aligning the corresponding grapheme and phoneme sequences in the input training dictionary. The approach adopted in this work is based on the EM driven multiple-to-multiple alignment algorithm proposed in Jiampojamarn et al (2007) and extended in Jiampojamarn and Kondrak (2010). This algorithm is capable of learning complex G↔P relationships like → //, and represents an improvement over earlier one-to-one stochastic alignment algorithms such as that introduced in Yianilos and Ristad (1998).…”
Section: Grapheme-to-phoneme Alignmentmentioning
confidence: 99%