Proceedings of the First Workshop on Building Linguistically Generalizable NLP Systems 2017
DOI: 10.18653/v1/w17-5403
|View full text |Cite
|
Sign up to set email alerts
|

Massively Multilingual Neural Grapheme-to-Phoneme Conversion

Abstract: Grapheme-to-phoneme conversion (g2p) is necessary for text-to-speech and automatic speech recognition systems. Most g2p systems are monolingual: they require language-specific data or handcrafting of rules. Such systems are difficult to extend to low resource languages, for which data and handcrafted rules are not available. As an alternative, we present a neural sequence-to-sequence approach to g2p which is trained on spelling-pronunciation pairs in hundreds of languages. The system shares a single encoder an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
36
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 37 publications
(37 citation statements)
references
References 17 publications
0
36
0
Order By: Relevance
“…Class We use grapheme-to-phoneme (G2P) as a proxy of a phonological task (Deri and Knight, 2016;Peters et al, 2017). The dataset contains over 650,000 such training instances, for a total of 311 languages (Deri and Knight, 2016).…”
Section: Datasetmentioning
confidence: 99%
“…Class We use grapheme-to-phoneme (G2P) as a proxy of a phonological task (Deri and Knight, 2016;Peters et al, 2017). The dataset contains over 650,000 such training instances, for a total of 311 languages (Deri and Knight, 2016).…”
Section: Datasetmentioning
confidence: 99%
“…They also show evidence that the model can handle intrasentence code-switching. Peters et al (2017) train a multilingual sequence-to-sequence translation architecture on grapheme-to-phoneme conversion using more than 300 languages. They report better performance when adding multiple languages, even those which are not present in the test data.…”
Section: Related Workmentioning
confidence: 99%
“…For zero-shot learning (Sec. 3.5), we use an architecture very similar to Johnson et al (2016), also used for grapheme-tophoneme mapping in Peters et al (2017).…”
Section: Related Workmentioning
confidence: 99%