2016
DOI: 10.1162/tacl_a_00114
|View full text |Cite
|
Sign up to set email alerts
|

Minimally Supervised Number Normalization

Abstract: We propose two models for verbalizing numbers, a key component in speech recognition and synthesis systems. The first model uses an end-to-end recurrent neural network. The second model, drawing inspiration from the linguistics literature, uses finite-state transducers constructed with a minimal amount of training data. While both models achieve near-perfect performance, the latter model can be trained using several orders of magnitude less data than the former, making it particularly useful for low-resource l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2
2

Relationship

3
5

Authors

Journals

citations
Cited by 19 publications
(23 citation statements)
references
References 17 publications
(23 reference statements)
0
23
0
Order By: Relevance
“…Silly errors This category consists of those "bizarre" errors which defy any purely linguistic characterization. In addition to the aforementioned case of *membled, such errors have also been reported for other language generation tasks such as machine translation (Arthur et al 2016) and text normalization (Gorman and Sproat 2016, Sproat and Jaitly 2017, Zhang et al 2019.…”
Section: Error Taxonomymentioning
confidence: 69%
“…Silly errors This category consists of those "bizarre" errors which defy any purely linguistic characterization. In addition to the aforementioned case of *membled, such errors have also been reported for other language generation tasks such as machine translation (Arthur et al 2016) and text normalization (Gorman and Sproat 2016, Sproat and Jaitly 2017, Zhang et al 2019.…”
Section: Error Taxonomymentioning
confidence: 69%
“…Our general process of minimally supervised number names induction is described in [4]. In essence, we use a set of training data consisting of digits mapped to their verbalization (like 123 → one hundred twenty three) to induce a finite state transducer (FST) which can produce the factorization for any number.…”
Section: Number Names Inductionmentioning
confidence: 99%
“…Verbalizers for most semiotic classes depend on underlying core number names grammars specifying the verbalization of numbers like English one, two, three. We first describe how we modify the induction algorithm in [4] to build these number names grammars across a wider range of languages. We then describe a system which builds on this algorithm to induce verbalization grammars for ASR and TTS systems alike.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…One solution to this problem is to use covering grammars, (usually) finite-state models that can constrain the neural models to a reasonable (context-independent) space of options so that 2mA could be read as two milliamperes or two m a, but not two million liters [3]. These covering grammars can be learned in whole or in part from data [7].…”
Section: Introductionmentioning
confidence: 99%