2021
DOI: 10.48550/arxiv.2102.06380
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Inverse Text Normalization

Abstract: While there have been several contributions exploring state of the art techniques for text normalization, the problem of inverse text normalization (ITN) remains relatively unexplored. The best known approaches leverage finite state transducer (FST) based models which rely on manually curated rules and are hence not scalable. We propose an efficient and robust neural solution for ITN leveraging transformer based seq2seq models and FST-based text normalization techniques for data preparation. We show that this … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 11 publications
(18 reference statements)
0
1
0
Order By: Relevance
“…Moreover, ASR systems output numbers, dates, times, and other numerical entities in a literal manner (e.g., 20 → twenty). Thus, inverse text normalization was proposed to format entities like numbers, dates, times, and addresses [9,17,18]. In addition, a neural generation model of ASR spelling correction [19] and grammar correction [20] was proposed to deal with the many spelling and grammar errors in the spoken text.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, ASR systems output numbers, dates, times, and other numerical entities in a literal manner (e.g., 20 → twenty). Thus, inverse text normalization was proposed to format entities like numbers, dates, times, and addresses [9,17,18]. In addition, a neural generation model of ASR spelling correction [19] and grammar correction [20] was proposed to deal with the many spelling and grammar errors in the spoken text.…”
Section: Related Workmentioning
confidence: 99%