2017
DOI: 10.48550/arxiv.1712.01769
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

State-of-the-art Speech Recognition With Sequence-to-Sequence Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
47
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
8
1
1

Relationship

1
9

Authors

Journals

citations
Cited by 29 publications
(47 citation statements)
references
References 0 publications
0
47
0
Order By: Relevance
“…The second encoder in the multistage model, which takes the transcript as input, also uses an embedding layer of the same size. All decoders use 4-headed additive attention [31,17,22]. Our Baseline is the multistage model in which the two stages that do ASR and NLU are trained independently, but using the same training data.…”
Section: Modelmentioning
confidence: 99%
“…The second encoder in the multistage model, which takes the transcript as input, also uses an embedding layer of the same size. All decoders use 4-headed additive attention [31,17,22]. Our Baseline is the multistage model in which the two stages that do ASR and NLU are trained independently, but using the same training data.…”
Section: Modelmentioning
confidence: 99%
“…The LAS architecture has achieved state-of-the-art word error rates (WER) on a task with two orders of magnitude more training data than here [9], but on smaller datasets hybrid TDNN-HMM ASR approaches are still considerably better. Table 1 shows the results of our ASR model contrasted with those reported by XNMT in [7], on the TED-LIUM development and test sets.…”
Section: Word Error Ratesmentioning
confidence: 92%
“…There has been a remarkable growth in the interest towards deep neural networks (DNNs) in the last decade, as they surpassed previous state-of-the-art machine learning models in many tasks, such as speech recognition [5] and natural language processing [3]. Aside from theoretical developments in DNN architectures and training methods, there has been two trends that still fuel this growth to date: increasing computing power, and availability of large data sets.…”
Section: Introductionmentioning
confidence: 99%