2021
DOI: 10.1016/j.neunet.2021.05.011
|View full text |Cite
|
Sign up to set email alerts
|

A distributed optimisation framework combining natural gradient with Hessian-free for discriminative sequence training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
1
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 56 publications
0
1
0
Order By: Relevance
“…Besides the low resource problem, how to train an ASR system with large amount of data is also dramatically important. Haider et al [13] presented a novel Natural Gradient and hessian-Free (NGHF) optimisation framework for neural network training that can operate efficiently in a distributed manner. Their experiments show that NGHF not only achieves larger word error rate reductions than standard stochastic gradient descent or Adam, but also requires orders of magnitude fewer parameter updates.…”
mentioning
confidence: 99%
“…Besides the low resource problem, how to train an ASR system with large amount of data is also dramatically important. Haider et al [13] presented a novel Natural Gradient and hessian-Free (NGHF) optimisation framework for neural network training that can operate efficiently in a distributed manner. Their experiments show that NGHF not only achieves larger word error rate reductions than standard stochastic gradient descent or Adam, but also requires orders of magnitude fewer parameter updates.…”
mentioning
confidence: 99%