2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018
DOI: 10.1109/icassp.2018.8461608
|View full text |Cite
|
Sign up to set email alerts
|

High Order Recurrent Neural Networks for Acoustic Modelling

Abstract: Vanishing long-term gradients are a major issue in training standard recurrent neural networks (RNNs), which can be alleviated by long short-term memory (LSTM) models with memory cells. However, the extra parameters associated with the memory cells mean an LSTM layer has four times as many parameters as an RNN with the same hidden vector size. This paper addresses the vanishing gradient problem using a high order RNN (HORNN) which has additional connections from multiple previous time steps. Speech recognition… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
15
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 19 publications
(17 citation statements)
references
References 29 publications
2
15
0
Order By: Relevance
“…Figure 6 shows 1) all RNNs with memory or gated structure outperforms vanila-RNN and high-order RNN by a large margin, which indicates the advantages of memory and gated structure for controlling information flow; 2) high-order RNN performs better than vanila-RNN which implies the necessary of the non-local operations since high-order connections can be considered as a simple non-local operation in a local area. It is also consistent with the existing conclusions [37,50]; 3) our NRNM outperforms LSTM significantly which demonstrates the superiority of our model over standard LSTM.…”
Section: Investigation On Nrnmsupporting
confidence: 92%
“…Figure 6 shows 1) all RNNs with memory or gated structure outperforms vanila-RNN and high-order RNN by a large margin, which indicates the advantages of memory and gated structure for controlling information flow; 2) high-order RNN performs better than vanila-RNN which implies the necessary of the non-local operations since high-order connections can be considered as a simple non-local operation in a local area. It is also consistent with the existing conclusions [37,50]; 3) our NRNM outperforms LSTM significantly which demonstrates the superiority of our model over standard LSTM.…”
Section: Investigation On Nrnmsupporting
confidence: 92%
“…Notable examples are the clockwork RNN (Koutnik et al, 2014), gated feedback RNN (Chung et al, 2015), hierarchical multi-scale RNN (Chung et al, 2016), fast-slow RNN (Mujika et al, 2017), and higher order RNNs (HORNNs) (Soltani and Jiang, 2016). These modern RNN architectures have found various applications in motion classification (Neverova et al, 2016;Yan et al, 2018), speech synthesis (Wu and King, 2016;Achanta and Gangashetty, 2017;Zhang and Woodland, 2018), recognition (Chan et al, 2016), and other related areas (Liu et al, 2015;Krause et al, 2017;Kurata et al, 2017). These applications of hierarchical RNN architectures further confirm the relevance of hierarchically organized sequence generators for capturing complex dynamics in our everyday environments.…”
Section: A Hierarchy Of Time Scales: Machine Learningmentioning
confidence: 68%
“…Weight decay factors were carefully tuned to maximise the performance of each system. More details about the LSTM implementation and training configuration can be found in [42,43].…”
Section: Methodsmentioning
confidence: 99%