2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS) 2017
DOI: 10.1109/mwscas.2017.8053243
|View full text |Cite
|
Sign up to set email alerts
|

Gate-variants of Gated Recurrent Unit (GRU) neural networks

Abstract: -The paper evaluates three variants of the Gated Recurrent Unit (GRU) in recurrent neural networks (RNN) by reducing parameters in the update and reset gates. We evaluate the three variant GRU models on MNIST and IMDB datasets and show that these GRU-RNN variant models perform as well as the original GRU RNN model while reducing the computational expense.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
521
0
4

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 1,064 publications
(525 citation statements)
references
References 11 publications
0
521
0
4
Order By: Relevance
“…By generating temporal representations from learning, LSTM has been successfully applied to speech recognition and machine translation. LSTM is similar to GRU we used in our residual block but LSTM has a higher computing cost [25].…”
Section: H a Comparative Studymentioning
confidence: 99%
“…By generating temporal representations from learning, LSTM has been successfully applied to speech recognition and machine translation. LSTM is similar to GRU we used in our residual block but LSTM has a higher computing cost [25].…”
Section: H a Comparative Studymentioning
confidence: 99%
“…GRU: The GRU is a simplified version of the more complex LSTM unit that combines the input and forgets gates into a single update gate. It then merges both the cell and hidden states for faster operation (28). Equations 13-15 describe the mathematical operations inside the GRU neurons.…”
Section: Simple Rnnmentioning
confidence: 99%
“…With regard to the process of feature extraction, most approaches use recurrent neural networks [26], typically implemented with LSTM [27] and GRU [28], to extract textual features. As for the visual features, convolutional neural networks [29] are used to obtain region features from image, among which VGG-net [30] and deep residual networks [31] are most common.…”
Section: A Feature Extraction and Representationmentioning
confidence: 99%