2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6855088
|View full text |Cite
|
Sign up to set email alerts
|

Investigation of maxout networks for speech recognition

Abstract: We explore the use of maxout neuron in various aspects of acoustic modelling for large vocabulary speech recognition systems; including low-resource scenario and multilingual knowledge transfers. Through the experiments on voice search and short message dictation datasets, we found that maxout networks are around three times faster to train and offer lower or comparable word error rates on several tasks, when compared to the networks with logistic nonlinearity. We also present a detailed study of the maxout un… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
24
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 38 publications
(26 citation statements)
references
References 25 publications
(14 reference statements)
2
24
0
Order By: Relevance
“…We have also tried to train sigmoid [32] network, but the initial loss never decreased. Finally, as proposed by Swietojanski et.al [33], we have tested combination of ReLU for first layers and maxout for the last layers of the network.…”
Section: Methodsmentioning
confidence: 99%
“…We have also tried to train sigmoid [32] network, but the initial loss never decreased. Finally, as proposed by Swietojanski et.al [33], we have tested combination of ReLU for first layers and maxout for the last layers of the network.…”
Section: Methodsmentioning
confidence: 99%
“…The dropout method was shown to improve the generalization ability of neural networks by preventing the co-adaptation of units [28]. Dropout is now routinely used in the training of DNNs for speech recognition, and some researchers have already reported that it works nicely with maxout units as well [19,29]. We also find it to yield a significant performance gain.…”
Section: Introductionmentioning
confidence: 56%
“…This activation function can be regarded as a generalization of the rectifier function [16], and so far, only a few studies have attempted to apply maxout networks to speech recognition tasks. These all found that maxout nets slightly outperformed ReLU networks, in particular under lowresource conditions [17][18][19]. Here, we show that the pooling procedure applied in CNNs and the pooling step of the maxout function are practically the same, and hence, it is trivial to combine the two techniques and construct convolutional networks out of maxout neurons.…”
Section: Introductionmentioning
confidence: 91%
See 1 more Smart Citation
“…Maxout units are adopted to improve the performance compared with ReLU except the first layer as suggested in [21], the bottom layers should be replaced by layers of a smaller number of ReLU units. The configuration is described as follows:…”
Section: Basic Experiments Of Dmcnmentioning
confidence: 99%