2021
DOI: 10.1007/s11760-021-01863-z
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing nonlinear activation function for convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(29 citation statements)
references
References 18 publications
0
29
0
Order By: Relevance
“…In Machine learning, saturated activation functions like Sigmoidal, tanh functions are used. The new unsaturated variant, the ReLU has been prominent in the CNN’s as per its best outcomes [ 18 ]. The voluminous literature applying the same activation function ReLU invariably to all types of data (right from text-based data mining to computer vision) and for all applications indicates a research gap.…”
Section: Literature Surveymentioning
confidence: 99%
See 1 more Smart Citation
“…In Machine learning, saturated activation functions like Sigmoidal, tanh functions are used. The new unsaturated variant, the ReLU has been prominent in the CNN’s as per its best outcomes [ 18 ]. The voluminous literature applying the same activation function ReLU invariably to all types of data (right from text-based data mining to computer vision) and for all applications indicates a research gap.…”
Section: Literature Surveymentioning
confidence: 99%
“…As there is a stack of spatial domain images, wavelet sub-bands (inclusive of LF scaling functions and HF wavelet functions) a need arises to formulate activation function for each “domain” and “sub-bands.” In the frequency domain, both positive and negative coefficients are significant [ 17 ]. Varshney et al, in [ 18 ], emphasize the negative spectral coefficients and the importance to save them from dismissal by activations. But the bottleneck with the conventional spatial domain activation functions is that they nullify the negative coefficients.…”
Section: Introductionmentioning
confidence: 99%
“…In the first batch of hidden layers, a Rectified Linear Unit (ReLU, Agarap, 2018) is used as the activation function for all layers, as is common in many contemporary neural networks (Varshney and Singh, 2021). In the second batch of hidden layers, Leaky ReLUs (Clevert et al, 2015) are used to combat the somewhat common dying ReLU (Lu et al, 2019) problem that was encountered when initially attempting to build the network with standard ReLUs.…”
Section: A1 Activation Functionsmentioning
confidence: 99%
“…The parameters can be reduced to a certain extent to avoid repeated convolution cores. Pooling [21] refers to the image after convolution which is downsampled to shorten the size of the image, ensure that there is no overfitting, and improve the efficiency of calculation. The processing flow of the convolution neural network is shown in Fig.…”
Section: Convolutional Neural Networkmentioning
confidence: 99%