2018
DOI: 10.1016/j.neucom.2017.06.070
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive activation functions in convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
67
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 144 publications
(77 citation statements)
references
References 9 publications
1
67
0
Order By: Relevance
“…Therefore, in this work, we are particularly focusing on adaptive activation functions, which adapt automatically such that the network can be trained faster. Various methods are proposed in the literature for adaptive activation function, like the adaptive sigmoidal activation function proposed by Yu et al [27] for multilayer feedforward NNs, while Qian et al [21] focuses on learning activation functions in convolutional NNs by combining basic activation functions in a data-driven way. Multiple activation functions per neuron are proposed by Dushkoff and Ptucha [7], where individual neurons select between a multitude of activation functions.…”
Section: Introductionmentioning
confidence: 99%
“…Therefore, in this work, we are particularly focusing on adaptive activation functions, which adapt automatically such that the network can be trained faster. Various methods are proposed in the literature for adaptive activation function, like the adaptive sigmoidal activation function proposed by Yu et al [27] for multilayer feedforward NNs, while Qian et al [21] focuses on learning activation functions in convolutional NNs by combining basic activation functions in a data-driven way. Multiple activation functions per neuron are proposed by Dushkoff and Ptucha [7], where individual neurons select between a multitude of activation functions.…”
Section: Introductionmentioning
confidence: 99%
“…In our case, the ReLU activation function is used inside each step of the convolutional layers until the last layer, since the ReLU clips negative values to zero while keeping positive values unchanged. This function acts as a filter that breaks up the linearity and increases the non‐linearity of the images (Qian, Liu, Liu, Wu, & SanWong, ). In the last layer, the sigmoid function is used, which is more appropriate for cases in which prediction of probability is requested as an output.…”
Section: Training the Networkmentioning
confidence: 99%
“…ELU (exponential linear unit) [20] is another popular activation function based on ReLU -it uses an exponential function for negative inputs instead of linear function. An adaptive ELU extension parametric ELU PELU together with mixing different activation functions using adaptive linear combination or hierarchical gated combination of activation function was shown to perform well [21].…”
Section: Adaptive Activation Functionsmentioning
confidence: 99%