2022
DOI: 10.3390/electronics11040540
|View full text |Cite
|
Sign up to set email alerts
|

Smish: A Novel Activation Function for Deep Learning Methods

Abstract: Activation functions are crucial in deep learning networks, given that the nonlinear ability of activation functions endows deep neural networks with real artificial intelligence. Nonlinear nonmonotonic activation functions, such as rectified linear units, Tan hyperbolic (tanh), Sigmoid, Swish, Mish, and Logish, perform well in deep learning models; however, only a few of them are widely used in mostly all applications due to their existing inconsistencies. Inspired by the MB-C-BSIF method, this study proposes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 42 publications
(19 citation statements)
references
References 30 publications
0
10
0
Order By: Relevance
“…The recent research made by Jin et al (2022) [28] focuses on solving the Time-varying Sylvester equation using Zeroing Neural Network based on the Versatile Activation Function (VAF) variations which include exponential convergence characteristic and adjustable parameter. Another recent work proposed a function called Smish by Wang et al (2022) [29]. It was experimented on the CIFAR-10 dataset with the EfficientNetB3 network, on the MNIST dataset with the EfficientNetB5 network, and on the SVHN dataset with the EfficientnetB7 network.…”
Section: Related Workmentioning
confidence: 99%
“…The recent research made by Jin et al (2022) [28] focuses on solving the Time-varying Sylvester equation using Zeroing Neural Network based on the Versatile Activation Function (VAF) variations which include exponential convergence characteristic and adjustable parameter. Another recent work proposed a function called Smish by Wang et al (2022) [29]. It was experimented on the CIFAR-10 dataset with the EfficientNetB3 network, on the MNIST dataset with the EfficientNetB5 network, and on the SVHN dataset with the EfficientnetB7 network.…”
Section: Related Workmentioning
confidence: 99%
“…Its trainable parameter offers fine tuning for maximized output of propagation with a smooth gradient. This generates a faster, easier, and efficient generalizability of results [72].…”
Section: Algorithm 2: Artificial Neural Networkmentioning
confidence: 99%
“…The Smish inherits the nonmonotonic properties of the Logish function. The Smish function is mostly suitable for machine learning networks which can be expressed in equation ( 2) [17],…”
Section: Egpnetmentioning
confidence: 99%