2018 26th Signal Processing and Communications Applications Conference (SIU) 2018
DOI: 10.1109/siu.2018.8404724
|View full text |Cite
|
Sign up to set email alerts
|

A comparison of activation functions in artificial neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
20
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 35 publications
(20 citation statements)
references
References 6 publications
0
20
0
Order By: Relevance
“…Irrespective of the heterogeneous population, ANN, showed only five misclassifications in the healthy young-middle age group, two were assigned to the healthy older group and three to the patient group showing an impressive performance for this group in particular. In contrast to both SVM and RF, ANN can analyze the complex structure among variables, by using various activation functions (e.g., Tanh, Sigmoid) even though, similar to SVM the variable interactions are not visible 30,33 . Although a large data set is required to find the optimal activation function and avoid overfitting 22 , ANN has the capacity to adapt for the limited dataset with suitable activation functions which may need to be adjusted depending on the type of data 33,39 , to build a small neural network.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Irrespective of the heterogeneous population, ANN, showed only five misclassifications in the healthy young-middle age group, two were assigned to the healthy older group and three to the patient group showing an impressive performance for this group in particular. In contrast to both SVM and RF, ANN can analyze the complex structure among variables, by using various activation functions (e.g., Tanh, Sigmoid) even though, similar to SVM the variable interactions are not visible 30,33 . Although a large data set is required to find the optimal activation function and avoid overfitting 22 , ANN has the capacity to adapt for the limited dataset with suitable activation functions which may need to be adjusted depending on the type of data 33,39 , to build a small neural network.…”
Section: Discussionmentioning
confidence: 99%
“…Nevertheless, RF disregard the intact interactions within and between trees, which might negatively impact the classification performance 32 . Although the black box problem also exists in the hidden layers of ANN 30 , the activation functions such as the tangents hyperbolic can properly analyze the complex interactions among the gait variables to improve the classification performance 33 . A recent study used deep learning to explain gait patterns based on kinematic and kinetic variables.…”
mentioning
confidence: 99%
“…"ReLU" was used as the active function [18]. The "batch normalization" [19] was applied to each hidden layer, and 20% of the training sets were used as the validation data sets to prevent overfitting issues. To prevent "overfitting" in the process of developing the ANN model, the "training loss" was compared against the "val loss".…”
Section: Simulationmentioning
confidence: 99%
“…An activation function is used to determine whether the sum of the input causes activation or not. According to the researchers [14,15], there are many activation functions and amongst them, rectified linear unit (ReLU) is one of the best activation functions to carry out the DNN. Especially, Pedamonti [14] mentioned that ReLU is a better neuron replacing sigmoid function, and Cent et al [15] concluded that ReLU is the best activation function after reviewing 10 activation functions.…”
Section: Activation Functionmentioning
confidence: 99%
“…According to the researchers [14,15], there are many activation functions and amongst them, rectified linear unit (ReLU) is one of the best activation functions to carry out the DNN. Especially, Pedamonti [14] mentioned that ReLU is a better neuron replacing sigmoid function, and Cent et al [15] concluded that ReLU is the best activation function after reviewing 10 activation functions. Therefore, the widely used ReLU activation function in recent years is adopted as an activation function due to the benefits of this function like faster computation and avoiding the vanishing gradient problem [14][15][16].…”
Section: Activation Functionmentioning
confidence: 99%