2010 Fourth Asia International Conference on Mathematical/Analytical Modelling and Computer Simulation 2010
DOI: 10.1109/ams.2010.27
|View full text |Cite
|
Sign up to set email alerts
|

Performance Comparison of Different Multilayer Perceptron Network Activation Functions in Automated Weather Classification

Abstract: Multilayer perceptron network (MLP) has been recognized as a powerful tool for many applications including classification. Selection of the activation functions in the multilayer perceptron (MLP) network plays an essential role on the network performance. This paper presents a comparison study of two commonly used MLP activation function; sigmoid and hyperbolic tangent for weather classification. Meteorological data such as solar radiation, ambient temperature, current, surface temperature, voltage, wind direc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0
1

Year Published

2013
2013
2018
2018

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 16 publications
0
6
0
1
Order By: Relevance
“…The transfer function in the hidden layer was a hyperbolic tangent and an identity function in the case of the output layer. The hyperbolic tangent gave optimal results in previous studies (Isa et al, 2010) in which a performance comparison was carried out to select the best MLP activation function (Olaya-Marín et al, 2012). The Levenberg-Marquardt (LM) optimization algorithm was used to train the candidate models because this algorithm is the fastest method to train neural networks of moderate size (Karul et al, 2000).…”
Section: Artificial Neural Network (Ann)mentioning
confidence: 99%
“…The transfer function in the hidden layer was a hyperbolic tangent and an identity function in the case of the output layer. The hyperbolic tangent gave optimal results in previous studies (Isa et al, 2010) in which a performance comparison was carried out to select the best MLP activation function (Olaya-Marín et al, 2012). The Levenberg-Marquardt (LM) optimization algorithm was used to train the candidate models because this algorithm is the fastest method to train neural networks of moderate size (Karul et al, 2000).…”
Section: Artificial Neural Network (Ann)mentioning
confidence: 99%
“…Several MLP models were built and tested, in order to establish (by systematic trial and error) the optimal number of neurons in the hidden layer and the optimal transfer function in the hidden and output layers. Commonly, transfer functions are nonlinear; they transform the weighted sum of inputs into an output signal (Zhang et al, 1998;Isa et al, 2010) and it is typical to use the same transfer function in hidden and output layers (Lek et al, 2005;Goethals et al, 2007). MLP results are very sensitive to the implemented transfer functions in their layers (Piekniewski & Rybicki, 2004;Isa et al, 2010).…”
Section: Artificial Neural Network Modellingmentioning
confidence: 99%
“…Commonly, transfer functions are nonlinear; they transform the weighted sum of inputs into an output signal (Zhang et al, 1998;Isa et al, 2010) and it is typical to use the same transfer function in hidden and output layers (Lek et al, 2005;Goethals et al, 2007). MLP results are very sensitive to the implemented transfer functions in their layers (Piekniewski & Rybicki, 2004;Isa et al, 2010). Generally, the selection of a transfer function is based on the best performance by trial and error (Isa et al, 2010), by comparing different transfer functions in the hidden and output layers.…”
Section: Artificial Neural Network Modellingmentioning
confidence: 99%
See 1 more Smart Citation
“…Researchers in various studies have corroborated that the polynomial and spline activation functions have the potential to estimate the highly non-linear systems [20][21][22][23]. Many other works imply that selection of activation functions is dependent on the application [24,25]. Consequently and as a new point of view, if we choose the arbitrary function f (x) as the activation function in the last layer, so that f (x) would meet the just mentioned conditions, then as the energy of the network approaches its minimum, the output of f (x) will converge to its desired objective.…”
Section: Basic Concepts: To Solve An Objective Function Using Unidimementioning
confidence: 99%