2020
DOI: 10.1140/epjst/e2020-000098-9
|View full text |Cite
|
Sign up to set email alerts
|

Evolution of novel activation functions in neural network training for astronomy data: habitability classification of exoplanets

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

3
6

Authors

Journals

citations
Cited by 21 publications
(23 citation statements)
references
References 66 publications
(107 reference statements)
0
23
0
Order By: Relevance
“…If there is no non-linear activation function, the deep CNN architecture will evolve into a single equivalent convolutional layer, and its performance will hardly be so. The ReLU activation function is used explicitly as a non-linear activation function, in contrast to other non-linear functions such as Sigmoid, because it has been observed from experience that the CNN using ReLU trains faster than the corresponding CNN [79]. Furthermore, the ReLU activation function is a one-to-one mathematical operation, as shown in Equation (6).…”
Section: An Overview Of Machine Learning In Agriculturementioning
confidence: 99%
“…If there is no non-linear activation function, the deep CNN architecture will evolve into a single equivalent convolutional layer, and its performance will hardly be so. The ReLU activation function is used explicitly as a non-linear activation function, in contrast to other non-linear functions such as Sigmoid, because it has been observed from experience that the CNN using ReLU trains faster than the corresponding CNN [79]. Furthermore, the ReLU activation function is a one-to-one mathematical operation, as shown in Equation (6).…”
Section: An Overview Of Machine Learning In Agriculturementioning
confidence: 99%
“…shown by Saha et al [57], sigmoid activation has a few limitations which impede its ability to reduce error in backpropagation some times. This includes not having a critical point and, therefore, being handicapped with the absence of global optima.…”
Section: Phase Ii: a Deeper Analysismentioning
confidence: 99%
“…For each of these cases, 80% of samples were used for training set and 20% for testing. Eliminating the need for parameters tuning, SBAF parameters were computed from the fixed point plots taken from [57] and were set to α = 0.5 and k = 0.91. The results of SBAF are compared with sigmoid and presented in Tables 9 and 10.…”
Section: Relevant Data: All Casesmentioning
confidence: 99%
“…ML has been extensively used to solve challenging problems of different domains. An extensive exercise is attempted by Saha et al [71] to perform a detailed experiment to categorize and classify new exoplanet samples. The method uses NN as base architecture, and plugs-in different activation functions to investigate how well they perform on the PHL-EC dataset.…”
Section: Neural Network On Phl-ec Datasetmentioning
confidence: 99%