2020
DOI: 10.25046/aj050570
|View full text |Cite
|
Sign up to set email alerts
|

Advances in Optimisation Algorithms and Techniques for Deep Learning

Abstract: In the last decade, deep learning(DL) has witnessed excellent performances on a variety of problems, including speech recognition, object recognition, detection, and natural language processing (NLP) among many others. Of these applications, one common challenge is to obtain ideal parameters during the training of the deep neural networks (DNN). These typical parameters are obtained by some optimisation techniques which have been studied extensively. These research have produced state-of-art(SOTA) results on s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(3 citation statements)
references
References 57 publications
(77 reference statements)
0
2
0
Order By: Relevance
“…The activation function [53] for the hidden dense layer is the rectified linear unit (ReLU). For binary classification, the output dense layer is the sigmoid function, and the models are compiled using binary cross-entropy as the loss function and the adaptive movement estimation optimizer Adam [54]. In the case of multi-class classification with 3 or 5 classes, the models are compiled using the sparse categorical cross-entropy function, and for the output dense layer, we use the so f tmax function.…”
Section: Models 441 Supervised Classificationmentioning
confidence: 99%
“…The activation function [53] for the hidden dense layer is the rectified linear unit (ReLU). For binary classification, the output dense layer is the sigmoid function, and the models are compiled using binary cross-entropy as the loss function and the adaptive movement estimation optimizer Adam [54]. In the case of multi-class classification with 3 or 5 classes, the models are compiled using the sparse categorical cross-entropy function, and for the output dense layer, we use the so f tmax function.…”
Section: Models 441 Supervised Classificationmentioning
confidence: 99%
“…This error function is iteratively minimized by optimization algorithms, which aims to find optimal weights that improve the learning process of the model and minimize the error between predicted and observable values. The MLP model uses several types of optimization algorithms or also called learning algorithms such as Gradient Descending (SGD), RMSprop and Adam, among others [26].…”
Section: The Multilayer Perceptron (Mlp) Modelmentioning
confidence: 99%
“…e psychological community has never stopped discussing the definition of emotions, and the criteria for classifying emotions have become a point of debate among scholars. Some of the more recognized theories of emotion classification are the basic emotion theory and the dimensional spatial emotion theory [18]. e basic emotion theory holds that emotions have archetypal forms in their occurrence; that is, there are several pan-human basic emotion types.…”
Section: Analysis Of the Chaotic Fish Swarm Algorithm Mental Health S...mentioning
confidence: 99%