2017 13th International Wireless Communications and Mobile Computing Conference (IWCMC) 2017
DOI: 10.1109/iwcmc.2017.7986470
|View full text |Cite
|
Sign up to set email alerts
|

Parameters optimization of deep learning models using Particle swarm optimization

Abstract: Deep learning has been successfully applied in several fields such as machine translation, manufacturing, and pattern recognition. However, successful application of deep learning depends upon appropriately setting its parameters to achieve high-quality results. The number of hidden layers and the number of neurons in each layer of a deep machine learning network are two key parameters, which have main influence on the performance of the algorithm. Manual parameter setting and grid search approaches somewhat e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
57
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 92 publications
(57 citation statements)
references
References 15 publications
0
57
0
Order By: Relevance
“…Former literature solely discusses on how to determine the number of hidden neurons (assuming by using one hidden layer), but does rarely discuss the way to determine the optimal number of hidden layers. This fact is due to the assumption that a network with only one hidden layer is sufficient to universally approve almost all functions [3]- [7]. However, several studies probed that the application of two hidden layers provides better performance compare to one hidden layer in some cases [3]- [7].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Former literature solely discusses on how to determine the number of hidden neurons (assuming by using one hidden layer), but does rarely discuss the way to determine the optimal number of hidden layers. This fact is due to the assumption that a network with only one hidden layer is sufficient to universally approve almost all functions [3]- [7]. However, several studies probed that the application of two hidden layers provides better performance compare to one hidden layer in some cases [3]- [7].…”
Section: Introductionmentioning
confidence: 99%
“…Since the increasing ability of computers, the application of neural networks with more than one hidden layer has become one of the attractions for researchers, especially since the use of deep neural networks (DNN) to solve problems in various fields. The utilization of DNN is defined as a technique applying neural networks by using numerous hidden layers between the input and output layers [3], [10]; or in other words it is considered as machine learning with DNN. One of the challenges in the successful implementation of deep neural networks lies on the determination of the network architecture, which is closely related to the number of hidden layers and hidden neurons.…”
Section: Introductionmentioning
confidence: 99%
“…Please refer to Table 2. [16,2048] With the objectives of reducing the amount of parameters (weights), the computational burden, and to control overfitting, for each convolutional layer, we used a pooling (downsampling) layer of a fixed size of 2 × 2 along with the max pooling option. Further, we utilized the ReLU (rectified linear unit) activation function and employed the Adam optimizer with a fixed learning rate of θ = 0.0001 and the mean squared error (MSE) loss function.…”
Section: Parameter Setupmentioning
confidence: 99%
“…GA have many successful applications in the domains of deep learning and CNNs 17,18 . Better results can be obtained by applying meta-heuristic methods such as genetic algorithm (GA) 17 and swarm intelligence 27 to the process of CNN hyperparameter optimization.…”
mentioning
confidence: 99%