Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
2019
DOI: 10.14311/nnw.2019.29.025
|View full text |Cite
|
Sign up to set email alerts
|

Towards an optimal set of initial weights for a Deep Neural Network architecture

Abstract: Modern neural network architectures are powerful models. They have been proven efficient in many fields, such as imaging and acoustic. However, these neural networks involve a long-running and time-consuming process. To accelerate the training process, we propose a two-stage approach based on data analysis and focus on the gravity center concept. The neural network is first trained on reduced data represented by a set of centroids of the original data points, and then the learned weights are used to initialize… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 12 publications
(16 reference statements)
0
3
0
Order By: Relevance
“…This refers to a reasonable ANN model with the condition that the training error should be small, and the difference between the training error and the predicted error should be also small. Otherwise, it is an overfitting or underfitting ANN model (Briscoe and Feldman, 2011;Chang et al, 2012;Xiao et al, 2013;Lee et al, 2016;Belkin et al, 2019;Mehta et al, 2019;Saadi and Belhadef, 2019;Doroudi, 2020). Therefore, to verify the rationality of the both BPNN models of this study, we analyzed the training errors in the training process and predicted errors in the predicted process.…”
Section: Resultsmentioning
confidence: 95%
“…This refers to a reasonable ANN model with the condition that the training error should be small, and the difference between the training error and the predicted error should be also small. Otherwise, it is an overfitting or underfitting ANN model (Briscoe and Feldman, 2011;Chang et al, 2012;Xiao et al, 2013;Lee et al, 2016;Belkin et al, 2019;Mehta et al, 2019;Saadi and Belhadef, 2019;Doroudi, 2020). Therefore, to verify the rationality of the both BPNN models of this study, we analyzed the training errors in the training process and predicted errors in the predicted process.…”
Section: Resultsmentioning
confidence: 95%
“…An IT project extension forecast is a tabular binary classification problem, with positive samples representing the completed projects and negative samples the uncompleted projects. We used a two-stage approach based on data analysis and the concept of the center of gravity [39]. e neural network was trained on a set of unbalanced data of the centers of mass from an a priori knowledge base (the first stage involved finding local minima close to the global minimum, and successful training was indicated when the local minima were relative to the global minimum).…”
Section: Experiments Setupmentioning
confidence: 99%
“…There are several architectures that can be implemented when it comes to deep learning (Saadi & Belhadef, 2019;Mansuri & Patel, 2021). Each of these architectures has its uses and compatibilities with certain applications.…”
Section: Introductionmentioning
confidence: 99%