2020
DOI: 10.1002/cpe.6143
|View full text |Cite
|
Sign up to set email alerts
|

Weight initialization based‐rectified linear unit activation function to improve the performance of a convolutional neural network model

Abstract: Convolutional Neural Networks (CNNs) have made a great impact on attaining state‐of‐the‐art results in image task classification. Weight initialization is one of the fundamental steps in formulating a CNN model. It determines the failure or success of the CNN model. In this paper, we conduct a research based on the mathematical background of different weight initialization strategies to determine the one with better performance. To have smooth training, we expect the activation of each layer of the CNN model f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
1
1

Relationship

2
7

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 17 publications
0
15
0
Order By: Relevance
“…After stage 1, the extracted features, that is, the landmark points ( ) per frame are flattened, concatenated and stored in a file to check and remove any null entries from the data. Data cleaning is important since it prevents failed detection of features 56 58 , which occurs when a blurred image is sent to the detector and leads to a null entry into the dataset. Thus, when training occurs with this noisy data, the prediction accuracy is reduced and bias may occur.…”
Section: Proposed Methodologymentioning
confidence: 99%
“…After stage 1, the extracted features, that is, the landmark points ( ) per frame are flattened, concatenated and stored in a file to check and remove any null entries from the data. Data cleaning is important since it prevents failed detection of features 56 58 , which occurs when a blurred image is sent to the detector and leads to a null entry into the dataset. Thus, when training occurs with this noisy data, the prediction accuracy is reduced and bias may occur.…”
Section: Proposed Methodologymentioning
confidence: 99%
“…The batch normalization reduces the internal co-variant shift and also regularizes the model. A rectified linear unit (ReLU) [ 24 , 25 , 26 , 27 ] activation function is applied. Two advantages accompany the ReLU activation function: (1) It realizes the sparse representation of the network.…”
Section: Methodsmentioning
confidence: 99%
“…5 ). The ReLU (Rectified Linear Unit) activation function 39 is assigned to neurons in all convolution layers and fully-connected layer, while the Softmax activation function is assigned to neurons in the last layer to output the classification results. The filter with the size of 3 × 3 is used to expand the number of channels to extract expressive and complex features, and the output data has the same size as the input data through the numeral zero padding.…”
Section: Network Architecture and Applied Strategiesmentioning
confidence: 99%