2022
DOI: 10.3389/fncom.2022.1004988
|View full text |Cite
|
Sign up to set email alerts
|

GAPCNN with HyPar: Global Average Pooling convolutional neural network with novel NNLU activation function and HYBRID parallelism

Abstract: With the increasing demand for deep learning in the last few years, CNNs have been widely used in many applications and have gained interest in classification, regression, and image recognition tasks. The training of these deep neural networks is compute-intensive and takes days or even weeks to train the model from scratch. The compute-intensive nature of these deep neural networks sometimes limits the practical implementation of CNNs in real-time applications. Therefore, the computational speedup in these ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 39 publications
0
5
0
Order By: Relevance
“…Afterwards, a final batch normalization layer was introduced to ensure the normalization of the data. Following this, a Global Average Pooling [ 41 ] layer is incorporated to generate a single vector from multiple time vectors. This is achieved by computing the average.…”
Section: Materials and Methodsmentioning
confidence: 99%
“…Afterwards, a final batch normalization layer was introduced to ensure the normalization of the data. Following this, a Global Average Pooling [ 41 ] layer is incorporated to generate a single vector from multiple time vectors. This is achieved by computing the average.…”
Section: Materials and Methodsmentioning
confidence: 99%
“…The features extracted by the convolutional layer are transformed into one-dimensional data by the global average pooling (GAP) layer, which is passed to the FC output layer. GAP converts the feature map into a one-dimensional vector, like the flatten layer but is known to improve the performance of the network compared to the flatten layer [ 21 ]. The FC output layer uses sigmoid as the activation function for the regression task, and the number of output nodes is one.…”
Section: Methodsmentioning
confidence: 99%
“…In [29], a streamlined sequential CNN architecture was proposed and compared with two pre-trained CNN models (VGG-16 and InceptionV3) for diagnosing diseases from chest X-ray images, achieving a peak accuracy of 89.89%. In the work conducted by [30], a hybrid parallelization strategy, incorporating both model and data parallelism, was applied to CNN to speed up its operation, with a normalized non-linearity activation function (NNLU) and replacing FC layers with Global Average Pooling (GAP) layers. These changes significantly improved accuracy and increased computation speed by 3.62 times with a batch size of (512,128), achieving a validation accuracy of approximately 98 with minimal training loss.…”
Section: Review Of Related Workmentioning
confidence: 99%