2020
DOI: 10.48550/arxiv.2001.01240
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Cooperative Initialization based Deep Neural Network Training

Abstract: Researchers have proposed various activation functions. These activation functions help the deep network to learn non-linear behavior with a significant effect on training dynamics and task performance. The performance of these activations also depends on the initial state of the weight parameters, i.e., different initial state leads to a difference in the performance of a network. In this paper, we have proposed a cooperative initialization for training the deep network using ReLU activation function to impro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 28 publications
0
3
0
Order By: Relevance
“…The memory requirement in CNNs can be viewed either as runtime CPU/GPU memory usage or storage space for the model. A number of recent works [7,1,53,54,49,50,64,14,17,48,33,55,52,51,31,25] have explored such possibilities for efficient deep learning.…”
Section: Introductionmentioning
confidence: 99%
“…The memory requirement in CNNs can be viewed either as runtime CPU/GPU memory usage or storage space for the model. A number of recent works [7,1,53,54,49,50,64,14,17,48,33,55,52,51,31,25] have explored such possibilities for efficient deep learning.…”
Section: Introductionmentioning
confidence: 99%
“…Convolutional neural networks (CNNs) have surpassed many traditional machine learning approaches in solving several computer vision tasks such as classification [9,18], segmentation [2], detection [16,12] and others. Various works [24,25,20,23,21,19,13,26,22] have been proposed for efficient deep learning. Researchers have recently been trying to improve CNN performance, by promoting channels (feature maps) that are more relevant [5].…”
Section: Introductionmentioning
confidence: 99%
“…This had led to considerable interest in making the model more efficient, in terms of storage as well as computation [7,15,50,25,47,45,33,52,13,49,48,59]. A popular approach to increase the efficiency of the model is via model compression.…”
Section: Introductionmentioning
confidence: 99%