2017
DOI: 10.1109/tnnls.2016.2541681
|View full text |Cite
|
Sign up to set email alerts
|

Sparseness Analysis in the Pretraining of Deep Neural Networks

Abstract: A major progress in deep multilayer neural networks (DNNs) is the invention of various unsupervised pretraining methods to initialize network parameters which lead to good prediction accuracy. This paper presents the sparseness analysis on the hidden unit in the pretraining process. In particular, we use the L -norm to measure sparseness and provide some sufficient conditions for that pretraining leads to sparseness with respect to the popular pretraining models-such as denoising autoencoders (DAEs) and restri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
26
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 63 publications
(27 citation statements)
references
References 24 publications
1
26
0
Order By: Relevance
“…For patch style loss network ϕ s , since existing networks are mainly trained for full images, instead of directly applying the existing pretrained discriminator network, we apply Generative Adversarial Network (GAN) [Goodfellow et al, 2014;Radford et al, 2015] for learning ϕ s and meanwhile initializing the parameters of decoder D e of the generator. We will describe it in the next subsection.…”
Section: Objective Function Of Discriminatormentioning
confidence: 99%
See 2 more Smart Citations
“…For patch style loss network ϕ s , since existing networks are mainly trained for full images, instead of directly applying the existing pretrained discriminator network, we apply Generative Adversarial Network (GAN) [Goodfellow et al, 2014;Radford et al, 2015] for learning ϕ s and meanwhile initializing the parameters of decoder D e of the generator. We will describe it in the next subsection.…”
Section: Objective Function Of Discriminatormentioning
confidence: 99%
“…We firstly describe utilizing GAN [Goodfellow et al, 2014;Radford et al, 2015] for learning patch style network ϕ s and meanwhile initializing the parameters of decoder D e . The inputs of this stage are image patches X (2) and the style image y s .…”
Section: Optimization Of Generatormentioning
confidence: 99%
See 1 more Smart Citation
“…Existing methods have shown satisfactory performance of visual feature representation using deep learning methods [46,82,83,84,16,85]. For example, Kiapour et al applied CNN features as image representation, and calculated the cross-entropy loss measuring whether two images are matched or non-matched.…”
Section: Background and Motivationmentioning
confidence: 99%
“…Therefore, we utilize a pretrained image classification network (i.e., VGG-19) [112,113] as the initialization of E n . Also, the VGG network is utilized as the global loss network φ and the patch content loss network ϕ c .…”
Section: Architecturementioning
confidence: 99%