2016
DOI: 10.1109/tpami.2015.2502579
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Very Deep Convolutional Networks for Classification and Detection

Abstract: Abstract-This paper aims to accelerate the test-time computation of convolutional neural networks (CNNs), especially very deep CNNs [1] that have substantially impacted the computer vision community. Unlike previous methods that are designed for approximating linear filters or linear responses, our method takes the nonlinear units into account. We develop an effective solution to the resulting nonlinear optimization problem without the need of stochastic gradient descent (SGD). More importantly, while previous… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
456
1

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 702 publications
(466 citation statements)
references
References 42 publications
4
456
1
Order By: Relevance
“…• CPD [17]: CP-Decomposition; • GBD [18]: Group-wise Brain Damage; • LANR [31]: Low-rank Approximation of Non-linear Responses.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…• CPD [17]: CP-Decomposition; • GBD [18]: Group-wise Brain Damage; • LANR [31]: Low-rank Approximation of Non-linear Responses.…”
Section: Methodsmentioning
confidence: 99%
“…In Table 5, we also report the comparison against LANR [31] on VGG-16. For the similar speed-up rate (4×), their approach outperforms ours in the top-5 classification error (an increase of 0.95% against 1.83%).…”
Section: Quantizing the Convolutional Layermentioning
confidence: 99%
“…Compared with traditional data mapping, He et al [35] found that residual mapping can acquire a more effective learning effect and rapidly reduce the training loss after passing through a multi-layer network, which has achieved a state-of-the-art performance in object detection [36], image super-resolution [37], and so on. Essentially, Szegedy et al [38] demonstrated that residual networks take full advantage of identity shortcut connections, which can efficiently transfer various levels of feature information between not directly connected layers without attenuation.…”
Section: Residual Learningmentioning
confidence: 99%
“…Examples include the CNN, widely used for image classi cation [25,26] and recently for text classi cation [27], and the DNN, used for speech recognition [28]. In our pilot work [29], we assumed that an important noti cation depends on the noti cation's contents and the user's context.…”
Section: Machine Learning Modelsmentioning
confidence: 99%