2018 25th IEEE International Conference on Image Processing (ICIP) 2018
DOI: 10.1109/icip.2018.8451339
|View full text |Cite
|
Sign up to set email alerts
|

Effnet: An Efficient Structure for Convolutional Neural Networks

Abstract: With the ever increasing application of Convolutional Neural Networks to customer products the need emerges for models to efficiently run on embedded, mobile hardware. Slimmer models have therefore become a hot research topic with various approaches which vary from binary networks to revised convolution layers. We offer our contribution to the latter and propose a novel convolution block which significantly reduces the computational burden while surpassing the current stateof-the-art. Our model, dubbed EffNet,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
66
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 101 publications
(78 citation statements)
references
References 18 publications
(44 reference statements)
2
66
0
2
Order By: Relevance
“…There are a few possibilities for decreasing the computational demand in our approach. First of all, a lot of research has been done on creating efficient Convolutional Neural Network architectures for low latency predictions on mobile hardware [34,35]. We did not focus on efficiency and it should be relatively easy to decrease the computational load by optimizing the architecture.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…There are a few possibilities for decreasing the computational demand in our approach. First of all, a lot of research has been done on creating efficient Convolutional Neural Network architectures for low latency predictions on mobile hardware [34,35]. We did not focus on efficiency and it should be relatively easy to decrease the computational load by optimizing the architecture.…”
Section: Conclusion and Discussionmentioning
confidence: 99%
“…Currently, the design of computationally efficient CNNs is moving from manual tuning [17][18][19] towards automatic algorithms [20][21][22][23][24][25][26]. The incorporation of specific platform constraints to such approaches involves modeling how the network architecture relates with the optimization target.…”
Section: Related Workmentioning
confidence: 99%
“…Following these observations, we define an image patch as hard-to-upscale based on the following criterion: (9) where the TV threshold TV thr is a tunable parameter whose value is automatically configured by MobiSR as discussed in Section 3.5.…”
Section: Difficulty Evaluation Unitmentioning
confidence: 99%
“…So far, substantial effort has been invested on developing network compression techniques, such as pruning [12,57], quantization [11], and knowledge distillation [18], for building efficient neural networks. In particular, a number of convolution approximations, such as low-rank tensor decomposition [47], have been successfully employed as a primary component in building fast and accurate discriminative vision models [9,19,58,61]. These techniques typically aim to express a convolution as a sequence of simpler tensor operations, reducing in this manner the storage and computation cost of the network.…”
Section: Introductionmentioning
confidence: 99%