2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00244
|View full text |Cite
|
Sign up to set email alerts
|

Resolution Adaptive Networks for Efficient Inference

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
123
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 161 publications
(124 citation statements)
references
References 20 publications
0
123
0
Order By: Relevance
“…The AdaFocusV2 network studied in this paper can be classified into this category as well. Many of the spatially adaptive networks are designed from the lens of inference efficiency [5,24,52,62,71]. For example, recent works have revealed that 2D images can be efficiently processed via attending to the task-relevant or more informative image regions [17,61,66,70].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The AdaFocusV2 network studied in this paper can be classified into this category as well. Many of the spatially adaptive networks are designed from the lens of inference efficiency [5,24,52,62,71]. For example, recent works have revealed that 2D images can be efficiently processed via attending to the task-relevant or more informative image regions [17,61,66,70].…”
Section: Related Workmentioning
confidence: 99%
“…In specific, we attach two linear classifiers, FC G (•) and FC L (•), to the outputs of f G and f L , and replace the loss function L in (9) by L : we assume that processing a subset of frames (from the beginning) rather than all may be sufficient for these "easier" samples. To implement this idea, at test time, we propose to compare the largest entry of the softmax prediction p t (defined as confidence in previous works [28,65,66,71]) at t th frame with a pre-defined threshold η t . Once max j p tj ≥ η t , the prediction will be postulated to be reliable enough, and the inference will be terminated by outputting p t .…”
Section: Training Techniquesmentioning
confidence: 99%
“…Special architectures. One way is to change the architecture of the model to support adaptive computations [4,14,15,18,25,27,30,37,42,51,54]. For example, models that represent a neural network as a fixed-point function can have the property of adaptive computation by default.…”
Section: Related Workmentioning
confidence: 99%
“…Using ODEs requires a specific solver, is often slower than fix depth models and requires adding extra constrains on the model design. [54] learns a set of classifiers with different resolu-tions executed in order; computation stops when confidence of the model is above the threshold. [27] proposed a residual variant with shared weights and a halting mechanism.…”
Section: Related Workmentioning
confidence: 99%
“…Such works have followed either multi-scale or HyperNet strategies. While the former redesigns network topology to encode features from shallow and deep layers [Yang et al 2020], the latter preserves network topology, encouraging application on off-the-shelf networks [Sindagi and Patel 2019]. Despite the positive results, both strategies increase the computational burden significantly since they insert time-consuming operations at multiple levels of the network.…”
Section: Introductionmentioning
confidence: 99%