2020
DOI: 10.1109/tmm.2019.2947352
|View full text |Cite
|
Sign up to set email alerts
|

A Dilated Inception Network for Visual Saliency Prediction

Abstract: Recently, with the advent of deep convolutional neural networks (DCNN), the improvements in visual saliency prediction research are impressive. One possible direction to approach the next improvement is to fully characterize the multiscale saliency-influential factors with a computationally-friendly module in DCNN architectures. In this work, we proposed an end-to-end dilated inception network (DINet) for visual saliency prediction. It captures multi-scale contextual features effectively with very limited extr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
97
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 122 publications
(97 citation statements)
references
References 59 publications
0
97
0
Order By: Relevance
“…We firstly clarify two similar concepts, attention model and attentionbased deep learning model. Attention models are a class of models that aim to predict task-free saliency, including human fixation prediction [8,12,17,33,40] and salient object detection [13,34,36,45,47]. While attention-based deep learning models are deep learning models that aim to enhance their representational power by predicting weights of intermediate features in a task-specific context.…”
Section: Related Work 21 Attention-based Deep Learning Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…We firstly clarify two similar concepts, attention model and attentionbased deep learning model. Attention models are a class of models that aim to predict task-free saliency, including human fixation prediction [8,12,17,33,40] and salient object detection [13,34,36,45,47]. While attention-based deep learning models are deep learning models that aim to enhance their representational power by predicting weights of intermediate features in a task-specific context.…”
Section: Related Work 21 Attention-based Deep Learning Modelmentioning
confidence: 99%
“…The attention weights are projected to the corresponding bounded regions in ascending order so that low-attention regions will be covered by high-attention regions when there is any overlapping. The fixation maps are generated with Yang et al's work [40]. Figure 6: IAA-driven object-level attention maps (labels are omitted) that do not fully align with human attention.…”
Section: Further Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…Liu et al 40 planned a multibranch residual module with dilated convolutions to extract multiscale features so that the classification and identification of spacecraft electronic load signals can be solved. Yang et al 41 designed an end-to-end dilated inception network (DINet) to predict visual saliency maps. The dilated inception module of the DINet used dilated convolutions with different dilation rates in parallel, which not only can significantly reduce the computational load but also can enrich the diversity of the receptive field in the features.…”
Section: Dilated Convolutionmentioning
confidence: 99%
“…Srinivas et al [36] devised a fully CNN for predicting human eye fixation. Yang et al [37] proposed a dilated inception network for visual saliency prediction. Lv et al [38] developed an attention-based fusion network for human eye-fixation prediction in 3-D images.…”
Section: Introductionmentioning
confidence: 99%