2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00514
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting Video Saliency: A Large-Scale Benchmark and a New Model

Abstract: In this work, we contribute to video saliency research in two ways. First, we introduce a new benchmark for predicting human eye movements during dynamic scene freeviewing, which is long-time urged in this field. Our dataset, named DHF1K (Dynamic Human Fixation), consists of 1K high-quality, elaborately selected video sequences spanning a large range of scenes, motions, object types and background complexity. Existing video saliency datasets lack variety and generality of common dynamic scenes and fall short i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
222
0
4

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 222 publications
(231 citation statements)
references
References 64 publications
2
222
0
4
Order By: Relevance
“…Metric NSS CC SIM AUC-J s-AUC GBVS 1.775 0.331 0.201 0.855 0.592 SALICON [20] 1.901 0.327 0.232 0.857 0.590 OM-CNN [19] 1.911 0.344 0.256 0.856 0.583 DVA [38] 2.013 0.358 0.262 0.860 0.595 SalGAN [29] 2.043 0.370 0.262 0.866 0.709 ACLNet [39] 2.354 0.434 0.315 0.890 0.601 TASED-Net 2.667 0.470 0.361 0.895 0.712 itative results of our model and ACLNet for the better and worse cases are given in Figure 5 (see Supplementary material for more examples of qualitative results). As shown in (a) and (b) in Figure 5, TASED-Net seems highly sensitive to salient moving objects and less sensitive to background objects, which is consistent with the goal of video saliency in general.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Metric NSS CC SIM AUC-J s-AUC GBVS 1.775 0.331 0.201 0.855 0.592 SALICON [20] 1.901 0.327 0.232 0.857 0.590 OM-CNN [19] 1.911 0.344 0.256 0.856 0.583 DVA [38] 2.013 0.358 0.262 0.860 0.595 SalGAN [29] 2.043 0.370 0.262 0.866 0.709 ACLNet [39] 2.354 0.434 0.315 0.890 0.601 TASED-Net 2.667 0.470 0.361 0.895 0.712 itative results of our model and ACLNet for the better and worse cases are given in Figure 5 (see Supplementary material for more examples of qualitative results). As shown in (a) and (b) in Figure 5, TASED-Net seems highly sensitive to salient moving objects and less sensitive to background objects, which is consistent with the goal of video saliency in general.…”
Section: Methodsmentioning
confidence: 99%
“…Since the ground-truth annotations for the test set of DHF1K [39] are hidden for fair comparison, we first evaluate variants of our model on the validation set. The performance of TASED-Net with different T and temporal aggregation strategies are compared in Table 1.…”
Section: Evaluation On Dhf1kmentioning
confidence: 99%
See 3 more Smart Citations