2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01268
|View full text |Cite
|
Sign up to set email alerts
|

DeepGaze IIE: Calibrated prediction in and out-of-domain for state-of-the-art saliency modeling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 45 publications
(34 citation statements)
references
References 24 publications
0
34
0
Order By: Relevance
“…In DeepGaze II [45], the fine-tuning of the VGG-19 features is used for saliency detection. The performance of saliency model is enhanced by using ResNet50 features instead of the VGG19 of DeepGaze II in [46]. In addition to the aforementioned algorithms which are mostly in the spatial domain, some models work in the frequency domain and benefit from Fourier transform, discrete cosine transform, etc.…”
Section: Related Workmentioning
confidence: 99%
“…In DeepGaze II [45], the fine-tuning of the VGG-19 features is used for saliency detection. The performance of saliency model is enhanced by using ResNet50 features instead of the VGG19 of DeepGaze II in [46]. In addition to the aforementioned algorithms which are mostly in the spatial domain, some models work in the frequency domain and benefit from Fourier transform, discrete cosine transform, etc.…”
Section: Related Workmentioning
confidence: 99%
“…• Saliency Baseline: As a complement to the baselines above, this baseline selects the query image whose original pixels hidden by the occlusion patch have a lower probability of being looked at by humans. This simulates that participants select the image with a patch that occludes less prominent information and is estimated with the saliency prediction model DeepGaze IIE (Linardos et al, 2021). Specifically, we pass the unoccluded query image through the saliency prediction model DeepGaze IIE (Linardos et al, 2021), which yields a probability density over the entire image.…”
Section: A5 Baselinesmentioning
confidence: 99%
“…This simulates that participants select the image with a patch that occludes less prominent information and is estimated with the saliency prediction model DeepGaze IIE (Linardos et al, 2021). Specifically, we pass the unoccluded query image through the saliency prediction model DeepGaze IIE (Linardos et al, 2021), which yields a probability density over the entire image. Next, we integrate said density over each of the two square patches.…”
Section: A5 Baselinesmentioning
confidence: 99%
“…Yang et al [43] approximated the foveated retina by using the segmentation maps of a full-resolution image and its blurred version, predicted by a pretrained Panoptic-FPN [21], to approximate high-resolution fovea and low-resolution peripheral, respectively. Like other models for predicting human attention [30,24,25,6,46], both approaches rely on pretrained networks to extract image features and train much smaller networks for the downstream tasks using transfer learning, usually due to the lack of human fixation data for training. Also noteworthy is that these approaches apply networks pretrained on full-resolution images (e.g., ResNets [14] trained on Im-ageNet [38]), expecting the pretrained networks to approximate how humans perceive blurred images.…”
Section: Introductionmentioning
confidence: 99%