2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2020
DOI: 10.1109/iros45743.2020.9341574
|View full text |Cite
|
Sign up to set email alerts
|

Tidying Deep Saliency Prediction Architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 34 publications
(12 citation statements)
references
References 24 publications
0
11
0
Order By: Relevance
“…With the development of deep learning, different model structures have been proposed to improve the capabilities of feature representation. For example, (Wang and Shen 2018;Kümmerer et al 2017;Reddy et al 2020;Kroner et al 2020) explored the combination of multi-resolution features, (Cornia et al 2018) applied recurrent architecture and (Lou et al 2022) integrated transformer to refine the learnt features.…”
Section: Human Attention Predictionmentioning
confidence: 99%
See 1 more Smart Citation
“…With the development of deep learning, different model structures have been proposed to improve the capabilities of feature representation. For example, (Wang and Shen 2018;Kümmerer et al 2017;Reddy et al 2020;Kroner et al 2020) explored the combination of multi-resolution features, (Cornia et al 2018) applied recurrent architecture and (Lou et al 2022) integrated transformer to refine the learnt features.…”
Section: Human Attention Predictionmentioning
confidence: 99%
“…Inspired by Reddy et al (2020); Jia and Bruce (2020), we adopt three most popular metrics, i.e., Kullback-Leibler Divergence (KLdiv), Linear Correlation Coefficient (CC) and Normalized Scanpath Saliency (NSS) to construct our loss function. Denote a predicted saliency map as P , the ground truth of human saliency map as Q, and the ground truth fixation map only contains binary values as F .…”
Section: P (Objmentioning
confidence: 99%
“…For CXR image saliency prediction, comparison was conducted with 3 state-of-the-art saliency prediction models, which are SimpleNet (Reddy et al, 2020), MSINet (Kroner et al, 2020) and VGGSSM (Cao et al, 2020). Saliency prediction using standard UNet (denoted as UNetS) is also included for reference.…”
Section: Benchmark Comparisonmentioning
confidence: 99%
“…GT GazeGAN [93] SalGAN [94] UAVDVSM [95] SimpleNet [96] LSR+ Fig. 11: Visualization of predictions of the re-constructed benchmark COL models and our COL base model ("LSR+").…”
Section: Imagementioning
confidence: 99%
“…Discriminative region localization: We introduce the first camouflaged object discriminative region localization task. Considering the same ground truth acquisition process of our task and the widely studied eye fixation prediction task [50] (where both ground truth maps are obtained with eye trackers), we re-train existing eye fixation prediction models (GazeGAN [93], SalGAN [94], UAVDVSM [95] and SimpleNet [96]) with our camouflaged object localization training dataset and construct the first camouflaged object localization benchmark models in Table 2. The better performance of our COL model ("LSR+") compared with the benchmark models validate superiority of our solution.…”
Section: Performance Comparisonmentioning
confidence: 99%