2022
DOI: 10.1016/j.cag.2022.06.002
|View full text |Cite
|
Sign up to set email alerts
|

SST-Sal: A spherical spatio-temporal approach for saliency prediction in 360 videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

2
6

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 42 publications
0
4
0
Order By: Relevance
“…in video games or other interactive environments, in order to improve user experience and performance; this can be done, for instance, by helping the user conduct the specific task if they are taking too long. Further, existing approaches that focus on modeling and predicting visual behavior in immersive environments [6,28,46] are primarily trained on datasets containing viewing data from free exploration tasks, making them less effective in modeling behaviors related to other tasks. Looking ahead, we believe that different gaze prediction models could leverage our insights and incorporate behavioral priors, be fine-tuned, or be directly trained with task-dependent gaze data.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…in video games or other interactive environments, in order to improve user experience and performance; this can be done, for instance, by helping the user conduct the specific task if they are taking too long. Further, existing approaches that focus on modeling and predicting visual behavior in immersive environments [6,28,46] are primarily trained on datasets containing viewing data from free exploration tasks, making them less effective in modeling behaviors related to other tasks. Looking ahead, we believe that different gaze prediction models could leverage our insights and incorporate behavioral priors, be fine-tuned, or be directly trained with task-dependent gaze data.…”
Section: Discussionmentioning
confidence: 99%
“…Since then, several studies have gathered large datasets of free-viewing behavior [9,54]. Subsequent works have leveraged them to model visual attention, usually based on mechanisms such as visual saliency [6,47] or scanpath prediction [2,46]. More recently, these models have also incorporated auditory cues to account for multimodal attention [10,66].…”
Section: Related Workmentioning
confidence: 99%
“…However, as VR environments are often dynamic, these models may not be sufficient for certain applications. To address this, some recent works have focused on attention prediction in 360 • videos [3,9,12]. Nevertheless, all these models only take visual stimuli as input, and therefore they do not take into account the potential influence of sound in VR environments [30].…”
Section: Analyzing and Predicting Viewing Behavior In Vrmentioning
confidence: 99%
“…We showcase this application scenario by evaluating the performance of recent state-of-the-art audiovisual 360 • saliency predictors on D-SAV360, specifically those proposed by Cokelek et al [10] and Chao et al [8]. We chose to implement Cokelek et al's method on top of SST-Sal [3], as it requires a video saliency predictor as a base architecture. Please refer to Section S.8 in the supplementary for implementation details.…”
Section: Applications Of Our Datasetmentioning
confidence: 99%
“…O terceiro grau de liberdade pode ser ajustado para colocar conteúdo de interesse à frente do panorama. Isso pode ser feito manualmente, onde o usuário ajusta a rotação de acordo com seus interesses pessoais, ou até de modo automático, através do uso de técnicas que identificam regiões "interessantes" na imagem esférica usando técnicas de saliência visual [Bernal-Berdun et al 2022].…”
Section: Correção De Orientaçãounclassified