2020
DOI: 10.1109/tits.2019.2915540
|View full text |Cite
|
Sign up to set email alerts
|

How Do Drivers Allocate Their Potential Attention? Driving Fixation Prediction via Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
48
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 83 publications
(48 citation statements)
references
References 34 publications
0
48
0
Order By: Relevance
“…We trained a convolutiondeconvolution gaze network (Palazzi et al 2018;Zhang et al 2018b;2018a;Deng et al 2019) with KL divergence ( = 1e − 10) as loss function to predict human gaze positions. A separate network is trained for each game.…”
Section: Baseline Model and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We trained a convolutiondeconvolution gaze network (Palazzi et al 2018;Zhang et al 2018b;2018a;Deng et al 2019) with KL divergence ( = 1e − 10) as loss function to predict human gaze positions. A separate network is trained for each game.…”
Section: Baseline Model and Resultsmentioning
confidence: 99%
“…Considerable evidence has shown that human gaze can be considered as an overt behavioral signal that encodes a wealth of information about both the motivation behind an action and the anticipated reward of an action (Hayhoe and Ballard 2005;Johnson et al 2014). Recent work has also proposed learning visual attention models from human gaze as an intermediate step towards learning the decision policy, and this intermediate signal has been shown to improve policy learning (Li, Liu, and Rehg 2018;Zhang et al 2018b;Xia et al 2019;Chen et al 2019;Liu et al 2019;Deng et al 2019) Addressing the demands and challenges described above, we collected a large-scale dataset of humans playing Atari video games -one of the most widely used task domain in RL and IL research. The dataset is named Atari-HEAD (Atari Human Eye-Tracking And Demonstration) 1 .…”
Section: Introductionmentioning
confidence: 99%
“…Ye Xia et al present in [51] an in-lab dataset called Berkeley DeepDrive Attention (BDD-A), together with a system which includes a convolutional LSTM network and a method to show relevant frames to the model more frequently. Tao Deng et al [52] provide a traffic driving video dataset with fixations, together with a saliency detection model based on compact convolutional-deconvolutional neural networks (CDNN). Either BDD-A [51] or CDNN [52] databases have more diverse fixations than DR(eye)VE [49], but recorded in a laboratory and not under real driving conditions.…”
Section: B Visual Attention For Autonomous Vehiclesmentioning
confidence: 99%
“…Tao Deng et al [52] provide a traffic driving video dataset with fixations, together with a saliency detection model based on compact convolutional-deconvolutional neural networks (CDNN). Either BDD-A [51] or CDNN [52] databases have more diverse fixations than DR(eye)VE [49], but recorded in a laboratory and not under real driving conditions. Moreover, DR(eye)VE [49] provides labels for contextual conditions, which enables us to demonstrate our model's capabilities.…”
Section: B Visual Attention For Autonomous Vehiclesmentioning
confidence: 99%
See 1 more Smart Citation