The driver gaze zone is an indicator of a driver’s attention and plays an important role in the driver’s activity monitoring. Due to the bad initialization of point-cloud transformation, gaze zone systems using RGB-D cameras and ICP (Iterative Closet Points) algorithm do not work well under long-time head motion. In this work, a solution for a continuous driver gaze zone estimation system in real-world driving situations is proposed, combining multi-zone ICP-based head pose tracking and appearance-based gaze estimation. To initiate and update the coarse transformation of ICP, a particle filter with auxiliary sampling is employed for head state tracking, which accelerates the iterative convergence of ICP. Multiple templates for different gaze zone are applied to balance the templates revision of ICP under large head movement. For the RGB information, an appearance-based gaze estimation method with two-stage neighbor selection is utilized, which treats the gaze prediction as the combination of neighbor query (in head pose and eye image feature space) and linear regression (between eye image feature space and gaze angle space). The experimental results show that the proposed method outperforms the baseline methods on gaze estimation, and can provide a stable head pose tracking for driver behavior analysis in real-world driving scenarios.
Driving behavior analysis is vital for the advanced driving assistance system, aiming to improve driving behavior and decrease traffic accidents. Most existing driving behavior learning methods focus on either vehicle sensor information or driver's attention information, and provide a classification result on the current time data samples. The visualization of driving behavior on time series data samples could give an understanding and review of the driver's continuous actions. However, there has been little progress in combining the multi-modal vehicle and driver information on driving behavior learning and visualization. A multi-information driving behavior learning and visualization method with natural gaze prediction is proposed in this paper, which automatically integrates driver's gaze direction estimated from face camera, and various vehicle sensor data collected from on-board diagnostics (OBD) system. To accurately estimate the eye gaze under large head movement, a novel head pose-free eye gaze prediction method without calibration is proposed based on global and local scale sparse encoding, which treats the direction mapping as small gaze region classification. To understand driving behavior more intuitively, the latent features that represent different driving behaviors are extracted by FastICA from the fused time series data, and mapped into RGB color space for distinguished visualization. Experimental results demonstrate the effectiveness of the proposed method, and show that the proposed method performs better than the compared methods.
We present a simple multi-scale learning network for image classification that is inspired by the MobileNet. The proposed method has two advantages: (1) It uses the multi-scale block with depthwise separable convolutions, which forms multiple sub-networks by increasing the width of the network while keeping the computational resources constant.(2) It combines the multi-scale block with residual connections and that accelerates the training of networks significantly. The experimental results show that the proposed method has strong performance compared to other popular models on different datasets.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.