2021
DOI: 10.3390/math9222908
|View full text |Cite
|
Sign up to set email alerts
|

Truck Driver Fatigue Detection Based on Video Sequences in Open-Pit Mines

Abstract: Due to complex background interference and weak space–time connection, traditional driver fatigue detection methods perform poorly for open-pit truck drivers. For these issues, this paper presents a driver fatigue detection method based on Libfacedetection and an LRCN. The method consists of three stages: (1) using a face detection module with a tracking method to quickly extract the ROI of the face; (2) extracting and coding the features; (3) combining the coding model to build a spatiotemporal classification… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 21 publications
(18 reference statements)
0
0
0
Order By: Relevance
“…Traditional driver fatigue detection methods encounter limitations in capturing the nuanced spatiotemporal dynamics of continuous video frames, crucial for identifying subtle fatigue-related facial movements [ 22 , 23 ]. Addressing these deficiencies, we introduce the Composite Action Recognition Network, merging the capabilities of 3D convolutional networks (3D CNN) and Long Short-Term Memory networks (LSTM).…”
Section: Our Approachmentioning
confidence: 99%
“…Traditional driver fatigue detection methods encounter limitations in capturing the nuanced spatiotemporal dynamics of continuous video frames, crucial for identifying subtle fatigue-related facial movements [ 22 , 23 ]. Addressing these deficiencies, we introduce the Composite Action Recognition Network, merging the capabilities of 3D convolutional networks (3D CNN) and Long Short-Term Memory networks (LSTM).…”
Section: Our Approachmentioning
confidence: 99%
“…In other words, convolutional layers are used to extract spatial features from the frames, and the spatial features are sent to LSTM layers at each time step to model temporal sequences, as shown in Figure 5. In this way, the network learns spatial and temporal features immediately in an end-to-end training process, which makes the model more stable [22][23][24][25][26]. This means that most of the time ϕ_v (v_t), the convolutional inference, and training can be completed in parallel over time.…”
Section: Long-term Recurrent Convolutional Modelmentioning
confidence: 99%