2019
DOI: 10.1007/978-3-030-12939-2_21
|View full text |Cite
|
Sign up to set email alerts
|

Context-driven Multi-stream LSTM (M-LSTM) for Recognizing Fine-Grained Activity of Drivers

Abstract: Automatic recognition of in-vehicle activities has significant impact on the next generation intelligent vehicles. In this paper, we present a novel Multi-stream Long Short-Term Memory (M-LSTM) network for recognizing driver activities. We bring together ideas from recent works on LSTMs, transfer learning for object detection and body pose by exploring the use of deep convolutional neural networks (CNN). Recent work has also shown that representations such as hand-object interactions are important cues in char… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
27
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1

Relationship

4
2

Authors

Journals

citations
Cited by 18 publications
(27 citation statements)
references
References 35 publications
0
27
0
Order By: Relevance
“…These models have the ability to long-term dependencies by incorporating memory units. These memory units allow the network to learn, forget previously hidden states, and update hidden states (Behera et al 2018). Figure 7b depicts the general arrangement of an LSTM memory cell.…”
Section: Dnn With Long Short-term Memory (Lstm) Layersmentioning
confidence: 99%
See 1 more Smart Citation
“…These models have the ability to long-term dependencies by incorporating memory units. These memory units allow the network to learn, forget previously hidden states, and update hidden states (Behera et al 2018). Figure 7b depicts the general arrangement of an LSTM memory cell.…”
Section: Dnn With Long Short-term Memory (Lstm) Layersmentioning
confidence: 99%
“…where w x ; b x ; ; i t ; j t ; f t ; o t are weight matrices, biases, element-wise vector product, input gate contributing to Connection solar panels to the weatherboard memory, input moderation gate contributing to memory, forget gate, and output gate as a multiplier between memory gates, respectively. The c t and h t are the two types of hidden layers to allow the LSTM to make complex decisions over a short period of time (Behera et al 2018;Jozefowicz et al 2015). The i t and f t gates are switching each other to selectively consider the current inputs or forget its previous memory.…”
Section: Dnn With Long Short-term Memory (Lstm) Layersmentioning
confidence: 99%
“…Martin et al [17] advocate a method to combine multiple streams involving body pose and contextual information in videos to recognise driver's activities. Similarly, Behera et al [13] describe a multi-stream LSTM for recognising driver's activities by combining high-level body pose and body-object interaction with CNN features. These models [16], [15], [17], [13] are similar to video classification methods, which require complete observation and is unsuitable for live activity recognition.…”
Section: Related Workmentioning
confidence: 99%
“…Similarly, Behera et al [13] describe a multi-stream LSTM for recognising driver's activities by combining high-level body pose and body-object interaction with CNN features. These models [16], [15], [17], [13] are similar to video classification methods, which require complete observation and is unsuitable for live activity recognition. Simialrly Alotaibi and Alotaibi [19] describe an approach that combines the inception module with a residual block and a hierarchical recurrent neural network to enhance the recognition performance of the distracted behaviours of drivers.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation