Proceedings of the 2018 ACM International Symposium on Wearable Computers 2018
DOI: 10.1145/3267242.3267286
|View full text |Cite
|
Sign up to set email alerts
|

Understanding and improving recurrent networks for human activity recognition by continuous attention

Abstract: Deep neural networks, including recurrent networks, have been successfully applied to human activity recognition. Unfortunately, the final representation learned by recurrent networks might encode some noise (irrelevant signal components, unimportant sensor modalities, etc.). Besides, it is difficult to interpret the recurrent networks to gain insight into the models' behavior. To address these issues, we propose two attention models for human activity recognition: temporal attention and sensor attention. Thes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
100
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 153 publications
(100 citation statements)
references
References 27 publications
0
100
0
Order By: Relevance
“…The Daphnet data was recorded during various walking tasks of 10 different participants and have three different annotations: 1) transient activities (which are discarded here), 2) freezing of gait, and 3) normal movements. The data were recorded with a sampling rate of 64Hz, however, we downsample it to 32Hz by decimation and discard the transient activities, following [40].…”
Section: ) Daphnetmentioning
confidence: 99%
See 2 more Smart Citations
“…The Daphnet data was recorded during various walking tasks of 10 different participants and have three different annotations: 1) transient activities (which are discarded here), 2) freezing of gait, and 3) normal movements. The data were recorded with a sampling rate of 64Hz, however, we downsample it to 32Hz by decimation and discard the transient activities, following [40].…”
Section: ) Daphnetmentioning
confidence: 99%
“…We choose a sliding window of approximately 2.5 seconds for the HARD, HARD2 and HARD3 datasets, with 50% of overlapping. Following other works [8], [26], [40], the PAMAP2 and Daphnet datasets have, respectively, a window size of approximately In these networks, the second FC layer contains the same number of neurons as the number of classes and is followed by the softmax function. The convolutional kernel for all CNN layers was set to 3 × 3, whereas the max-pooling kernel size was 2 × 2.…”
Section: B Data Pre-processingmentioning
confidence: 99%
See 1 more Smart Citation
“…Targeting the spatial and temporal subset selection within such data, attention mechanisms have been recently explored to improve HAR performances on MoCap data. To enable understanding of the relevance of each sensor in such scenario, Zeng et al [19] proposed an attentionbased LSTM framework, where a sensor attention module was used at the input level and for each timestep, with an additional temporal attention module at a later layer. Their sensor attention module was implemented with input from different sensors at single timestep, while temporal attention was computed based on the output of the LSTM layer.…”
Section: B Attention Mechanism Adapted For Har On Mocap Datamentioning
confidence: 99%
“…We also compare with a variant of the BANet with a fullyconnected layer used in the temporal attention computation (BANet-dense) instead of a 1 × 1 convolution layer. In addition, we compare our work with the approach used in related HAR studies [19] [20] [21], where the sensor attention was computed before the extraction of temporal information. As such, we create a variant (BANet-compat, for BANet compatibility version) where the computation of body attention was done at input level instead of at feature fusion level with the same attention algorithms presented in last section.…”
Section: A Comparison Experimentsmentioning
confidence: 99%