2023
DOI: 10.1109/access.2023.3318563
|View full text |Cite
|
Sign up to set email alerts
|

Accuracy Comparison of CNN, LSTM, and Transformer for Activity Recognition Using IMU and Visual Markers

María Fernanda Trujillo-Guerrero,
Stadyn Román-Niemes,
Milagros Jaén-Vargas
et al.

Abstract: Human activity recognition (HAR) has applications ranging from security to healthcare. Typically these systems are composed of data acquisition and activity recognition models. In this work, we compared the accuracy of two acquisition systems: Inertial Measurement Units (IMUs) vs Movement Analysis Systems (MAS). We trained models to recognize arm exercises using state-of-the-art deep learning architectures and compared their accuracy. MAS uses a camera array and reflective markers. IMU uses accelerometers, gyr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
1
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 47 publications
0
1
0
Order By: Relevance
“…The scenarios using a large window (long time series) for OEP recognition could take advantage of CNN-BiLSTM. Similar results were also reported by a previous study, reporting that CNN-LSTM outperformed CNN, LSTM, and Transformer for IMU-based activity recognition [18]. However, deep learning models require more training examples than classical models.…”
Section: A Stagesupporting
confidence: 89%
See 2 more Smart Citations
“…The scenarios using a large window (long time series) for OEP recognition could take advantage of CNN-BiLSTM. Similar results were also reported by a previous study, reporting that CNN-LSTM outperformed CNN, LSTM, and Transformer for IMU-based activity recognition [18]. However, deep learning models require more training examples than classical models.…”
Section: A Stagesupporting
confidence: 89%
“…This architecture has been validated for many datasets with better recognition performance than both CNN and LSTM [19], [20]. It also outperformed Transformers in recent research for HAR [18]. Based on this architecture and Bi-directional-LSTM (BiLSTM) layers, the CNN-BiLSTM was further developed [21], [22].…”
Section: B Machine Learning Modelsmentioning
confidence: 96%
See 1 more Smart Citation
“…Nevertheless, the superior performance of deep learning, such as convolutional neural networks (CNNs), long short-term memory (LSTM), and attention mechanisms, makes them attractive for PAR and EE estimation applications [33,34]. In [35], CNN and LSTM with self-attention mechanisms improved PAR performance, and a similar experiment in [36] enhanced arm exercise activity recognition. CNN-LSTM models showed improved performance and were used by other recent researchers [37,38].…”
Section: Previous Studies and Our Contributionmentioning
confidence: 99%