2020
DOI: 10.1145/3411841
|View full text |Cite
|
Sign up to set email alerts
|

IMUTube

Abstract: The lack of large-scale, labeled data sets impedes progress in developing robust and generalized predictive models for on-body sensor-based human activity recognition (HAR). Labeled data in human activity recognition is scarce and hard to come by, as sensor data collection is expensive, and the annotation is time-consuming and error-prone. To address this problem, we introduce IMUTube, an automated processing pipeline that integrates existing computer vision and signal processing techniques to convert videos o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 92 publications
(14 citation statements)
references
References 29 publications
0
10
0
Order By: Relevance
“…Using e.g. a feature extraction tool as seen in [20] could be used to create such larger datasets. Further, we will analyse how 1-layered LSTMs learn compared to 2-layered ones within the setting of HAR by replicating a similar analysis as performed by Karparthy et al [17].…”
Section: Discussionmentioning
confidence: 99%
“…Using e.g. a feature extraction tool as seen in [20] could be used to create such larger datasets. Further, we will analyse how 1-layered LSTMs learn compared to 2-layered ones within the setting of HAR by replicating a similar analysis as performed by Karparthy et al [17].…”
Section: Discussionmentioning
confidence: 99%
“…Zhang and Alshurafa (2020) proposed two deep generative cross-modal architectures for synthesising accelerometer data from video sequences. Kwon et al (2020Kwon et al ( , 2021 presented IMUTube, an automated processing pipeline for human activity recognition (HAR) that integrates existing computer vision and signal processing techniques to convert video of human activity into virtual streams of IMU data. A similar approach was chosen by Lämsä et al (2022), who used neural networks with VIDEO2IMU to generate IMU signals and features from monocular videos of human activities.…”
Section: Data Simulation and Synthesismentioning
confidence: 99%
“…Cross-modality data translation generates synthetic data in the target modality from real data in the source modality. The studies [18,21,35] generate synthetic IMU data from videos of human activities. The work [2] generates mmWave radar data from videos.…”
Section: Multimentioning
confidence: 99%