2020
DOI: 10.48550/arxiv.2006.05675
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

IMUTube: Automatic Extraction of Virtual on-body Accelerometry from Video for Human Activity Recognition

Abstract: The lack of large-scale, labeled data sets impedes progress in developing robust and generalized predictive models for on-body sensor-based human activity recognition (HAR). Labeled data in human activity recognition is scarce and hard to come by, as sensor data collection is expensive, and the annotation is time-consuming and error-prone. To address this problem, we introduce IMUTube, an automated processing pipeline that integrates existing computer vision and signal processing techniques to convert videos o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 64 publications
0
4
0
Order By: Relevance
“…Recent approaches addressing the lack of annotated datasets involve extraction of virtual movement data from modalities such as mocap [37,72] and videos [44], which contain a large number of participants and diverse set of actions and environments. Such techniques have the advantage of being capable of extracting large quantities of virtual movement data at arbitrary positions on the body.…”
Section: Research Agenda: Representation Learning For Human Activity ...mentioning
confidence: 99%
“…Recent approaches addressing the lack of annotated datasets involve extraction of virtual movement data from modalities such as mocap [37,72] and videos [44], which contain a large number of participants and diverse set of actions and environments. Such techniques have the advantage of being capable of extracting large quantities of virtual movement data at arbitrary positions on the body.…”
Section: Research Agenda: Representation Learning For Human Activity ...mentioning
confidence: 99%
“…Similar to BMTL, the loss function of TMTL is also calculated on each individual task. We then take the summation of losses over all source domains as the final objective function (1).…”
Section: Invariant Feature Learningmentioning
confidence: 99%
“…Human activity recognition (HAR) is the foundation to realize remote health services and in-home mobility monitoring. Although deep learning has seen many successes in this field, training deep models often requires a large amount of sensory data that is not always available [1]. For research ethics compliance, it often takes months to design study protocols, recruit volunteers and collect customized sensory datasets.…”
Section: Introductionmentioning
confidence: 99%
“…For example, the smartphonebased speech assistant (e.g. ProxiTalk [57]), and wearable sensor-enabled activity recognition (IMUTube [34], MITIER [8]). Notably, there is a growing trend to bring deep learning (e.g.…”
Section: Introductionmentioning
confidence: 99%