2016
DOI: 10.1109/jsen.2015.2487358
|View full text |Cite
|
Sign up to set email alerts
|

A Real-Time Human Action Recognition System Using Depth and Inertial Sensor Fusion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
108
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 154 publications
(108 citation statements)
references
References 19 publications
0
108
0
Order By: Relevance
“…The label information (the forth column) is valuable because it aids in the process of annotation. The data modalities (the fifth column) include color, depth, skeleton and accelerometer, which are helpful for researchers to quickly identify the datasets especially when they work on multi-modal fusion [15,16,58]. Accelerometer data is able to indicate the potential impact of the object and starts an analysis of depth information, at the same time, it simplifies complexity of the motion feature and increases its reliability.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The label information (the forth column) is valuable because it aids in the process of annotation. The data modalities (the fifth column) include color, depth, skeleton and accelerometer, which are helpful for researchers to quickly identify the datasets especially when they work on multi-modal fusion [15,16,58]. Accelerometer data is able to indicate the potential impact of the object and starts an analysis of depth information, at the same time, it simplifies complexity of the motion feature and increases its reliability.…”
Section: Discussionmentioning
confidence: 99%
“…The collection procedure for one object takes roughly 5 mins. In this procedure, four adjustable lights are put in 16 Sampled RGB (top row) and depth images (bottom row) from Kinect FaceDB [64]. Left to right: neutral face with normal illumination, smiling, mouth open, strong illumination, occlusion by sunglasses, occlusion by hand, occlusion by paper, turn face right and turn face left different places, illuminating the recording environment.…”
Section: Big Bird (Berkeley Instance Recognition Dataset)mentioning
confidence: 99%
“…The comparison of the existing work with our approach using the UTD-MHAD dataset for the half-subject experiment is illustrated in Table 3. The second experiments are the subject-specific settings in [19]. As each subject performs an action four times, the first two repetitions are used for training and the two remaining repetitions for testing.…”
Section: Utd-mhad Datasetmentioning
confidence: 99%
“…To get action prediction outputs from the feature variables, we train three classifiers: Collaborative Representation Classifier (CRC) [10,14], Sparse Representation Classifier (SRC) [15,16] and Kernel based Extreme Learning Machine (KELM) [17]. These techniques are among the most widely used methods in the literature [9,10,14,15,18,19], as they have shown good performances for activity recognition systems, but as far we know, this is the first time that these three classifiers are fused together to classify action. Finally, we consider a Naive-Bayes approach to combine the classification scores, that shows an improvement in the accuracy of human action recognition when tested on publicly available datasets [14] [20].…”
Section: Introductionmentioning
confidence: 99%
“…However, the data collected by the Kinect camera not only include RGB data, but also include 3D skeletal data, depth map sequences and infrared videos. Chen et al [4] used a depth camera and an inertial sensor to built a real time action recognition system by a decision-level fusion. Wang et al [5] proposed a algorithm to mine a set of key-pose-motifs to recognize actions from skeletal data by matching a sequence to the motifs of each class and selecting the class that maximizes the matching score.…”
Section: Introductionmentioning
confidence: 99%