Abstract:The technology for human activity recognition has become an active research topic in recent years as it has many potential applications, such as surveillance systems, healthcare systems, and human-computer interaction. In the research of activity recognition, supervised machine learning approaches have been widely used for activity recognition. However, the cost of collecting labeled sensor data in new environments is high. Furthermore, these methods do not work well in a crossdomain environment using conventi… Show more
“…In Table A.5 (“Appendix 1 ”), the description of sensor-datasets is illustrated with attributes such as data source, #factors, sensor location, and activity type. It includes wearable sensor-based datasets (Alsheikh et al 2016 ; Asteriadis and Daras 2017 ; Zhang et al 2012 ; Chavarriaga et al 2013 ; Munoz-Organero 2019 ; Roggen et al 2010 ; Qin et al 2019 ), as well as smart-device sensor-based datasets (Ravi et al 2016 ; Cui and Xu 2013 ; Weiss et al 2019 ; Miu et al 2015 ; Reiss and Stricker 2012a , b ; Lv et al 2020 ; Gani et al 2019 ; Stisen et al 2015 ; Röcker et al 2017 ; Micucci et al 2017 ) Apart from datasets mentioned in Table A.5 , there are few more datasets worth mentioning such as Kasteren dataset (Kasteren et al 2011 ; Chen et al 2017 ), which is also very popular. (2) Vision-based HAR: Devices for collecting 3D data are CCTV cameras (Koppula and Saxena 2016 ; Devanne et al 2015 ; Zhang and Parker 2016 ; Li et al 2010 ; Duan et al 2020 ; Kalfaoglu et al 2020 ; Gorelick et al 2007 ; Mahadevan et al 2010 ), depth cameras (Cippitelli et al 2016 ; Gaglio et al 2015 ; Neili Boualia and Essoukri Ben Amara 2021 ; Ding et al 2016 ; Cornell Activity Datasets: CAD-60 & CAD-120 2021 ), and videos from public domains like YouTube and Hollywood movie scenes (Gu et al 2018 ; Soomro et al 2012 ; Kuehne et al 2011 ; Sigurdsson et al 2016 ; Kay et al 2017 ; Carreira et al 2018 ; Goyal et al 2017 ).…”
Human activity recognition (HAR) has multifaceted applications due to its worldly usage of acquisition devices such as smartphones, video cameras, and its ability to capture human activity data. While electronic devices and their applications are steadily growing, the advances in Artificial intelligence (AI) have revolutionized the ability to extract deep hidden information for accurate detection and its interpretation. This yields a better understanding of rapidly growing acquisition devices, AI, and applications, the three pillars of HAR under one roof. There are many review articles published on the general characteristics of HAR, a few have compared all the HAR devices at the same time, and few have explored the impact of evolving AI architecture. In our proposed review, a detailed narration on the three pillars of HAR is presented covering the period from 2011 to 2021. Further, the review presents the recommendations for an improved HAR design, its reliability, and stability. Five major findings were: (1) HAR constitutes three major pillars such as devices, AI and applications; (2) HAR has dominated the healthcare industry; (3) Hybrid AI models are in their infancy stage and needs considerable work for providing the stable and reliable design. Further, these trained models need solid prediction, high accuracy, generalization, and finally, meeting the objectives of the applications without bias; (4) little work was observed in abnormality detection during actions; and (5) almost no work has been done in forecasting actions. We conclude that: (a) HAR industry will evolve in terms of the three pillars of electronic devices, applications and the type of AI. (b) AI will provide a powerful impetus to the HAR industry in future.
Supplementary Information
The online version contains supplementary material available at 10.1007/s10462-021-10116-x.
“…In Table A.5 (“Appendix 1 ”), the description of sensor-datasets is illustrated with attributes such as data source, #factors, sensor location, and activity type. It includes wearable sensor-based datasets (Alsheikh et al 2016 ; Asteriadis and Daras 2017 ; Zhang et al 2012 ; Chavarriaga et al 2013 ; Munoz-Organero 2019 ; Roggen et al 2010 ; Qin et al 2019 ), as well as smart-device sensor-based datasets (Ravi et al 2016 ; Cui and Xu 2013 ; Weiss et al 2019 ; Miu et al 2015 ; Reiss and Stricker 2012a , b ; Lv et al 2020 ; Gani et al 2019 ; Stisen et al 2015 ; Röcker et al 2017 ; Micucci et al 2017 ) Apart from datasets mentioned in Table A.5 , there are few more datasets worth mentioning such as Kasteren dataset (Kasteren et al 2011 ; Chen et al 2017 ), which is also very popular. (2) Vision-based HAR: Devices for collecting 3D data are CCTV cameras (Koppula and Saxena 2016 ; Devanne et al 2015 ; Zhang and Parker 2016 ; Li et al 2010 ; Duan et al 2020 ; Kalfaoglu et al 2020 ; Gorelick et al 2007 ; Mahadevan et al 2010 ), depth cameras (Cippitelli et al 2016 ; Gaglio et al 2015 ; Neili Boualia and Essoukri Ben Amara 2021 ; Ding et al 2016 ; Cornell Activity Datasets: CAD-60 & CAD-120 2021 ), and videos from public domains like YouTube and Hollywood movie scenes (Gu et al 2018 ; Soomro et al 2012 ; Kuehne et al 2011 ; Sigurdsson et al 2016 ; Kay et al 2017 ; Carreira et al 2018 ; Goyal et al 2017 ).…”
Human activity recognition (HAR) has multifaceted applications due to its worldly usage of acquisition devices such as smartphones, video cameras, and its ability to capture human activity data. While electronic devices and their applications are steadily growing, the advances in Artificial intelligence (AI) have revolutionized the ability to extract deep hidden information for accurate detection and its interpretation. This yields a better understanding of rapidly growing acquisition devices, AI, and applications, the three pillars of HAR under one roof. There are many review articles published on the general characteristics of HAR, a few have compared all the HAR devices at the same time, and few have explored the impact of evolving AI architecture. In our proposed review, a detailed narration on the three pillars of HAR is presented covering the period from 2011 to 2021. Further, the review presents the recommendations for an improved HAR design, its reliability, and stability. Five major findings were: (1) HAR constitutes three major pillars such as devices, AI and applications; (2) HAR has dominated the healthcare industry; (3) Hybrid AI models are in their infancy stage and needs considerable work for providing the stable and reliable design. Further, these trained models need solid prediction, high accuracy, generalization, and finally, meeting the objectives of the applications without bias; (4) little work was observed in abnormality detection during actions; and (5) almost no work has been done in forecasting actions. We conclude that: (a) HAR industry will evolve in terms of the three pillars of electronic devices, applications and the type of AI. (b) AI will provide a powerful impetus to the HAR industry in future.
Supplementary Information
The online version contains supplementary material available at 10.1007/s10462-021-10116-x.
“…(25) Chen et al, however, proposed a transfer learning framework based on principal component analysis (PCA) transformation, Gale-Shapley similarity measurement, and Jensen-Shannon divergence (JSD) feature mapping. (26) Semisupervised learning makes use of only a small amount of labeled training data and a substantial amount of unlabeled training data. (27) For example, self-training, cotraining, (28) and En-Co-Training are some typical semisupervised techniques whereas a special case of semisupervised learning, (29) i.e., active learning, mainly focuses on labeling the most profitable instances, but human intervention is necessary to some extent for a small amount of labeled data.…”
Section: Related Workmentioning
confidence: 99%
“…Precision, recall, and F-measure are usually used in a binary classification setting and consist of three scores including true positive (TP), false positive (FP), and false negative (FN). (26) In our work, nonzero labels (i.e., 17 target activities) are treated as positive, and zero labels (i.e., transition activities and RP) are treated as negative. In this manner, the multilabel labeling issue in this work is turned into a binary labeling issue.…”
Labeled datasets are one of the key factors for obtaining a good and robust classifier using supervised learning methods. However, labeling raw data is a tedious and labor-intensive process, which is usually done manually. Many efforts were proposed to utilize a small amount of labeled data to train a classifier that is sufficiently robust to label more data for training or make a prediction on unlabeled data. Unlike previous studies, we proposed an automatic labeling framework without labeling a small amount of data in advance, to directly annotate unlabeled time series data regarding body-worn sensor-based human activity recognition (HAR) in laboratory settings. The framework automatically labels collected time series activity data by transforming the original data into its corresponding absolute wavelet energy entropy and detects activity endpoints based on constraints and information extracted from a predefined human activity sequence. The performance of the proposed framework was evaluated on the collected dataset and the UCI HAR Dataset. In both cases, the average precision and recall scores are above 81.9% and the average F-measure scores are above 88.9%. Results showed that the proposed framework can be adopted as a rapid and reliable way of generating labeled datasets from unlabeled data.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.