In recent decades, gait recognition has garnered a lot of attention from the researchers in the IT era. Gait recognition signifies verifying or identifying the individuals by their walking style. Gait supports in surveillance system by identifying people when they are at a distance from the camera and can be used in numerous computer vision and surveillance applications. This paper proposes a stupendous Color-mapped Contour Gait Image (CCGI) for varying factors of Cross-View Gait Recognition (CVGR). The first contour in each gait image sequence is extracted using a Combination of Receptive Fields (CORF) contour tracing algorithm which extracts the contour image using Difference of Gaussians (DoG) and hysteresis thresholding. Moreover, hysteresis thresholding detects the weak edges from the total pixel information and provides more well-balanced smooth features compared to an absolute one. Second CCGI encodes the spatial and temporal information via color mapping to attain the regularized contour images with fewer outliers. Based on the front view of a human walking pattern, the appearance of cross-view variations would reduce drastically with respect to a change of view angles. This proposed work evaluates the performance analysis of CVGR using Deep Convolutional Neural Network (CNN) framework. CCGI is considered a gait feature for comparing and evaluating the robustness of our proposed model. Experiments conducted on CASIA-B database show the comparisons of previous methods with the proposed method and achieved 94.65% accuracy with a better recognition rate.
PurposeThe paper aims to introduce an intelligent recognition system for viewpoint variations of gait and speech. It proposes a convolutional neural network-based capsule network (CNN-CapsNet) model and outlining the performance of the system in recognition of gait and speech variations. The proposed intelligent system mainly focuses on relative spatial hierarchies between gait features in the entities of the image due to translational invariances in sub-sampling and speech variations.Design/methodology/approachThis proposed work CNN-CapsNet is mainly used for automatic learning of feature representations based on CNN and used capsule vectors as neurons to encode all the spatial information of an image by adapting equal variances to change in viewpoint. The proposed study will resolve the discrepancies caused by cofactors and gait recognition between opinions based on a model of CNN-CapsNet.FindingsThis research work provides recognition of signal, biometric-based gait recognition and sound/speech analysis. Empirical evaluations are conducted on three aspects of scenarios, namely fixed-view, cross-view and multi-view conditions. The main parameters for recognition of gait are speed, change in clothes, subjects walking with carrying object and intensity of light.Research limitations/implicationsThe proposed CNN-CapsNet has some limitations when considering for detecting the walking targets from surveillance videos considering multimodal fusion approaches using hardware sensor devices. It can also act as a pre-requisite tool to analyze, identify, detect and verify the malware practices.Practical implicationsThis research work includes for detecting the walking targets from surveillance videos considering multimodal fusion approaches using hardware sensor devices. It can also act as a pre-requisite tool to analyze, identify, detect and verify the malware practices.Originality/valueThis proposed research work proves to be performing better for the recognition of gait and speech when compared with other techniques.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.