Identifying key frames is the first and necessary step before solving the variety of other Bharatanatyam problems. The paper aims to partition the momentarily stationary frames (key frames) from this dance video's motion frames. The proposed key frames (KFs) localization is novel, simple, and effective compared to the existing dance video analysis methods. It is distinctive from standard KFs detection algorithms as used in other human motion videos. In the dance's basic structure, the occurrence of KFs during performances is often not completely stationary and varies with the dance form and the performer. Hence, it is not easy to decide a global threshold (on the quantum of motion) to work across dancers and performances. The earlier approaches try to compute the threshold iteratively. However, the novelty of the paper is: (a) formulating an adaptive threshold, (b) adopting Machine Learning (ML) approach and, (c) generating the effective feature by combining three frame differencing and bit-plane technique for the KF detection. In ML, we use Support Vector Machine (SVM) and Convolutional Neural Network (CNN) as the classifiers. The proposed approaches are also compared and analyzed with the earlier approaches. Finally, the proposed ML techniques emerge as a winner with around 90% accuracy.
This paper provides a method to understand the underlying semantics of Bharatnatyam dance motion and classifies it. Each dance performance is audio-driven and spans over space and time. The dance is captured and analyzed, which is helpful in cultural heritage preservation, and tutoring systems to assist the naive learner. This paper attempts to solve the fundamental problem; recognizing the motions during a dance performance based on motion-pattern. The used dataset is the video recordings of an Indian Classical Dance form known as Bharatanatyam. The different Adavus (The basic unit of Bharatanatyam) of Bharatanatyam dance are captured using Kinect. We choose RGB from various forms of captured data (RGB, Depth, and Skeleton). Motion History Image (MHI) and Histogram of Gradient of MHI (HoGMHI) are computed for each motion and used as an input for the Machine Learning (ML) algorithms to recognize motion. The paper explores two ML techniques; Support Vector Machine (SVM) and Convolutional Neural Network (CNN). The overall accuracy of both the classifiers is more than 90%. The novelties of the work are (a) analysing all possible involved motions based on the motion-patterns rather than the joint velocities or pose, (b)exploring the impact of training data and the different features on the classifiers' accuracy, (c) not restricting the number of frames in a motion during recognition and formulate a method to deal with the variable number of frames in the motions.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.