Abstract:Background
Video‐based tic detection and scoring is useful to independently and objectively assess tic frequency and severity in patients with Tourette syndrome. In trained raters, interrater reliability is good. However, video ratings are time‐consuming and cumbersome, particularly in large‐scale studies. Therefore, we developed two machine learning (ML) algorithms for automatic tic detection.
Objective
The aim of this study was to evaluate the performances of state‐of‐the‐art ML approaches for automatic vide… Show more
“…Previous studies on tic detection from videos have utilized diverse algorithms to learn spatial and temporal properties for identifying tics. Similar to our methodology, other studies [24], [25] have emphasized the significance of incorporating tic-related features, such as facial landmarks or facial action unit intensities, into the input data of the model. Selecting tic-related features contributes to achieving more interpretable results compared to the visual features extracted by deep neural networks [24], [25].…”
Section: Discussionmentioning
confidence: 97%
“…In comparison, close-up video recordings have the capacity to capture facial movements and thus are well-suited for facial tic identification. Previous studies detected facial tics from short video recordings conducted in clinical settings [23]- [25]. However, these videos provide only snapshots of the current state and lack a comprehensive representation of the natural expression of tics in everyday environments.…”
Section: Discussionmentioning
confidence: 99%
“…Diverse machine learning algorithms have been proposed to address the challenge of detecting facial tics from video data. While some approaches involve architectures incorporating unsupervised visual feature learning and deep feature extraction [23], [26], others have shown the contribution of tic-related feature extraction (e.g., facial landmarks, facial action unit intensities) for enhancing network performance [24], [25].…”
The intrinsic nature of tic disorders, characterized by symptom
variability and fluctuation, poses challenges in clinical evaluations.
Currently, tic assessments predominantly rely on subjective
questionnaires administered periodically during clinical visits, thus
lacking continuous quantitative evaluation. This study aims to establish
an automatic objective measure of tic expression in natural behavioral
settings. A custom-developed smartphone application was used to record
selfie-videos of children and adolescents with tic disorders exhibiting
facial motor tics. Facial landmarks were utilized to extract tic-related
features from video segments. These features were then passed through a
tandem of custom deep neural networks to learn spatial and temporal
properties for tic classification. The model achieved a mean accuracy of
95% when trained on data across all subjects, and consistently exceeded
90% accuracy in leave-one-session-out and leave-one-subject-out cross
validation training schemes. This automatic tic identification measure
may provide a valuable tool for clinicians in facilitating diagnosis,
patient follow-up, and treatment efficacy evaluation. Combining this
measure with standard smartphone technology has the potential to
revolutionize large-scale clinical studies, thereby expediting the
development and testing of novel interventions.
“…Previous studies on tic detection from videos have utilized diverse algorithms to learn spatial and temporal properties for identifying tics. Similar to our methodology, other studies [24], [25] have emphasized the significance of incorporating tic-related features, such as facial landmarks or facial action unit intensities, into the input data of the model. Selecting tic-related features contributes to achieving more interpretable results compared to the visual features extracted by deep neural networks [24], [25].…”
Section: Discussionmentioning
confidence: 97%
“…In comparison, close-up video recordings have the capacity to capture facial movements and thus are well-suited for facial tic identification. Previous studies detected facial tics from short video recordings conducted in clinical settings [23]- [25]. However, these videos provide only snapshots of the current state and lack a comprehensive representation of the natural expression of tics in everyday environments.…”
Section: Discussionmentioning
confidence: 99%
“…Diverse machine learning algorithms have been proposed to address the challenge of detecting facial tics from video data. While some approaches involve architectures incorporating unsupervised visual feature learning and deep feature extraction [23], [26], others have shown the contribution of tic-related feature extraction (e.g., facial landmarks, facial action unit intensities) for enhancing network performance [24], [25].…”
The intrinsic nature of tic disorders, characterized by symptom
variability and fluctuation, poses challenges in clinical evaluations.
Currently, tic assessments predominantly rely on subjective
questionnaires administered periodically during clinical visits, thus
lacking continuous quantitative evaluation. This study aims to establish
an automatic objective measure of tic expression in natural behavioral
settings. A custom-developed smartphone application was used to record
selfie-videos of children and adolescents with tic disorders exhibiting
facial motor tics. Facial landmarks were utilized to extract tic-related
features from video segments. These features were then passed through a
tandem of custom deep neural networks to learn spatial and temporal
properties for tic classification. The model achieved a mean accuracy of
95% when trained on data across all subjects, and consistently exceeded
90% accuracy in leave-one-session-out and leave-one-subject-out cross
validation training schemes. This automatic tic identification measure
may provide a valuable tool for clinicians in facilitating diagnosis,
patient follow-up, and treatment efficacy evaluation. Combining this
measure with standard smartphone technology has the potential to
revolutionize large-scale clinical studies, thereby expediting the
development and testing of novel interventions.
“…Previous comparable studies made use of classical supervised machine learning approach such as k-Nearest Neighbors, Support Vector Machine, Random Forests, and Quadratic Discriminant Analysis using handcrafted features as well as deep learning approaches. [10][11][12][13] A comparison between HIVE-COTE 2.0 and the use of different algorithms on the current use case is deemed necessary in the future.…”
Section: Discussionmentioning
confidence: 99%
“…Key point detection by markerless motion analysis-also referred to as human pose estimation in the field of computer vision-can be used to process video data as an input for subsequent machine learning or deep learning models. [10][11][12][13][14] Several open-source toolbox codes facilitate key point detection based on convolutional neural networks. 15 Examples of these toolbox codes are OpenPose 16 and DeepLabCut.…”
Background Movement disorders in children and adolescents with dyskinetic cerebral palsy (CP) are commonly assessed from video recordings, however scoring is time-consuming and expert knowledge is required for an appropriate assessment. Objective To explore a machine learning approach for automated classification of amplitude and duration of distal leg dystonia and choreoathetosis within short video sequences. Methods Available videos of a heel-toe tapping task were preprocessed to optimize key point extraction using markerless motion analysis. Postprocessed key point data were passed to a time series classification ensemble algorithm to classify dystonia and choreoathetosis duration and amplitude classes (scores 0, 1, 2, 3, and 4), respectively. As ground truth clinical scoring of dystonia and choreoathetosis by the Dyskinesia Impairment Scale was used. Multiclass performance metrics as well as metrics for summarized scores: absence (score 0) and presence (score 1-4) were determined. Results Thirty-three participants were included: 29 with dyskinetic CP and 4 typically developing, age 14 years:6 months ± 5 years:15 months. The multiclass accuracy results for dystonia were 77% for duration and 68% for amplitude; for choreoathetosis 30% for duration and 38% for amplitude. The metrics for score 0 versus score 1 to 4 revealed an accuracy of 81% for dystonia duration, 77% for dystonia amplitude, 53% for choreoathetosis duration and amplitude. Conclusions This methodology study yielded encouraging results in distinguishing between presence and absence of dystonia, but not for choreoathetosis. A larger dataset is required for models to accurately represent distinct classes/scores. This study presents a novel methodology of automated assessment of movement disorders solely from video data.
BackgroundThe occurrence of tics is the main basis for the diagnosis of Gilles de la Tourette syndrome (GTS). Video‐based tic assessments are time consuming.ObjectiveThe aim was to assess the potential of automated video‐based tic detection for discriminating between videos of adults with GTS and healthy control (HC) participants.MethodsThe quantity and temporal structure of automatically detected tics/extra movements in videos from adults with GTS (107 videos from 42 participants) and matched HCs were used to classify videos using cross‐validated logistic regression.ResultsVideos were classified with high accuracy both from the quantity of tics (balanced accuracy of 87.9%) and the number of tic clusters (90.2%). Logistic regression prediction probability provides a graded measure of diagnostic confidence. Expert review of about 25% of lower‐confidence predictions could ensure an overall classification accuracy above 95%.ConclusionsAutomated video‐based methods have a great potential to support quantitative assessment and clinical decision‐making in tic disorders.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.