Dysarthria may present during the natural course of many degenerative neurological conditions. Hypokinetic and ataxic dysarthria are common in movement disorders and represent the underlying neuropathology. We developed an artificial intelligence (AI) model to distinguish ataxic dysarthria and hypokinetic dysarthria from normal speech and differentiate ataxic and hypokinetic speech in parkinsonian diseases and cerebellar ataxia. We screened 804 perceptual speech analyses performed in the Samsung Medical Center Neurology Department between January 2017 and December 2020. The data of patients diagnosed with parkinsonian disorders or cerebellar ataxia were included. Two speech tasks (numbering from 1 to 50 and reading nine sentences) were analyzed. We adopted convolutional neural networks and developed a patch-wise wave splitting and integrating AI system for audio classification (PWSI-AI-AC) to differentiate between ataxic and hypokinetic speech. Of the 395 speech recordings for the reading task, 76, 112, and 207 were from normal, ataxic dysarthria, and hypokinetic dysarthria subjects, respectively. Of the 409 recordings of the numbering task, 82, 111, and 216 were from normal, ataxic dysarthria, and hypokinetic dysarthria subjects, respectively. The reading and numbering task recordings were classified with 5-fold cross-validation using PWSI-AI-AC as follows: hypokinetic dysarthria vs. others (area under the curve: 0.92 ± 0.01 and 0.92 ± 0.02), ataxia vs. others (0.93 ± 0.04 and 0.89 ± 0.02), hypokinetic dysarthria vs. ataxia (0.96 ± 0.02 and 0.95 ± 0.01), hypokinetic dysarthria vs. none (0.86 ± 0.03 and 0.87 ± 0.05), and ataxia vs. none (0.87 ± 0.07 and 0.87 ± 0.09), respectively. PWSI-AI-AC showed reliable performance in differentiating ataxic and hypokinetic dysarthria and effectively augmented data to classify the types even with limited training samples. The proposed fully automatic AI system outperforms neurology residents. Our model can provide effective guidelines for screening related diseases and differential diagnosis of neurodegenerative diseases.
Background This study aims to develop artificial intelligence (AI) system to automatically classify patients with maxillary sinus fungal ball (MFB), chronic rhinosinusitis (CRS), and healthy controls (HCs). Methods We collected 512 coronal image sets from ostiomeatal unit computed tomography (OMU CT) performed on subjects who visited a single tertiary hospital. These data included 254 MFB, 128 CRS, and 130 HC subjects and were used for training the proposed AI system. The AI system takes these 1024 sets of half CT images as input and classifies these as MFB, CRS, or HC. To optimize the classification performance, we adopted a 3-D convolutional neural network of ResNet 18. We also collected 64 coronal OMU CT image sets for external validation, including 26 MFB, 18 CRS, and 20 HCs from subjects from another referral hospital. Finally, the performance of the developed AI system was compared with that of the otolaryngology resident physicians. Results Classification performance was evaluated using internal 5-fold cross-validation (818 training and 206 internal validation data) and external validation (128 data). The area under the receiver operating characteristic over the internal 5-fold cross-validation and the external validation was 0.96 ±0.006 and 0.97 ±0.006, respectively. The accuracy of the internal 5-fold cross-validation and the external validation was 87.5 ±2.3% and 88.4 ±3.1%, respectively. As a result of performing a classification test on external validation data from six otolaryngology resident physicians, the accuracy was obtained as 84.6 ±11.3%. Conclusions This AI system is the first study to classify MFB, CRS, and HC using deep neural networks to the best of our knowledge. The proposed system is fully automatic but performs similarly to or better than otolaryngology resident physicians. Therefore, we believe that in regions where otolaryngology specialists are scarce, the proposed AI will perform sufficiently effective diagnosis on behalf of doctors.
We consider the problem of sparse phase retrieval from Fourier transform magnitudes to recover the k-sparse signal vector and its support T . We exploit extended support estimate E with size larger than k satisfying E ⊇ T and obtained by a trained deep neural network (DNN). To make the DNN learnable, it provides E as the union of equivalent solutions of T by utilizing modulo Fourier invariances. Set E can be estimated with short running time via the DNN, and support T can be determined from the DNN output rather than from the full index set by applying hard thresholding to E. Thus, the DNN-based extended support estimation improves the reconstruction performance of the signal with a low complexity burden dependent on k. Numerical results verify that the proposed scheme has a superior performance with lower complexity compared to local search-based greedy sparse phase retrieval and a state-of-the-art variant of the Fienup method. Index TermsDeep neural network, extended support estimation, Fourier transform, sparse phase retrieval. 1 Signal vector x • is called k-sparse if it has k nonzero elements.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.