Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-1078
|View full text |Cite
|
Sign up to set email alerts
|

Convolutional Neural Network to Model Articulation Impairments in Patients with Parkinson’s Disease

Abstract: Speech impairments are one of the earliest manifestations in patients with Parkinson's disease. Particularly, articulation deficits related to the capability of the speaker to start/stop the vibration of the vocal folds have been observed in the patients. Those difficulties can be assessed by modeling the transitions between voiced and unvoiced segments from speech. A robust strategy to model the articulatory deficits related to the starting or stopping vibration of the vocal folds is proposed in this study. T… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
30
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
1

Relationship

3
3

Authors

Journals

citations
Cited by 56 publications
(31 citation statements)
references
References 13 publications
(20 reference statements)
1
30
0
Order By: Relevance
“…The results found here are similar to others reported in the state-of-the-art, and show to be complementary to those reported previously, where we performed an acoustic analysis of the daily routine monologues performed by the patients [15]. We believe that they could improve if the participants were evaluated in conversational speech samples or in recordings while describing a scene or a story (retelling tasks).…”
Section: Resultssupporting
confidence: 90%
“…The results found here are similar to others reported in the state-of-the-art, and show to be complementary to those reported previously, where we performed an acoustic analysis of the daily routine monologues performed by the patients [15]. We believe that they could improve if the participants were evaluated in conversational speech samples or in recordings while describing a scene or a story (retelling tasks).…”
Section: Resultssupporting
confidence: 90%
“…The transition is detected, and 80 ms of the signal are taken to the left and to the right of each border, forming "chunks" of signals with 160 ms length. Each chunk is transformed into a time-frequency representation using the short-time Fourier transform (STFT) and used as input to a CNN [15]. The CNN extracts the most suitable features from the STFT and makes the final decision about whether the utterance corresponds to a PD patient or a HC speaker, or classify the speaker according to the level of the speech item in the third part of the Movement Disorder Society-Unified Parkinson's Disease Rating Scale (MDS-UPDRS-III).…”
Section: Cnn Modelingmentioning
confidence: 99%
“…Recently in [3] automatic classification using Support Vector Machines (SVM) between 20 CI users and 20 healthy speakers was performed in order to evaluate articulation disorders considering acoustic features. For the case of pathological speech detection, CNNs have outperformed classical machine learning methods [4][5][6]. In these studies, the conventional method is to perform time-frequency analysis by computing spectrograms over the speech signals to feed the CNNs with single channel inputs.…”
Section: Introductionmentioning
confidence: 99%