2016
DOI: 10.1155/2016/2618265
|View full text |Cite
|
Sign up to set email alerts
|

Vowel Imagery Decoding toward Silent Speech BCI Using Extreme Learning Machine with Electroencephalogram

Abstract: The purpose of this study is to classify EEG data on imagined speech in a single trial. We recorded EEG data while five subjects imagined different vowels, /a/, /e/, /i/, /o/, and /u/. We divided each single trial dataset into thirty segments and extracted features (mean, variance, standard deviation, and skewness) from all segments. To reduce the dimension of the feature vector, we applied a feature selection algorithm based on the sparse regression model. These features were classified using a support vector… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
40
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 70 publications
(44 citation statements)
references
References 47 publications
1
40
0
Order By: Relevance
“…In the presented work, we have used a SVM classification on CSP extracted features from EEG data corresponding to VIm. A cross-validation accuracy of above 95% is shown among individual participants and close to 89% across average of the 17 participants compared to other studies showing accuracy around 75% classifying VInt [19]. These accuracy scores are quite impressive considering no pre-processing was performed to prime the data; no artifact removal, frequency filtration or noise reduction.…”
Section: Discussionmentioning
confidence: 52%
See 1 more Smart Citation
“…In the presented work, we have used a SVM classification on CSP extracted features from EEG data corresponding to VIm. A cross-validation accuracy of above 95% is shown among individual participants and close to 89% across average of the 17 participants compared to other studies showing accuracy around 75% classifying VInt [19]. These accuracy scores are quite impressive considering no pre-processing was performed to prime the data; no artifact removal, frequency filtration or noise reduction.…”
Section: Discussionmentioning
confidence: 52%
“…In 2016, Min et. al., have proposed a vowel (/a/, /e/, /i/, /o/, and /u/) imagery decoding towards developing a silent speech BCI using extreme learning machine with a single-trial EEG [19]. In the same year, Yoshimura et.…”
Section: Introductionmentioning
confidence: 99%
“…Most of the researchers, including Zhao and Rudzicz ( 2015 ), Min et al ( 2016 ), Nguyen et al ( 2017 ), Koizumi et al ( 2018 ), and Sereshkeh et al ( 2017a ) have used a 64-electrode EEG system with a sampling rate of 1 KHz for acquiring the EEG data corresponding to imagined speech. In the case of the work reported by Deng et al ( 2010 ) and Brigham and Kumar ( 2010 ), 128-electrode EEG data has been recorded at a sampling rate of 1 KHz.…”
Section: Data Acquisitionmentioning
confidence: 99%
“…Three methods have been primarily followed by researchers to cue the participant as to what the prompt is and when to start imagining speaking the prompt. These are (1) auditory (Brigham and Kumar, 2010 ; Deng et al, 2010 ; Min et al, 2016 ; Koizumi et al, 2018 ); (2) visual (Wang et al, 2014 ; Sereshkeh et al, 2017a ; Jahangiri et al, 2018 ; Koizumi et al, 2018 ); and (3) a combination of auditory and visual cues (Zhao and Rudzicz, 2015 ; Coretto et al, 2017 ; Nguyen et al, 2017 ; Watanabe et al, 2020 ). Although somatosensory cues have been used for motor imagery (Panachakel et al, 2020b ), no such work has been reported for speech imagery.…”
Section: Data Acquisitionmentioning
confidence: 99%
“…To leverage the temporal dynamics of the MEG signal, we extracted four feature sets at each time point across all the sensors to train the LSTM-RNN deep learning algorithm for continuous temporal prediction of 'Speech' vs. 'Non-Speech'. The first three features were the 'absolute magnitude', 'root-mean-square' (RMS), and 'standard deviation' which have been shown to be effective in previous studies [7,[26][27][28]. The 4th feature was the index of the most active sensor, i.e., the sensor with maximum magnitude value among all sensors at each time point.…”
Section: Feature Extractionmentioning
confidence: 99%