2017
DOI: 10.1002/brb3.665
|View full text |Cite
|
Sign up to set email alerts
|

Vowel decoding from single‐trial speech‐evoked electrophysiological responses: A feature‐based machine learning approach

Abstract: IntroductionScalp‐recorded electrophysiological responses to complex, periodic auditory signals reflect phase‐locked activity from neural ensembles within the auditory system. These responses, referred to as frequency‐following responses (FFRs), have been widely utilized to index typical and atypical representation of speech signals in the auditory system. One of the major limitations in FFR is the low signal‐to‐noise ratio at the level of single trials. For this reason, the analysis relies on averaging across… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 32 publications
(37 citation statements)
references
References 45 publications
0
37
0
Order By: Relevance
“…While recent studies have brought substantial advances in decoding auditory features, studies using discreet stimuli have focused on decoding complex features such as pitch/rate modulation based on spectral information in MEG signals (Herrmann et al, 2013b) or bistable percepts based on evoked MEG responses (Billig et al, 2018). In the domain of speech decoding, speech-evoked responses can be used to decode vowel categories (Yi et al, 2017), but typically a combination of complex spectral features is used to decode the speech envelope (Luo and Poeppel, 2007;Ng et al, 2013;de Cheveigné et al, 2018). Here, robust decoding of pure tone frequency was achieved based on relatively early M/EEG response latencies (Ͻ100 ms) evoked by very brief tones (ϳ33 ms), despite their presentation in gapless streams.…”
Section: Discussionmentioning
confidence: 99%
“…While recent studies have brought substantial advances in decoding auditory features, studies using discreet stimuli have focused on decoding complex features such as pitch/rate modulation based on spectral information in MEG signals (Herrmann et al, 2013b) or bistable percepts based on evoked MEG responses (Billig et al, 2018). In the domain of speech decoding, speech-evoked responses can be used to decode vowel categories (Yi et al, 2017), but typically a combination of complex spectral features is used to decode the speech envelope (Luo and Poeppel, 2007;Ng et al, 2013;de Cheveigné et al, 2018). Here, robust decoding of pure tone frequency was achieved based on relatively early M/EEG response latencies (Ͻ100 ms) evoked by very brief tones (ϳ33 ms), despite their presentation in gapless streams.…”
Section: Discussionmentioning
confidence: 99%
“…1) Greenberg et al, 1987). Previous research has taken advantage of this property of the FFR to decode sound categories from FFR responses (Llanos, Xie, & Chandrasekaran, 2017;Reetzke et al, 2018;Yi et al, 2017). Building upon this finding, we aimed to decode listeners from FFRs as a proxy to assess the biometric specificity of the FFR.…”
Section: The Frequency Following Responsementioning
confidence: 96%
“…Results show that the spectral amplitude of the FFR in F0 and F1 bands can be used to correctly classify vowels with up to 70–80% accuracy (Sadeghian et al, 2015) and that spectral information related to the F2 band can be used to classify cortical evoked responses to vowels on the basis of single-trial data (Kim et al, 2014). Data acquired with an innovative combination of single-trial classification and machine-learning methods support the notion that the spectral amplitude of the FFR may be used to correctly predict vowel categorization into learned and novel vowel categories (Yi et al, 2017). Moreover, temporal information contained in the phase of theta oscillations (2–9 Hz) could correctly classify eight phonetic categories such that confusion matrices from phase and perceptual responses were not statistically distinguishable from one another (R.…”
Section: Introductionmentioning
confidence: 97%