2018
DOI: 10.1101/475673
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

EEG-based detection of the locus of auditory attention with convolutional neural networks

Abstract: When multiple people talk simultaneously, the healthy human auditory system is able to attend to one particular speaker of interest. Recently, it has been demonstrated that it is possible to infer to which speaker someone is attending by relating the neural activity, recorded by electroencephalography (EEG), with the speech signals. This is relevant for an effective noise suppression in hearing devices, in order to detect the target speaker in a multi-speaker scenario. Most auditory attention detection algorit… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

16
96
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1
1

Relationship

2
4

Authors

Journals

citations
Cited by 32 publications
(112 citation statements)
references
References 58 publications
16
96
0
Order By: Relevance
“…This indicates that the β-band is the most important band, motivating the choice of this band in Section IV-G. Similar conclusions have been drawn in [19], [34]. Furthermore, note that the performance does not degrade over time when the attention is sustained (see Supplementary Material), which has been reported in the context of α-power lateralization [17].…”
Section: H Decoding Mechanismssupporting
confidence: 72%
See 4 more Smart Citations
“…This indicates that the β-band is the most important band, motivating the choice of this band in Section IV-G. Similar conclusions have been drawn in [19], [34]. Furthermore, note that the performance does not degrade over time when the attention is sustained (see Supplementary Material), which has been reported in the context of α-power lateralization [17].…”
Section: H Decoding Mechanismssupporting
confidence: 72%
“…This confirms that, as opposed to the CNN method, our FB-CSP method does not overfit on speakers or stories, which could occur when using random crossvalidation. For the CNN method, the results were significantly better when not leaving out the speaker and/or story in the training set, which could be a sign of overfitting [19]. Furthermore, our FB-CSP method does not perform worse than the CNN method, as a Wilcoxon signed-rank test (W = 56, n = 15, p = 0.85, one outlier subject removed) shows no significant difference based on the MESD (Fig.…”
Section: B Comparison With Convolutional Neural Network Approachmentioning
confidence: 87%
See 3 more Smart Citations