Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-66
|View full text |Cite
|
Sign up to set email alerts
|

Time-Frequency Masking for Blind Source Separation with Preserved Spatial Cues

Abstract: In this paper, we address the problem of speech source separation by relying on time-frequency binary masks to segregate binaural mixtures. We describe an algorithm which can tackle reverberant mixtures and can extract the original sources while preserving their original spatial locations. The performance of the proposed algorithm is evaluated objectively and subjectively, by assessing the estimated interaural time differences versus their theoretical values and by testing for localization acuity in normal-hea… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…To address this problem, speech enhancement is usually performed to separate the clean speech from background noises with reverberation. Thanks to the advancement of computational auditory scene analysis [3,4], speech enhancement has been viewed as a supervised learning problem. Existing speech enhancement methods mainly learn a deep-learning-based neural network, which can effectively model the relationship between a noisecorrupted signal and the clean speech.…”
Section: Introductionmentioning
confidence: 99%
“…To address this problem, speech enhancement is usually performed to separate the clean speech from background noises with reverberation. Thanks to the advancement of computational auditory scene analysis [3,4], speech enhancement has been viewed as a supervised learning problem. Existing speech enhancement methods mainly learn a deep-learning-based neural network, which can effectively model the relationship between a noisecorrupted signal and the clean speech.…”
Section: Introductionmentioning
confidence: 99%
“…For decades, monaural speech separation has been studied for speech processing applications. Inspired by Time-Frequency (T-F) masking based on computational auditory scene analysis (CASA) [1,2], monaural speech separation has been recently treated as a supervised learning problem. The choice of training model, acoustic features, and training target are all important factors for accurate supervised speech separation.…”
Section: Introductionmentioning
confidence: 99%