2017
DOI: 10.1109/taslp.2016.2632528
|View full text |Cite
|
Sign up to set email alerts
|

On-the-Fly Audio Source Separation—A Novel User-Friendly Framework

Abstract: Abstract-This article addresses the challenging problem of single-channel audio source separation. We introduce a novel user-guided framework where source models that govern the separation process are learned on-the-fly from audio examples retrieved online. The user only provides the search keywords that describe the sources in the mixture. In this framework, the generic spectral characteristics of each source are modeled by a universal sound class model learned from the retrieved examples via non-negative mat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
6
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 9 publications
(8 citation statements)
references
References 31 publications
1
6
0
Order By: Relevance
“…The number of NMF components in W l j for each speech example was set to 32, while that for noise example was 16. These values were found to be reasonable in [15] and our work on single-channel case [18]. Each W l j were obtained by optimizing (17) with 20 MU iterations.…”
Section: A Dataset and Parameter Settingssupporting
confidence: 62%
See 3 more Smart Citations
“…The number of NMF components in W l j for each speech example was set to 32, while that for noise example was 16. These values were found to be reasonable in [15] and our work on single-channel case [18]. Each W l j were obtained by optimizing (17) with 20 MU iterations.…”
Section: A Dataset and Parameter Settingssupporting
confidence: 62%
“…. This leads to a straightforward extension of the conventional optimization criterion described by (15) where H j is now estimated by optimizing the criterion:…”
Section: B Proposed Source Variance Fitting With Gssm and Mixed Groumentioning
confidence: 99%
See 2 more Smart Citations
“…We also explore what type of covariance model is most effective for musical source separation (tied vs. untied across classes, diagonal vs. spherical). Furthermore, we discuss a simple modification of our pre-trained embedding networks for query-by-example separation [18,19,20], where given an isolated example of a sound we want to separate, we can extract the portion of a mixture most like the query without supervision.…”
Section: Introductionmentioning
confidence: 99%