Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-10362
|View full text |Cite
|
Sign up to set email alerts
|

Separating Long-Form Speech with Group-wise Permutation Invariant Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 22 publications
0
2
0
Order By: Relevance
“…Many studies have been proposed to improve different aspects of the CSS framework [3,4,5,6,7,8,9,10]. We introduced a modulation factor based on segment overlap ratio to dynamically adjust the separation loss [3].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Many studies have been proposed to improve different aspects of the CSS framework [3,4,5,6,7,8,9,10]. We introduced a modulation factor based on segment overlap ratio to dynamically adjust the separation loss [3].…”
Section: Introductionmentioning
confidence: 99%
“…In [4], a recurrent selective attention network is used to separate one speaker at a time. The work in [5] and [6] proposed new training criteria that generalizes PIT to capture long speech contexts. Li et al [7] proposed a dual-path separation model to leverage inter-segment information from a memory embedding pool.…”
Section: Introductionmentioning
confidence: 99%