2021
DOI: 10.1109/taslp.2021.3078640
|View full text |Cite
|
Sign up to set email alerts
|

Group Communication With Context Codec for Lightweight Source Separation

Abstract: Despite the recent progress on neural network architectures for speech separation, the balance between the model size, model complexity and model performance is still an important and challenging problem for the deployment of such models to low-resource platforms. In this paper, we propose two simple modules, group communication and context codec, that can be easily applied to a wide range of architectures to jointly decrease the model size and complexity without sacrificing the performance. A group communicat… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 26 publications
(18 citation statements)
references
References 57 publications
0
18
0
Order By: Relevance
“…It reduces the model size and complexity by weight sharing across all groups (group communication), and further decrease the number of multiplyaccumulate operation using encoder-decoder-based temporal compression method (context codec). In the encoder part of the context codec, the temporal context of local feature is summarized into a single feature representing the global characteristics of the context [37]. After passing the group communication-equipped separation module, the compressed feature is transformed back to the context feature through the decoder part of the context codec and reconstructed to the estimated waveforms through a decoding transformation.…”
Section: Noise Suppression Modelmentioning
confidence: 99%
See 4 more Smart Citations
“…It reduces the model size and complexity by weight sharing across all groups (group communication), and further decrease the number of multiplyaccumulate operation using encoder-decoder-based temporal compression method (context codec). In the encoder part of the context codec, the temporal context of local feature is summarized into a single feature representing the global characteristics of the context [37]. After passing the group communication-equipped separation module, the compressed feature is transformed back to the context feature through the decoder part of the context codec and reconstructed to the estimated waveforms through a decoding transformation.…”
Section: Noise Suppression Modelmentioning
confidence: 99%
“…Considering the model size of the joint model consisted of NS and SED, we used the GC3-TCN for the NS model. More details are described in [37].…”
Section: Noise Suppression Modelmentioning
confidence: 99%
See 3 more Smart Citations