ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2023
DOI: 10.1109/icassp49357.2023.10095126
|View full text |Cite
|
Sign up to set email alerts
|

Speech Signal Improvement Using Causal Generative Diffusion Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 6 publications
0
2
0
Order By: Relevance
“…Moreover, the AudioSet pretraining outperformed the Lib-riTTS one in all variations, which implies that the diverse audio sources and recording environments in AudioSet have been beneficial for generalization. SGMSE+ is reported to have SOTA generalization ability [32], and while it worked similarly to the pretrained ViT-AE in our study, it does not work for classification tasks.…”
Section: Speech Enhancementmentioning
confidence: 67%
“…Moreover, the AudioSet pretraining outperformed the Lib-riTTS one in all variations, which implies that the diverse audio sources and recording environments in AudioSet have been beneficial for generalization. SGMSE+ is reported to have SOTA generalization ability [32], and while it worked similarly to the pretrained ViT-AE in our study, it does not work for classification tasks.…”
Section: Speech Enhancementmentioning
confidence: 67%
“…Legends-tencent [35] AGC → GSM-GAN (Restore) → Enhance Genius-team [36] TRGAN (Restore) → MTFAA-Lite (Enhance) Hitot [37] Half temporal, half frequency attention U-Net N&B [38] GateDCCRN (Repairing) → GateDCCRN, S-DCCRN (Denoising) Hamburg [39] Generative diffusion model (modified NCSN++) tencent [35] and Genius-team [36] perform particularly well. • There is still significant room for improvement in this test set for OVRL and SIG.…”
Section: Team Modelmentioning
confidence: 99%