2019
DOI: 10.1109/taslp.2019.2940662
|View full text |Cite
|
Sign up to set email alerts
|

Speech Enhancement Based on Teacher–Student Deep Learning Using Improved Speech Presence Probability for Noise-Robust Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 79 publications
(36 citation statements)
references
References 41 publications
0
30
0
Order By: Relevance
“…To solve over-smoothing and speech information loss problems in a low SNR, Chai et al propose a prediction error model by considering a generalized Gaussian distribution (GGD) for DNN-based SE [17]. Tu et al propose a novel teacher-student learning framework for the preprocessing of a speech recognizer, leveraging the online noise tracking capabilities of improved MCRA and DNN network of nonlinear interactions between speech and noise [18]. To leverage long-term contexts for tracking a target speaker, Tan et al present a novel convolutional neural network (CNN) architecture for monaural SE [19].…”
Section: In the 1980s Ephraim And Malah Proposed The Minimummentioning
confidence: 99%
“…To solve over-smoothing and speech information loss problems in a low SNR, Chai et al propose a prediction error model by considering a generalized Gaussian distribution (GGD) for DNN-based SE [17]. Tu et al propose a novel teacher-student learning framework for the preprocessing of a speech recognizer, leveraging the online noise tracking capabilities of improved MCRA and DNN network of nonlinear interactions between speech and noise [18]. To leverage long-term contexts for tracking a target speaker, Tan et al present a novel convolutional neural network (CNN) architecture for monaural SE [19].…”
Section: In the 1980s Ephraim And Malah Proposed The Minimummentioning
confidence: 99%
“…Specifically, in the training process, when the value of the loss function is smaller, the WER is not necessarily lower. Discrimination training can alleviate this problem by using the solution of the traditional speech recognition system for reference [41]- [43]. In this paper, the Maximum Mutual Information (MMI) criterion is used for discriminative training.…”
Section: F Discriminative Trainingmentioning
confidence: 99%
“…Deep learning has bridged the gap between what human perceive and what a computer understands. It has significantly improved speech recognition [33][34][35][36][37][38][39][40][41][42]. These approaches are to make the computer think like human.…”
Section: Imentioning
confidence: 99%