2018 IEEE International Work Conference on Bioinspired Intelligence (IWOBI) 2018
DOI: 10.1109/iwobi.2018.8464132
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid Speech Enhancement with Wiener filters and Deep LSTM Denoising Autoencoders

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 15 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…Two layers of a bi-directional Long Short-Term Memory (bi-LSTM) with 64 filters and 32 filters are included in the encoder network to reconstruct a clean MFCC representation, followed by a temporal Fully Connected (FC) in a decoder network. An LSTM neural network was previously used to extract speaker representation and remove noise from audio [30] [32] . In this study, Deep-MASKS employs bi-LSTM, a variation of LSTM that considers the backward sequence in addition to the forward sequence, to learn temporal patterns over the MFCC sequence.…”
Section: Methodsmentioning
confidence: 99%
“…Two layers of a bi-directional Long Short-Term Memory (bi-LSTM) with 64 filters and 32 filters are included in the encoder network to reconstruct a clean MFCC representation, followed by a temporal Fully Connected (FC) in a decoder network. An LSTM neural network was previously used to extract speaker representation and remove noise from audio [30] [32] . In this study, Deep-MASKS employs bi-LSTM, a variation of LSTM that considers the backward sequence in addition to the forward sequence, to learn temporal patterns over the MFCC sequence.…”
Section: Methodsmentioning
confidence: 99%
“…An autoencoder for noise reduction is a neural network architecture that has been successful in various tasks related to speech [29]. This architecture consists of an encoder that transforms an input vector s into a representation in the hidden layers h through a f mapping.…”
Section: Autoencoders Of Blstm Networkmentioning
confidence: 99%
“…One of the main architectures applied for regression tasks (including speech enhancement) using deep neural networks are the autoencoders. An autoencoder for speech enhancement is a neural network architecture that has been successful in various tasks related to speech [33]. This architecture consists of an encoder that transforms an input vector s into a representation in the hidden layers h through a f mapping.…”
Section: Autoencoders Of Blstm Networkmentioning
confidence: 99%