2020
DOI: 10.1007/978-3-030-51935-3_24
|View full text |Cite
|
Sign up to set email alerts
|

Speech Enhancement Based on Deep AutoEncoder for Remote Arabic Speech Recognition

Abstract: Remote applications that deal with speech need the speech signal to be compressed. First, speech coding transforms the continuous waveform into a numerical form. Then, the digitized signal is compressed with or without loss of information. This transformation affects the original waveform and degrades performances for further recognition of the speech signal. Meanwhile, the transmission is another source of speech degradation. To restore the original "clean" speech, speech enhancement (SE) is widely used, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 25 publications
0
3
0
Order By: Relevance
“…After the denoising step with the wavelet technique, the next step is feature extraction. MFCC is one of the successful techniques for feature extraction [11,14,15]. Experiments revealed that MFCC is a commonly utilized technique, especially for a noisy dataset like the collected speech dataset produced by an EL device.…”
Section: Preprocessing and Feature Extractionmentioning
confidence: 99%
See 2 more Smart Citations
“…After the denoising step with the wavelet technique, the next step is feature extraction. MFCC is one of the successful techniques for feature extraction [11,14,15]. Experiments revealed that MFCC is a commonly utilized technique, especially for a noisy dataset like the collected speech dataset produced by an EL device.…”
Section: Preprocessing and Feature Extractionmentioning
confidence: 99%
“…Model's name Accuracy Word error rate Jaber and Abdulbaqi [28] Autoencoder (CNN) 93% -Eljawad et al [7] Fuzzy neural network 94.5% -Dendani et al [11] Autoencoder (MLP) 65.72% -Alsayadi et al [14] Autoencoder (LSTM) 71.58 28.42% Alsayadi et al [15] CNN-LSTM -13.52% Proposed model Autoencoder (GRU) 95.31% 4.69%…”
Section: Referencementioning
confidence: 99%
See 1 more Smart Citation
“…Long short-term memory (LSTM) and gated recurrent unit (GRU) are used for Arabic speech recognition [35]. Deep auto-encoder was used in [36], for remote Arabic speech recognition system. The authors used isolated words Arabic speech database for their experiments, where the database contained only a recording of 20 words.…”
Section: Literature Reviewsmentioning
confidence: 99%