2019
DOI: 10.48550/arxiv.1908.08044
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Coarse-to-fine Optimization for Speech Enhancement

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 27 publications
0
8
0
Order By: Relevance
“…To evaluate MASnet, we use a standard dataset [29,30] and compare with LLASnet-8, LLASnet-15 and the CRMRN U-net from [9,8]. All of our experiments are implemented with Py-Torch [31].…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…To evaluate MASnet, we use a standard dataset [29,30] and compare with LLASnet-8, LLASnet-15 and the CRMRN U-net from [9,8]. All of our experiments are implemented with Py-Torch [31].…”
Section: Methodsmentioning
confidence: 99%
“…Spectrogram-based speech enhancement methods often treat STFT representations as images [9,8]. This is fine at trainingtime, but at test-time and especially in low-latency applications such as telephony it is important to be able to output processed audio immediately, without having to record several seconds that could be transformed into a spectrogram.…”
Section: Low-latency Speech Enhancementmentioning
confidence: 99%
See 1 more Smart Citation
“…The idea of solving tasks in a coarse-to-fine order has previously been explored in computer vision and signal processing [e.g., for object detection and recognition 14,3,38,45], head pose estimation [62], or more general computer vision tasks [32,36,47,65,70]. Similarly, coarse-to-fine ideas have also been used for various tasks in natural language processing [e.g., 30,12,68]. Our approach is different in that it is widely applicable; it does not depend on the problem space at hand, but can rather be applied as-is to any classification problem.…”
Section: Related Workmentioning
confidence: 99%

Coarse-to-Fine Curriculum Learning

Stretcu,
Platanios,
Mitchell
et al. 2021
Preprint
“…Second, we propose a new time-domain loss function, an emphasized multi-scale cos similarity loss function. A time-domain loss function has recently been used as a popular loss function [6,7,8,9,10]. To better design the time-domain cos similarity loss function proposed in [6], we change it into a multi-scale version of it with proper emphasis functions and show the effectiveness of it.…”
Section: Introductionmentioning
confidence: 99%