2016
DOI: 10.1016/j.specom.2015.09.013
|View full text |Cite
|
Sign up to set email alerts
|

Fast algorithms for high-order sparse linear prediction with applications to speech processing

Abstract: In speech processing applications, imposing sparsity constraints on high-order linear prediction coefficients and prediction residuals has proven successful in overcoming some of the limitation of conventional linear predictive modeling. However, this modeling scheme, named sparse linear prediction, is generally formulated as a linear programming problem that comes at the expenses of a much higher computational burden compared to the conventional approach. In this paper, we propose to solve the optimization pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
19
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 17 publications
(21 citation statements)
references
References 66 publications
(99 reference statements)
0
19
0
Order By: Relevance
“…Similarly to what was done in [38], we processed only the vowel and semivowel phones [45] from the TIMIT database (sampled at 16 kHz), belonging to 3696 sentences from 462 speakers. We chose the ones of duration of at least 640 samples (40 ms) for a total of about 40,000 voiced speech frames.…”
Section: Prediction Gainmentioning
confidence: 99%
See 3 more Smart Citations
“…Similarly to what was done in [38], we processed only the vowel and semivowel phones [45] from the TIMIT database (sampled at 16 kHz), belonging to 3696 sentences from 462 speakers. We chose the ones of duration of at least 640 samples (40 ms) for a total of about 40,000 voiced speech frames.…”
Section: Prediction Gainmentioning
confidence: 99%
“…We chose the ones of duration of at least 640 samples (40 ms) for a total of about 40,000 voiced speech frames. We extend the analysis in [38] by investigating the prediction gain from the ADMM solution with a different number of iterations and compare with the IP solution obtained through the CVX+SeDuMi interface and solver. In both formulation of the SLP problem (5), we chose γ = 0.12 obtained through a modified version of the L-curve [46] by using all except 50 frames picked randomly that will be used as a test set.…”
Section: Prediction Gainmentioning
confidence: 99%
See 2 more Smart Citations
“…Motivated by the compressive sensing research, a least 1-norm criterion is proposed for voiced speech analysis [9], where sparse priors on both the excitation signals and prediction coefficients are utilized. Fast methods and the stability of the 1-norm cost function for spectral envelope estimation are further investigated in [10], [11]. More recently, in [12], the excitation signal of speech is formulated as a combination of block sparse and white noise components to capture the block sparse or white This work was funded by the Danish Council for Independent Research, grant ID: DFF 4184-00056 noise excitation separately or simultaneously.…”
Section: Introductionmentioning
confidence: 99%