2017
DOI: 10.1109/tsp.2017.2669903
|View full text |Cite
|
Sign up to set email alerts
|

Kernel Risk-Sensitive Loss: Definition, Properties and Application to Robust Adaptive Filtering

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
74
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 146 publications
(76 citation statements)
references
References 38 publications
1
74
0
Order By: Relevance
“…The MCC as a measure of the information content in information theoretic learning was developed by principe with his team to deal with error distributions with non-Gaussian characteristics, and it has been widely used in pattern classification, feature selection, dimension reduction and adaptive filtering [21][22][23][24][25]. The MCC indicates the similarity between the predicted output and the real sample in the correntropy sense; it shows good robustness for nonlinear and non-Gaussian data processing, such as determining whether electricity is suitable for the prediction of time series non-stationary and time-varying predictions [26].…”
Section: Maximum Correntropy Criterionmentioning
confidence: 99%
“…The MCC as a measure of the information content in information theoretic learning was developed by principe with his team to deal with error distributions with non-Gaussian characteristics, and it has been widely used in pattern classification, feature selection, dimension reduction and adaptive filtering [21][22][23][24][25]. The MCC indicates the similarity between the predicted output and the real sample in the correntropy sense; it shows good robustness for nonlinear and non-Gaussian data processing, such as determining whether electricity is suitable for the prediction of time series non-stationary and time-varying predictions [26].…”
Section: Maximum Correntropy Criterionmentioning
confidence: 99%
“…where J 1 , J 2 have been shown in Equation 8. The optimal decision function under our ensemble loss function is obtained by setting the derivative of J to zero, 5 shows an example of two convex functions J 1 and J 2 with the optimal point f * 1 and f * 2 respectively.…”
Section: Bayes Consistencymentioning
confidence: 99%
“…On the other hand, when kernel bandwidth becomes larger, the robustness will be significantly reduced when outliers appear. To achieve better performance, a new similarity measure, i.e., KRSL, was proposed in [37], which can be more "convex" and thus can achieve faster convergence rate and higher filtering accuracy while maintaining the robustness to outliers.…”
Section: The Kernel Risk-sensitive Loss (Krsl) Algorithmmentioning
confidence: 99%
“…However, correntropic loss (C-Loss) is a non-convex function, and may converge slowly, especially when the initial value is far away from the optimal value. Recently, a modified similarity measure called kernel risk-sensitive loss (KRSL) was derived in [37], which can be more "convex" and can achieve faster convergence speed and higher filtering accuracy, while maintaining the robustness to outliers. After applying KRSL to adaptive filtering, a new robust KAF algorithm was accordingly proposed [37].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation