2020
DOI: 10.1016/j.heliyon.2020.e03176
|View full text |Cite
|
Sign up to set email alerts
|

Kernel partial diagnostic robust potential to handle high-dimensional and irregular data space on near infrared spectral data

Abstract: In practice, the collected spectra are very often composes of complex overtone and many overlapping peaks which may lead to misinterpretation because of its significant nonlinear characteristics. Using linear solution might not be appropriate. In addition, with a high-dimension of dataset due to large number of observations and data points the classical multiple regressions will neglect to fit. These complexities commonly will impact to multicollinearity problem, furthermore the risk of contamination of multip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 37 publications
(55 reference statements)
0
5
0
Order By: Relevance
“…In the case of nonlinearity, x is transformed into high-dimensional feature space through nonlinearity, so as to find the optimal classification surface. In the highdimensional space, the inner product operation is carried out by using the function of the original space, and the nonlinear problem in the linear space [32] is mapped to the high-dimensional feature space, so as to fundamentally solve the nonlinear problem. Therefore, assuming that the kernel function satisfying the Mercer condition is Kðx i , x j Þ, the linear classification can be obtained through nonlinear transformation without increasing the computational complexity.…”
Section: Teaching Planningmentioning
confidence: 99%
“…In the case of nonlinearity, x is transformed into high-dimensional feature space through nonlinearity, so as to find the optimal classification surface. In the highdimensional space, the inner product operation is carried out by using the function of the original space, and the nonlinear problem in the linear space [32] is mapped to the high-dimensional feature space, so as to fundamentally solve the nonlinear problem. Therefore, assuming that the kernel function satisfying the Mercer condition is Kðx i , x j Þ, the linear classification can be obtained through nonlinear transformation without increasing the computational complexity.…”
Section: Teaching Planningmentioning
confidence: 99%
“…PLS discriminant analysis (PLS-DA) is widely used for analysis of metabolomic data, and is a variant of PLS used for classification problems, ie when y is categorical. 49 The importance of variables may be assessed through the loadings and variable importance scores such as variable importance in projection (VIP) scores and the selectivity ratio (SR), 48,50 which summarize the contributions of the original variables to the model. However, PLS models are prone to overfitting and the number of latent variables should be carefully selected by validation, which is described in Section 7.…”
Section: Partial Least Squares (Pls)mentioning
confidence: 99%
“…The 100(IV)% of the predictor variables were randomly selected as relevant variables, and the remaining were considered as less relevant. The formulation of this simulation can be defined as follows: 5,20) ( j e = 1, 2, . .…”
Section: Monte Carlo Simulation Studymentioning
confidence: 99%
“…However, despite having these benefits, several studies have reported its weakness due to its robustness. The fitted model performs poorly when outliers and leverage points are present in a dataset [17,18], as it fails to fit the nonlinear behavior in the input space [19,20]. In addition, the contamination of irrelevant variables involves during the fitting process [21][22][23] is a popular topic in most discussions.…”
Section: Introductionmentioning
confidence: 99%