2020
DOI: 10.1016/j.softx.2020.100630
|View full text |Cite|
|
Sign up to set email alerts
|

PCAfold: Python software to generate, analyze and improve PCA-derived low-dimensional manifolds

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
1
1

Relationship

3
5

Authors

Journals

citations
Cited by 19 publications
(12 citation statements)
references
References 20 publications
0
12
0
Order By: Relevance
“…Several reports have shown that the SVM classifier is a crucial tool for classification problems in sports medicine and lower limb biomechanics ( Xiang et al, 2022 ; Phinyomark et al, 2015 ; Christian et al, 2016 ). In the previous studies, an unsupervised PCA algorithm was commonly used in data preprocessing for discovering the underlying low-dimensional manifolds in high-dimensional datasets ( Zdybał et al, 2020 ) and for data extraction consideration ( Wu and Wang, 2008 ; Taylor et al, 2013 ; Clermont et al, 2017 ; Suda et al, 2020 ). Eleven low-dimensional features were extracted in our study, explaining 90% cumulative variance of the original data.…”
Section: Discussionmentioning
confidence: 99%
“…Several reports have shown that the SVM classifier is a crucial tool for classification problems in sports medicine and lower limb biomechanics ( Xiang et al, 2022 ; Phinyomark et al, 2015 ; Christian et al, 2016 ). In the previous studies, an unsupervised PCA algorithm was commonly used in data preprocessing for discovering the underlying low-dimensional manifolds in high-dimensional datasets ( Zdybał et al, 2020 ) and for data extraction consideration ( Wu and Wang, 2008 ; Taylor et al, 2013 ; Clermont et al, 2017 ; Suda et al, 2020 ). Eleven low-dimensional features were extracted in our study, explaining 90% cumulative variance of the original data.…”
Section: Discussionmentioning
confidence: 99%
“…1 ). The proposed algorithm is available in the PCAfold Python library [30] . The supplementary material includes an example code snippet for running the algorithm on a combustion dataset.…”
Section: Manifold-informed Subset Of State Variablesmentioning
confidence: 99%
“…In order to determine the best scaling for the input parameters, we use a recently proposed cost function, , that can help assess the quality of data parameterization for regression tasks [45,53]. By minimizing the cost value, we determine which scaling factor shall be applied on the design variables (independent variables) to yield good regressibility of each objective function (dependent variables).…”
Section: Regression Model Performancementioning
confidence: 99%
“…The closer 𝑅 2 is to unity, the better the regression model predictions. Since the overall 𝑅 2 computed for all points in the dataset can be misleading, the data matrix is divided into a number of bins (15)(16)(17)(18)(19)(20), according to the objective function under investigation, and 𝑅 2 is computed in each bin to investigate the prediction performance thoroughly [45]. With this approach, we make sure that the trained model does not focus on any specific region but represents the entire dataset well.…”
Section: Regression Model Performancementioning
confidence: 99%