2023
DOI: 10.1038/s41598-023-37232-8
|View full text |Cite
|
Sign up to set email alerts
|

Computational intelligence modeling of hyoscine drug solubility and solvent density in supercritical processing: gradient boosting, extra trees, and random forest models

Abstract: This work presents the results of using tree-based models, including Gradient Boosting, Extra Trees, and Random Forest, to model the solubility of hyoscine drug and solvent density based on pressure and temperature as inputs. The models were trained on a dataset of hyoscine drug with known solubility and density values, optimized with WCA algorithm, and their accuracy was evaluated using R2, MSE, MAPE, and Max Error metrics. The results showed that Gradient Boosting and Extra Trees models had high accuracy, wi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(12 citation statements)
references
References 35 publications
0
6
0
Order By: Relevance
“…The model performance is examined from two perspectives: the model interpretation ability and the prediction error. In terms of model interpretation ability, Chen et al 45 and Ghazwani and Begum 46 illustrate that ensemble learning can adjust itself based on the deviation between the model fitting value and the observation value in the previous calculation and can self-check the accuracy of the model. Therefore, we believe that the difference between the estimated values of the model and the observations can be used as a standard to evaluate the interpretation ability of the prediction model.…”
Section: Methodsmentioning
confidence: 99%
“…The model performance is examined from two perspectives: the model interpretation ability and the prediction error. In terms of model interpretation ability, Chen et al 45 and Ghazwani and Begum 46 illustrate that ensemble learning can adjust itself based on the deviation between the model fitting value and the observation value in the previous calculation and can self-check the accuracy of the model. Therefore, we believe that the difference between the estimated values of the model and the observations can be used as a standard to evaluate the interpretation ability of the prediction model.…”
Section: Methodsmentioning
confidence: 99%
“…Boosting learners learn in a sequential manner to correct errors from the previous learner and create a robust model to reduce model bias. Therefore, a gradient boosting ML increases accuracy more than other ML algorithms, such as random forest 49 , 51 . In this study, the LightGBM 26 was adopted as a gradient boosting ML, which significantly outperforms other gradient boosting algorithms in terms of computational speed and memory consumption 52 .…”
Section: Methodsmentioning
confidence: 99%
“…The process entails quantifying the average reduction in predictive accuracy resulting from altering a single input variable while holding all other variables constant. This process entails assigning a score that represents the relative relevance of each variable, which then aids in selecting the most impactful features for the ultimate model 47 .…”
Section: Theoretical Backgroundmentioning
confidence: 99%
“…It was made to prevent overfitting. Training each base estimator with a random subset of features is fundamental to the ETR algorithm's success, just as in the RF 47 . ETR uses the whole training dataset to train each regression tree.…”
Section: Theoretical Backgroundmentioning
confidence: 99%