2022
DOI: 10.21037/jgo-22-1238
|View full text |Cite
|
Sign up to set email alerts
|

Development and validation of an XGBoost model to predict 5-year survival in elderly patients with intrahepatic cholangiocarcinoma after surgery: a SEER-based study

Abstract: Background: Nomograms have been established to predict survival in postoperative or elderly intrahepatic cholangiocarcinoma (ICC) patients. There are no models to predict postoperative survival in elderly ICC patients. Extreme gradient boosting (XGBoost) can adjust the errors generated by existing models. This retrospective cohort study aimed to develop and validate an XGBoost model to predict postoperative 5-year survival in elderly ICC patients. Methods: The Surveillance, Epidemiology, and End Results (SEER)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
9
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(10 citation statements)
references
References 38 publications
1
9
0
Order By: Relevance
“…In addition, Zhong et al [ 16 ] applied the XGBoost algorithm to create a prognostic model for patients with breast cancer with bone metastasis and showed AUC values of 0.88 and 0.80 in the training and test sets. Consistent with the previous studies [ 14 16 ], our present study also revealed that the XGBoost model showed good performance in prognostic survival prediction models, showing AUCs greater than 0.7 and even the 5-year AUC value over 0.8 (training data). Generally, an AUC ≥ 0.7 indicates that the model has an adequate predictive ability [ 22 ].…”
Section: Discussionsupporting
confidence: 92%
See 2 more Smart Citations
“…In addition, Zhong et al [ 16 ] applied the XGBoost algorithm to create a prognostic model for patients with breast cancer with bone metastasis and showed AUC values of 0.88 and 0.80 in the training and test sets. Consistent with the previous studies [ 14 16 ], our present study also revealed that the XGBoost model showed good performance in prognostic survival prediction models, showing AUCs greater than 0.7 and even the 5-year AUC value over 0.8 (training data). Generally, an AUC ≥ 0.7 indicates that the model has an adequate predictive ability [ 22 ].…”
Section: Discussionsupporting
confidence: 92%
“…In recent years, machine learning-based AI models attracted increasing attention in clinical practice [ 14 , 20 , 21 ]. Especially, AI-based technologies have made a significant contribution to the field of cancer research [ 21 ].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Four filling methods and three feature screening methods were used to obtain 12 datasets. Eighteen machine learning algorithms, including logistic regression [27, 28], Latent Dirichlet allocation [29], Quadratic Discriminant Analysis [30], Stochastic Gradient Descent [31], k-Nearest Neighbor [32], Decision Tree [33], Naive Bayes [34], Gaussian Naïve Bayes [35], Multinomial Naive Bayes [36], Bernoulli Naïve Bayes [37], Support Vector Machine [38], passive-aggressive [39], AdaBoost [40], bagging, Random Forest [41], Extremely Randomized Trees [42], gradient boosting [41], XGBoost [43], and ensemble learning [44], were used to train 216 models. The process of building the models was as follows:The dataset was randomly divided into a training and a test set in a ratio of 8:2.The training set data were entered into the machine learning model, and the 10-fold cross-validation method was used to continuously adjust the model parameters, so that the parameters had the largest area under the receiver operating characteristic curve (AUC) value on the training set.…”
Section: Methodsmentioning
confidence: 99%
“…The predictive performance of each trained model (for each of the 7 prediction windows, as well as for the averaging approach) on its corresponding test set was evaluated using several standard performance measures [ 45 , 46 ]. These performance metrics were AUROC (the probability that a classifier will be able to distinguish between an instance of the positive class and one of the negative class), sensitivity (true positive rate), specificity (true negative rate), positive predictive value (PPV, the proportion of true positives among all positive predictions), negative predictive value (NPV, the proportion of true negatives among all negative predictions), and accuracy (the proportion of correct predictions among all predictions).…”
Section: Methodsmentioning
confidence: 99%