Silent AR is common during DGF. Prolonged DGF is associated with reduced graft survival after kidney transplantation, and much of this association can be explained by silent AR. In the absence of data from randomized trials, protocol biopsies and treatment of silent AR during prolonged DGF appear to be warranted.
BACKGROUND Hepatorenal syndrome (HRS) is a life-threatening condition among patients with advanced liver disease. Data trends specific to hospital mortality and hospital admission resource utilization for HRS remain limited. AIM To assess the temporal trend in mortality and identify the predictors for mortality among hospital admissions for HRS in the United States. METHODS We used the National Inpatient Sample database to identify an unweighted sample of 4938 hospital admissions for HRS from 2005 to 2014 (weighted sample of 23973 admissions). The primary outcomes were temporal trends in mortality as well as predictors for hospital mortality. We estimated odds ratios from multi-level mixed effect logistic regression to identify patient characteristics and treatments associated with hospital mortality. RESULTS Overall hospital mortality was 32%. Hospital mortality decreased from 44% in 2005 to 24% in 2014 ( P < 0.001), while there was an increase in the rate of liver transplantation ( P = 0.02), renal replacement therapy ( P < 0.001), length of hospital stay ( P < 0.001), and hospitalization cost ( P < 0.001). On multivariable analysis, older age, alcohol use, coagulopathy, neurological disorder, and need for mechanical ventilation predicted higher hospital mortality, whereas liver transplantation, transjugular intrahepatic portosystemic shunt, and abdominal paracentesis were associated with lower hospital mortality. CONCLUSION Although there was an increase in resource utilizations, hospital mortality among patients admitted for HRS significantly improved. Several predictors for hospital mortality were identified.
This study aims to evaluate the risk factors and the association of acute kidney injury with treatments, complications, outcomes, and resource utilization in patients hospitalized for heat stroke in the United States. Hospitalized patients from years 2003 to 2014 with a primary diagnosis of heat stroke were identified in the National Inpatient Sample dataset. End stage kidney disease patients were excluded. The occurrence of acute kidney injury during hospitalization was identified using the hospital diagnosis code. The associations between acute kidney injury and clinical characteristics, in-hospital treatments, outcomes, and resource utilization were assessed using multivariable analyses. A total of 3346 hospital admissions were included in the analysis. Acute kidney injury occurred in 1206 (36%) admissions, of which 49 (1.5%) required dialysis. The risk factors for acute kidney injury included age 20–39 years, African American race, obesity, chronic kidney disease, congestive heart failure, and rhabdomyolysis, whereas age <20 or ≥60 years were associated with lower risk of acute kidney injury. The need for mechanical ventilation and blood transfusion was higher when acute kidney injury occurred. Acute kidney injury was associated with electrolyte and acid-base derangements, sepsis, acute myocardial infarction, ventricular arrhythmia or cardiac arrest, respiratory, circulatory, liver, neurological, hematological failure, and in-hospital mortality. Length of hospital stay and hospitalization cost were higher in acute kidney injury patients. Approximately one third of heat stroke patients developed acute kidney injury during hospitalization. Acute kidney injury was associated with several complications, and higher mortality and resource utilization.
Introduction We aimed to assess the association between serum potassium and mortality in patients receiving continuous renal replacement therapy (CRRT). Methods We studied 1279 acute kidney injury patients receiving CRRT in a tertiary referral hospital in the United States. We used logistic regression to assess the association of serum potassium before CRRT and mean serum potassium during CRRT with 90‐day mortality after CRRT initiation, using serum potassium 4.0–4.4 mmol/L as reference group. Results Before CRRT, there was a U‐shaped association between serum potassium and 90‐day mortality. There was a significant increase in mortality when serum potassium before CRRT was ≤3.4 and ≥4.5 mmol/L. During CRRT, progressively increased mortality was noted when mean serum potassium was ≥4.5 mmol/L. The odds ratio of 90‐day mortality was significantly higher when mean serum potassium was ≥4.5 mmol/L. Conclusion Hypokalemia and hyperkalemia before CRRT and hyperkalemia during CRRT predicts 90‐day mortality.
Background: This study aimed to better characterize morbidly obese kidney transplant recipients, their clinical characteristics, and outcomes by using an unsupervised machine learning approach. Methods: Consensus cluster analysis was applied to OPTN/UNOS data from 2010 to 2019 based on recipient, donor, and transplant characteristics in kidney transplant recipients with a pre-transplant BMI ≥ 40 kg/m2. Key cluster characteristics were identified using the standardized mean difference. Post-transplant outcomes, including death-censored graft failure, patient death, and acute allograft rejection, were compared among the clusters. Results: Consensus clustering analysis identified 3204 kidney transplant recipients with a BMI ≥ 40 kg/m2. In this cohort, five clinically distinct clusters were identified. Cluster 1 recipients were predominantly white and non-sensitized, had a short dialysis time or were preemptive, and were more likely to receive living donor kidney transplants. Cluster 2 recipients were older and diabetic. They were likely to have been on dialysis >3 years and receive a standard KDPI deceased donor kidney. Cluster 3 recipients were young, black, and had kidney disease secondary to hypertension or glomerular disease. Cluster 3 recipients had >3 years of dialysis and received non-ECD, young, deceased donor kidney transplants with a KDPI < 85%. Cluster 4 recipients were diabetic with variable dialysis duration who either received non-ECD standard KDPI kidneys or living donor kidney transplants. Cluster 5 recipients were young retransplants that were sensitized. One-year patient survival in clusters 1, 2, 3, 4, and 5 was 98.0%, 94.4%, 98.5%, 98.7%, and 97%, and one-year death-censored graft survival was 98.1%, 93.0%, 96.1%, 98.8%, and 93.0%, respectively. Cluster 2 had the worst one-year patient survival. Clusters 2 and 5 had the worst one-year death-censored graft survival. Conclusions: With the application of unsupervised machine learning, variable post-transplant outcomes are observed among morbidly obese kidney transplant recipients. Recipients with earlier access to transplant and living donation show superior outcomes. Unexpectedly, reduced graft survival in cluster 3 recipients perhaps underscores socioeconomic access to post-transplant support and minorities being disadvantaged in access to preemptive and living donor transplants. Despite obesity-related concerns, one-year patient and graft survival were favorable in all clusters, and obesity itself should be reconsidered as a hard barrier to kidney transplantation.
Background: Lactic acidosis is the most common cause of anion gap metabolic acidosis in the intensive care unit (ICU), associated with poor outcomes including mortality. We sought to compare machine learning (ML) approaches versus logistic regression analysis for prediction of mortality in lactic acidosis patients admitted to the ICU. Methods: We used the Medical Information Mart for Intensive Care (MIMIC-III) database to identify ICU adult patients with lactic acidosis (serum lactate ≥4 mmol/L). The outcome of interest was hospital mortality. We developed prediction models using four ML approaches consisting of random forest (RF), decision tree (DT), extreme gradient boosting (XGBoost), artificial neural network (ANN), and statistical modeling with forward stepwise logistic regression using the testing dataset. We then assessed model performance using area under the receiver operating characteristic curve (AUROC), accuracy, precision, error rate, Matthews correlation coefficient (MCC), F1 score, and assessed model calibration using the Brier score, in the independent testing dataset. Results: Of 1919 lactic acidosis ICU patients, 1535 and 384 were included in the training and testing dataset, respectively. Hospital mortality was 30%. RF had the highest AUROC at 0.83, followed by logistic regression 0.81, XGBoost 0.81, ANN 0.79, and DT 0.71. In addition, RF also had the highest accuracy (0.79), MCC (0.45), F1 score (0.56), and lowest error rate (21.4%). The RF model was the most well-calibrated. The Brier score for RF, DT, XGBoost, ANN, and multivariable logistic regression was 0.15, 0.19, 0.18, 0.19, and 0.16, respectively. The RF model outperformed multivariable logistic regression model, SOFA score (AUROC 0.74), SAP II score (AUROC 0.77), and Charlson score (AUROC 0.69). Conclusion: The ML prediction model using RF algorithm provided the highest predictive performance for hospital mortality among ICU patient with lactic acidosis.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.