Objectives To assess fairness and bias of a previously validated machine learning opioid misuse classifier. Materials & Methods Two experiments were conducted with the classifier’s original (n = 1000) and external validation (n = 53 974) datasets from 2 health systems. Bias was assessed via testing for differences in type II error rates across racial/ethnic subgroups (Black, Hispanic/Latinx, White, Other) using bootstrapped 95% confidence intervals. A local surrogate model was estimated to interpret the classifier’s predictions by race and averaged globally from the datasets. Subgroup analyses and post-hoc recalibrations were conducted to attempt to mitigate biased metrics. Results We identified bias in the false negative rate (FNR = 0.32) of the Black subgroup compared to the FNR (0.17) of the White subgroup. Top features included “heroin” and “substance abuse” across subgroups. Post-hoc recalibrations eliminated bias in FNR with minimal changes in other subgroup error metrics. The Black FNR subgroup had higher risk scores for readmission and mortality than the White FNR subgroup, and a higher mortality risk score than the Black true positive subgroup (P < .05). Discussion The Black FNR subgroup had the greatest severity of disease and risk for poor outcomes. Similar features were present between subgroups for predicting opioid misuse, but inequities were present. Post-hoc mitigation techniques mitigated bias in type II error rate without creating substantial type I error rates. From model design through deployment, bias and data disadvantages should be systematically addressed. Conclusion Standardized, transparent bias assessments are needed to improve trustworthiness in clinical machine learning models.
Background Approaches are needed to better delineate the continuum of opioid misuse that occurs in hospitalized patients. A prognostic enrichment strategy with latent class analysis (LCA) may facilitate treatment strategies in subtypes of opioid misuse. We aim to identify subtypes of patients with opioid misuse and examine the distinctions between the subtypes by examining patient characteristics, topic models from clinical notes, and clinical outcomes. Methods This was an observational study of inpatient hospitalizations at a tertiary care center between 2007 and 2017. Patients with opioid misuse were identified using an operational definition applied to all inpatient encounters. LCA with eight class-defining variables from the electronic health record (EHR) was applied to identify subtypes in the cohort of patients with opioid misuse. Comparisons between subtypes were made using the following approaches: (1) descriptive statistics on patient characteristics and healthcare utilization using EHR data and census-level data; (2) topic models with natural language processing (NLP) from clinical notes; (3) association with hospital outcomes. Findings The analysis cohort was 6,224 (2.7% of all hospitalizations) patient encounters with opioid misuse with a data corpus of 422,147 clinical notes. LCA identified four subtypes with differing patient characteristics, topics from the clinical notes, and hospital outcomes. Class 1
Background: Automated de-identification methods for removing protected health information (PHI) from the source notes of the electronic health record (EHR) rely on building systems to recognize mentions of PHI in text, but they remain inadequate at ensuring perfect PHI removal. As an alternative to relying on de-identification systems, we propose the following solutions: (1) Mapping the corpus of documents to standardized medical vocabulary (concept unique identifier [CUI] codes mapped from the Unified Medical Language System) thus eliminating PHI as inputs to a machine learning model; and (2) training character-based machine learning models that obviate the need for a dictionary containing input words/n-grams. We aim to test the performance of models with and without PHI in a use-case for an opioid misuse classifier. Methods: An observational cohort sampled from adult hospital inpatient encounters at a health system between 2007 and 2017. A case-control stratified sampling (n = 1000) was performed to build an annotated dataset for a reference standard of cases and non-cases of opioid misuse. Models for training and testing included CUI codes, character-based, and n-gram features. Models applied were machine learning with neural network and logistic regression as well as expert consensus with a rule-based model for opioid misuse. The area under the receiver operating characteristic curves (AUROC) were compared between models for discrimination. The Hosmer-Lemeshow test and visual plots measured model fit and calibration. Results: Machine learning models with CUI codes performed similarly to n-gram models with PHI. The top performing models with AUROCs > 0.90 included CUI codes as inputs to a convolutional neural network, max pooling network, and logistic regression model. The top calibrated models with the best model fit were the CUIbased convolutional neural network and max pooling network. The top weighted CUI codes in logistic regression has the related terms 'Heroin' and 'Victim of abuse'.
Background Unhealthy alcohol use (UAU) is known to disrupt pulmonary immune mechanisms and increase the risk of acute respiratory distress syndrome in patients with pneumonia; however, little is known about the effects of UAU on outcomes in patients with COVID-19 pneumonia. To our knowledge, this is the first observational cross-sectional study that aims to understand the effect of UAU on the severity of COVID-19. Objective We aim to determine if UAU is associated with more severe clinical presentation and worse health outcomes related to COVID-19 and if socioeconomic status, smoking, age, BMI, race/ethnicity, and pattern of alcohol use modify the risk. Methods In this observational cross-sectional study that took place between January 1, 2020, and December 31, 2020, we ran a digital machine learning classifier on the electronic health record of patients who tested positive for SARS-CoV-2 via nasopharyngeal swab or had two COVID-19 International Classification of Disease, 10th Revision (ICD-10) codes to identify patients with UAU. After controlling for age, sex, ethnicity, BMI, smoking status, insurance status, and presence of ICD-10 codes for cancer, cardiovascular disease, and diabetes, we then performed a multivariable regression to examine the relationship between UAU and COVID-19 severity as measured by hospital care level (ie, emergency department admission, emergency department admission with ventilator, or death). We used a predefined cutoff with optimal sensitivity and specificity on the digital classifier to compare disease severity in patients with and without UAU. Models were adjusted for age, sex, race/ethnicity, BMI, smoking status, and insurance status. Results Each incremental increase in the predicted probability from the digital alcohol classifier was associated with a greater odds risk for more severe COVID-19 disease (odds ratio 1.15, 95% CI 1.10-1.20). We found that patients in the unhealthy alcohol group had a greater odds risk to develop more severe disease (odds ratio 1.89, 95% CI 1.17-3.06), suggesting that UAU was associated with an 89% increase in the odds of being in a higher severity category. Conclusions In patients infected with SARS-CoV-2, UAU is an independent risk factor associated with greater disease severity and/or death.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.