2020
DOI: 10.1038/s41598-020-62729-x
|View full text |Cite
|
Sign up to set email alerts
|

Predictive Analytics for Retention in Care in an Urban HIV Clinic

Abstract: consistent medical care among people living with HiV is essential for both individual and public health. HIV-positive individuals who are 'retained in care' are more likely to be prescribed antiretroviral medication and achieve HIV viral suppression, effectively eliminating the risk of transmitting HIV to others. However, in the United States, less than half of HIV-positive individuals are retained in care. Interventions to improve retention in care are resource intensive, and there is currently no systematic … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
17
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
2

Relationship

2
8

Authors

Journals

citations
Cited by 28 publications
(19 citation statements)
references
References 34 publications
1
17
1
Order By: Relevance
“…Results from NLP of clinical notes could potentially augment such electronic alerts. Recent studies have used structured EMR fields, including documentation of substance use and mental illness, to create predictive models of HIV appointment adherence [ 12 , 40 ]. However, if mental illness and substance use are not adequately documented in structured EMR fields, inclusion of NLP of clinician notes may improve such predictive models by identifying additional risk factors for appointment nonadherence.…”
Section: Discussionmentioning
confidence: 99%
“…Results from NLP of clinical notes could potentially augment such electronic alerts. Recent studies have used structured EMR fields, including documentation of substance use and mental illness, to create predictive models of HIV appointment adherence [ 12 , 40 ]. However, if mental illness and substance use are not adequately documented in structured EMR fields, inclusion of NLP of clinician notes may improve such predictive models by identifying additional risk factors for appointment nonadherence.…”
Section: Discussionmentioning
confidence: 99%
“…We trained the logistic regression classifier on a dataset consisting of roughly 400,000 pings, with 20% of the data held out as a test set. We used an 2 penalty to regularize the classifier, and chose the regularization strength via temporal cross-validation [5,6] while optimizing for the area under the curve (AUC) on the validation data. The dataset was heavily imbalanced, with about 6% of samples in the total dataset occurring in the positive class (translator responding "yes").…”
Section: Resultsmentioning
confidence: 99%
“…We evaluated the overall prediction accuracy (proportion with correctly predicted positive or negative disengagement status out of the total population) and area under the diagnostic curve (AUC). Additionally, we assessed the efficiency of each model for targeting the "highest risk" individuals in the context of limited resources where not all individuals can receive an intervention [14,21]. For set risk score thresholds corresponding to proportions of the population flagged as "high risk" (ranging from 10-50% targeted to simulate different intervention scenarios), we evaluated each model's sensitivity (proportion correctly categorized as "high risk"-i.e., truly went on to disengage from care-out of the total population who disengaged from care in the prediction period) and positive predictive value (PPV, the proportion correctly categorized as "high risk" out of the total population categorized as high risk).…”
Section: Plos Global Public Healthmentioning
confidence: 99%