2018
DOI: 10.1016/j.jbi.2018.06.016
|View full text |Cite
|
Sign up to set email alerts
|

Patient representation learning and interpretable evaluation using clinical notes

Abstract: We have three contributions in this work: 1. We explore the utility of a stacked denoising autoencoder and a paragraph vector model to learn task-independent dense patient representations directly from clinical notes. To analyze if these representations are transferable across tasks, we evaluate them in multiple supervised setups to predict patient mortality, primary diagnostic and procedural category, and gender. We compare their performance with sparse representations obtained from a bag-of-words model. We o… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
24
0
3

Year Published

2019
2019
2023
2023

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(32 citation statements)
references
References 24 publications
0
24
0
3
Order By: Relevance
“…Liu’s model [ 15 ] forecasts the onset of 3 kinds of diseases using medical notes. Sushil [ 16 ] utilizes a stacked denoised autoencoder and a paragraph vector model to learn generalized patient representation directly from clinical notes and the learned representation is used to predict mortality.…”
Section: Introductionmentioning
confidence: 99%
“…Liu’s model [ 15 ] forecasts the onset of 3 kinds of diseases using medical notes. Sushil [ 16 ] utilizes a stacked denoised autoencoder and a paragraph vector model to learn generalized patient representation directly from clinical notes and the learned representation is used to predict mortality.…”
Section: Introductionmentioning
confidence: 99%
“…Miotto et al [25] adopted SDAs to generate patient representations. Furthermore, Sushil et al [26] derived task-independent patient representations directly from clinical notes by applying SDAs and a paragraph vector model. The above two methods only consider the frequency of medical events.…”
Section: Related Workmentioning
confidence: 99%
“…The learned representations are used to predict ICD codes occurring in the next 30, 60, 90, and 180 days. In contrast to the previous works, Sushil et al 20 focuses exclusively on EHR text to learn patient representations using unsupervised methods, such as stacked denoising autoencoders and doc2vec. 7 They find that the learned representations outperform traditional bag-of-words representations when few training examples are available and that the target task does not rely on strong lexical features.…”
Section: Introductionmentioning
confidence: 99%
“…7 They find that the learned representations outperform traditional bag-of-words representations when few training examples are available and that the target task does not rely on strong lexical features. Like Sushil et al, 20 our work uses text variables only.…”
Section: Introductionmentioning
confidence: 99%