2019
DOI: 10.48550/arxiv.1904.05342
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission

Abstract: Clinical notes contain information about patients that goes beyond structured data like lab values and medications. However, clinical notes have been underused relative to structured data, because notes are high-dimensional and sparse. This work develops and evaluates representations of clinical notes using bidirectional transformers (ClinicalBert). Clini-calBert uncovers high-quality relationships between medical concepts as judged by humans. ClinicalBert outperforms baselines on 30-day hospital readmission p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
186
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 217 publications
(221 citation statements)
references
References 36 publications
0
186
0
Order By: Relevance
“…Similarly to [15,9], we leveraged approximately 2 million clinical notes extracted from the MIMIC-III [21] dataset, which is the largest publicly available EHR dataset that contains clinical narratives of over 40,000 patients admitted to the intensive care units. We only applied minimal pre-processing steps, including 1) to remove all de-identification placeholders that were generated to protect the PHI (protected health information); 2) to replace all characters other than alphanumericals and punctuation marks; 3) to convert all alphabetical characters to lower cases; and 4) to strip extra white spaces.…”
Section: Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…Similarly to [15,9], we leveraged approximately 2 million clinical notes extracted from the MIMIC-III [21] dataset, which is the largest publicly available EHR dataset that contains clinical narratives of over 40,000 patients admitted to the intensive care units. We only applied minimal pre-processing steps, including 1) to remove all de-identification placeholders that were generated to protect the PHI (protected health information); 2) to replace all characters other than alphanumericals and punctuation marks; 3) to convert all alphabetical characters to lower cases; and 4) to strip extra white spaces.…”
Section: Datasetsmentioning
confidence: 99%
“…In clinical NLP, the models that are applied with the transformer-based approaches also encounter this limitation [8]. For example, the discharge summaries in MIMIC-III, which are always used to predict hospital re-admission [9] or mortality [10], have 1,435 words in average, far exceeding the 512 tokens limits of BERT like models.…”
Section: Introductionmentioning
confidence: 99%
“…The large amount of data generated in this process offers an opportunity for * Corresponding Author. deep learning technology to improve healthcare, such as diagnoses prediction (Choi et al, 2016), medication recommendation (Shang et al, 2019), mortality prediction (Tang et al, 2020), and readmission prediction (Huang et al, 2019). However, comparing to common academic datasets, such as ImageNet (Deng et al, 2009) and WMT (Macháček and Bojar, 2014), real-world EHR data is longitudinal, heterogeneous, and multimodal, which proposes big challenges to leverage the information included in it.…”
Section: Introductionmentioning
confidence: 99%
“…Due to the importance of clinical notes, it is necessary to combine them with other data sources for the integrity of clinical features. Since the common pre-trained language models such as BERT (Devlin et al, 2019) do not consider the specific complexity of clinical notes, we apply the ClinicalBERT (Huang et al, 2019) that are pre-trained on clinical notes for handling the notes data in this paper.…”
Section: Introductionmentioning
confidence: 99%
“…Recent years have witnessed the great success of pre-trained language models (PLMs), such as BERT [7], in a broad range of natural language processing (NLP) tasks. Moreover, several domainoriented PLMs have been proposed to adapt to specific domains [4,8,10]. For instance, BioBERT [13] and SciBERT [2] are pretrained leveraging large-scale domain-specific corpora for biomedical and scientific domain tasks respectively.…”
Section: Introductionmentioning
confidence: 99%