Proceedings of the 2nd Clinical Natural Language Processing Workshop 2019
DOI: 10.18653/v1/w19-1909
|View full text |Cite
|
Sign up to set email alerts
|

Publicly Available Clinical

Abstract: Contextual word embedding models such as ELMo (Peters et al., 2018) and BERT (Devlin et al., 2018) have dramatically improved performance for many natural language processing (NLP) tasks in recent months. However, these models have been minimally explored on specialty corpora, such as clinical text; moreover, in the clinical domain, no publicly-available pre-trained BERT models yet exist. In this work, we address this need by exploring and releasing BERT models for clinical text: one for generic clinical text … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

11
850
1
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 1,054 publications
(863 citation statements)
references
References 26 publications
11
850
1
1
Order By: Relevance
“…Contextual word embeddings are often pretrained on a large dataset through self-supervised tasks, then released for fine-tuned use in downstream tasks. This pretraining can be domain specific, such as in the various clinical-text specific BERT models which have been released [3,29,60]. This pretraining task can be another source in which bias present in training text can be hard-coded into a word embedding model.…”
Section: Background and Related Work 21 Contextual Embeddingsmentioning
confidence: 99%
See 3 more Smart Citations
“…Contextual word embeddings are often pretrained on a large dataset through self-supervised tasks, then released for fine-tuned use in downstream tasks. This pretraining can be domain specific, such as in the various clinical-text specific BERT models which have been released [3,29,60]. This pretraining task can be another source in which bias present in training text can be hard-coded into a word embedding model.…”
Section: Background and Related Work 21 Contextual Embeddingsmentioning
confidence: 99%
“…Several BERT models pretrained on MIMIC notes are publicly available [3,29,60]. However, to address several limitations, we choose to train our own clinical BERT model in this work.…”
Section: Pretrained Clinical Embeddingsmentioning
confidence: 99%
See 2 more Smart Citations
“…Alsentzer et al used approximately 2 million clinical notes from the MIMIC-III v1.4 database [8] and pre-trained a Clinical BERT model [9]. They made it publicly available; otherwise it originally took about 17 days of computational runtime by a single GeForce GTX TITAN X 12 GB GPU.…”
Section: Clinical Bertmentioning
confidence: 99%