2021
DOI: 10.48550/arxiv.2110.05006
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pre-trained Language Models in Biomedical Domain: A Systematic Survey

Abstract: Pre-trained language models (PLMs) have been the de facto paradigm for most natural language processing (NLP) tasks. This also benefits biomedical domain: researchers from informatics, medicine, and computer science (CS) communities proposes various PLMs trained on biomedical datasets, e.g., biomedical text, electronic health records, protein, and DNA sequences for various biomedical tasks. However, the cross-discipline characteristics of biomedical PLMs hinder their spreading among communities; some existing … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(13 citation statements)
references
References 288 publications
0
11
0
Order By: Relevance
“…domain such as BioBERT , Clini-calXLNET (Huang et al, 2020), and others (Alrowili and Shanker, 2021; Kraljevic et al, 2021;Phan et al, 2021), this paradigm is widely used for creating many task-specific models (Wang et al, 2021a;Banerjee et al, 2021). However, task-specific models have limitations to real-world applications because this approach is computationally expensive (i.e., require large computational resources) and time-consuming (Strubell et al, 2019;Schwartz et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…domain such as BioBERT , Clini-calXLNET (Huang et al, 2020), and others (Alrowili and Shanker, 2021; Kraljevic et al, 2021;Phan et al, 2021), this paradigm is widely used for creating many task-specific models (Wang et al, 2021a;Banerjee et al, 2021). However, task-specific models have limitations to real-world applications because this approach is computationally expensive (i.e., require large computational resources) and time-consuming (Strubell et al, 2019;Schwartz et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…3) Our dataset only includes papers in the biomedical domain, while SSN consists of papers from several different fields including mathematical, physics, and computer science, our dataset can help the evaluation of domainspecific tasks in the research community. Moreover, biomedical scientific papers are laden with terminology and have complex syntactic structures [21,31]. This makes our dataset a challenging benchmark for automatic summarization methods.…”
Section: Citation-aware Pubmed Datasetmentioning
confidence: 99%
“…[42], [43] as an image extraction approach, following the recent breakthrough of Transformers [27] in handling natural language processing tasks [44]. ViT mainly consists of the following parts: Linear Projection of Flattened Patches (Embedding layer), Transformer Encoder, and MLP Head.…”
Section: Sota Vit and Cnn Modelsmentioning
confidence: 99%