2021
DOI: 10.48550/arxiv.2112.07869
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fine-Tuning Large Neural Language Models for Biomedical Natural Language Processing

Abstract: Motivation: A perennial challenge for biomedical researchers and clinical practitioners is to stay abreast with the rapid growth of publications and medical notes. Natural language processing (NLP) has emerged as a promising direction for taming information overload. In particular, large neural language models facilitate transfer learning by pretraining on unlabeled text, as exemplified by the successes of BERT models in various NLP applications. However, fine-tuning such models for an end task remains challen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…We, therefore, introduce a tokenizer training stage that generates a new vocabulary file, by combining the biomedical and clinical corpora used to train and fine-tune the final model. The main difference between our approach and most BERT variants is that other LMs are based technically on a continuous training approach, whereby the source model is fine-tuned on a specific domain corpus [13], whereas our model adopts a similar approach to that of Tinn et al [8] by including a dedicated tokenizer in the process.…”
Section: • We Propose a Modeling Optimization Technique Using Amentioning
confidence: 99%
See 2 more Smart Citations
“…We, therefore, introduce a tokenizer training stage that generates a new vocabulary file, by combining the biomedical and clinical corpora used to train and fine-tune the final model. The main difference between our approach and most BERT variants is that other LMs are based technically on a continuous training approach, whereby the source model is fine-tuned on a specific domain corpus [13], whereas our model adopts a similar approach to that of Tinn et al [8] by including a dedicated tokenizer in the process.…”
Section: • We Propose a Modeling Optimization Technique Using Amentioning
confidence: 99%
“…• training epoch i = [5 : 100] • per device training batch size β = [4,8,12,16,32,64] By running 200 trials for each search, the resultant customized hyperparameters helped boost our performance to an average 3.11% F1 score for all downstream tasks.…”
Section: A Fine-tuning Hyperparametersmentioning
confidence: 99%
See 1 more Smart Citation
“…EHRs contain a tremendous amount of structured and unstructured data, which can be used to fine-tune predictive algorithms and drug compatibilities and help to understand the course of diseases and patients. Various researchers have proposed adapted NLP models to address better biomedical documents [4], [5], [7], [8]. With medical texts representing 80% of the EHR data [1], it's imperative to develop more robust and efficient language models which can be used to understand and extract relevant information contained in those texts.…”
Section: Introductionmentioning
confidence: 99%
“…We thus conducted a tokenizer training that generated a new vocabulary file, combining biomedical and clinical corpora used to train and fine-tune the final model. The main difference between our approach and most BERT variants is that those LM are technically based on a continuous training approach where the source model is fine-tuned on a specific domain corpus [13], while ours adopts a similar approach as [8] by including in the process a dedicated tokenizer.…”
Section: Introductionmentioning
confidence: 99%