2023
DOI: 10.1016/j.patter.2023.100729
|View full text |Cite
|
Sign up to set email alerts
|

Fine-tuning large neural language models for biomedical natural language processing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 42 publications
(17 citation statements)
references
References 14 publications
0
17
0
Order By: Relevance
“…The overall low mDISCERN scores observed for the Bloomz model should not be interpreted as a definitive disqualification for patient recommendation. Instead, these findings should motivate the scientific community to explore and enhance the potential of this model through advanced fine-tuning techniques [25] and more effective prompting strategies, especially given that it is the sole open-source model within the examined cohort. Other general factors that might have an impact on model performance are the complexity and diversity of the training data, the presence of inherent biases in the data, the computational resources available during training, general model architecture, and the ongoing adjustments and updates made to the model postdeployment to respond to real-world feedback.…”
Section: Discussionmentioning
confidence: 99%
“…The overall low mDISCERN scores observed for the Bloomz model should not be interpreted as a definitive disqualification for patient recommendation. Instead, these findings should motivate the scientific community to explore and enhance the potential of this model through advanced fine-tuning techniques [25] and more effective prompting strategies, especially given that it is the sole open-source model within the examined cohort. Other general factors that might have an impact on model performance are the complexity and diversity of the training data, the presence of inherent biases in the data, the computational resources available during training, general model architecture, and the ongoing adjustments and updates made to the model postdeployment to respond to real-world feedback.…”
Section: Discussionmentioning
confidence: 99%
“…This operation requires a costly manual curation, operated by highly expert users. This is an inevitable effort to handle data scarcity, analyzed in [ 65 ] in general terms, becoming even more relevant in biomedical fields [ 66 , 67 ]. To minimize such effort in our case, we implemented a process that supports the building of small high-quality training datasets.…”
Section: Methodsmentioning
confidence: 99%
“…Recent studies have provided compelling evidence that the fine-tuning of Large Language Models (LLMs) with specialized medical data sources can facilitate their adaptation to specific tasks within clinical settings [19, 20, 21]. For instance, Yang et al successfully developed an LLM model from fine-tuning BERT and ChatGPT to extract and recognize HPO phenotypes in clinical texts within the presence of non-HPO phenotypes, typos, and semantic differences with the model’s original training data [22].…”
Section: Background and Significancementioning
confidence: 99%
“…Recent studies have provided compelling evidence that the fine-tuning of LLMs with specialized medical data sources can facilitate their adaptation to specific tasks within clinical settings [20][21][22]. Fine-tuning is a process in which an unsupervised pre-trained LLM is further trained on a smaller, task-specific dataset to adapt its parameters to a specific task.…”
Section: Introductionmentioning
confidence: 99%