Proceedings of the 20th Workshop on Biomedical Language Processing 2021
DOI: 10.18653/v1/2021.bionlp-1.5
|View full text |Cite
|
Sign up to set email alerts
|

Are we there yet? Exploring clinical domain knowledge of BERT models

Abstract: We explore whether state-of-the-art BERT models encode sufficient domain knowledge to correctly perform domain-specific inference. Although BERT implementations such as BioBERT are better at domain-based reasoning than those trained on general-domain corpora, there is still a wide margin compared to human performance on these tasks. To bridge this gap, we explore whether supplementing textual domain knowledge in the medical NLI task: a) by further language model pretraining on the medical domain corpora, b) by… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 44 publications
0
6
0
Order By: Relevance
“…Overall, although these models show benefits in their respective domains, they did not incorporate clinical knowledge to address challenges in clinical applications. 23 In practice, NR for clinical context has the following challenges. First, there can be accumulation of multiple numeric examples in a condensed context, such as "Physical examination: temperature 97.5, blood pressure 124/55, pulse 79, respirations 18, O 2 saturation 99% on room air."…”
Section: Impact Statementmentioning
confidence: 99%
“…Overall, although these models show benefits in their respective domains, they did not incorporate clinical knowledge to address challenges in clinical applications. 23 In practice, NR for clinical context has the following challenges. First, there can be accumulation of multiple numeric examples in a condensed context, such as "Physical examination: temperature 97.5, blood pressure 124/55, pulse 79, respirations 18, O 2 saturation 99% on room air."…”
Section: Impact Statementmentioning
confidence: 99%
“…More generally, however, there is some evidence that the effectiveness of augmenting questions with textual knowledge is limited in the biomedical domain. For instance, Sushil et al [62] evaluated the effect of such augmentation strategies and failed to obtain any statistically significant improvements for MedNLI [55], a well-known benchmark for Natural Language Inference (NLI) in the biomedical domain. These findings were also corroborated by our own initial analysis.…”
Section: Related Workmentioning
confidence: 99%
“…When it comes to interpreting patient descriptions, however, the potential of such strategies is less clear. For instance, Sushil et al [62] used an information retrieval engine to find relevant sentences in biomedical corpora, which were then added to the premise of Natural Language Inference (NLI) instances. In experiments on MedNLI [55], they found no statistically significant improvements as a result Table 1: Example of a question from MedQA, along with the answer candidates.…”
Section: Introductionmentioning
confidence: 99%
“…Other works [8,12,28] designed special modules for numerical reasoning in text which were then integrated with neural networks. Overall, these models have shown advancements in the respective domains for specialized problems but they did not incorporate clinical knowledge with specific extensive reasoning for clinical applications [33].…”
Section: Related Workmentioning
confidence: 99%