2021
DOI: 10.1371/journal.pone.0257092
|View full text |Cite
|
Sign up to set email alerts
|

Relation classification via BERT with piecewise convolution and focal loss

Abstract: Recent relation extraction models’ architecture are evolved from the shallow neural networks to natural language model, such as convolutional neural networks or recurrent neural networks to Bert. However, these methods did not consider the semantic information in the sequence or the distance dependence problem, the internal semantic information may contain the useful knowledge which can help relation classification. Focus on these problems, this paper proposed a BERT-based relation classification method. Compa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(9 citation statements)
references
References 27 publications
(31 reference statements)
0
6
0
Order By: Relevance
“…The current leading NLP models such as BERT [20], GPT [21], and T5 [22] announced later are all based on this transformer block. In particular, BERT is commonly used in biomedical text mining research because it is built on multiple transformers encoder blocks, which has the advantage of compressing the sentence and mining semantic information from it [8,[23][24][25].…”
Section: Deep Learning-based Semantic Relation Classification Modelmentioning
confidence: 99%
“…The current leading NLP models such as BERT [20], GPT [21], and T5 [22] announced later are all based on this transformer block. In particular, BERT is commonly used in biomedical text mining research because it is built on multiple transformers encoder blocks, which has the advantage of compressing the sentence and mining semantic information from it [8,[23][24][25].…”
Section: Deep Learning-based Semantic Relation Classification Modelmentioning
confidence: 99%
“…They also leveraged the entity information in their proposed model to improve performance. In a recent study, Liu et al further extend this architecture in [29]. To capture the latent information around the target entities, the authors utilize piecewise convolution [30].…”
Section: Related Workmentioning
confidence: 99%
“…We also use another BERT-based architecture proposed in [29] with minor changes. This architecture is an extension of the previous BERT-based architecture proposed by [28].…”
Section: ) Multilingual-r-bert + Pcnnmentioning
confidence: 99%
See 1 more Smart Citation
“…To avoid potential overfitting, we used an early stop when the learning rate drops below 10 -6 or 1000 epochs were exceeded. Focal loss function was applied (33,34). Note that the deep learning model used only the image information where clinical features were not included.…”
Section: Deep Learning Model Constructionmentioning
confidence: 99%