2020
DOI: 10.3390/ijerph17082687
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Named Entity Recognition of Chinese Electronic Medical Records by Combining Domain Dictionary and Rules

Abstract: Electronic medical records are an integral part of medical texts. Entity recognition of electronic medical records has triggered many studies that propose many entity extraction methods. In this paper, an entity extraction model is proposed to extract entities from Chinese Electronic Medical Records (CEMR). In the input layer of the model, we use word embedding and dictionary features embedding as input vectors, where word embedding consists of a character representation and a word representation. Then, the in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 30 publications
(35 reference statements)
0
6
0
Order By: Relevance
“…Based on the evaluation tasks of the previous two years, CCKS 2019 was interested in disease and diagnosis, examination, inspection, surgery, medication, and anatomy. From the related studies of clinical information extraction, it can be found that admission notes, disease course notes, and discharge summaries are the main objects of clinical information extraction [ 20 , 21 ], which contain a large number of clinical entities.…”
Section: Related Workmentioning
confidence: 99%
“…Based on the evaluation tasks of the previous two years, CCKS 2019 was interested in disease and diagnosis, examination, inspection, surgery, medication, and anatomy. From the related studies of clinical information extraction, it can be found that admission notes, disease course notes, and discharge summaries are the main objects of clinical information extraction [ 20 , 21 ], which contain a large number of clinical entities.…”
Section: Related Workmentioning
confidence: 99%
“…Most of the current text classification methods combine models such as CNN and LSTM [ 15 , 16 , 17 ], which prioritize sequential and local information and can capture semantic and syntactic information in continuous word sequences very well. However, they ignore global word co-occurrence, which carries discontinuous as well as long-distance semantic information.…”
Section: Related Workmentioning
confidence: 99%
“…e reason why the BERT pre-training language model is chosen is that the text vector is used as the input of the model, and the granularity of Chinese division is divided into character-level and word-level. Existing research shows that character-level pretraining schemes show better results [37,39], while the BERT pre-training language model is a character-level pretraining program.…”
Section: Bert-bigru-crf Model Constructionmentioning
confidence: 99%