2021
DOI: 10.2196/23104
|View full text |Cite
|
Sign up to set email alerts
|

Clinical Term Normalization Using Learned Edit Patterns and Subconcept Matching: System Development and Evaluation

Abstract: Background Clinical terms mentioned in clinical text are often not in their standardized forms as listed in clinical terminologies because of linguistic and stylistic variations. However, many automated downstream applications require clinical terms mapped to their corresponding concepts in clinical terminologies, thus necessitating the task of clinical term normalization. Objective In this paper, a system for clinical term normalization is presented that utilizes edit … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 21 publications
0
3
0
Order By: Relevance
“…Our effectiveness results agree with the literature [ 83 , 88 ], in which a Macro-F1 score >80% is considered a successful extraction of medical records. Even though there is still a need to cover more tasks related to ICHOM patient-reported outcome measures [ 3 , 74 , 76 , 85 ], we hypothesized that these tasks comprise a feeling state, and the lack of normalization of data contained in EMRs may explain the fact that these task categories did not perform very well [ 70 , 89 ]. Medical records related to baseline characteristics and care processes typically contain much more structured data (eg, numerical values for tasks) than medical patient-reported outcomes, which focus more on unstructured data [ 83 , 90 ].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Our effectiveness results agree with the literature [ 83 , 88 ], in which a Macro-F1 score >80% is considered a successful extraction of medical records. Even though there is still a need to cover more tasks related to ICHOM patient-reported outcome measures [ 3 , 74 , 76 , 85 ], we hypothesized that these tasks comprise a feeling state, and the lack of normalization of data contained in EMRs may explain the fact that these task categories did not perform very well [ 70 , 89 ]. Medical records related to baseline characteristics and care processes typically contain much more structured data (eg, numerical values for tasks) than medical patient-reported outcomes, which focus more on unstructured data [ 83 , 90 ].…”
Section: Discussionmentioning
confidence: 99%
“…Considering that the model's results can influence health decision-making in some way, the F1 score thresholds may vary depending on the type of class and the imbalance of the data. We reported the results by means of a heatmap, adopting a red color for F1<20%, a gradual color scale from orange to yellow for 21%<F1<79%, and green for F1>80% [68][69][70][71]. Tasks (represented by the lines) were ordered by the average of the performed models, whereas the ordering of the columns shows the rank position of each method according to the statistical analysis.…”
Section: Experimental Analysismentioning
confidence: 99%
“…The authors used a BiLSTM-CRF tagger for NER and adopted a graph-based model to rank concept candidates for each entity mention. Many other recent works have also tackled hybrid approaches ( 59 ) and edit patterns ( 60 ), analyzed the problem of ambiguity ( 61 ), explored transformer networks trained via a triplet objective ( 62 ) and multi-task frameworks ( 63 ) and experimented using large-scale datasets ( 64 ).…”
Section: Related Workmentioning
confidence: 99%