2021
DOI: 10.1186/s13244-021-01018-1
|View full text |Cite
|
Sign up to set email alerts
|

T-staging pulmonary oncology from radiological reports using natural language processing: translating into a multi-language setting

Abstract: Background In the era of datafication, it is important that medical data are accurate and structured for multiple applications. Especially data for oncological staging need to be accurate to stage and treat a patient, as well as population-level surveillance and outcome assessment. To support data extraction from free-text radiological reports, Dutch natural language processing (NLP) algorithm was built to quantify T-stage of pulmonary tumors according to the tumor node metastasis (TNM) classif… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 12 publications
(11 citation statements)
references
References 19 publications
0
11
0
Order By: Relevance
“…The pre-trained language model (PLM) is driven by a large amount of corpus and can use these data to realize the semantic representation of knowledge contained in a large amount of text to realize downstream tasks. The downstream tasks include natural language processing tasks such as classification (Li et al, 2019b ; Maltoudoglou et al, 2022 ; Ni et al, 2020a , 2020b ), sequence labeling (Dai et al, 2019 ; Li et al, 2020b ), summarization (Chintagunta et al, 2021 ; Lacson et al, 2006 ; Yuan et al, 2021 ), translation (Névéol et al, 2018 ; Nobel et al, 2021 ; Wang et al, 2019 ), generation (Melamud & Shivade, 2019 ; Peng et al, 2019 ; Xiong et al, 2019 ), etc. As one of the new downstream tasks, the translation task, Zhu et al ( 2020 ) previously found that using the pre-trained language model as contextual embedding instead of direct fine-tuning will produce better results.…”
Section: Related Workmentioning
confidence: 99%
“…The pre-trained language model (PLM) is driven by a large amount of corpus and can use these data to realize the semantic representation of knowledge contained in a large amount of text to realize downstream tasks. The downstream tasks include natural language processing tasks such as classification (Li et al, 2019b ; Maltoudoglou et al, 2022 ; Ni et al, 2020a , 2020b ), sequence labeling (Dai et al, 2019 ; Li et al, 2020b ), summarization (Chintagunta et al, 2021 ; Lacson et al, 2006 ; Yuan et al, 2021 ), translation (Névéol et al, 2018 ; Nobel et al, 2021 ; Wang et al, 2019 ), generation (Melamud & Shivade, 2019 ; Peng et al, 2019 ; Xiong et al, 2019 ), etc. As one of the new downstream tasks, the translation task, Zhu et al ( 2020 ) previously found that using the pre-trained language model as contextual embedding instead of direct fine-tuning will produce better results.…”
Section: Related Workmentioning
confidence: 99%
“…For both the training and the validation sets, the substage accuracy scores were calculated separately for the T-stage and the N-stage. T-substage is a subdivision of the T-stage to provide more detail, for example, stage T1 (≤3 cm) contains substage T1c (2 to ≤3 cm) [ 1 , 18 ]. Next to the T- and N-stage, the combined accuracy score (TN-stage) was scored for the training and validation sets.…”
Section: Methodsmentioning
confidence: 99%
“…NLP has also been used in a recent and ongoing transnational project to extract the stage in pulmonary oncology from free-text radiological chest CT scan reports [ 17 , 18 ]. The overall goal is to build a language-independent algorithm that can extract pulmonary oncology staging according to the TNM classification.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…We can observe that, in 2021, researchers mainly concentrated on studying English-language data. Indeed, compared to previous years, a fewer number of languages were covered: Chinese [3][4][5][6][7][8][9][10], Dutch [11], French [12,13], Italian [14][15][16], Japanese [17], Korean [18,19], Norwegian [20], and Spanish . Besides, except for Chinese, there were also very few works done for the languages represented in publications.…”
Section: Languages Addressedmentioning
confidence: 99%