2020
DOI: 10.48550/arxiv.2005.00574
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Clinical Reading Comprehension: A Thorough Analysis of the emrQA Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
6
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(6 citation statements)
references
References 35 publications
0
6
0
Order By: Relevance
“…emrQA falls into the category of extractive question answering, aiming to identify answer spans from reference texts instead of generating new answers in a word-by-word fashion. Researchers have attempted to solve emrQA tasks by using word embedding models [28], conditional random fields (CRFs) [29] and transformer-based models [30], among which transformer-based models performed best. In our experiments, we investigate the performance of our pre-trained models using the three largest emrQA subsets: Medication, Relation, and Heart Disease.…”
Section: Question Answeringmentioning
confidence: 99%
See 2 more Smart Citations
“…emrQA falls into the category of extractive question answering, aiming to identify answer spans from reference texts instead of generating new answers in a word-by-word fashion. Researchers have attempted to solve emrQA tasks by using word embedding models [28], conditional random fields (CRFs) [29] and transformer-based models [30], among which transformer-based models performed best. In our experiments, we investigate the performance of our pre-trained models using the three largest emrQA subsets: Medication, Relation, and Heart Disease.…”
Section: Question Answeringmentioning
confidence: 99%
“…F1score is a looser metric derived from token-level precision and recall, which measures the overlap between the predictions and the targets. We generate train-dev-test splits by following the instruction of Yue et al [28]. The training set of relation and medication subsets are randomly under-sampled to reduce training time.…”
Section: Question Answeringmentioning
confidence: 99%
See 1 more Smart Citation
“…emrQA falls into the category of extractive question answering that aims at identifying the answer spans from reference contexts instead of generating answers in a word-by-word fashion. Researchers have attempted to solve emrQA tasks by using word embedding models [25], conditional random fields (CRFs) [26] and transformer-based models [27], among which transformer-based models defeated their competitors in terms of performance. In our experiments, we investigated the performance of our pre-trained models using the three largest emrQA subsets: Medication, Relation and Heart Disease.…”
Section: Question Answeringmentioning
confidence: 99%
“…F1-score is a looser metric derived from token-level precision and recall, which aims at measuring the overlap between the predictions and the targets. We generated train-dev-test splits by following the instruction of [25], where the training set of relation and medication subsets were randomly under-sampled to reduce training time. Based on their experience, the performance would not be compromised after under-sampling.…”
Section: Question Answeringmentioning
confidence: 99%