2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM) 2019
DOI: 10.1109/bibm47256.2019.8983212
|View full text |Cite
|
Sign up to set email alerts
|

Distantly Supervised Biomedical Named Entity Recognition with Dictionary Expansion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
1
1

Relationship

4
4

Authors

Journals

citations
Cited by 19 publications
(11 citation statements)
references
References 27 publications
0
11
0
Order By: Relevance
“…To answer these questions, we apply our fine-grained entity extraction system CORD-NER (Wang et al, 2020c) to extract 75 types of entities to enrich the KG, including many COVID-19 specific new entity types (e.g., coronaviruses, viral proteins, evolution, materials, substrates, and immune responses). CORD-NER relies on distantly-and weakly-supervised methods (Wang et al, 2019b;, with no need for expensive human annotation. Its entity annotation quality surpasses SciSpacy (up to 93.95% F-score, over 10% higher on the F1 score based on a sample set of documents), a fully supervised BioNER tool.…”
Section: Fine-grained Text Entity Extractionmentioning
confidence: 99%
“…To answer these questions, we apply our fine-grained entity extraction system CORD-NER (Wang et al, 2020c) to extract 75 types of entities to enrich the KG, including many COVID-19 specific new entity types (e.g., coronaviruses, viral proteins, evolution, materials, substrates, and immune responses). CORD-NER relies on distantly-and weakly-supervised methods (Wang et al, 2019b;, with no need for expensive human annotation. Its entity annotation quality surpasses SciSpacy (up to 93.95% F-score, over 10% higher on the F1 score based on a sample set of documents), a fully supervised BioNER tool.…”
Section: Fine-grained Text Entity Extractionmentioning
confidence: 99%
“…Aiming to reduce expensive manual annotation, distant supervision has been used to generate training labels automatically by utilizing the entity information from existing KBs. The major research efforts lie in dealing with the incomplete annotation problem caused by an incomplete coverage of the KBs (Fries et al, 2017;Shang et al, 2018b;Peng et al, 2019;Wang et al, 2019aWang et al, , 2020aLiang et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…The open information extraction pipeline extracts entities with distant supervision from knowledge bases and relations with automatic meta-pattern discovery methods. In particular, to extract highquality entities and relations, we design noiserobust neural models for distantly supervised named entity recognition (Shang et al, 2018b;Wang et al, 2019) and wide-window meta-pattern discovery methods to deal with the long and complex sentences in biomedical literature (Wang et al, 2018a;. Data Collection.…”
Section: Open Information Extractionmentioning
confidence: 99%
“…To tackle the problem of limited coverage of the input dictionary, we first apply a data-driven phrase mining algorithm, AutoPhrase (Shang et al, 2018a), to extract high-quality phrases as additional entity candidates. Then we automatically expand the dictionary with a novel dictionary expansion method (Wang et al, 2019). The expanded dictionary is used to label the input corpora with the 17 finegrained entity types to train a neural model.…”
Section: Open Information Extractionmentioning
confidence: 99%
See 1 more Smart Citation