2019
DOI: 10.1007/978-3-030-22734-0_29
|View full text |Cite
|
Sign up to set email alerts
|

Creating Training Data for Scientific Named Entity Recognition with Minimal Human Effort

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 31 publications
0
7
0
Order By: Relevance
“…For the social science dataset we apply only FastText word embeddings and lexicon features, as the previous experiment demonstrated that this configuration performed best of the configurations studied. Table 3 compares the precision, recall, and F1 scores of SciNER to our previous work, in which we used a KNN classifier and many manually created rules [29]. Even in this quite different environment, SciNER achieves an F1 score of 0.847, significantly outperforming our rule-based approach that achieves an F1 score of 0.592.…”
Section: Experiments On the Social Science Corpusmentioning
confidence: 99%
See 1 more Smart Citation
“…For the social science dataset we apply only FastText word embeddings and lexicon features, as the previous experiment demonstrated that this configuration performed best of the configurations studied. Table 3 compares the precision, recall, and F1 scores of SciNER to our previous work, in which we used a KNN classifier and many manually created rules [29]. Even in this quite different environment, SciNER achieves an F1 score of 0.847, significantly outperforming our rule-based approach that achieves an F1 score of 0.592.…”
Section: Experiments On the Social Science Corpusmentioning
confidence: 99%
“…Each paper was reviewed by two expert reviewers to label polymer names, and when disagreement arose a third more senior domain expert made the final decision. The result is a list of 495 polymer names identified by experts in the 100 papers [29].…”
Section: Data Preparationmentioning
confidence: 99%
“…The early applications of chemical NER were mainly focused on extraction of drugs and biochemical information to perform more effective document searches ( Corbett and Copestake, 2008 ; Jessop et al., 2011 ; Rocktäschel et al., 2012 ; García-Remesal et al, 2013 ). Recently, chemical NER has shifted toward (in)organic materials and their characteristics ( Swain and Cole, 2016 ; He et al., 2020 ; Weston et al., 2019 ; Shah et al., 2018 ), polymers ( Tchoua et al., 2019 ), nanoparticles ( Hiszpanski et al., 2020 ), synthesis actions and conditions ( Vaucher et al., 2020 ; Hawizy et al., 2011 ; Kim et al., 2017c ; Kononova et al., 2019 ). The methods used for NER vary from traditional rule-based and dictionary look-up approaches to modern methodology built around advanced ML and NLP techniques, including conditional random field (CRF) ( Lafferty et al., 2001 ), long short-term memory (LSTM) neural networks ( Hochreiter and Schmidhuber, 1997 ), and others.…”
Section: Text Mining Of Scientific Literaturementioning
confidence: 99%
“…Therefore, the recognition of polymer names is not easy. For such variations, machine learning techniques are effective and have been used for this type of recognition [15,16]. However, because manual preparation of the training data requires high cost, remarkable results have not yet been obtained.…”
Section: Introductionmentioning
confidence: 99%