2015
DOI: 10.1016/j.joi.2015.08.004
|View full text |Cite
|
Sign up to set email alerts
|

Exploring author name disambiguation on PubMed-scale

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
64
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 37 publications
(64 citation statements)
references
References 16 publications
0
64
0
Order By: Relevance
“…Coauthorship has been reported to be the most effective (Liu et al, ; Onodera et al, ; Torvik, Weeber, Swanson, & Smalheiser, ; Wang et al, ). Conversely, affiliation information has been found to be more important than coauthorship (Song et al, ; Wu & Ding, ).…”
Section: Related Workmentioning
confidence: 99%
“…Coauthorship has been reported to be the most effective (Liu et al, ; Onodera et al, ; Torvik, Weeber, Swanson, & Smalheiser, ; Wang et al, ). Conversely, affiliation information has been found to be more important than coauthorship (Song et al, ; Wu & Ding, ).…”
Section: Related Workmentioning
confidence: 99%
“…Journal of Data and Information Science A Multi-match Approach to the Author Uncertainty Problem http://www.jdis.org https://www.degruyter.com/view/j/jdis user is left with a much smaller (and more manageable) dataset to manually inspect (and/or apply additional disambiguation techniques to). Other studies (e.g., Song et al, 2015;Amancio, 2015) advocate using a combination of approaches, and doing so in conjunction with the procedure advanced in this paper is expected to produce results of both high precision and recall.…”
Section: Research Papermentioning
confidence: 96%
“…This model can only be used on MEDLINE, not on other bibliographic databases. Tran et al in [20] suggested a deep neural network-based approach to automatically learn the weights of the features and disambiguate the authors. Determining optimal number of hidden layers, data representations for the first layer, and number of units is a complex task and requires skill and experience.…”
Section: Related Workmentioning
confidence: 99%
“…These methods have improved the situation somewhat, but there is still a need for improvement in current solutions. Supervised methods [2,5,6,[20][21]25] require much clean and representative data for the training of the model and give poor results when a model is trained on noisy data or nonrepresentative data [1,3]. Most unsupervised AND methods assume that a number of ambiguous authors/clusters "K" are known in advance [4,7,10,12].…”
Section: Introductionmentioning
confidence: 99%