Proceedings of the 2017 ACM on Conference on Information and Knowledge Management 2017
DOI: 10.1145/3132847.3132873
|View full text |Cite
|
Sign up to set email alerts
|

Name Disambiguation in Anonymized Graphs using Network Embedding

Abstract: In real-world, our DNA is unique but many people share names.is phenomenon o en causes erroneous aggregation of documents of multiple persons who are namesake of one another. Such mistakes deteriorate the performance of document retrieval, web search, and more seriously, cause improper a ribution of credit or blame in digital forensic. To resolve this issue, the name disambiguation task is designed which aims to partition the documents associated with a name reference such that each partition contains document… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
96
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 104 publications
(97 citation statements)
references
References 25 publications
0
96
0
1
Order By: Relevance
“…Since there is no implementation details, we speculate that it might be because the merging rules are too loose. By incorporating both rules and neural networks to model co-authorships, affiliations and titles explicitly, our PNP model outperforms all baselines in terms of F1score (+3.54% over Zhang et al [25], +11.75% over Zhang and Al Hasan [24], +11.23% over Louppe et al [13] and +39.74% over Fan et al [3] relatively). In the bottom half part of Table 1, some incremental results of our method are presented.…”
Section: Comparison Methodsmentioning
confidence: 98%
See 1 more Smart Citation
“…Since there is no implementation details, we speculate that it might be because the merging rules are too loose. By incorporating both rules and neural networks to model co-authorships, affiliations and titles explicitly, our PNP model outperforms all baselines in terms of F1score (+3.54% over Zhang et al [25], +11.75% over Zhang and Al Hasan [24], +11.23% over Louppe et al [13] and +39.74% over Fan et al [3] relatively). In the bottom half part of Table 1, some incremental results of our method are presented.…”
Section: Comparison Methodsmentioning
confidence: 98%
“…-Louppe et al [13]: It trains a pairwise distance function based on carefully designed similarity features, and uses semi-supervised Hierarchical Agglomerative Clustering (HAC) algorithm to determine clusters. -Zhang and Al Hasan [24]: It constructs graphs for each author name based on co-author and document similarity. Embeddings are learned for each name and the final results are also obtained by HAC.…”
Section: Comparison Methodsmentioning
confidence: 99%
“…With the development of unsupervised feature learning techniques [3], deep learning methods proved successful in natural language processing tasks through neural language models [10,33,36]. These models have been used to capture the semantic and syntactic structures of human language [8], and even logical analogies [20], by embedding words as vectors.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, as the node embedding representations are often learned in a task-agnostic fashion, they are generalizable to a number of downstream learning tasks such as node classification [33], community detection [44], link prediction [15], and visualization [37]. On top of that, it also has broader impacts in advancing many real-world applications, ranging from recommendation [45], polypharmacy side effects prediction [53] to name disambiguation [49]. The basic idea of network embedding is to represent each node by a lowdimensional vector in which the relativity information among nodes in the original network is maximally transcribed.…”
Section: Introductionmentioning
confidence: 99%