Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.667
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Calibration of Knowledge Graph Embeddings for Trustworthy Link Prediction

Abstract: Little is known about the trustworthiness of predictions made by knowledge graph embedding (KGE) models. In this paper we take initial steps toward this direction by investigating the calibration of KGE models, or the extent to which they output confidence scores that reflect the expected correctness of predicted knowledge graph triples. We first conduct an evaluation under the standard closed-world assumption (CWA), in which predicted triples not already in the knowledge graph are considered false, and show t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(13 citation statements)
references
References 21 publications
0
12
0
Order By: Relevance
“…In future work, we are seeking to improve UEs quality obtained using the DPP dropout with the help of calibration (Safavi et al, 2020)…”
Section: Discussionmentioning
confidence: 99%
“…In future work, we are seeking to improve UEs quality obtained using the DPP dropout with the help of calibration (Safavi et al, 2020)…”
Section: Discussionmentioning
confidence: 99%
“…In fact, it is often good practise to use so-called "hard" negative samples, which are similar to entities in T . A better alternative for finding entities not in T would be using more advanced techniques as proposed in [16].…”
Section: Closed-world Assumptionmentioning
confidence: 99%
“…We use Recall (R@k), the fraction of known missing/future links that are in the size-k set returned by the method, and Precision (P@k), the fraction of the k pairs that are known to be missing/future links. Recall is a more important metric, since (1) the returned set of pairs P does not contain final predictions, but rather pairs for a LP method to make final decisions about, and (2) our real-world graphs are inherently incomplete, and thus pairs returned that are not known to be missing links, could nonetheless be missing in the original dataset prior to ground-truth removal (i.e., the openworld assumption [25]). We report both in Table III.…”
Section: B Recall and Precision (Rq1)mentioning
confidence: 99%
“…Link prediction is a long-studied problem that attempts to predict either missing links in an incomplete graph, or links that are likely to form in the future. This has applications in discovering unknown protein interactions to speed up the discovery of new drugs, friend recommendation in social networks, knowledge graph completion, and more [1], [15], [16], [25]. Techniques range from heuristics, such as predicting links based on the number of common neighbors between a pair of nodes, to machine learning techniques, which formulate the link prediction problem as a binary classification problem over node pairs [7], [29].…”
Section: Introductionmentioning
confidence: 99%