Vol-3194/paper23
Jump to navigation
Jump to search
Paper
Paper | |
---|---|
edit | |
description | |
id | Vol-3194/paper23 |
wikidataid | Q117344884→Q117344884 |
title | Explaining Link Prediction with Kelpie |
pdfUrl | https://ceur-ws.org/Vol-3194/paper23.pdf |
dblpUrl | https://dblp.org/rec/conf/sebd/0002FMT22 |
volume | Vol-3194→Vol-3194 |
session | → |
Explaining Link Prediction with Kelpie
Explaining Link Prediction with Kelpie Andrea Rossi1 , Donatella Firmani2 , Paolo Merialdo1 and Tommaso Teofili1 1 Roma Tre University, Rome, Italy 2 Sapienza University, Rome, Italy Abstract Link Prediction (LP) is the problem of inferring new facts in a Knowledge Graph from the already known ones. Recent advances in Machine Learning have led researchers to develop LP models that represent Knowledge Graph elements as vectors in an embedding space. This approach has led to greatly encouraging results, often outperforming traditional approaches; its main shortcoming is its opaqueness, hindering both our understanding and our trust in these models. In this context, we discuss the Kelpie explainability framework. Kelpie can be applied to any embedding-based LP model and can identify the combinations of training facts that have enabled the prediction of a given link. Kelpie can extract two complementary types of explanations, that we dub necessary and sufficient. Keywords Knowledge Graph Embeddings, Link Prediction, Explainable AI 1. Introduction Knowledge Graphs (KGs) provide a natural way to model structured real-world information. The nodes of a KG represent real-world entities, and they are connected by directed edges denoted by labels conveying semantic relations. Each edge connects two entities, forming a triple that is called a fact. In time, several KGs have achieved web-scale size, e.g., DBPedia, Yago, Wikidata, and, in industry, the Google KG; nonetheless, all KGs suffer from incompleteness, as they miss large portions of the information they should contain. For instance, it has been observed that, in the FreeBase KG, over 70% of the person entities have unknown birthplace, and over 99% have unknown ethnicity [1]. Link Prediction (LP) aims at inferring new, missing facts by analyzing those already present in the KG. For instance, in the graph in Figure 1 (left), knowing ⟨Barack_Obama, born_in, Honolulu⟩ and ⟨Honolulu, located_in, USA⟩, we can probably infer ⟨Barack_Obama, nationality, USA⟩. In the last decade, LP researchers have achieved extremely promising results with the use of KG embeddings, i.e., numerical vectors mapped to each element in the KG, and learned with Machine Learning (ML) techniques leveraging the known facts. KG Embeddings have rapidly become the dominant approach to LP in literature, with dozens of new models proposed every year (see [2] for a survey). Like most ML systems, LP models based on KG Embeddings are inherently opaque, as they do not provide any insights on the reasons why they predict specific links. We argue that interpretability in LP systems is vital. Explanations reveal which patterns our models SEBD 2022: The 30th Italian Symposium on Advanced Database Systems, June 19-22, 2022, Tirrenia (PI), Italy $ andrea.rossi3@uniroma3.it (A. Rossi); donatella.firmani@uniroma1.it (D. Firmani); paolo.merialdo@uniroma3.it (P. Merialdo); tommaso.teofili@uniroma3.it (T. Teofili) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) � supported supported supported Barack Obama Bill Gates Homologous Bill Gates Non-Homologous Bill Gates Mimic Mimic born_in born_in born_in born_in nationality nationality born_in nationality born_in Honolulu Seattle Honolulu Seattle Honolulu Seattle president_of president_of president_of located_in located_in located_in located_in located_in located_in Figure 1: A small KG example (left), together with an example of homologous (center) and non- homologous (right) mimic. The missing fact 〈Barack_Obama, nationality, USA〉 is denoted in red. are leveraging to perform their predictions: in the LP field, this would provide a pathway to gain a deeper understanding of our models, to identify biases or errors in our KGs, and most importantly to decide whether our models can be trusted or not. In this paper we discuss the Kelpie framework for explaining embedding-based link predic- tions, that we originally present in [3]. Accordingly to the taxonomy for explainable AI methods in [4], Kelpie is a local post-hoc interpretability method; it can be applied to any LP model based on embeddings, independently of its architecture. Given a prediction, Kelpie explains it by identifying the subset of training facts that have enabled it. Given any prediction to explain, Kelpie can interpret it in two scenarios, dubbed necessary and sufficient. A necessary explanation is the minimal set of facts in absence of which the model becomes unable to yield the prediction: in doing so, necessary explanations can suggest why a specific entity has been predicted in a certain way. A sufficient explanation is the minimal set of facts that, if added to other entities, lead the model to yield the same prediction for those entities too: sufficient explanations can thus suggest the rules that would extend the same prediction to any entity. As observed in [5], necessity and sufficiency are the building blocks of any successful explanation. The Kelpie code and resources are publicly available. 1 2. Problem definition A Knowledge Graph (KG) is a labeled directed graph 𝐾𝐺 = (ℰ, ℛ, 𝒢) where ℰ is the set of nodes, or entities, in the graph; ℛ is its set of labels, or relations; and 𝒢 ⊆ ℰ × ℛ × ℰ is its set of edges linking entities via relations, i.e., its set of facts. In any fact 〈h, r, t〉 the first entity h is the head, r is the relation, and the second entity t is the tail. Models for embedding-based Link Prediction (LP) typically define a scoring function 𝜑 that estimates the plausibility of any fact using the embeddings of its head, relation and tail. In training, models learn embeddings to optimize the plausibility of the known facts: in prediction, new links are identified by finding which entities, if added to incomplete triples as heads or tails, yield the best scores. A tail prediction ⟨ℎ, 𝑟, 𝑡⟩ is the process that finds 𝑡 to be the best scoring tail for the incomplete triple ⟨ℎ, 𝑟, ?⟩, i.e., 𝑡 = argmax𝑒∈ℰ 𝜑(ℎ, 𝑟, 𝑒). Head predictions are defined symmetrically. In the following we focus on tail predictions; all our formulations 1 https://github.com/AndRossi/Kelpie �can be adapted for head predictions analogously. LP literature usually relies on datasets sampled from real-world KGs, and in which the set of extracted facts 𝒢 is split into a training set 𝒢𝑡𝑟𝑎𝑖𝑛 , a validation set 𝒢𝑣𝑎𝑙𝑖𝑑 , and a test set 𝒢𝑡𝑒𝑠𝑡 . In evaluation, head and tail predictions are performed on each fact in 𝒢𝑡𝑒𝑠𝑡 ; in each prediction, the target entity, i.e, the expected answer, is ranked against all the others in ℰ. The obtained ranks are then aggregated into standard global metrics: Hits@K (H@K), that measures the fraction of ranks 𝑇 with value ≤ 𝑘, and Mean Reciprocal Rank (MRR), that averages the inverse of all the obtained ranks 𝑇 . Both H@K and MRR are always between 0 and 1; higher values convey better results. Discussions on the pros and cons of such methodologies can be found in [6, 7, 8]. For some predictions there can exist more than one correct answer (e.g., in case of 1-to-𝑛 relations). In this event we follow the common practice called filtered setting of excluding such alternative answers when computing the rank of the target entity. For a given prediction Kelpie defines two explanation types: necessary and sufficient. We indicate with 𝑋 the generic candidate explanation (either necessary or sufficient) and with 𝑋 * the explanation we want to identify and return. We use the subscripts 𝑛 (e.g., 𝑋𝑛 ) and 𝑠 to denote necessary and sufficient explanations respectively. We give definitions of 𝑋𝑛* and 𝑋𝑠* w.r.t. tail predictions 〈ℎ, 𝑟, 𝑡〉, as head predictions can be handled analogously. ∙ 𝑋𝑛* is the smallest set of training facts featuring ℎ such that, removing 𝑋𝑛* from 𝒢𝑡𝑟𝑎𝑖𝑛 and retraining the model from scratch, the top ranking tail for ⟨ℎ, 𝑟, ?⟩ changes to any 𝑒 ̸= 𝑡. ∙ 𝑋𝑠* is the smallest set of training facts featuring ℎ such that, given a set of random entities 𝐶 ⊆ ℰ such that 𝑡 is not the top ranking tail for any ⟨𝑐, 𝑟, ?⟩ prediction, if we add the facts in 𝑋𝑠* to any 𝑐 ∈ 𝐶 and retrain the model from scratch, the top ranking tail for ⟨𝑐, 𝑟, ?⟩ changes to 𝑡. When this happens, we say that the 𝑐 entities have been converted. For instance, in the previous example of the tail prediction ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑛𝑎𝑡𝑖𝑜𝑛𝑎𝑙𝑖𝑡𝑦, 𝑈 𝑆𝐴⟩, the set {⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑏𝑜𝑟𝑛_𝑖𝑛, 𝐻𝑜𝑛𝑜𝑙𝑢𝑙𝑢⟩, ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑝𝑟𝑒𝑠𝑖𝑑𝑒𝑛𝑡, 𝑈 𝑆𝐴⟩} is a necessary explanation if removing these facts from Barack_Obama the pre- dicted 𝑛𝑎𝑡𝑖𝑜𝑛𝑎𝑙𝑖𝑡𝑦 changes from 𝑈 𝑆𝐴 to any other entity. Analogously, the single-sized set {⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑝𝑟𝑒𝑠𝑖𝑑𝑒𝑛𝑡_𝑜𝑓, 𝑈 𝑆𝐴⟩} constitutes a sufficient explanation if adding it to any non-American entities, e.g., Édith_Piaf, changes their predicted nationality to USA. 3. System Overview The high-level architecture of Kelpie is shown in Figure 2. We briefly describe the Kelpie modules w.r.t. tail predictions ⟨ℎ, 𝑟, 𝑡⟩; head predictions can be handled analogously. Pre-filter. This module analyzes the set of training facts mentioning ℎ, that we call 𝒢𝑡𝑟𝑎𝑖𝑛ℎ , in order identify and discard the least promising ones: its overall goal is to reduce the search space for the following steps, preventing combinatorial explosion when ℎ is featured in too many facts. The Pre-Filter achieves this goal by computing for each fact in 𝒢𝑡𝑟𝑎𝑖𝑛 ℎ a promisingness value based on the graph topology, and by returning the subset ℱ𝑡𝑟𝑎𝑖𝑛 of the top promising facts. Intuitively, ℎ topologically closer entities often bear a stronger semantic relationship: so, for any ⟨ℎ, 𝑠, 𝑞⟩ (or ⟨𝑞, 𝑠, ℎ⟩) connecting ℎ to an entity 𝑞 close to 𝑡 we compute promisingness as the length of the shortest non-oriented path connecting 𝑞 to 𝑡; lower values convey better promisingness � & 𝒢!"#$% & ℱ!"#$% 𝑋∗ Pre-Filter Explanation Builder 𝑟𝑒𝑙𝑒𝑣𝑎𝑛𝑐𝑒 𝑋 Relevance Engine Figure 2: High-level Kelpie architecture, highlighting the interactions among its three modules. (and thus higher priority in the filtering). Consider again the example in Figure 1 (left) and suppose we want to explain the tail prediction ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑛𝑎𝑡𝑖𝑜𝑛𝑎𝑙𝑖𝑡𝑦, 𝑈 𝑆𝐴⟩: ∙ ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑝𝑟𝑒𝑠𝑖𝑑𝑒𝑛𝑡_𝑜𝑓, 𝑈 𝑆𝐴⟩ has promisingness 0 (the best possible value). ∙ ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑏𝑜𝑟𝑛_𝑖𝑛, 𝐻𝑜𝑛𝑜𝑙𝑢𝑙𝑢⟩ has promisingness 1. ∙ ⟨𝐵𝑖𝑙𝑙_𝐺𝑎𝑡𝑒𝑠, 𝑠𝑢𝑝𝑝𝑜𝑟𝑡𝑒𝑑, 𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎⟩ has promisingness 2, as the shortest path from 𝐵𝑖𝑙𝑙_𝐺𝑎𝑡𝑒𝑠 to 𝑈 𝑆𝐴 has length 2: [〈Bill_Gates, born_in, Seattle〉, 〈Seattle, located_in, USA〉]. Explanation Builder. This module explores the space of the candidate explanations 𝑋 that can be obtained by combining the Pre-Filtered facts, with the goal of identifying the smallest 𝑋 that is relevant enough for the prediction to explain. The notion of relevance of an explanation embodies the likelihood of yielding changes in predictions, and it is quantified by the Relevance Engine described in the next paragraph. Suppose that in the example in Figure 1 (left) only the facts 𝑓1 = ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑝𝑟𝑒𝑠𝑖𝑑𝑒𝑛𝑡_𝑜𝑓, 𝑈 𝑆𝐴⟩ and 𝑓2 = ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑏𝑜𝑟𝑛_𝑖𝑛, 𝐻𝑜𝑛𝑜𝑙𝑢𝑙𝑢⟩ survive the Pre-Filtering. The Explanation Builder starts by exploring short explanations, such as, 𝑋1 = {𝑓1 } and 𝑋2 = {𝑓2 }; then, it iteratively builds longer ones, such as 𝑋3 = {𝑓1 , 𝑓2 }, until either 𝑋 * is found with relevance higher than a user-specified threshold, or early stopping policies are enacted. In the latter case, the best explanation seen so far is returned. Among same-sized explanations (e.g., 𝑋1 and 𝑋2 ), a prioritization mechanism based on an off-line heuristic is used. For details about early stopping and exploration priority, we refer the reader to Algorithm 3 in [3]. Relevance Engine. This module aims at estimating how adding or removing training facts from ℎ would affect the prediction to explain. Ideally, the relevance of any candidate explanation 𝑋 could be computed by retraining the model after adding or removing its facts from 𝒢𝑡𝑟𝑎𝑖𝑛 ; in practice this is unfeasible, so Kelpie introduces the novel ML process of Post-Training (PT). PT consists in using an already trained model to obtain an alternate version, i.e., a mimic, of an entity 𝑒 and of the corresponding embedding. Given 𝑒, PT amounts to: (i) replicating the training facts mentioning 𝑒, 𝒢𝑡𝑟𝑎𝑖𝑛 𝑒 , and potentially injecting additions/removals; (ii) training a new embedding to optimize the plausibility of those facts, while keeping all the other embeddings and parameters in the model frozen. Compared to a full training, PT is a lightweight process as it optimizes only one embedding and uses a training set that is several orders of magnitude smaller than the entire 𝐺𝑡𝑟𝑎𝑖𝑛 . For any entity 𝑒 Kelpie can post-train two types of mimics: ∙ A homologous mimic 𝑒′ has its embedding initialized randomly, and then post-trained on an exact replica of 𝒢𝑡𝑟𝑎𝑖𝑛 𝑒 . As a result, we expect 𝑒′ to behave similarly to the original entity 𝑒. ∙ A non-homologous mimic 𝑒′−𝑀 (or 𝑒′+𝑀 ) initialized randomly as well, and then post-trained � Necessary Explanations Effectiveness Sufficient Explanations Effectiveness FB15k WN18 FB15k-237 WN18RR YAGO3-10 FB15k WN18 FB15k-237 WN18RR YAGO3-10 ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR ΔH@1 ΔMRR TransE Kelpie -0.490 -0.321 -0.920 -0.857 -0.540 -0.356 -0.920 -0.904 -0.740 -0.580 +0.319 +0.516 +0.259 +0.434 +0.128 +0.218 +0.117 +0.169 +0.048 +0.109 DP -0.380 -0.258 -0.900 -0.859 -0.460 -0.303 -0.770 -0.701 -0.670 -0.533 +0.273 +0.461 +0.183 +0.350 +0.051 +0.080 +0.082 +0.116 +0.036 +0.117 Kelpie -0.850 -0.695 -0.910 -0.827 -0.590 -0.413 -0.980 -0.913 -0.960 -0.858 +0.945 +0.920 +0.953 +0.962 +0.377 +0.490 +0.834 +0.877 +0.858 +0.892 ComplEx DP -0.540 -0.458 -0.800 -0.742 -0.340 -0.185 -0.750 -0.650 -0.810 -0.714 +0.910 +0.888 +0.931 +0.940 +0.245 +0.334 +0.836 +0.878 +0.835 +0.878 Criage -0.030 -0.020 -0.050 -0.045 -0.090 -0.050 -0.180 -0.150 -0.05 -0.030 +0.068 +0.069 +0.105 +0.137 +0.035 +0.038 +0.110 +0.147 +0.000 +0.000 Kelpie -0.710 -0.516 -0.930 -0.860 -0.430 -0.284 -0.980 -0.914 -0.980 -0.884 +0.677 +0.649 +0.903 +0.900 +0.225 +0.203 +0.827 +0.856 +0.799 +0.848 ConvE DP -0.350 -0.232 -0.790 -0.752 -0.290 -0.195 -0.850 -0.750 -0.880 -0.799 +0.234 +0.199 +0.396 +0.412 +0.202 +0.169 +0.373 +0.419 +0.366 +0.391 Criage -0.080 -0.056 -0.160 -0.146 -0.290 -0.196 -0.170 -0.156 -0.070 -0.042 +0.106 +0.065 +0.164 +0.166 +0.132 +0.094 +0.150 +0.164 +0.019 +0.020 Table 1 Effectiveness of necessary and sufficient explanations on a set of facts that replicates 𝒢𝑡𝑟𝑎𝑖𝑛 𝑒 with the removal (or addition) of a set 𝑀 of facts. After the post-training is over, the mimic is expected to approximate the behaviour that the original entity 𝑒 would have displayed if the injected variation had been present since the beginning. Examples of homologous and non-homologous mimics are in Figure 1 (center and right respec- tively). When explaining any tail prediction ⟨ℎ, 𝑟, 𝑡⟩, we can thus estimate the effect of adding (or removing) from ℎ the facts of any candidate explanation 𝑋 by comparing the outcomes obtained using a homologous mimic ℎ′ and a non-homologous mimic mimic ℎ′+𝑋 (or ℎ′+𝑋 ). 4. Experiments In this section we briefly report the main experimental results of Kelpie. For a deeper report on the experimental evaluation of the system, see [3]. Setup. Our experiments have been run on a server with 88 CPUs Intel Core(TM) i7-3820, 516GB RAM and 4 16GB NVIDIA Tesla P100 GPUs. We consider the 5 best-established dataset in LP literature: FB15k and FB15k-237, sampled from Freebase; WN18 and WN18RR, sampled from WordNet; and YAGO3-10, sampled from YAGO3. We consider 3 models representative for the 3 different families of the taxonomy in [2]: the geometric model TransE [9], the matrix factorization model ComplEx [10], and the deep learning model ConvE [11]. End-to-end performance. We compute the performance of Kelpie as the H@1 and MRR variation caused by applying the extracted explanation to the predictions to disable (in the necessary scenario) or to enable (in the sufficient scenario). More specifically: ∙ In our necessary scenario, for each model and dataset we randomly sample a set of 100 correct test tail predictions ⟨ℎ, 𝑟, 𝑡⟩. We extract their necessary explanations and remove the explanation facts from 𝒢𝑡𝑟𝑎𝑖𝑛 ; we then re-train the model and check how the removals have worsened the ranks of the target tails 𝑡. We measure such worsenings in terms of ∆𝐻@1 and ∆𝑀 𝑅𝑅: more negative values correspond to higher effectiveness. ∙ In our sufficient scenario, we once again sample 100 correct test tail predictions ⟨ℎ, 𝑟, 𝑡⟩ to explain for each model and dataset. We extract their sufficient explanations, each with the � Necessary Explanations Sufficient Explanations FB15k WN18 FB15k-237 WN18RR YAGO3-10 FB15k WN18 FB15k-237 WN18RR YAGO3-10 AVG STD AVG STD AVG STD AVG STD AVG STD AVG STD AVG STD AVG STD AVG STD AVG STD TransE 2.78 1.13 2.17 1.24 2.5 1.15 1.67 0.96 1.99 1.20 1.89 1.03 1.02 0.14 3.43 0.83 1.67 0.79 1.45 0.73 ComplEx 3.40 1.10 3.39 1.00 3.83 0.57 3.16 1.08 2.27 1.31 1.40 0.75 1.00 0.00 2.51 1.20 1.00 0.00 1.04 0.31 ConvE 3.36 0.89 2.91 1.20 2.26 1.24 2.66 1.21 1.73 0.97 1.83 1.23 1.01 0.10 3.09 1.10 1.04 0.24 1.18 0.48 Table 2 Lengths of the necessary and sufficient explanations goal of converting a different set 𝐶 of 10 random entities. Then, for each prediction ⟨ℎ, 𝑟, 𝑡⟩ we add the explanation facts to all 𝑐 entities in the corresponding 𝐶 set, we re-train the model, and we measure how the addition of the explanation facts has improved the tail rank of 𝑡 in each ⟨𝑐, 𝑟, 𝑡⟩ prediction. We measure such improvements in terms of ∆𝐻@1 and ∆𝑀 𝑅𝑅: more positive values correspond to higher effectiveness. We use as baselines two data poisoning systems, namely Criage [12] and the framework by Zhang et al. [13], that we denote as DP. The goal of these systems is to disable the selected ⟨ℎ, 𝑟, 𝑡⟩ predictions by removing/adding individual training facts. In the necessary scenario we compare directly to their removal setting. In the sufficient scenario, that they do not support natively, we adapt their formulations to identify which facts would not only disable ⟨ℎ, 𝑟, 𝑡⟩ but also enable the selected ⟨𝑐, 𝑟, 𝑡⟩ predictions. We do not run Criage on TransE, as the code provided by the authors only supports multiplicative models (e.g., ConvE and ComplEx). Results by Criage, DP and Kelpie are reported in Table 1. Kelpie almost always outperforms baselines, by tackling predictions that others fail to explain. Since Criage and DP are limited to explanations with only 1 fact, we have also experimented with a single-fact version of Kelpie; we have obtained results comparable to the best baselines but almost always worse than "full" Kelpie. This demonstrates the effectiveness of post-training at identifying the most relevant facts, as well as the utility of supporting fact combinations. Explanations Lengths. We report in Table 2, for each scenario, model and dataset, the average (AVG) and the standard deviation (STD) of the lengths of our explanations. We observe that necessary explanations tend to always be longer than sufficient ones, for the same datasets and models. Necessary explanations should encompass all the facts supporting a prediction, so that removing them disables the prediction. Sufficient explanations, on the other hand, just need to to extend the same prediction to other entities: so, it is usually enough for them to include just a few facts (or even just one) as pieces of evidence for the prediction. We also observe that, in many cases, the average explanation lengths are lesser than 2, suggesting that Kelpie can identify minimal explanations, as expected (extensive experiments on minimality are in [3]). Explanations and Bias. LP datasets have been found to suffer from various forms of data bias [14]; we now report a brief example showing how Kelpie explanations can unveil previously unknown instances of bias in the training data. We have observed that the correctly predicted YAGO3-10 test facts that convey that a person born_in a city are generally explained by that person being a soccer player playing for a football team from the same city. This is surprising: in the real world, being member of a football team does not imply having been born in its city. � Guided by our explanations, we have found that, in YAGO3-10, people playing for a football team are born in same city unnaturally often (thus providing data bias) and that personal data are significantly scarce, making birthplace prediction very challenging: the correct predictions, in this regard, seem to be affected even by the slight preference that football players may have towards teams from their birthplace. This type of observations can allow researchers to intervene on LP datasets to make them better adhere to the semantics of the real world. 5. Related Works So far few works have addressed directly the interpretability of embedding-based LP models. The works most related to ours are Criage [12] and the framework by Zhang et al. [13]. Both of them follow a data poisoning approach: given a prediction ⟨ℎ, 𝑟, 𝑡⟩, their goal is to find which is the individual fact that, if added to or removed from 𝒢𝑡𝑟𝑎𝑖𝑛 , worsens the score 𝜑(ℎ, 𝑟, 𝑡) the most. Criage [12] uses Influence Functions [15] to estimate how the addition or removal of any fact would affect 𝜑(ℎ, 𝑟, 𝑡); unfortunately, the adaptability of their formulation to the scoring functions of non-multiplicative models is unclear. The framework by Zhang et al. [13] uses gradient analysis to perturb the embedding of ℎ in the direction that worsens 𝜑(ℎ, 𝑟, 𝑡) the most. Other than having empirical differences with Kelpie (as illustrated in Section 4) these methods have an inherently different goal, as they do not aim directly at explaining predictions, but rather they investigate how models withstand adversarial modifications. One of the aspects that make the interpretability of LP models so challenging is the large variety of ML architectures. In time, researchers have tried to bypass this problem in a variety of ways. Some authors have chosen to support specific architectures [16], while other have proposed inherently interpretable LP models [17]. Finally, a few frameworks explain predictions by focusing on the dataset topology more than on the model and its embeddings [18]. We refer the reader to [3] for further discussion about the application of general purpose explanation methods like LIME [19], SHAP [20] and ANCHOR [21] to the LP setting, and their shortcomings. 6. Conclusions We have discussed Kelpie, a full-fledged explainability framework for embedding-based LP models. Kelpie explanations highlight the most relevant training samples that have enabled our predictions; they are based on the fundamental notions of necessity and sufficiency, and they can encompass combinations of multiple training samples. In the sparkling topic of LP on Knowledge Graphs, interpretability is a strikingly desirable property; yet, it is hardly ever guaranteed by current state-of-the-art systems and frameworks. The effectiveness of Kelpie explanations, measured through extensive experiments, shows that Kelpie largely surpasses pre-existing methods across almost all scenarios in literature; furthermore, as demonstrated, Kelpie explanations can be precious in identifying unbalances and biases in LP datasets. Acknowledgments The work by Donatella Firmani has been supported in part by SEED PNR FLOWER grant 2021. �References [1] R. West, E. Gabrilovich, K. Murphy, S. Sun, R. Gupta, D. Lin, Knowledge base completion via search-based question answering, in: WWW, 2014. [2] A. Rossi, D. Barbosa, D. Firmani, A. Matinata, P. Merialdo, Knowledge graph embedding for link prediction: A comparative analysis, TKDD (2021). [3] A. Rossi, D. Firmani, P. Merialdo, T. Teofili, Explaining link prediction systems based on knowledge graph embeddings, in: SIGMOD, 2022. [4] R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, D. Pedreschi, A survey of methods for explaining black box models, ACM Comput. Surv. (2018). [5] D. S. Watson, L. Gultchin, A. Taly, L. Floridi, Local explanations via necessity and suffi- ciency: Unifying theory and practice, in: UAI, 2021. [6] A. Rossi, A. Matinata, Knowledge Graph Embeddings: Are Relation-Learning Models Learning Relations?, in: PIE, 2020. [7] Y. Wang, D. Ruffinelli, R. Gemulla, S. Broscheit, C. Meilicke, On Evaluating Embedding Models for Knowledge Base Completion, in: RepL4NLP@ACL, 2019. [8] A. Rossi, Interpreting link prediction on knowledge graphs., in: SEBD, 2020. [9] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, O. Yakhnenko, Translating embeddings for modeling multi-relational data, in: NIPS, 2013. [10] T. Trouillon, J. Welbl, S. Riedel, É. Gaussier, G. Bouchard, Complex Embeddings for Simple Link Prediction, in: ICML, 2016. [11] T. Dettmers, P. Minervini, P. Stenetorp, S. Riedel, Convolutional 2d knowledge graph embeddings, in: AAAI, 2018. [12] P. Pezeshkpour, Y. Tian, S. Singh, Investigating robustness and interpretability of link prediction via adversarial modifications, in: NAACL-HLT, 2019. [13] H. Zhang, T. Zheng, J. Gao, C. Miao, L. Su, Y. Li, K. Ren, Data poisoning attack against knowledge graph embedding, in: IJCAI, 2019. [14] A. Rossi, D. Firmani, P. Merialdo, Knowledge graph embeddings or bias graph embeddings? a study of bias in link prediction models, in: DL4KG, 2021. [15] P. W. Koh, P. Liang, Understanding black-box predictions via influence functions, in: D. Precup, Y. W. Teh (Eds.), ICML, 2017. [16] Z. Ying, D. Bourgeois, J. You, M. Zitnik, J. Leskovec, Gnnexplainer: Generating explanations for graph neural networks, in: NIPS, 2019. [17] W. Zhang, S. Deng, H. Wang, Q. Chen, W. Zhang, H. Chen, Xtranse: Explainable knowledge graph embedding for link prediction with lifestyles in e-commerce, in: JIST, 2019. [18] W. Zhang, B. Paudel, W. Zhang, A. Bernstein, H. Chen, Interaction embeddings for prediction and explanation in knowledge graphs, in: WSDM, 2019. [19] M. T. Ribeiro, S. Singh, C. Guestrin, "Why Should I Trust You?": Explaining the Predictions of Any Classifier, in: SIGKDD, 2016. [20] S. M. Lundberg, S.-I. Lee, A unified approach to interpreting model predictions, in: NIPS, 2017. [21] M. T. Ribeiro, S. Singh, C. Guestrin, Anchors: High-precision model-agnostic explanations, in: AAAI, 2018. �