Review Comment:
This paper benchmarks the ability of comparing knowledge graphs using whole graph embedding techniques as well as graph kernels. The comparison is evaluated by changing (1) the size of the KG (2) the entity labels and (3) the KG schema.
There have been numerous approaches proposed in the literature that find entity and relation embeddings or triple embeddings in a KG. However, this paper’s focus is on representing the entire graph in a low dimensional space. More specifically, as this paper uses the embedding techniques designed for Homogeneous graphs rather than knowledge graphs, they first apply the conversion technique proposed by Vries et al - Journal of Web semantics 2015.
To evaluate the impact of the aforementioned changes on the embeddings, first the gradual changes algorithm is proposed. In this setting, two random graphs are chosen from the dataset and gradually one is converted to the other by performing deletion and addition of triples all the while guaranteeing that the connectivity is preserved. It is proposed in the paper that they expect to see the distance between the graphs to gradually increase as more iterations are performed. There has been no explanation given with this regard. I would expect the exact opposite as the two graphs are becoming more and more alike, the distance between their embeddings are expected to decrease. Moreover, the pseudocode provided can get stuck in an infinite loop despite what is mentioned in footnote 6 and that is due to the main while loop.
To see how the schema affects the embeddings, two scenarios are envisioned. It must be noted that both of these two scenarios are much simplified with regards to what can be changed in the schema of the KGs. First, the entity types are gradually replaced by their superclass. In this scenario, it’s expected that deviation in the embeddings will be small. Whereas in the second scenario, the entity types are replaced with a random class and it’s expected that we see an increase in the distance of embeddings. As shown in the results neither of the graph embeddings used for the experiments are able to capture the modifications that have been imposed. The question is, why would we in the first place expect that the used graph embedding techniques and kernels to have such capability? Neither of these techniques are designed for the KGs nor are able to take into account the semantics of the graph.
Finally, to see how dependent embedding techniques are to the name (label) of an entity, they are gradually substituted with another name. It is vague with respect to what is written in section 4.3.3 and in 4.6 what is the expected behavior. At one point, the authors wish the embeddings not to change when the name of the entities change and in another, they report it as a success if the embeddings have been able to capture this difference.
The idea to see how much prone to change the embeddings are in a KG with respect to change in the size as well as semantics is indeed an interesting task. However, it might be more meaningful to capture this in the context of works proposed for entity and relation embeddings in a KG. Indeed, the proposed scenarios are not convincing and the datasets that are used are not a good representation of real KG.
The paper lacks thorough and broad analysis of the problem as well as state of the art KG embedding techniques in its related work. The reported results are not convincing (see the paper “how powerful are graph neural networks” by Xu et al -ICLR 2019 for a comparison of the expressiveness of graph embedding techniques). The paper's writing in its current version as well as the structure of the paper can be improved.
Strengths :
-The idea of comparing supervised and unsupervised embedding techniques wrt their capacity of taking the schema into account is novel for KG.
- The related work about existing approaches that can be used to compare labelled graphs (not KG) is well structured and covers many kinds of supervised and unsupervised approaches.
Weaknesses:
- Scenarios that are presented to argue on whole KG comparison are not convincing
- The expectations and the results are not clear and not well illustrated
- The considered schema changes are over-simplified
- The used datasets are small and are not good examples of real KGs
---- Some Minor remarks----
Related work section: Ontologie alignment techniques can also be seen as methods that can lead to KG comparison (see OAEI competition results).
Table 2: The number of classes, properties that belong to the original schema and to a graph on average have not been mentioned
Abstract: they have not been tailor towards → tailored
to some extend → extent
Section 4.3.1: removing the triples in G1 that are in G1 but not in G2 → removing the triples in G1 that are not in G2
Section 4: Subgraph changes: The KG grows or schrinks → shrinks
Section 4.7: because next to embeddings → in addition to embeddings
Section 5.6: bad performance → poor performance
Tables 3,4,5: Best results indicated in bolt → in bold
|