Review Comment:
---- SUMMARY -----
This paper propose a survey on knowledge graph embedding approaches that also take into consideration the fact that KGs contain literals. The authors not only present a comprehensive summary of the approaches, but they also experiment some (i.e., those for which models are available) on different tasks; providing insightful results on the state-of-the-art of these approaches. The knowledge graph embedding topic is getting much attention lately. Despite some issues (that are not directly related to the content) and some questions that I have, I really like the paper and I think it might become a valuable contribution.
SWJ Guidelines:
(1) Suitability as introductory text, targeted at researchers, PhD students, or practitioners, to get started on the covered topic.
Answer: There are fixes to do and I have some questions. Also, I suggest the authors to add some details on standard knowledge graph embeddings to make their work more accessible to a broader audience.
(2) How comprehensive and how balanced is the presentation and coverage.
Answer: Good review of papers. With also some experimental evaluations.
(3) Readability and clarity of the presentation.
Answer: Well written in general, equations should be aligned to a common scheme.
(4) Importance of the covered material to the broader Semantic Web community.
Answer: Very important, literals are less considered in standard knowledge graph embedding approaches.
---- REVIEW -----
This paper reviews that state-of-the-art knowledge graph embeddings by giving a closer look to those approaches that treat literals. I like this paper and I think it is really a nice contribution. I like the summaries at the end of each section as I think they provide a nice way to recap each section.
Points 1 and 2 in the Introduction motivate well why literal values are important and should be considered in embedding approaches.
I quickly skimmed over ref [57] authors' own work and I think the extension the authors have deeply extended their previous work.
Knowledge graph embedding approaches for literals that extend standard knowledge graph embedding inherit their limits. And I'd do a brief summary of the problems of knowledge graph embeddings and how this might affect the representation of literals. (page 19) line 16. DistMult does not model well anti-symmetric relations because of its score function. I'd say this in a previous section.
Notation changes along with the paper and it would be easier to have a consistent scheme of symbols and the experiment should probably be better described (see some points on these two things in later sections of the review).
- QUESTIONS -
Sometimes the notation is different for different methods. Is this intentional? e.g., Equation 2, 21, 27 and 28 seem to express very similar loss functions (that should be the original TransE loss). But one uses the square bracket positive notation []+ to get the positive parts and another one uses the max function. One uses the set S' for the negatives and one uses the set T' for the negatives. I'd try to align the notation to a common scheme where it is posssible.
Equation 8) what are s and s*?
Section 6.2, "Extended Dataset:" did you create this dataset? will you share this? I think it's a valuable contribution given the recent discussion on knowledge graph embedding evaluation [1]. "as mentioned by the authors in the paper" which paper? are you referring to TransEA? (you mention this a page later).
In the link prediction with numeric literals, are you trying to predict the exact numerical label? I wonder if it is correct to use HITS@k to evaluate the predictions. I know this is standard, but still, I'm not sure if it is the best way to do this (I'm not asking for an experiment, but if this is true it might be worth discussing it in the paper -i.e., predicting 15 instead of 16 is different than predicting 1021 -).
(page 15, line 12): in the evaluation procedure section you should probably add on which list of results you compute the measures (e.g., the ordered set of corrupted triples and you look for the rank of the correct one). Moreover, the filtering setting should also remove the other correct triples from the ranking list if I remember correctly.
There is also probably the need to further describing the extended datasets. Is it well balanced?
I'd also name these datasets differently if they are an extended version of FB15K and FB15K-237, as reading the paper I was getting confused because I was thinking about the old FB15K and FB15K-237. Table 3 shows the details for the standard FB15K* datasets, I'd also include those that come from extended datasets.
Page 16) "Experimental Setup", embeddings of 100 and 200 dimensions. Why these two? is this because sometimes you use a ComplEx-based model that uses complex numbers and they thus require 2xN dimensions to be represented?
Since you already run lots of experiments, could you also had some analysis on the runtime of the algorithms? under a more "production-related" point of view, this might be very interesting.
"Note that the reason for DistMult-LiteralEg model to achieve the best result on FB15K-237 dataset is the fact that this dataset does not have any symmetric relation. " why? shouldn't DistMult favor symmetric relationships? am I missing something here? please correct me if I'm wrong. You also mention that "FB15K-237 achieves slightly better result compared to FB15K" is this because of the fact that the dataset has been extended? Because in structured knowledge graph embeddings [2] FB15k-237 is the one that is more difficult to solve. I suggest you to add more details in this section because I think it is important.
When you train TransEA, do you re-normalize values (since you have retrieved the original ones) or keep the original ones? if the latter is the case, how does this impact the models?
Are there specific categories of literals that are easier to predict than others? It would be nice, if you can generate it, to have a small summary table on this (but I guess it will be biased by the proportion of this information in the training data).
- SUGGESTIONS -
Table 1 shows a really good summary of the various models in the field by also defining the categories. I wonder if you could add some text to describe this table. Otherwise, just looking at the table, I do not get the difference between translational models and bilinear/the others. In my opinion, without some text that explains the categorization it might be not really informative and the categorization could become: 1) models that do not use literals, models that use literals.
I understand the paper focuses on literals, but I'd extend a bit the Introduction or the Related Work Sections by also explaining how approaches that do not focus on literal work.
It'd be easier for a reader to understand the paper if there was a short introduction on kg embeddings in general (e.g., it might be better to introduce what negative sampling and scoring function are in the context of KG embeddings). You could explain briefly explain some properties of TransE (and its limits, like 1-N representations), this might come useful because at (Page 6, line 13) you mention that DKRL is an extension of TransE. The same goes for EAKGAE.
On the same line, it might not be too clear to someone new to the field what a "complex conjugate" is, but I'm not sure how much you can give details about this. I'd just underline the fact that some of these methods are often mapped in different spaces (i.e., R and C).
This paper might be of interest to you [1] and I suggest the author cite this in their paper (RESCAL, in this paper, is reported as one of the models that is competitive and that achieves good performances ).
In section 7, I'd restate the research questions and I'd answer them.
page 2, line 22. "handling different challenges." => I'd name a few of those challenges here
page 3, line 23. I cannot understand the sentence "... has been given with experiments conducted ..."
It would be interesting to see some predictions of the models in the paper. Can you add some examples of the prediction of a literal for an entity?
- TYPOS -
The results from the ComplEx-LiteralEg model shows => show
"some approaches have been proposed which incorporate the information underlying literals to generate KG embedding" => could you rephrase this sentence?
[1] https://openreview.net/forum?id=BkxSmlBFvr
[2] https://arxiv.org/pdf/1902.10197.pdf
|