Review Comment:
The submission proposes two main contributions: a dataset for Entity Linking with out-of-KG references,
and an entity linking approach which detects out-of-kg entities and clusters them to
automatically create entity descriptions for these entities not yet present in the KG.
While the first contribution is a nice-to-have (as there is already existing NILK and other datasets), I will focus
in the review on the second contribution.
On Section 2.2:
what kind of aliases do you mean when using them for an entity dictionary?
what does that mean for your specific evaluation use case? how can alternative labels be extracted/generated for random KGs?
The authors mention here that they generate a candidate set of size 100. Is this the maximum size or do they always retrieve 100 candidates for each mention?
I do not understand the sentence "our method does not rely on a parallel text corpus"?
On 2.3.2:
The entity definition is defined as the schema:description information of the Wikidata KG. This makes the authors' approach very specific to Wikidata.
What about entity definitions in other KGs?
On 2.3.3:
Apparently, the ranking method includes a popularity measure using the outdegree of the entity within the KG.
In network/graph theory, nodes with a high outdegree are hubs -> nodes with a high distribution level and very low specificity.
Taking this into account, the outdegree is not a metric to measure the popularity of a node. Rather, the indegree should be used.
The main contribution of the submission is described here briefly:
the decision if an entity is out-of-KG is dependent on the similarity of the entity candidate definition and the input context. This is it. What does it mean, that the similarity is not enough?
What if the context is not enough? In this case, this approach would often detect out-of-KG entity candidates.
This simple decision is not convincing for all-purpose scenarios.
Or for KGs of specific domains where entities might be very similar and the differentiation of in-KG entities and out-of-KG entities might be too small.
On 2.4:
Here, the authors describe the clustering of out-of-KG entity embeddings. Unfortunately, they only mention the use of DBSCAN and do not discuss different clustering approaches.
DBSCAN is a density-based clustering approach. When this approach is applied on embeddings very close embeddings are clustered together which makes sense in this case.
But, embeddings are not semantic. Here again I see challenges/difficulties for domains with a very specific domain where the vocabulary is very restricted and similarities are more common than in all-purpose KGs as Wikidata.
On the experiments:
On entity linking:
The authors argument that they do not compare their approach on entity linking to others because they first want to examine different parameter settings.
I don't think this is a legible argument. Why don't compare the results for the best parameter setting to competitive approaches?
At the end, the performance of their overall approach is the most interesting part and the lack of sota comparisons is clearly one of the main drawbacks of the submission.
On out-of-KG detection:
Here again, the authors only show a survey of how different parameter ratios are affecting the quality of the approach.
How is there contribution performing compared to others? The discussion regarding the reasons for the different achieved recall and precision results is basic and trivial.
On the out-of-KG entity clustering:
The section would benefit from a repeated brief introduction which of the shown approaches in Table 6 are the ones of the authors. Also, the different approaches should be named constantly within the text and both tables.
In this comparative analysis, the author show that their approach might be competitive to sota approaches, but the contribution overall and results are not convincing enough to be published in a journal.
Overall, I think the authors work on an interesting topic, but the submission leaves too many questions unanswered and the performance of the proposed approach is not compared to competitive approaches, at least only in parts.
At this stage of the work, the submission is not eligible to be published in SWJ.
|