Review Comment:
Generally speaking I'm very in favour of accepting a paper that describes LOV. Nonetheless, I think that this paper, in its current form, is not ready yet for publication and would need a serious revision in both form and content. I hope that my review can help in improving the paper for a final acceptance.
I would recommend the authors to insert a section just after the introduction to give an initial overview of LOV in term of its content. I would expect basic statistics: number of vocabularies, terms, properties and classes and their trend over time (there are some details in section 4, but they come too late I think), most frequent knowledge domains represented in vocabularies, min/max number of properties or classes in a vocabulary, statistics on the different inter-vocabulary relationships (e.g. is extension more frequent than import?), etc. In short, I think that the some more details about the actual content of LOV should make the paper more comprehensive.
The other missing bit in the paper is the explanation of the "LOD popularity": in some places (and in evidence in LOV search results) there are references to the occurrences of a term in LOD datasets. Where do those numbers come from? How are they computed? How often are they updated? Are they included in the LOV dumps? Are those occurrence numbers "reliable" (i.e. are they actual indicators of vocabulary reuse)? More importantly, is this sort of "popularity" measure included in the LOV ranking when returning search results?
More specifically regarding the current content of the various sections of the paper:
- Figure 1: this is a minor comment, but it would have been more natural to me if the figure was "upside down", with the Web at the bottom and the community on top, since this is the usual way to represent things that gets data from some sources and elaborate them to present them to final users. As a consequence, I would also revert somehow the order of the following sections.
- Section 2.1: while I understand that the community plays an important role, it sounds a bit odd to consider it a "component" also identified by a squared box in Figure 1 like software components. At least a different graphical representation would be advisable.
- Section 2.2, curators: please, add more information about curators and their work. Who are the curators? How many? What's their background? Are they sufficient or you envision an increment of this "editorial board" the more vocabularies are inserted in LOV? What is the process they follow? What are the criteria for inserting or rejecting the insertion of a vocabulary in the catalogue? How long does it take on average for a new vocabulary to be added to LOV? What happens if a vocabulary suggested for insertion is about a domain that is unknown/unfamiliar for the curators (e.g. a very specialized biomedical vocabulary)? What if a suggested vocabulary contains hundred thousand or million terms? I recommend the authors to add those (and maybe other) details; maybe the LOV curation team is worth a "component" in Figure 1 as well.
- Section 2.2, inlinks/outlinks: I may be wrong, but it seems to me that the definitions of inlinks and outlinks are reverted (I assume that inlinks are those on the left and outlinks those on the right in Figure 2).
- Section 2.2, associated metadata: the authors say that the curators added creator information in about 85% of the cases (which is a lot). Do they also contact back the creators to suggest them to complement/complete their vocabulary metadata to improve future versions? If not, I think this would be a very useful service.
- Section 2.2, last sentence: "one needs to know in which vocabularies and datasets a particular vocabulary term is referenced", I understand that in LOV each term has a link to its respective vocabulary but not to datasets that use it. Or am I missing something? (This is also linked to the LOD popularity comment above.)
- Section 2.3: can you add a link to the LOV code and indicate its respective software license? Creative Commons are data/content licenses, not usually applicable to code.
- Section 2.3.1 and Figure 3: what are duplicate terms? If it simply means same term searched by another LOV user, it seems to me that it is an irrelevant information. Moreover, are multiple terms queries included in the count? Can we consider Figure 3 as a display of the total number of queries on LOV over time?
- Section 2.3.1 and Table 2: I would recommend adding a line with numbers for single term queries, just for reference and comparison. Moreover, can the authors comment on the numbers in Table 2? What do those numbers tell us about user searches?
- Section 2.3.2: do the dumps also contain the full vocabularies with all the terms or only their description according to VOAF? Can the authors add some statistics on number of triples, dimensions of dump files, etc.? This is another place where they mention the number of occurrences in LOD without fully explaining what this mean.
- Section 3, ontology assessment: here the authors mention a term score. Is this the same score explained in Section 2.3.1 with respect to the local name and various labels or is it something else?
- Section 3, listings: both SPARQL examples miss the WHERE clause.
- Section 3, ontology localization: while I am fully convinced of the importance of localization, the example provided by the authors sounds misleading, since it seems that they are saying that words in different languages sharing the same "root" have the same meaning. This is of course false (try with "burro" in Italian and in Spanish...).
- Section 3 and Figure 9: that figure is too complex; therefore either the authors insert a full explanation or they remove it since it does not convey much more that what is already written in the section (I suggest the latter option).
- Section 4: as said above, some of the things written in the section introduction should be anticipated and expanded. Minor comment on Figure 10: what is the reason for the tiny decrease in vocabulary number in October 2014?
- Section 4.2, last paragraph: "Giovanni" doesn't seem an author of reference [10].
- Section 5: when comparing the number of results between Swoogle and LOV, the authors wrote that in LOV "the term only appears in 1,562 vocabularies". I think this is incorrect since LOV has a total of less than 500 vocabularies, probably that is the number of terms matching that query.
- Section 6: again the authors present as a contribution the "terms search scoring". While it is indeed useful to distinguish between primary labels and other properties, I'd say that an average vocabulary "seeker" would also like to know who used a specific vocabulary and in which context, as a metrics/indicator/clue to choose between competing vocabularies. Therefore, as a possible future extension and improvement of LOV, I would recommend to think about ways to support users in vocabulary selection: besides the current scoring/ranking system (which is about relevance of the result w.r.t. the user query), LOV could offer other "scores" or additional supporting information for a more informed selection of vocabularies.
As final remark, I would recommend the authors to have their paper checked by a native speaker (I'm not...), because some expressions sound a bit weird to me. For example I think that LOV "alone" should not have the article ("LOV is" instead of "the LOV is"); on the other hand, when used in attributive manner it should need the article ("the LOV architecture/the LOV curators/..." instead of "LOV architecture/LOV curators/..."). Native speakers can of course prove me wrong.
|