Review Comment:
Some remaining issues
---------------------
* The difference between related work in 2.1 and in 2.2 is not clear to me. In 2.2 you say: "Moreover, they allow to evaluate the final results (answers extracted from the KBs) and to provide precise evaluation figures." But this is true also for work mentioned in 2.1, because once systems have constructed queries, these queries can be executed to get answers. Maybe simply merge 2.1 and 2.2 into one section?
* Footnote 3: The link you provide is not persistent, please use http://sc.cit-ec.uni-bielefeld.de/qald/ instead (although it's not possible to point to QALD-3 Task 2 directly).
* The provided URLs (e.g. in Footnote 4 and 5) are ok, but the underlying links in the PDF are messed up due to the line breaks. Please fix this.
* In 4.2 you define the question topic as the type of semantic entity which is the major context of the question. It is not clear what you mean by "major context". Then, on page 8, you take the first semantic entity as question topic. This means first in a left-to-right sense?
* At the very end of the conclusion, you point to where the implementation and test questions can be found. I think that a much better place for this would be to put it in Section 7.4 (Reproducability).
Typos and suggestions
---------------------
Abstract:
* A recent and intensive research -> Recent and intensive research
* using SPARQL language. -> using the SPARQL language. or: using SPARQL.
* RDF triples description -> RDF triple descriptions
* Natural Language question-answering -> natural language question answering
Page 1:
* A recent and intensive research -> Recent and intensive research
* life-science bases -> life science knowledge bases
* In the enumeration (ClinicalTrials.gov, Sider, DrugBank), use commas instead of semicolons.
Page 2:
* cannot manage the syntactic and semantic requirements of the SPARQL language neither can they manage the structure of various knowledge bases -> can manage neither the syntactic and semantic requirements of the SPARQL language, nor the structure of various knowledge bases
(Actually, "cannot manage" is quite a strong claim; I would rather write something like: are usually familiar neither with the query language SPARQL nor the structure of various knowledge bases)
* mediate technical and semantic complexity --> I think you don't mean "mediate" but "lower".
* some related works -> some related work
* linked data -> Linked Data
* Please use \cite{paper1,paper2,paper3}, which will give [5,15,31] instead of [5][15][31].
* languages KBs -> KBs
* querry -> query
* One Question-Answering system (AutoSPARQL) is based on -> The question answering system AutoSPARQL
* dependent from -> dependent of
And I would suggest the following (non-)capitalizations:
* Knowledge-Based Specific Interface -> knowledge-base specific interface
* Question-Answering System -> question answering system
* Question-Answering system -> question answering system
* Natural Language interfaces -> natural language interfaces
* Kind -> kind, Entity -> entity, Property -> property, Relation -> relation
Page 3:
* manually-written grammar -> manually written grammar
* Questioning over Linked Data -> Querying Linked Data
* main objective of the related set of works is more complex than in works presented -> main objective of the following related work is more complex than work presented
* first -> First
* question the KB -> query the KBthis kind of works -> this kind of work
* a 0.62 precision -> a precision of 0.62
* activation of query graph -> activation of the query graph
* boolean -> Boolean
* retrieval of precise biomedical information in linked KBs -> retrieval of precise biomedical information from linked KBs
* [17], processing -> [17], or processing
* enriching question with -> enriching questions with
* define the frames -> define frames
* and to build the SPARQL queries -> build SPARQL queries
Page 4:
* see Fig. 1 -> see Figure 1 (You always write Figure, not Fig., so you should also be consistent here.)
* RDF triple description -> RDF triple descriptions
* generating the SPARQL queries -> generating SPARQL queries
* proposed by the task 2 -> provided by task 2
* and Sider described -> and Sider, described
* on the several questions -> using the following questions
* 4. Question translation -> 4. Question Translation
Page 5:
* see Fig.2 -> see Figure 2
* such as disease names, side effects -> such as disease names and side effects
* semantic entities recognition -> semantic entity recognition
* Here you print vocabulary elements (e.g. diseasome/disease/1154) using typewriter font, whereas in the rest of the paper you use bold face or italics (in Figure 4). Please stick to one. I would prefer typewrite font, but it's up to you.
Page 7:
* e.g. side effects drugs -> e.g. side effects of drugs
* expressing the negation -> expressing negation
* building the representation -> building a representation
* semantic annotation -> semantic annotations
* semantic entities like lead is removed theyr are part of larger entities -> semantic entities like lead are remove if they are part of larger entities
* On the whole -> In total
* Definition of the Result form: negated -> Definition of the Result Form: Negated
* coordination marks -> coordination markers
* boolean -> Boolean
* Identification of the Question topic: we -> Identification of the Question Topic: We
* corresponds the question topic -> corresponds to the question topic
* Arguments: we -> Arguments: We
* Scope of coordination: arguments -> Scope of Coordination: Arguments
* graph representation and abstraction -> graph representations and abstractions
* The objective of the query construction step is to associate previously identified elements --> Here something is missing, as you always associate something with something else. Or you don't mean "associate" but something else?
* to build representation of the -> to build a representation of the
Page 9:
Caption of Figure 6:
* You need to swap "left part" and "right part".
* displays graph representation -> displays the graph representation
Page 10:
* during the question abstraction: they concern -> during the question abstraction, as they concern (Or some other suitable conjunction...)
Page 11:
* In the enumeration, please use commas instead of semicolons.
* The predicates are associated between them through their subjects and objects --> This doesn't work. You associate something with something else, not between and also not through something. Please fix this.
* ?v0 which is -> ?v0, which is
Page 12:
* in Figure 8a: the -> In Figure 8a: The
* In 5., remove the line break.
* are processed: predicates -> are processed: Predicates
* the examplified questions -> the example questions (This occurs twice.)
* the result form SELECT -> the result form is SELECT
* each RDF triple and the filtering -> each RDF triple and the filters
* examples from Figures 9a to 9e and Figure 9g -> examples in Figures 9a-e and 9g
* SPARQL end-point -> SPARQL endpoint (also in Footnote 8)
* relies on: (1) the -> relies on (1) the
* ; (2) -> , (2)
* the three following -> the following three
* Footnote 8: This is also not a persisent URL, please simply leave it out, saying "For our experiments, we use the SPARQL endpoint provided by the QALD-4 challenge."
Page 13:
* Figure 9: You could consider using namespaces that you define in the caption, then the URLs and the queries would be easier to read.
Page 14:
* named entity recognition; -> named entity recognition.
* disease gene associations -> disease/gene associations
* At the end of all bullet points, please use "." instead of ";"
* drug-target relations -> drug/target relations (or change "disorder/gene associations" above in to "disorder-gener associations")
* subject predicate object RDF triples -> RDF triples of form
* for the Question Annotation -> for Question Annotation
* questions: in this way -> questions. This way
* IRIs -> URIs
* training test gather training and test sets -> Please fix this.
Page 15:
* 4 Gb -> 4 GB
* 2.7GHz -> 2.7 GHz (or write "4GB" above)
* running time -> run time (occurs twice)
* sub-steps -> substeps
* TermTagger which -> TermTagger, which
* in 2 seconds on the average -> in two seconds on average
* with 0.78 F-measure -> with an F-measure of 0.78
* for three questions, the -> for three questions the
* SPARQL end-point -> SPARQL endpoint
Page 16:
* were returned then. -> were returned.
* Comparison with Existing Works -> Comparison with Existing Work
* with the existing ones -> with existing ones
* for Precision, Recall and F-measure -> for precision, recall and F-measure
* exploits Grammatical Framework grammar based on formal syntax -> exploits a Grammatical Framework grammar
* POS-tagging -> POS tagging
* On the whole -> In general
* Similar works -> Similar work
* For instance, comparable approach -> For instance, a comparable approach
* provide the possibility -> provides the possibility
* descrive -> describe
* scenarii with more detail -> scenarios in more detail
Page 17:
* performance of automatic systems -> the performance of automatic systems
* in a previous work -> in previous work
* specific entities while -> specific entities, while
* to the general entities -> to general entities
* construction step: the -> construction step: The
* define regular expression -> define a regular expression
* calcium supplement while -> calcium supplement, while
Page 18:
* DailyMed, is also -> DailyMed is also
* relies on linguistic and semantic annotation -> relies on the linguistic and semantic annotation
* RDF triples description -> RDF triple descriptions
* with 0.78 F-measure -> with an F-measure of 0.78
|