Review Comment:
Very interesting position paper on the integration and divide of symbolic and subsymbolic systems with a particular focus on benefits from their integration within the context of the Semantic Web or vice versa how SW can be beneifical to deep learning. Overall, I think it is a well-written position paper with an excellent line of argumentation and many very interesting pointers to get people interested in the field started.
When discussing the possibilities of how to train an NN for a specific ontology or knowledge graph, I think it would be interesting to consider the scenario of transfer learning and domain adaptation. Currently, Section 2 hints at those phenomena by talking about training on one domain/resource and than re-using this pre-trained model. However, I think it would be helpful to directly address the terms used for those processes, i.e., transfer learning and domain adaptation. It would actually also be interesting to discuss whether transfer learning in its current format could even be applied to knowledge bases with relatively small (compared to DL datasets) and very often conflicting contents.
In terms of knowledge graph embeddings, it might be interesting to note that the majority of approaches focused on encoding triples as stand-alone entities and only recently encoding contexts, that is, entire KG paths, has become a research topic. A class of algorithms called path-ranking based models has started investigating this phenomenon, such as Yin et al. (2018) below (also Das and Palumbo relate to this).
If there is still space, it might be interesting to talk about integrations of knowledge graphs and NNs directly in the architecture itself, as it has been attempted in the case of Graph CNNs (see references below).
Additional references:
Even though many KG embedding approaches are provided, one that is interesting in terms of treatment of graph neighborhoods and which is frequently being re-used nowadays is missing, that is, the Node2Vec approach:
"node2vec: Scalable Feature Learning for Networks". A. Grover, J. Leskovec. ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), 2016.
path-ranking:
Wenpeng Yin, Yadollah Yaghoobzadeh, and Hinrich Schütze. 2018. Recurrent one-hop predictions for reasoning over knowledge graphs. In COLING.
Rajarshi Das, Arvind Neelakantan, David Belanger,and Andrew McCallum. 2017. Chains of reasoning over entities, relations, and text using recurrent neural networks. In EACL.
Enrico Palumbo, Giuseppe Rizzo, Raphael Troncy, Elena Baralis, Michele Osella, and Enrico Ferro. 2018. Knowledge graph embeddings with node2vec for item recommendation. In European Semantic Web Conference. Springer.
Graph CNN:
Defferrard, M., Bresson, X., & Vandergheynst, P. (2016). Convolutional neural networks on graphs with fast localized spectral filtering. In Advances in neural information processing systems (pp. 3844-3852).
Kipf & Welling (ICLR 2017), Semi-Supervised Classification with Graph Convolutional Networks (disclaimer: I'm the first author)
Hamilton, W., Ying, Z., & Leskovec, J. (2017). Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems (pp. 1024-1034).
Maybe cross-referencing papers from the issue could be included here, such as Freddy Lecue's submission on Explainable AI.
MINOR COMMENTS:
p. 3 Semantic Web Technologies => Semantic Web technologies
p. 5 While deep learning ...., but it => omit "but"
p. 5 "is producing" => produces
|