Review Comment:
Thank you for addressing the points of our previous review but there are still some open points that remain and need revising:
"While there are many graph embedding approaches, the approaches based on translating embeddings have shown to outperform the rest of the approaches on the task of link predictions." This is not correct when you have a look at newer approaches like ComplEx [1] and HolE [2] (which we already mentioned in our previous review).
While you included the three translational embedding approaches TransE, TransH, and TransR, some information is still missing: the dimension parameter should be mentioned and different variations could be shown. It also would have been beneficial if you would have chosen different types of graph embedding approaches like the aforementioned ComplEx [1] and HolE [2] or RESCAL and NTN which are already mentioned in the paper. At least mention them and explain why you do not compare to them or why they are not applicable.
Please make it explicit that the knowledge graph embeddings from the other approaches are trained for link prediction and not for the different evaluation tasks like classification.
Below are further open points:
-- Missing explanation / reference --
* It is not discussed why the edge information was used for the Weisfeiler-Lehmann graph kernel. There have been link prediction approaches which ignore the predicate so that a comparison of the exclusion of the edge information could be interesting.
p.10: "The dataset contains three smaller-scale RDF datasets (i.e., AIFB, MUTAG, and BGS) [...] Details on the datasets can be found in [71]" Please mention why you left out the AM dataset which was the fourth dataset described in [71].
-- Parameter Settings & Results --
p.11: "We use two pairs of settings, d = 2; h = 2 (WL_2_2) and d = 4; h = 3 (WL_4_3)." and the same for "the length of the paths l" The values changed from the previous version but we still couldn't find any explanation or insights for choosing these values.
p.14: Any explanation for the surprisingly low accuracy of SVM combined with DB2vec on AAUP, especially in respect to the WD2vec combination?
p.23ff: Are there any insights of any structural differences or coverage of Wikidata and DBpedia which explain the difference in performance e.g. for DB2vec SG 200w 200v 4d compared to WD2vec SG 200w 200v 4d?
-- Notation, Style & Errata--
p.13: "On the other hand, “Facebook” and “Mark Zuckerberg” are not similar at all, but are highly related, while “Google” and “Mark Zuckerberg” are not similar at all, and have somehow lower relatedness value." The word "somehow" does not fit in this context. Lower relatedness value compared to what?
p.15: table is too big -> in top margin
p.21: https://github.com/sisinflab/LODrecsys-datasets should be in the footnote like the rest of the URIs
p.21: "versions of Movielens, LibraryThing" is in the margin
p.25: https://github.com/sisinflab/lodreclib should be in the footnote like the rest of the URIs
* By addressing the reviewers' comments, you introduced some long bothersome sentences which destroy the reading flow. Please proof-read the paper again and recreate a good reading flow.
[1] Trouillon, Théo, et al. "Complex Embeddings for Simple Link Prediction." Proceedings of The 33rd International Conference on Machine Learning. 2016.
[2] Nickel, Maximilian, Lorenzo Rosasco, and Tomaso Poggio. "Holographic Embeddings of Knowledge Graphs.”
|