Review Comment:
The authors have significantly improved upon their original paper where they demonstrated many of the common features and differences between publicly available knowledge graphs. The number of measures used for analysis has reduced by one, but a number of additional statistics and more in depth explanations have been provided.
All of the metrics are now well described with explicit quantification which should allow a reader to reproduce the results exactly and apply the 34 metrics to any new KG.
Metrics such as ability to rank data, and existence/support of, e.g., owl:disjoint, could easily skew the result in favour of a particular data set with many ‘features’ without regard for quality, of course it would be up to the reader to choose a sensible weighting for such measures. This is a situation where the metrics now being valued continuously between 0 and 1 rather than simply true/false allows for better flexibility.
I feel that the gold standard is at the most risk of being subjective, while the metrics derived from it (completeness) are some of the more important values. I trust that the authors have truly chosen a suitable gold standard set, their descriptions would suggest so and the resulting values appear fair.
I enjoyed the expanded discussions and evaluations of each metric, in addition to the summary table of references. I would be happy using the information within this paper and being able to back up each decision and only ask the authors to perform a final spell-check, e.g., p5. firectly => directly.
I thank the authors for answering my previous questions and addressing mine and the other reviewers’ concerns.
|