Review Comment:
This is a survey paper, focusing on ontologies for modelling data about ICT devices, the materials they contain, in order to facilitate data sharing for the Circular Economy (CE). The survey results consists of a set of ontologies for modelling (i) ICT devices, (ii) materials and material composition, and (iii) CE concepts and strategies. I find this topic really interesting, and I agree with the authors that semantic technologies are a key enabled in the CE domain, since CE requires data sharing not only across organisations, but also across domains and for various unforeseen purposes. As such I find this paper highly valuable and timely, and I am sure it would interest many readers both in the semantic web community and beyond.
However, the paper also has several weaknesses, that I think need to be addressed before it can be published. It starts already with the paper title and the motivation and introduction in section The paper title claims that the paper will analyse the role of the semantic web for data sharing in CE, while the paper is in fact specifically targeting ONE specific technology, i.e. ontologies, and does not even for this technology really analyse its role for data sharing in CE, but merely surveys existing models. A better title could be something like “A survey of ontologies for facilitating data sharing about ICT devices and components in the Circular Economy”. The current title, together with the long motivation and explanation of what CE is and how it is important in the ICT domain, gives the reader wrong expectations when reading the rest of the paper. I acknowledge that at the end of the abstract, and at the end of section 1 the authors do briefly mention that this is a survey of ontologies only, but this is easy to miss when the title says something else.
Further, I think the other main weakness of the paper is the methodology. I am even not entirely sure if this is a problem with how the actual survey was carried out, or merely with how it is presented in the paper, but at least the presentation is a problem. Section 2 of the paper presents the survey methodology, but this section is entirely detached from the rest of the paper. Figure 2 is nice, but it is not clear how each sub-step was performed, in what order, by whom, on what sources etc. Also, there is no explicit connection between any of the steps and the results presented in the paper, although for some parts the reader can guess (e.g. step 2.3 probably resulted in Table 2). But for some parts it also seems that I cannot find any discussion on it in the paper, such as for the step 3.2, which sounds as if that would result in a survey of tools (potentially using ontologies but not necessarily) in CE. Also, the authors refer to PRISMA for guidelines on reporting surveys, however, first of all the current PRISMA site suggests to refer to their updated version from 2020 and not the 2009 version that is cited there (see [1]). Second, I do not agree that the reporting of this survey actually follows the PRISMA guidelines. For instance, in PRISMA 2020 you can find a 27-point checklist for the paper, listing things to be reported in various sections, such as that the methods section needs to make explicit the inclusion and exclusion criteria used in the survey, selection process, risk of biases and limitations of the study. Of course, some items are not really applicable, since PRISMA is mainly for the medical domain, but I would like to see the items that are in fact applicable actually being included and discussed in this paper. This would clarify several vague statements in the paper, such as that parts of the survey is “based on” other surveys - what does that actually mean? That you included all the results they found and then added your own? Or that you simply reuse their survey results and did not add your own? (See for instance the introduction sentence to section 4.2 for such as statement.) In the analysis section, the authors also need to describe the assessment method much more in detail, and the criteria for arriving at tables 6-8 are missing, e.g. what does it mean to “evaluate against” a competency question? What if a more general concept and property is available, but not exactly the one in the CQ? Does that give a “yes” or a “no” for the CQ? Granularity (in the text mistakenly called expressivity) is discussed briefly, but only stating that the ontologies differ in their level of detail and granularity, but no actually saying how this was handled when producing the tables. Also the delimitation to only focus on laptop computers needs to be more clear in the paper - how does that affect the results? Is this survey even applicable to the whole ICT domain then? Why/why not? Overall, the paper is methodologically weak, and this is main thing that needs to be improved in order to make the paper publishable.
The standards section is also a bit detached from the rest of the paper. How is this connected to the ontology survey? Do the ontologies follow the standards? Did the standards come after the ontologies? Are there clashes?
Finally, I also find the ontology survey to include some dubious assessments. On thing that is mentioned several times is the tool used to develop the ontologies - how is this relevant? Would you in a survey of software libraries include the IDE that was used to develop them, e.g. “this was coded using NetBeans IDE”? Probably not, unless that has some impact on the resulting artefact. However I fail to see how this is relevant for the ontologies. Further, there seems to be some confusion on the ontology languages and the notion of “expressivity”. First of all, RDFS and OWL are ontology languages (and there are others) but RDF is not. Overall I fail to see really how you could even express an ontology using RDF only, you would basically loose the whole idea of a formal semantics, since RDF is only plain triples, without any further possibility to express meaning. However, since both OWL and RDFS build on top of RDF and can be expressed as RDF graphs, e.g. for sharing on the web or in a triple store, what I suspect is that when some ontologies in the survey are listed as RDF ontologies, this actually just means they are available in some RDF serialisation format. Similar, there seems to be some confusion on the OWL versions. OWL2 is simply the current version of OWL, so any current ontology in OWL can probably be said to be in OWL2, it is not some specific language separate from OWL itself. OWL DL on the other hand is a certain subset of OWL that limits the expressivity of OWL(2), so that says something different. Also the term “expressivity” seems to be used in a strange way in the paper, i.e. in some cases I interpret it as the authors mean the level of detail of the ontology (e.g. depth of taxonomy, or granularity of modelling) rather than the actually expressivity used for the logical modelling, where the latter should instead indicate an OWL profile (or a specific DL perhaps if you want to be more precise). While the confusion of the term expressivity should definitely be resolved in the text, I acknowledge that it will probably be very difficult for the authors to determine the actual expressivity of each ontology in this survey (especially since many are not even available online). I would therefore suggest to merely state as much information as possible of each ontology, but be clear on the cases where this is unknown. I think the most relevant information is whether it is an RDFS or OWL ontology, if it adheres to a specific OWL profile (e.g. EL, QL, RL) or if it is simply in OWL DL. For the accessible ontologies the authors should be able to determine this by examining the ontology itself, and/or using some tool to assess it.
Detailed comments/questions and minor issues, in order of appearance:
- Footnotes 1 and 2 are more or less repetitions of each other.
- Does table 1 come directly from reference [15]? There has been a lot of discussion on how FAIR is actually to be implemented for ontologies, see for instance [2] below. A bit more discussion on this would benefit the paper. On the other hand, FAIR is not really used to assess the ontologies later on (apart from “are they online or not”), so this could also be better connected to the survey itself.
- Table 2 is nice and valuable, but quite long, so potentially it would fit better in an appendix and only a few examples included in the main text. I also have some doubts when reading parts of the table, such as how the key concepts were determined? Why is “device” not a key concept of Q1 and Q2, but of Q3 and several others? In this section it is also not so clear that you are only focusing on laptops, because you talk about ICT devices in general, and additionally some of the key concepts are NOT laptops, such as the router and switch listed for Q1. Why are they key for laptops, if phones and tables are not? Further, what is the difference between device and hardware in this context? Q6 uses the term device, while Q7 uses hardware, do they refer to the same thing? What is status and grade in Q8 and Q9? It is interesting that you consider software to be a component of a device, i.e. you are not only covering physical objects as components here, or am I misinterpreting Q10? Why is location not a key concept of Q15? Q19-22 state “is used” - is used in what, the component or the device? Q31 - seems random to only focus on USB-ports, why not other ports? Q38 - how is warranty a physical property? I am also not sure about memory capacity in Q41, I would have intuitively put that as a computational property. On the other hand, how is CO2 footprint (Q44) a computational property? Then comes a set of CQs that relate to cost, stock etc. that seem to be totally missing the contextual nature of these properties. In my opinion it doesn’t make sense (or at least it will be practically infeasible to get the data) to model the total world-wide stock of an item, instead I guess what could be captured in the stock of the type of item at some certain reseller, manufacturer etc. Similarly the cost is also not a universal thing, something may cost a certain amount to manufacture in one factory, and another amount in another factory. The product will cost x at a certain retailer, and y at another store etc. Similarly the material cost, is that the actual amount of money spent, or the potential cost based on what components are included? Q56-57 also seem context-dependent, in this case related to an actor, i.e. recommended by someone, or selected by some organisation? Q59 is unclear, what does this mean? Is it per device? Per type of device? Per organisation?
- In section 4.1.11 you state that only 4 ontologies are available, but this is actually a bit ambiguous since some of the papers you found presented ontology networks consisting of up to 9 ontologies. I get what you mean, but the number is I guess higher if you could individual ontologies.
- In the next sentence you say “rarely available online” when you discuss that the other things that are described in papers are not accessible - this is vague, are they online or not?
- In section 4.3.2 you mention that the authors have developed a library that somehow extracts parts of an ontology if I understand correctly? The implications of this should be discussed further. If non-standard methods of reuse are applied, then that may affect the FAIR-ness and reusability of the ontologies.
- Section 4.2.5: I am not sure that you really mean that one has to have studied philosophy to reuse EMMO? Do you mean that it is too abstract perhaps, and that you need a certain level of knowledge on modelling principles and basic ontological distinctions to use it?
- Section 4.2.8 mentions MDO but you have not described MDO yet, better put this section later, after MDO.
- The first paragraph of 4.2.11 is a bit strange. The second sentence seems to state the obvious - what would be the alternative? Not modelling materials in i material ontologies? And I am not sure how EWC comes into the picture, it was not in the list of surveyed knowledge structures.
- What do you mean with “build in an ‘ad hoc’ manner” at the bottom of page 19? Without following a proper methodology? Without having a clear goal? Without having a verified set of requirements? Or something else?
- Page 20, “The work of Sauter et al. [38] has a limited expressivity… (see Table 8)” - Table 8 does not say anything about expressivity.
- I am not a domain expert so I may be wrong, but isn’t the standard called BS 8001 (and not BIS 8001) and the publisher BSI?
- Some references are incomplete, even missing the year (e.g. [40]), online references missing access dates, and the format is not consistent (e.g. sometimes years in parenthesis, sometimes not).
Language issues:
- Overall “build” is used in many places in the text, where the correct inflected form should be “built”.
- An additional issue is that quotes are used in many places where you do not really need them - remove the quotes and reformulate in your own words instead!
- Section 4.1.5: Paatent -> Patent
- Section 4.1.6: It can be also be -> It can also be
- Bottom of page 12: exiting -> existing
- Section 4.3.1: three level of -> three levels of
- Section 4.3.3: adoptation - do you mean adoption or adaptation?
[1] http://prisma-statement.org/PRISMAStatement/PRISMAStatement.aspx
[2] Poveda-Villalón, M., Espinoza-Arias, P., Garijo, D., & Corcho, O. (2020). Coming to terms with FAIR ontologies. In Knowledge Engineering and Knowledge Management: 22nd International Conference, EKAW 2020, Bolzano, Italy, September 16–20, 2020, Proceedings 22 (pp. 255-270). Springer International Publishing.
|