Review Comment:
The paper presents an ontology (Explanation Ontology or EO) aimed at representing user-centred explanations in domain applications. This is mostly meant to help system designers in preparing their systems, i.e. modelling the explanations that can be obtained from the method(s) they develop. The EO model is described in terms of classes and properties, which are well liked to existing vocabularies, and further showcased in 5 real-world scenarios, where AI systems created through the IBM AIX-360 explanation toolkit provide user-centred explanations. The evaluation of the ontology is achieved through competency questions, aimed at assessing the ontology quality both from the tasks and the applications perspective.
In terms of the review criteria:
- originality: the work is original in what little work is devoted to structuring explanations and the processes that generate them. While it is indeed an extension of an already published ontology, the paper presents a new version of it which includes new classes, properties and vocabulary alignment. This is mostly driven by the use-cases that need(ed) to be represented with the EO.
- significance of the results : the paper is mostly an engineering exercise, so it does make little sense to talk about significance of the results. The use-cases and the competency questions however show the validity of the ontology, particularly the ability to represent multiple explanation types (and consequently designers needs) across domains. More in one of my points below.
- quality of writing : the paper is well-written and well-structured.
- reproducibility : the ontology is open source and publicly available on GitHub. The EO ontology uses PURL's domain which ensure longevity. It is well-documented, and I was able to check the data without issue. Replication of the experiments should not be a problem.
I recommend for acceptance, but I would like a few points to be further addressed. Apologies for the random order:
- The work is focused on the concept of user-centred explanations, but a bit more space should be dedicated in characterising them. One could argue that all explanations are somehow user-centred, as explanations are part of a communication process involving some sort of users (humans or machines). The paper seems to suggest (or, I wonder if) that user-centred explanations are those that come as outputs of the modern AI methods, but in the real world explanations might come also as part of other (non-AI) communication systems. Also, AI methods are mentioned, but all scenarios are ultimately reduced to explaining recommendations. So the ontology is only meant for systems providing recommendations? Perhaps I would suggest the authors to take a bit of time to discuss what do they mean by user-centred explanations, whether these are focused around AI methods and which (and why not on the rest), and which explanations might not be user-centred. This also relates to Section 3.2, where explanation types are discussed : where does the specification/granularity of explanation types end? How do you establish it? (again, perhaps the motivation is the answer to this?
- The point above could also be explained through a motivation section or paragraph, i.e. the concrete reason bringing the authors to tackle the given research problem. The introduction of Section 3 seems to suggest something in this direction, but could be expanded and should definitely come much earlier in the text (when i read the related work I am already wondering what user-centred explanations are). Overall, I would like to understand if the use-cases are the motivation that made the work happen, or the other way round.
- A few points regarding the model. First, while classes and properties are well described, as well as the instances in the use-cases, I would still be interested in having some concrete numbers of the ontology wrapped up in a table. How many classes and properties? instances per class? What are the most useful classes in your scenarios/properties? This is also information that could be useful to system designers, as well as for yourselves for ontology maintenance and evolution. Regarding Figure 2, it would be good to take some time and talk the reader through it (currently it is only quickly discussed). This would put the reader at ease with the classes you are manipulating (for instance, I am still confused on what an "object record" is). Also, it seems to me that only few of the classes in Figures 2 (and 1?) are used in Section 4, and some classes seem rather ad-hoc per scenario (e.g. ingredient, Recipe, column, cell, or classes I found on the Widoco documentation such as Income or Clinical Pearls). Are these meant to be part of the ontology? Shall they not be instances given that they are domain-specific? In addition to that, it would be good to see more classes in common among Figures 3 -- 8 (I could only see some). Finally, the EO is defined as upper-level in Section 7. From Wikipedia: an upper ontology [...] is an ontology [...] which consists of very general terms (such as "object", "property", "relation") that are common across all domains. Are all the classes from the EO fitting this definition?
- The authors claim initially that a broad set of competency questions are presented, but then only a few (6 + 7 ) are presented in the evaluations of Section 6. I would perhaps rephrase the initial statement. In general, the competency questions should be organised in terms of complexity, i.e. from easy to complex things to answers. I would also add what the ontology is *not* able to answer. The authors also mention that the first 6 Qs were the ones they managed to come up with. Did they try checking these with their intended users, i.e. the system designers? Could they not help identifying more simple and more complex queries, and also the CQs that the ontology was not able to cover? Was the ontology evaluated at all with the intended users? If yes, this should be better specified (how many, etc)
- How do the authors plan for the system designers to use the ontology? Is it currently only accessible through Protege, or is there an annotation tool/webapp they can use?
Minors:
- p. 3 l.22 : point > pointed
- subject matter experts
- p 3 l. 45 : don't > do not (there's several others to be changed)
- p4 l 26 : in explanation methods papers > in papers
- p6 l 24 : "we introduced classes if they were not a part of well-used ontologies" > should the *not* be removed?
- p 6 l47 : These include such things as as > These include things as contrastive
- p 9 l 18 : made and provides > made and provide
- p 12 l 42 : unlabelled > unlabeled
- In the figures, the point of the arrows (empty/non-empty triangle) should be consistent between legend and the graph
- p 16 l 40 : support system designers who are looking to support... > rephrase
- p 17 l 47 : . In use cases where the details of the explanation consumers are 47 present ... >> I am not sure I can syntactically parse this sentence
- p 19 l 25 : For the task-based abilities we aim to > For the task-based abilities, we
|