Review Comment:
This manuscript was submitted as 'full paper' and should be reviewed along the usual dimensions for research contributions which include (1) originality, (2) significance of the results, and (3) quality of writing.
This paper describes a new tool (or, rather, pair of tools) to help with reading and editing RDF constraints. The two tools are evaluated against Moody's Physics of Notations, and then undergo an empirical evaluation involving 12 participants. The paper is original, and the new tools have the potential to be significant in the field. Furthermore the paper is generally well-written (a list of typos is included at the end of this review). There are some significant short-comings with the article, however, which occur mainly in the data analysis (which can be rectified) and one particular issue with the study design (which it will be too late to rectify, with data collection already having occurred). I think that with some changes the paper could be accepted for publication, although it will perhaps not be as wide-in-scope as the authors intended it to be.
The first three sections of the paper are very strong, and on their own would merit some form of publication. Things start to go slightly wrong in section 4, however. On page 14, section 4.1, semiotic clarity is discussed. However, this is never properly introduced. We have the definition from [8] (the correspondence between symbols and their reference concepts), but then we have two further concepts of symbol redundancy/overload, and symbol excess/deficit, introduced. Importantly, it is never stated how semiotic clarity relates to these two new concepts. These new concepts are measured against the competing notations, but then semiotic clarity is re-introduced as a conclusion. In section 4.2, about perceptual discriminability, it is claimed that ShapeUML uses shape as a variable, but then immediately says only one kind of shape is used. Which means it is not variable. I find the claim (page 14, right column, line 42) that two different shapes can be distinguished because of the incoming edge to be spurious. I can distinguish between the collection of line-shape pairs, but not the shapes themselves. Furthermore, this could easily be obviated by using more shapes. Why use just rectangles and ellipses? In section 4.3, the claim that ShapeVOWL has high semantic transparency does not seem supportable, even with the references. It may have higher semantic transparency than ShapeUML, but that does not mean that there is high transparency in absolute terms.
In general, for section 4 there is too much reliance on just one framework (Moody's physics of notations), whilst other frameworks complement it. For example, Gestalt principles, Bertin's semiology of graphics, the use of pre-attentive features, to name but three. A stronger theoretical comparison could be made if more than one framework was utilised.
Section 6 details the analysis of an empirical study. The number of participants, whilst low, is appropriate given the target user-group, and the within-design used. However, it does not seem like a full within design, as participants did not see each example in both notations. That is a minor issue, however, as large portions of text after 6.4.2 cannot be supported. There was no statistical difference found between the notations (not even remotely close to significance) in 6.4.2, but then section 6.4.3 proceeds to explain various "differences" in the data, and give reasons for why these "differences" occur. As an example, on page 24, right column, line 2, we are told that there were "slightly better scores" in ShapeVOWL, and gives a potential reason why. However, we've already been told there was no statistical significance, and further the rates of 21% versus 25% probably equates to one participant. Unfortunately, this pattern repeats: despite the differences in scores being more plausible as random variation than any actual difference, minor "improvements" of ShapeVOWL over ShapeUML are explained in both 6.4.3 and 6.4.4. The conclusion drawn at the end is then that ShapeVOWL is better. Which, as mentioned, is entirely unsupported by the data.
The qualitative data analysis is more compelling, but taken together they still do not provide remotely enough evidence for the claim on page 28, right column, lines 14, 16-18 that "ShapeVOWL is preferred" and that the "work strongly suggest[s] that ShapeVOWL will find more user acceptance than ShapeUML".
In order for this work to be published, the entire analysis section needs to be rewritten to properly reflect the results of the analysis. In essence, it would become much shorter (effectively excising sections 6.4.3 and 6.4.4), but would be much more robust. I would like to see a more rounded theoretical comparison, too. I did like this work, but only what is in the data should be explained.
Other minor issues:
Page 5, right column, line 29: is redefining the compartments of a UML diagram not going to confuse people who are used to reading UML diagrams?
Page 9, figure 4: the Venn diagram strongly suggests inclusive or, not exclusive or, given that the intersection is shaded the same as (A-B) and (B-A). This could be a typo, as exclusive or is available in figure 5 as "one of".
Minor typos:
Page 1, left column, line 49: de described
Page 3, left column, line 9: cognitive effective --> cognitively effective
Page 7, right column, line 46: striked through --> struck through
Page 10, left column, line 21: suggest the red 6 is incorrect
|