Review Comment:
In this paper, the authors focus on the answer selection task in Community Question Answering and propose a Siamese architecture-based model to classify candidate answers into three categories (i.e., Good, Potential, and Bad) in accordance with a given question. With the following review, my suggested decision on this paper is Major Revisions Required, and the revision should address all the negative aspects listed below.
The proposed model consists of three modules: Initial Representation, Attention Layer, and Multi-task Learning. 1) The Initial Representation module is to obtain pre-trained vector representations of the given question, candidate answers, and the question context (i.e., question category and subject). The focus of this module is to disambiguate fragment semantics with knowledge graph-based entity linking. 2) The Attention Layer addresses the redundancy and noise problem of questions/answers by computing attentional question/answer representations based on the question context representation. 3) The Multi-task Learning module employs a pair of Siamese variational autoencoders to encode question/answer representations and compare them to classify candidate answers. The decoding of latent question/answer representations and the latent representation-based classification are jointly trained.
Strengths: 1) the ambiguity issue of questions/answers is addressed with information from the knowledge graph, and the Siamese architecture-based model achieved superior performances in comparison with baseline methods, which demonstrate the significance of the results; 2) an ablation study was conducted to scrutinize the contribution of each module.
Weaknesses:
1) several existing methods, e.g., Babelfy and NASARI, are directly utilized in the model. However, there is a lack of sufficient introduction to the used methods, which hampers the readability of the paper.
2) the introduction to the attention layer only takes two short paragraphs without any formal definitions or equations which makes this module totally unclear.
3) the given "implementation details" are not detailed at all. Only the layer numbers of the autoencoder and the dimensions of latent representations are given. What are the parameter settings of the initial representations, the attention layer, the MLP classifier, and the convolutional filters?
These two issues make it difficult to reproduce the system as well as the experimental setup. Please provide both the formal definitions as well as the parameter settings used for each component to enable reproducibility.
4) In the ablation study, when any one of the key components (e.g., the knowledge graph-based disambiguation, or the attention layer) was removed, the model still outperformed most of the baseline models. However, there was only one component removed every time in the ablation study. The baseline of the proposed model (i.e., the vanilla version without any component) is not evaluated. Is it possible to start with the evaluation of the baseline, incrementally add one component at a time, and analyze how the performance could be increased with more components added? Also, it would be appreciated to have more information about the comparison set-up as well as the implementation details of the proposed model and the other compared models.
5) the originality of the paper is limited because: first, the main modules of the proposed model, e.g., knowledge graph-based disambiguation, and the Siamese autoencoder, have been already widely utilized in the literature; second, given the lack of details of the model design and implementation (e.g., how the existing methods are integrated into the model, and how the attention layer is customized in this model), the originality of each proposed module is difficult to be assessed.
6) The quality of writing does not meet the requirement of this journal due to the aforementioned lack of readability and minor errors such as:
In Section I - Paragraph 1, several applications, e.g., recommender systems, are listed without adequate references.
In Section I - Paragraph 3, there is no reference for SemEval2015.
In Section I - Paragraph 7, "unable to encode" instead of "unable to encoding", and it should be a period instead of a semicolon at the end of the paragraph.
In Section I - the last contribution, there should be references for the three listed datasets.
In Section II, please use adequate mathematical expressions instead of English characters when denoting variables and parameters.
In Section II - Paragraph 2, "as follows" instead of "as follow".
In Section III, it is claimed that "none of existing methods have considered the context in question-answer representation". However, after reading the related works introduced in the paper itself, I am skeptical of this claim.
Please thoroughly check the writing of the paper.
|