Review Comment:
The authors of this paper propose TableMiner, a mechanism for semantically annotating an extracted data table (which likely comes from the Web but, strictly speaking, does not have to). Extracting table annotations is an active area of research, and can be useful in applications such as improving open linked databases, improving structured components of web search, and as an aid to data integration applications.
This is an interesting area where it is possible to make contributions. The authors do a good job of reviewing related work, and have assembled an admirable array of table datasets to evaluate. However, there are three serious problems with the article as written, one of which may be hard to address.
1) The authors decide that runtime is a crucial performance criterion of table annotation systems. This is the fundamental motive for their algorithms that attempt to iteratively sample from target tables, scoring candidate extractions' semantic labels row-by-row.
I think this is a serious mistake. Almost all work to date has focused on the markup quality, for good reason. The runtime overhead is entirely dependent on the features being used. But even in the Freebase-style case discussed here, there is likely no heavy computation going on; rather, it is likely that the observed Freebase lookup times are driven primarily by network latency. If the users had a local copy of the Knowledge Base data (which is possible for some some KBs) then the overhead time here would likely be minimal. (Note that experiments evaluate runtime entirely with a single thread, maximizing the problems with latency.) It is a stretch for me to believe that this is a serious problem: it can be fixed by simply downloading a dataset and building a local index.
2) The users don't actually compare against baselines from the literature. Rather, they compare against some simple --- and somewhat artificial --- baselines. As a result, it is hard to say whether their approximation results are actually as good as they should be.
3) There is a real lack of useful examples in the first half of the paper. The authors have a few one-line examples, but the paper could really benefit from a longer running example that helps to illustrate every extraction target.
I think #2 could be addressed with a revision of the paper, but #1 seems pretty core to the paper. As a result, I think the authors should either rethink the goals here, or should make a more serious effort to argue that performance is a serious evaluation metric for such a system.
Detailed comments:
The abstract could really use a stronger example
The use of "concept" to mean "type" should be better described in the intro. Also: use examples!
The link between your stated goals and the iterative method that is introduced at the end of page 2 is unclear. It eventually becomes clear why you're doing this, but iterative methods can serve multiple ends: minimizing human labels, for example. Your main goal is performance, but that doesn't come through right away.
"the largest collection of datasets for this task known to date". What is "this task"?
"most representative datasets" is unclear. Representative of what?
The evaluation criteria in the middle paragraph of left-column page 3 come from nowhere. I thought the whole point was increased performance, not accuracy. You have not told us that these other goals are things you're also trying to address.
There is a lot of related work from Kaushik Chakrabarti, on the InfoGather system. It should be cited and discussed.
The discussion on "Completeness" on page 7 could use a stronger example, esp on the point about "relations between non-subject columns"
I can't parse this important-seeming sentence from Section 4.1, "A candidate label for an NE-column…" This section could also use examples.
Algorithm 1 I-Inf is unnecessarily high level. It communicates the "meta-algorithm" but it is so high level that I can't understand if it's doing something useful.
Section 6 is unclear what is the authors' work, and what is due to [30].
Section 8: "it has been noted that…" by whom? By the authors, or does there need to be a citation here?
The List of peaks named Bear Mountain" example is too sketchily described. Authors should provide the concrete example in the paper and discuss it more carefully.
In the experiments, the authors permit a confusing statement about how the bootstrapping process and how it relates to interdependence among components. Paragraph begins, "In all baselines…" I am not sure whether they are trying to say something new here or not.
|