GeoKG-Enabled Similarity Computation with Defeasible Reasoning

Tracking #: 3839-5053

Authors: 
Bongjae Kwon
Kiyun Yu

Responsible editor: 
Guest Editors Geospatial Knowledge Graphs 2025

Submission type: 
Full Paper
Abstract: 
This paper investigates the application of defeasible reasoning within geospatial knowledge graphs (GeoKGs) for geospatial similarity computation. Motivated by the need for accurate and interpretable similarity assessments in domains such as urban planning and location-based services, this study proposes a novel approach that combines the structured data representation of GeoKGs with the uncertainty-aware inference capabilities of defeasible logic. A GeoKG is constructed by integrating data from OSMnx, Wikipedia, and GeoNames. Defeasible rules are generated to capture contextual and functional similarities, and a reasoning engine infers similarity scores through priority-based conflict resolution. The proposed method is benchmarked against knowledge graph embedding (KGE) models and a large foundation model (Gemini gemini flash 2.0) using an expert-annotated dataset. While the KGE model achieved 72.3% accuracy and the LFM 68.1%, defeasible reasoning achieved 67.2%. Despite its lower accuracy, it offers superior interpretability by explicitly representing the rationale behind similarity assessments. This transparency is critical in decision-making scenarios where trust and justification are paramount. The study also highlights the impact of rule refinement and conflict resolution strategies on performance, suggesting potential for further improvement. By introducing defeasible reasoning into GeoKG-based similarity computation, this work provides a promising, explainable alternative to black-box models, paving the way for future hybrid approaches that balance accuracy and interpretability.
Full PDF Version: 
Tags: 
Reviewed

Decision/Status: 
Reject

Solicited Reviews:
Click to Expand/Collapse
Review #1
Anonymous submitted on 02/Jun/2025
Suggestion:
Reject
Review Comment:

This paper proposed a framework that leverages defeasible reasoning with GeoKGs for geospatial similarity computation. The paper clearly written and generally well-organized. While the topic is timely and relevant, and I appreciate authors' efforts in addressing the challenges of geospatial similarity computation, I have significant concern regarding the clarity, rigor, and fairness of the experimental evaluation. Therefore, I do not recommend acceptance of this paper at its current stage.

1. Originality
The idea of introducing defeasible reasoning into geospatial similarity computation is conceptually interesting. However, the proposed approach relies primarily on predefined rules embedded in the GeoKG. While authors compare their method to KGE models and LFM-based models (e.g., Google Gemini), the comparison raises fairness concerns. In particular, LFM-based models are pretrained on vast external datasets, including geographic knowledge, while the proposed method relies on a limited internal rule base. This discrepancy in background knowledge is not sufficiently addressed, calling into question the comparability of the models. That's also one of the potential reasons why LFM-based models outperform. Furthermore, the paper does not clearly articulate what fundamentally new insights or technical contributions it introduces beyond existing rule-based KG systems.

The idea of introducing defeasible reasoning into geospatial similarity computation is conceptually interesting. However, the proposed approach relies primarily on predefined rules embedded in the GeoKG. While authors compare their method to KGE models and LFM-based models (e.g., Google Gemini), the comparison raises fairness concerns. In particular, LFM-based models are pretrained on vast external datasets, including geographic knowledge, while the proposed method relies on a limited internal rule base. This discrepancy in background knowledge is not sufficiently addressed, calling into question the comparability of the models. That's also one of the potential reasons why LFM-based models outperform. Furthermore, the paper does not clearly articulate what fundamentally new insights or technical contributions it introduces beyond existing rule-based KG systems.

2. Significance of the results
The reported results show that the proposed defeasible reasoning approach achieves lower accuracy than KGE models and the LFM-based models. While the authors argue that the interpretability compensates for lower accuracy, this trade-off is not compelling without a more substantial demonstration. There is no quantitative comparison of interpretability across models, which makes it difficult to assess whether the proposed method truly offers superior explainability. Tables 2 and 3 demonstrate the incremental performance gains through rule refinement and alternative conflict resolution strategies, but the overall accuracy remains lower than baseline models. Morever, the evaluation is limited to a single city (Amsterdam), and it's unclear how generalizable the system is to other regions.

There are also some minor things need to be addressed:
1.there should be full name of LFM when it's firstly mentioned.
2.Rule 3 introduces rating similarity as additional supporting evidence for semantic similarity. However, I find this connection somewhat unclear. Ratings primarly reflect customer satisfaction or preference, which may be influenced by many factors beyond cuisine - such as service or pricing. Therefore, it is not direct evidence how a small difference in rating meaningfully contributes to assessing whether two restaurants are semantically similar, especially in terms of their cuision type. In other words, likeness in customer ratings does not necessarily imply similarity in function or semantics, and this assumption should be better justified or empirically supported in the paper.

Review #2
By Yuhan Ji submitted on 06/Jun/2025
Suggestion:
Reject
Review Comment:

This manuscript was submitted as 'full paper' and should be reviewed along the usual dimensions for research contributions which include (1) originality, (2) significance of the results, and (3) quality of writing. Please also assess the data file provided by the authors under “Long-term stable URL for resources”. In particular, assess (A) whether the data file is well organized and in particular contains a README file which makes it easy for you to assess the data, (B) whether the provided resources appear to be complete for replication of experiments, and if not, why, (C) whether the chosen repository, if it is not GitHub, Figshare or Zenodo, is appropriate for long-term repository discoverability, and (4) whether the provided data artifacts are complete. Please refer to the reviewer instructions and the FAQ for further information.