Abstract

Computational measures of semantic similarity between geographic terms provide valuable support across geographic information retrieval, data mining, and information integration. To date, a wide variety of approaches to geo-semantic similarity have been devised. A judgment of similarity is not intrinsically right or wrong, but obtains a certain degree of cognitive plausibility, depending on how closely it mimics human behavior. Thus selecting the most appropriate measure for a specific task is a significant challenge. To address this issue, we make an analogy between computational similarity measures and soliciting domain expert opinions, which incorporate a subjective set of beliefs, perceptions, hypotheses, and epistemic biases. Following this analogy, we define the semantic similarity ensemble (SSE) as a composition of different similarity measures, acting as a panel of experts having to reach a decision on the semantic similarity of a set of geographic terms. The approach is evaluated in comparison to human judgments, and results indicate that an SSE performs better than the average of its parts. Although the best member tends to outperform the ensemble, all ensembles outperform the average performance of each ensemble's member. Hence, in contexts where the best measure is unknown, the ensemble provides a more cognitively plausible approach.

Highlights

  • The importance of semantic similarity in geographical information science (GIScience) is widely acknowledged [21]

  • Research in natural language processing and computational linguistics has produced a wide variety of approaches, classifiable as knowledge-based, corpus-based, or hybrid [29, 32]

  • Ten semantic similarity measures are tested on a set of pairs of geographic terms utilized in OpenStreetMap

Read more

Summary

Introduction

The importance of semantic similarity in geographical information science (GIScience) is widely acknowledged [21]. Selecting the most appropriate measure for a specific task is nontrivial, and represents in itself a challenge From this perspective, a semantic similarity measure bears resemblance with a human expert being summoned to give her opinion on a complex semantic problem. Complex computational problems in machine learning are often tackled with ensemble methods, which achieve higher accuracy by combining heterogeneous models, regressors, or classifiers [34] This idea was first explored in our previous work under the analogy of the similarity jury [7]. To measure the cognitive plausibility of each measure and ensemble, a set of 50 geographic term pairs including 97 unique terms, selected from OpenStreetMap and ranked by 203 human subjects, was adopted as ground truth.

Related work
WordNet similarity measures
Evaluation
Evaluation criteria
Ground truth
Experiment setup
Experiment results
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call