Abstract

A number of studies on network analysis have focused on language networks based on free word association, which reflects human lexical knowledge, and have demonstrated the small-world and scale-free properties in the word association network. Nevertheless, there have been very few attempts at applying network analysis to distributional semantic models, despite the fact that these models have been studied extensively as computational or cognitive models of human lexical knowledge. In this paper, we analyze three network properties, namely, small-world, scale-free, and hierarchical properties, of semantic networks created by distributional semantic models. We demonstrate that the created networks generally exhibit the same properties as word association networks. In particular, we show that the distribution of the number of connections in these networks follows the truncated power law, which is also observed in an association network. This indicates that distributional semantic models can provide a plausible model of lexical knowledge. Additionally, the observed differences in the network properties of various implementations of distributional semantic models are consistently explained or predicted by considering the intrinsic semantic features of a word-context matrix and the functions of matrix weighting and smoothing. Furthermore, to simulate a semantic network with the observed network properties, we propose a new growing network model based on the model of Steyvers and Tenenbaum. The idea underlying the proposed model is that both preferential and random attachments are required to reflect different types of semantic relations in network growth process. We demonstrate that this model provides a better explanation of network behaviors generated by distributional semantic models.

Highlights

  • How word meaning is represented in human memory is a longstanding problem that has attracted the interest of linguists, philosophers, psychologists and other scholars

  • It should be noted that, in this paper, we address only the indegree distribution to examine the scale-free property of semantic networks, because using the out-degree or the degree of the undirected network would introduce bias that stems from the task characteristics such as the number of associations [25]

  • The results indicate that a distributional semantic model (DSM) has the ability to produce semantic networks with a degree distribution that is the same as or similar to that of the association network

Read more

Summary

Introduction

How word meaning is represented in human memory is a longstanding problem that has attracted the interest of linguists, philosophers, psychologists and other scholars. Network analysis takes as input a network produced from the observable data (i.e., word cooccurrence and word association) about the mental lexicon or human semantic memory, and reveals the properties of the network. These network properties provide information about the structure of the mental lexicon and the cognitive mechanism underlying the semantic structure, neither of which is directly observable from the data [12]. Despite their simplicity, network models that simulate the observed network properties can provide valuable insight into the process of lexical development by which these network properties emerge. The most studied phenomenon among these is free word association [23, 24], because it reflects human lexical knowledge acquired through world experience, thereby revealing the structure of human semantic memory or the mental lexicon more directly and efficiently than other lexical phenomena

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call