Abstract
An enigma for human languages is that children learn to understand words in their mother tongue extremely fast. The cognitive sciences have not been able to fully understand the mechanisms behind this highly efficient learning process. In order to provide at least a partial answer to this problem, I have developed a cognitive model of the semantics of natural language in terms of conceptual spaces. I present a background to conceptual spaces and provide a brief summary of their main features, in particular how it handles learning of concepts. I then apply the model to give a geometric account of the semantics of different word classes. In particular, I propose a "single-domain hypotheses" for the semantics of all word classes except nouns. These hypotheses provide a partial answer to the enigma of how words are learned. Next, a dynamic cognitive model of events is introduced that replaces and extends the function of thematic roles. I apply it to analyze the meanings of different kinds of verbs. I argue that the model also explains some aspects of syntactic structure. In particular, I propose that a sentence typically refers to an event. Some further applications of conceptual spaces are briefly presented.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.