Abstract

The core question that spoken word recognition research attempts to address is: How does a phonological word-form activate the corresponding lexical representation that is stored in the mental lexicon? While speech perception research (see the separate Oxford Bibliographies in Linguistics article “Speech Perception”) focuses on the mapping of highly variable acoustic signal onto more abstract phonological units, spoken word recognition focuses on the mapping of phonological information onto lexical and semantic representations—the repository of linguistic knowledge stored in a “mental dictionary” or the mental lexicon (see the separate Oxford Bibliographies in Linguistics article “Mental Lexicon”). Earlier theoretical work considers the three following stages as being fundamental to spoken word recognition. First, there is activation of multiple word forms that share some phonological similarity to the auditory input. Second, there is a selection stage whereby activated word forms compete with each other for recognition. Finally, when a single lexical candidate remains, its meaning is accessed and is then integrated with higher levels of processing (e.g., with sentential or discourse information). Although these stages of spoken word recognition are presented as being part of a serial process, it is important to note that current theoretical and empirical work in the field emphasize the highly parallel, incremental, and continuous nature of spoken word recognition—even though theories of spoken word recognition continue to differ greatly in their description and conceptualization of these “stages,” and in the computational implementation of competition and lexical selection mechanisms. The temporal, fleeting nature of acoustic input creates unique theoretical and empirical challenges for the field, for instance, the challenge of word segmentation in continuous speech and for embedded words, which has traditionally progressed at a more gradual pace relative to research in visual word recognition (see the separate Oxford Bibliographies in Linguistics article “Visual Word Recognition”). Nevertheless, in the almost sixty years of its history, spoken word recognition research has led to the discovery of a number of lexical-semantic and contextual factors that influence the speed and accuracy of spoken word recognition. Lexical-semantic factors refer to the lexical and semantic properties of individual words, for instance, its frequency of occurrence in the language or its extent of phonological similarity to other words in the language. Contextual factors refer to how characteristics of the talker and listener, as well as environmental features or noise, can create suboptimal conditions for spoken word recognition. In addition, the robust top-down influences of lexical knowledge on sublexical representations highlight how the integration of top-down information and bottom-up perceptual input forms a crucial feature of models of spoken word recognition. These empirical findings provide important constraints on the development of models and theories that attempt to explain the cognitive mechanisms that support the retrieval of spoken words from the lexicon.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call