Abstract

Our brains are wired in such a way that we can recognize a friend or loved one almost as easily whether we hear their voice or see their face. Specialized areas of the brain—in this case, the visual and auditory networks—are specially tuned to different properties of physical objects. These properties can be represented by multiple sensory modalities, so that a voice conveys nearly as much information about a person’s identity as a face. This redundancy allows rapid, automatic recognition of multimodal stimuli. It may also underlie “unimodal” perception—hearing a voice on the phone, for example—by automatically reproducing cues that are usually provided by other senses. In this view, as you listen to the caller’s voice, you imagine their face to try to identify the speaker. In a new study, Katharina von Kriegstein and Anne-Lise Giraud used functional magnetic resonance imaging (fMRI) to explore this possibility and understand how multimodal features like voices and faces are integrated in the human brain. Studies using fMRI have established that when someone hears a familiar person’s voice, an area of the temporal lobe called the fusiform face area (FFA) is activated through temporal voice areas (TVAs), suggesting early interactions between these cortical sensory areas. von Kriegstein and Giraud wondered whether these cortical ensembles might lay the foundation for general “multisensory representation” templates that enhance unimodal voice recognition. To explore this question, the researchers analyzed changes in brain activity and behavior after people learned to associate voices with an unfamiliar speaker. One group of participants learned to associate voices with faces and a second group learned to associate voices with names. Though both types of learning involve multimodal associations, voices and faces provide redundant information about a person’s identity (such as gender and age), while voices and names provide arbitrary relationships (since any name could be associated with any voice). To further explore the contribution of redundant stimuli from the same source, the researchers added an additional set of conditions in which participants associated cellular phones with either ringtones or brand names. In this case, both cell phone and brand name were arbitrarily associated with a ringtone. In the first phase of the fMRI experiment, participants listened to and identified either voices or ringtones. In the second phase, one group of participants learned to associate the voices and ringtones with faces and cell phones, while another group learned voice–name and ringtone–brand name associations. In the third phase, participants again heard only the auditory signals and identified either voices or ringtones as in phase one. The brain scans showed that learning multisensory associations enhanced those brain regions involved in subsequent unimodal processing for both voice–face and voice–name association. But at the behavioral level, participants could recognize voices that they had paired with faces much more easily than they could recognize voices they had linked to names. Participants who had learned to associate voices with faces were the only ones to show increased FFA activation during unimodal voice recognition. The fMRI results show that even a brief association between voices and faces is enough to enhance functional connectivity between the TVA and FFA, which interact when a person recognizes a familiar voice. In contrast, voice–name association did not increase interactions between voice and written name sensory regions. Similarly, people did not recognize ringtones any better whether they had learned to associate them with cell phones or brand names. Nor did their brain scans reveal any interactions between auditory and visual areas during ringtone recognition. Altogether, these results show that learning voice–face associations generates a multimodal sensory representation that involves increased functional connectivity between auditory (TVA) and visual (FFA) regions in the brain and improves unimodal voice recognition performance. When only one sensory modality of a stimulus is available, the researchers conclude, one can optimally identify a natural object by automatically tapping into multisensory representations in the brain—cross-modal ensembles that are normally coactivated—as long as the stimulus provides redundant information about the object. Given that faces and voices are the major means of social communication for nonhuman primates as well as for humans, the reliance of multiple, redundant sensory modalities likely has deep roots in our evolutionary history.

Highlights

  • Fusion of two plasma membranes is central to exocytosis, the process by which a cell secretes neurotransmitters, digestive enzymes, and other products

  • When the others bind to their respective SNAREs, they leave the SNAREs in an open conformation, available for interacting with others and forming the complexes that drive membrane fusion

  • Synaptobrevin is a known partner for syntaxin, and it has been shown that the addition of synaptobrevin drives syntaxin in conjunction with SNAP-25 into SNARE complexes

Read more

Summary

Synopses of Research Articles

Fusion of two plasma membranes is central to exocytosis, the process by which a cell secretes neurotransmitters, digestive enzymes, and other products. By deleting SNAP-25, the authors verified its essential role in displacing Munc181 from syntaxin, suggesting there is an intermediate complex formed by Munc, syntaxin, synaptobrevin, and SNAP-25 These results shed light on the actual function of Munc, but allow the development of a more coherent picture of SM proteins, in which Munc is no longer the oddball. HIV-1 exploits retrograde transport of tRNAs in human cells to promote nuclear import of its reverse transcription complex. Using a high-speed centrifuge, they separated cytoplasm from cultured cells into fractions and tested each fraction for its ability to promote nuclear import of the RTCs in the gutted cells As expected, when they added in the fraction containing several known nuclear transport factors, the RTCs could enter the nucleus.

Chance Fluctuations in mRNA Output in Mammalian Cells
Transposon Silencing Keeps Jumping Genes in Their Place
Offspring of plants that had both
Protein Kinases and Plant Pores
Infection Status Drives Interspecies Mating Choices in Fruit Fly Females
Gut Bacteria Cospeciating with Insects
Predicting Species Abundance in the Face of Habitat Loss
Disrupted Intercellular Communication Causes a Disfiguring Birth Defect
Learning New Movements Depends on the Statistics of Your Prior Actions
Findings
Islands Spark Accelerated Evolution
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call