Abstract
Paralinguistic features of speech communicate emotion in the human voice. In addition to semantic content, speakers imbue their messages with prosodic features comprised of acoustic variations that listeners decode to extract meaning. Psychological science refers to these acoustic variations as affective prosody. Most studies of affective prosody obscure semantic content, although the resulting stimuli are less representative of naturally occurring emotional speech. The presented works details the creation of a naturalistic emotional speech database on which both acoustical analysis and a listening study were conducted. To this end, 55 adults were recorded speaking the same semantic content in happy, angry, and sad voices. Based on prior acoustic analyses of affective prosody, classic parameters were extracted including pitch, loudness, timing, as well as other low-level descriptors, and compared the acoustic features of each emotion with published evidence and theory. Preliminary results indicate that this naturalistic speech samples yielded acoustic features that are congruent with prior experimental stimuli of anger and happiness, but was less consistent with sadness. The results of the listening study indicated that listeners discriminated the intended emotions with 92% accuracy. The dataset therefore yielded a database of emotionally salient acoustical information for further analyses. [Work supported by NIH-R21-104547.]
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.