Abstract

Micro-task crowdsourcing has emerged as a powerful approach for rapid collection of user input from a large set of participants at low cost. While previous studies have investigated the acceptability of the crowdsourcing paradigm for obtaining reliable perceptual scores of audio or video quality, this work examines the suitability of crowdsourcing to collect voice likability ratings. Voice likability, or voice pleasantness, can be viewed as a speaker social characteristic that can determine the listener's attitudes and decisions towards the speaker and their message. The collection of valid voice likability labels is crucial for a susccessful automatic prediction of likability from speech features. This work presents different auditory tests that collect likability ratings of a common set of 30 voices. These tests are based on direct scaling and on paired-comparisons, and were conducted in the laboratory under controlled conditions—the typical approach—and via crowdsourcing using micro-tasks. Design considerations are proposed for adapting the laboratory listening tests to a mobile-based crowdsourcing platform. The likability scores obtained by the different test approaches are highly correlated. This outcome motivates the use of crowdsourcing for future listening tests, reducing the costs involved in engaging participants and administering the test on-site.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.