Abstract
3D audio is commonplace augmented or virtual reality environments, and is used simulate the direction of arrival for sounds presented over headphones. The Head-Related Transfer Function (HRTF) enables 3D audio by characterizing how sound is received from a sound source in every direction. Optimal sound localization performance with HRTF-based 3D audio is thought to require personalized HRTFs because the localization cues are dependent on the size and shape of the head and pinnae. In this work, we propose a prototype pipeline to generate HRTFs from high resolution smartphone videos of the head and ears, and we validate the localization performance of the resulting HRTFs through human-subject behavioral experiments. To create a personalized 3D HRTF, a video was taken of a subject’s torso, left ear and right ear using a smartphone camera. Acoustic simulation with Boundary Element Modeling was performed in COMSOL, taking an approximately 1 h to complete. Preliminary virtual sound localization results were obtained on a pilot subject for each of 3 different HRTF types: A generic HRTF, a gold standard acoustically measured HRTF, and the smartphone video HRTF. We found a mean angle error of 36.6, 21.1, and 14.0 deg, respectively, indicating that the smartphone-based HRTF developed in this study provides higher localization accuracy than other methods on this pilot subject. Further testing will add additional subjects, and compare to free-field localization over loudspeakers.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.