Abstract
Estimating the directions of arrival (DOAs) of multiple simultaneous mobile sound sources is an important step for various audio signal processing applications. In this contribution, we present an approach that improves upon our previous work that is now able to estimate the DOAs of multiple mobile speech sources, while being light in resources, both hardware-wise (only using three microphones) and software-wise. This approach takes advantage of the fact that simultaneous speech sources do not completely overlap each other. To evaluate the performance of this approach, a multi-DOA estimation evaluation system was developed based on a corpus collected from different acoustic scenarios named Acoustic Interactions for Robot Audition (AIRA).
Highlights
Estimating the direction of arrival (DOA) of a sound source is a well written-about topic in signal processing and has found a considerable amount of areas of application where its usefulness ranges from a complementary source of information to an essential part of the application
In [29], for the authors to be able to carry out real-time spatial rendering of different acoustic scenarios with a wide variety of sound sources, their DOA was required to be estimated via first calculating the inter-microphone time difference (ITD) using the generalized cross-correlation (GCC)-PHAT methodology
This is the result of the fact that the resulting θ from the single-DOA estimator is calculated from the ITD of the microphone pair most perpendicular to the sound direction
Summary
Estimating the direction of arrival (DOA) of a sound source is a well written-about topic in signal processing and has found a considerable amount of areas of application where its usefulness ranges from a complementary source of information to an essential part of the application. In [29], for the authors to be able to carry out real-time spatial rendering of different acoustic scenarios with a wide variety of sound sources, their DOA was required to be estimated via first calculating the ITD using the GCC-PHAT methodology.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: EURASIP Journal on Audio, Speech, and Music Processing
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.