Abstract
Auralization, the final step in computational room acoustic simulations, aims to make audible the acoustics of complex virtual architectural spaces in a realistic and accurate manner. This paper presents a novel real-time auralization system comprising a geometry engine, a beam-tracer, and an audio renderer. The computation of early reflection paths is based on an efficient beam-tracing algorithm capable of real-time detection of specular reflection paths in a static geometry with one or several moving listener(s). For simpler rooms, the real-time performance is maintained even with dynamic geometries and sources. Results of the beam-tracer, sent to the audio renderer, consist of visible reflection paths and their accumulated material attenuation. From this geometrical and acoustical data, listener position-related 3D room impulse responses are generated applying a higher-order virtual Ambisonics approach. Final rendering of the binaural room impulse response (BRIR) is made taking into account the listener's head-orientation. As higher order reflections are more diffuse in nature, they may be encoded using lower Ambisonic orders, thereby reducing computational load. The environment combines high quality audio with visual rendering realized using the open source platforms Pure Data and VirChor respectively. This auralization framework provides direct audio-visual feedback in real-time for VR environments.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.