Abstract

We propose a time-frequency processing method that localizes and enhances a target sound by exploiting spectral and spatial characteristics of the ego-noise captured by a microphone array mounted on a multi-rotor micro aerial vehicle. We first exploit the time-frequency sparsity of the acoustic signal to estimate at each individual time-frequency bin the local direction of arrival (DOA) of the sound and formulate spatial filters pointing at a set of candidate directions. Then, we combine a kurtosis measure based on the spatial filtering outputs and a histogram measure based on the local DOA estimation to calculate a spatial likelihood function for source localization. Finally, we enhance the target sound by formulating a time-frequency spatial filter pointing at the estimated direction. As the ego-noise generally originates from specific directions, we propose a DOA-weighted spatial likelihood function that improves source localization performance by identifying noiseless sectors in the DOA circle. The DOA weighting scheme localizes the target sound even in extremely low signal-to-noise conditions when the target sound comes from a noiseless sector. We experimentally validate the performance of the proposed method with two array placements.

Highlights

  • M ULTI-MICROPHONE acoustic sensing from a multirotor drone aims to record, localize and analyze sounds emitted by aerial or ground objects [1], [2]

  • Based on the observation that a target sound and the ego-noise usually have concentrated energy at sparsely isolated time-frequency bins, we proposed a timefrequency filtering approach [12], which formulates a spatial filter to enhance a target direction based on local direction of arrival (DOA) estimates at individual time-frequency bins

  • We proposed a time-frequency processing method to localize and enhance a target sound captured by an MAV by exploiting the spectral and spatial characteristics of the ego-noise

Read more

Summary

Introduction

M ULTI-MICROPHONE acoustic sensing from a multirotor drone (or MAV: micro aerial vehicle) aims to record, localize and analyze sounds emitted by aerial or ground objects [1], [2]. The rotating motors and propellers generate strong ego-noise [9], which masks the target sound, degrades the sound quality and leads to extremely-low signal-to-noise ratios (e.g. SNR < −15 dB). The nonstationary spectrum of the ego-noise depends on the rotation speed of each motor, which changes over time [10]. The microphones move with the MAV leading to a dynamic acoustic mixing network. The natural and motion-induced wind increases the noise captured by the microphones. All these issues make MAV-based acoustic sensing a very challenging task

Objectives
Methods
Findings
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.