Abstract

We present a survey of available frameworks for developing acoustical signal processing models based on deep neural networks. Given that this is a dynamic space with new frameworks, libraries, and even companies appearing on timescales measured in months, we provide an up-to-date assessment of the strength, popularity, and near-future directions of several tools and platforms available for research and product deployment for deep learning models of audio signal processing. Similarly, those new to these spaces may be unaware of software systems that will allow them to obtain and interrogate results more quickly and easily, while also integrating the nearly state-of-the-art optimization methods. Included tools, packages and platforms include PyTorch, Tensorflow, Keras, JAX, fastai, PyTorch Lightning, Julia, nbdev, HuggingFace, Weights and Biases, and Gradio. Examples will be drawn from the speaker's recent research publications in musical signal processing and computer vision applied to musical acoustics, as well as recent work by others. The goal of the talk is to provide acoustics researchers, educators, students with a set of helpful possibilities for pursuing and improving their understanding, research practices, and communications.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.