Year Year arrow
arrow-active-down-0
Publisher Publisher arrow
arrow-active-down-1
Journal
1
Journal arrow
arrow-active-down-2
Institution Institution arrow
arrow-active-down-3
Institution Country Institution Country arrow
arrow-active-down-4
Publication Type Publication Type arrow
arrow-active-down-5
Field Of Study Field Of Study arrow
arrow-active-down-6
Topics Topics arrow
arrow-active-down-7
Open Access Open Access arrow
arrow-active-down-8
Language Language arrow
arrow-active-down-9
Filter Icon Filter 1
Year Year arrow
arrow-active-down-0
Publisher Publisher arrow
arrow-active-down-1
Journal
1
Journal arrow
arrow-active-down-2
Institution Institution arrow
arrow-active-down-3
Institution Country Institution Country arrow
arrow-active-down-4
Publication Type Publication Type arrow
arrow-active-down-5
Field Of Study Field Of Study arrow
arrow-active-down-6
Topics Topics arrow
arrow-active-down-7
Open Access Open Access arrow
arrow-active-down-8
Language Language arrow
arrow-active-down-9
Filter Icon Filter 1
Export
Sort by: Relevance
  • Research Article
  • 10.1186/s13636-025-00429-y
MIDI music plagiarism detection method based on feature similarity learning
  • Dec 30, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Xun Jin + 3 more

  • Open Access Icon
  • Research Article
  • 10.1186/s13636-025-00433-2
Diffraction perception in L-shaped rooms using virtual reality
  • Dec 20, 2025
  • Eurasip Journal on Audio, Speech, and Music Processing
  • Joshua Mannall + 5 more

Outside of shoebox rooms, acoustic diffraction phenomena are present and can influence important aspects of auditory perception, such as localisation. A simple extension of a shoebox room is an L-shaped room as it introduces a single diffracting edge. This paper presents two experiments carried out in L-shaped rooms in virtual reality. The first investigated whether the inclusion of diffraction modelling influences the perceived plausibility of the acoustic simulation, and the second to what extent newly developed efficient IIR filter diffraction models are equally plausible to the physically accurate Biot-Tolstoy-Medwin-Svensson (BTMS) model. The study compared diffraction of only the direct sound and diffraction of both direct and reflected sound. The results show that the inclusion of diffraction increased the perceived plausibility of the acoustic simulation. A statistically significant increase in plausibility was found by the addition of diffracted reflection paths, but only in the so-called shadow zone. The second experiment determined that the IIR filter diffraction models were similarly plausible to BTMS in 14 of 18 cases with a threshold of 0.5 on a 6-point Likert scale.

  • Open Access Icon
  • Research Article
  • 10.1186/s13636-025-00434-1
A data-driven exploration of elevation cues in HRTFs: an explainable AI perspective across multiple datasets
  • Dec 15, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Juan A De Rus + 4 more

Abstract Precise elevation perception in binaural audio remains a challenge, despite extensive research on head-related transfer functions (HRTFs) and spectral cues. While prior studies have advanced our understanding of sound localization cues, the interplay between spectral features and elevation perception is still not fully understood. This paper presents a comprehensive analysis of over 600 subjects from 11 diverse public HRTF datasets, employing a convolutional neural network (CNN) model combined with explainable artificial intelligence (XAI) techniques to investigate elevation cues. In addition to testing various HRTF pre-processing methods, we focus on both within-dataset and inter-dataset generalization and explainability, assessing the model’s robustness across different HRTF variations stemming from subjects and measurement setups. By leveraging class activation mapping (CAM) saliency maps, we identify key frequency bands that may contribute to elevation perception, providing deeper insights into the spectral features that drive elevation-specific classification. This study offers new perspectives on HRTF modeling and elevation perception by analyzing diverse datasets and pre-processing techniques, expanding our understanding of these cues across a wide range of conditions.

  • Research Article
  • 10.1186/s13636-025-00439-w
Estimating depression and anxiety scores from conversational speech in females with and without comorbidity
  • Dec 13, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Aslı BeĹźirli + 2 more

  • Research Article
  • Cite Count Icon 1
  • 10.1186/s13636-025-00428-z
Sound and music biases in deep music transcription models: a systematic analysis
  • Dec 11, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Lukáš Samuel Marták + 2 more

  • Research Article
  • 10.1186/s13636-025-00437-y
Parametric virtual microphone techniques for sound field reconstruction with early reflection modeling
  • Dec 4, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Gioele Greco + 3 more

  • Research Article
  • Cite Count Icon 2
  • 10.1186/s13636-025-00436-z
AudioSet-tools: a Python framework for taxonomy-aware AudioSet curation and reproducible audio research
  • Dec 2, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Stefano Giacomelli + 3 more

  • Research Article
  • 10.1186/s13636-025-00432-3
Enhanced acoustic monitoring for industrial predictive maintenance using semi-coprime microphone array and joint spatial-frequency filtering
  • Nov 28, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Gonzalo Corral-GarcĂ­a + 5 more

  • Research Article
  • Cite Count Icon 1
  • 10.1186/s13636-025-00431-4
Speaker embedding loss for end-to-end speaker diarization without external embedding networks
  • Nov 21, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Jaehee Jung + 1 more

  • Research Article
  • Cite Count Icon 1
  • 10.1186/s13636-025-00430-5
Enhanced U-Net architectures for accurate room impulse response generation via differential-phase learning
  • Nov 17, 2025
  • EURASIP Journal on Audio, Speech, and Music Processing
  • Ignacio Martin-Salinas + 3 more