Abstract
In this paper, the explicit and implicit modelling of the subsegmental excitation information are experimentally compared. For explicit modelling, the static and dynamic values of the standard Liljencrants–Fant (LF) parameters that model the glottal flow derivative (GFD) are used. A simplified approximation method is proposed to compute these LF parameters by locating the glottal closing and opening instants. The proposed approach significantly reduces the computation needed to implement the LF model. For implicit modelling, linear prediction (LP) residual samples considered in blocks of 5 ms with shift of 2.5 ms are used. Different speaker recognition studies are performed using NIST-99 and NIST-03 databases. In case of speaker identification, the implicit modelling provides significantly better performance compared to explicit modelling. Alternatively, the explicit modelling seem to be providing better performance in case of speaker verification. This indicates that explicit modelling seem to have relatively less intra and inter-speaker variability. The implicit modelling on the other hand, has more intra and inter-speaker variability. What is desirable is less intra and more inter-speaker variability. Therefore, for speaker verification task explicit modelling may be used and for speaker identification task implicit modelling may be used. Further, for both speaker identification and verification tasks the explicit modelling provides relatively more complimentary information to the state-of-the-art vocal tract features. The contribution of the explicit features is relatively more robust against noise. We suggest that the explicit approach can be used to model the subsegmental excitation information for speaker recognition.
Full Text
Topics from this Paper
Explicit Modelling
Speaker Identification
Inter-speaker Variability
Implicit Modelling
Liljencrants–Fant
+ Show 5 more
Create a personalized feed of these topics
Get StartedTalk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Similar Papers
arXiv: Sound
Oct 12, 2020
The Journal of Contents Computing
Jun 30, 2020
Tectonophysics
Nov 1, 2020
Dec 1, 2015
Jan 10, 2021
Software: Practice and Experience
Sep 1, 1997
Journal of Coastal Research
Mar 3, 2022
EURASIP Journal on Audio, Speech, and Music Processing
Nov 25, 2020
Journal of Chemical Theory and Computation
Nov 24, 2015
IEEE Transactions on Audio, Speech, and Language Processing
Feb 1, 2013
Fire Safety Journal
Apr 1, 2015
International Journal of Speech Technology
Mar 1, 2013
Journal of the Japanese Association for Petroleum Technology
Jan 1, 1995
Forensic Science International
May 10, 2006
Sadhana
Sadhana
Apr 4, 2017
Sadhana
Mar 14, 2016
Sadhana
Mar 12, 2016
Sadhana
Mar 9, 2016
Sadhana
Mar 9, 2016
Sadhana
Mar 3, 2016
Sadhana
Feb 18, 2016
Sadhana
Feb 11, 2016
Sadhana
Feb 9, 2016