Please use this identifier to cite or link to this item:
Title: Fusion of bottleneck, spectral and modulation spectral features for improved speaker verification of neutral and whispered speech
Authors: Sarria Paja, Milton
Falk, Tiago H.
Keywords: Whispered speech;Speaker verification;Modulation spectrum;Deep neural networks;Bottleneck features
Issue Date: 27-Jul-2018
Publisher: Elsevier B.V.
Abstract: Speech based biometrics is becoming a preferred method of identity management amongst users and companies. Current state-of-the-art speaker verification (SV) systems, however, are known to be strongly dependent on the condition of the speech material provided as input, and can be affected by unexpected variability presented during testing, such as with environmental noise or changes in vocal effort. In this paper, SV using whispered speech is explored, as whispered speech is known to be a natural speaking style with reduced perceptibility but containing relevant information regarding speaker identity and gender. We propose to fuse information from spectral, modulation spectral and so-called bottleneck features computed via deep neural networks at the feature- and score-levels. Bottleneck features have been recently shown to provide robustness against train/test mismatch conditions and have yet to be tested for whispered speech. Experimental results showed that relative improvements as high as 79% and 60% could be achieved for neutral and whispered speech, respectively, relative to a baseline system trained with i-vectors extracted from mel frequency cepstral coefficients. Results from our fusion experiments, show that the proposed strategies allow to efficiently use the limited resources available and to result in whispered speech performance inline with that obtained with normal speech.
ISSN: 01676393
Appears in Collections:Artículos Científicos

Files in This Item:
File Description SizeFormat 
Fusion of bottleneck, spectral and modulation spectral features.jpg163,6 kBJPEGView/Open

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.