Intersession Variability in Speaker Recognition: A behind the Scene Analysis Daniel Garcia-Romero, Carol Y. Espy-Wilson Department of Electrical and Computer Engineering, University of Maryland, Collage Park, MD [email protected], [email protected]

Abstract The representation of a speaker’s identity by means of Gaussian supervectors (GSV) is at the heart of most of the state-of-the-art recognition s*ystems. In this paper we present a novel procedure for the visualization of GSV by which qualitative insight about the information being captured can be obtained. Based on this visualization approach, the Switchboard-I database (SWB-I) is used to study the relationship between a data-driven partition of the acoustic space and a knowledge based partition (i.e., broad phonetic classes). Moreover, the structure of an intersession variability subspace (IVS), computed from the SWB-I database, is analyzed by displaying the projection of a speaker’s GSV into the set of eigenvectors with highest eigenvalues. This analysis reveals a strong presence of linguistic information in the IVS components with highest energy. Finally, after projecting away the information contained in the IVS from the speaker’s GSV, a visualization of the resulting GSV provides information about the characteristic patterns of spectral allocation of energy of a speaker. Index Terms: speaker recognition, intersession variability, model compensation, intersession variability subspace.

1. Introduction The state-of-the-art in speaker recognition is mostly dominated by the use of Gaussian Mixture Models (GMMs) and Support Vector Machines (SVMs) as classifiers working with short-time cepstral features [1], [2]. A fundamental problem with this feature extraction approach is that not only is the speaker information captured, but also other sources of variability (e.g., channel characteristics, environment, speaker emotional state, linguistic content) which severely affect the performance of the recognition systems. These undesired sources of information are commonly referred to as a whole under the name of intersession variability. A wide variety of techniques have been proposed in the past ten years to address this issue (see [3] for an overview). Recently, featureand model-level intersession compensation techniques have been proposed in the form of factor analysis for GMMs [4]. These methods work by modeling the UBM-MAP adapted means of a speaker’s GMM using latent factors. The key idea behind these approaches is the construction of a Gaussian supervector by stacking the means of the mixture components. Since the MAP adaptation process only updates the means of the speaker’s GMM with respect to the UBM, all the discriminant information of the speaker captured by the GMM is contained in this high-dimensional GSV. In this way, a speaker’s utterance is represented by a point in a highdimensional vector space (i.e., dimension ~ 50k). Moreover, this space is assumed to be spanned by the direct sum of two

*

This research was supported by NSF grant # BCS-0519256

subspaces, namely the speaker information subspace (SIS) and the intersession variability subspace (IVS). Hence, any GSV belonging to the vector space has a unique decomposition in terms of speaker information components and intersession variability. An important property of the GSV paradigm is that it provides a fixed-length feature vector to represent variablelength speech utterances. Due to this property, a new class of recognizers has been proposed in which the UBM-MAP adapted GMM paradigm is used as a front-end and a SVM classifier is used in the back-end [2]. In this way, the representation of a speaker’s identity by means of GSV plays an important role in GMM systems, as a way to compensate for session variability, and also in SVM systems as input feature vectors. The excellent performance of ISV-compensated speaker recognition systems based on GSV motivates our interest in the structure of this representation space. It is for this reason that we propose a procedure to display GSV and use it to obtain a qualitative insight about the speaker information subspace and the intersession variability subspace.

2. Experimental setup 2.1. Switchboard-I database (SWB-I) The Switchboard-I database is comprised of conversational speech between two speakers recorded over landline telephone channels with a sampling rate of 8 KHz [5]. The average duration of each conversation is 5 minutes (approx. 2.5 min per speaker) and each conversation side is recorded in a different file. The total number of speakers in the database is 520 with a balance in gender and recorded into 4856 speech files. The telephone handsets were either electret or carbon bottom with an approximate proportion of 70% and 30% respectively. The availability of manual phonetic transcriptions [6] along with a fairly limited amount of channel/handset variability makes this database a good candidate for the experiments in this paper.

2.2. UBM-GMM system A UBM-GMM system with 2048 mixtures and diagonal covariance matrices [1] has been used throughout this work with two purposes. First to compute GSVs for the analysis presented in sections 3, 4 and 5. Second, to construct a GMM speaker recognition system to quantify the effects of intersession variability compensation.

3. Gaussian supervectors (GSVs) 3.1. GSV computation A Gaussian supervector is constructed by stacking the means of the mixture components of a MAP mean-adapted GMM from a UBM. In the following we provide a detailed

C_1 C_2 C_1 C_2

µ mix. 1 Cluster mean vectors µ mix. 2048

Matrix of mean column vectors: 19 x 2048

Number of clusters

C_N

C_N

Pseudo inverse each column

Matrix of clustered mean column vectors: 19 x 2048

GSV (19 x 2048) x 1

Matrix of clustered FFT coeffs: 128 x 2048

Figure 1: Diagram of the steps followed to construct a meaningful visualization of a GSV. description of the parametrization used as well as the construction of the UBM. Each file in the database was parameterized into a sequence of 19-dimensional MFCC vectors using a 20ms Hamming window with a 10ms shift. The MFCC vectors were computed using a simulated triangular filterbank on the FFT spectrum. Prior to projecting the Mel-frequency band (MFB) energies into a DCT basis, bandlimiting was performed by discarding the filterbank outputs outside of the frequency range 300Hz-3138Hz. Finally, after projecting the MFB energies into a DCT basis and discarding C0, the 19MFCC vectors were processed with RASTA filtering to reduce linear channel bias effects. No delta features were computed since we wanted to focus our analysis on static information only. Two UBMs were trained based on a partition of SWB-I into two sets, P1 and P2, of 260 speakers each with a balance in gender and handset type. The UBM trained on P1 was used to obtain GSVs for the files in P2 and vice versa. The resulting dimension of the GSV’s was 2048 x 19 = 38912.

MFB pseudo inverse

DCT-1

19 MFCC

20 MFB energies

128 FFT coeffs

Figure 2: Pseudo-inversion of the MFCC coefficients back to an approximation of the original FFT coefficients.

3.2. GSV visualization The speech technology community has greatly benefited from the ability to visualize spectro-temporal representations of the speech signal. A trained eye can gain a lot of qualitative insight by a simple inspection of a spectrogram. Unfortunately, what has proven very useful for information displaying (i.e., temporal sequences of FFT coefficients) is not optimal for other task unless further post-processing is applied. In the particular case of speaker recognition, examples of such post-processing include high-frequency resolution decrease, projection into orthogonal basis and dimensionality reduction. These standard signal processing techniques have tremendously improved the performance of the recognition systems. However, once the information has been processed in this way, it is extremely hard to make sense of what is really happening. One way to cope with this issue is to obtain a useful representation for the task at hand (i.e., speaker recognition) and then try to transform such

representation to a domain in which qualitative knowledge can be obtained. In this way, Figure 1 shows a diagram in which a GSV is transformed into a matrix of clustered sets of FFT coefficients. The transformation process starts by reshaping the GSV into a matrix with each mixture mean as a column. Subsequently, a number of clusters is specified and the mean vectors are grouped together by a simple K-means algorithm. As a result, the mean vectors corresponding to the Gaussian mixtures that are close together (i.e., Euclidean sense) in the acoustic space are clustered together. Up to this point, no meaningful transformation has being accomplished. The key of the process lies in the next step that we have denoted as “pseudo-inversion”. The reason for this name will become apparent in a few lines. Figure 2 depicts the steps followed in the pseudo-inversion. It is clear that it attempts to inverse the orthogonalization of the DCT basis as well as the effect of the simulated triangular filterbank. However, since we dropped the C0 coefficient in the computation of the 19 MFCCs, the result of the DCT inversion will be a vector of 20 MFB normalized energies. Moreover, the triangular filterbank processing is not an invertible mapping since it is many-to-one. It is for this reason that the prefix pseudo is attached to this inversion process. Hence, the pseudo-inversion of this process is performed by constructing a matrix whose columns are the weights of each one of the triangular filters (i.e., dimensions 128 x 20). Finally, it is important to note that since the spectrum was bandlimited during the feature extraction process, the resulting FFT coefficients only span the frequency range 300Hz-3138Hz. The second panel of Figure 3 shows the result of processing the GSV of the UBM of the partition P1 of SWB-I.

4. Relation between acoustic clustering and broad phonetic classes A GSV can be understood as a summary of the average patterns of spectral allocation of energy of a particular speaker. However, the linguistic content of the speech signal imposes some constraints in these patterns (e.g., relative position of formants). In this way, it seems natural to think that the elements (i.e., mean vectors) of the GSV will exhibit some kind of clustering. To check this, a simple K-means procedure was used to partition the elements of the two UBMs GSVs into a set of classes. The Euclidean distance between the mean vectors (i.e., 19 MFCC vectors) was used and the number of classes was set to 16. We followed the visualization methodology of Figure 1 to display the GSV of the UBM for P1. The second panel of Figure 3 shows the result. The same behavior was observed for the UBM of P2.

(a)

(b)

(c)

(d)

Clusters of Gaussian mixture’s mean vectors

(e)

Figure 3: (a) Broad-phonetic class alignment with the data-driven partition of the acoustic space. (b) GSV of the UBM used for partition 1 of SWB-I. (c) GSV of a female speaker adapted from the UBM. (d) Projection of the speaker’s GSV into the intersession variability subspace. (e) Projection of the speaker’s GSV into the orthogonal complement of the IVS.

Also, it is important to note, that the clustering was done prior to the pseudo-inversion stage and therefore no imprecision was introduce in the process. A quick inspection of the UBM’s GSV reveals that the mean vectors that get grouped together share in common their most predominant regions of spectral allocation of energy (i.e., formants). This raises the following question: Is there any relationship between a data-drive partition of the acoustic space and a knowledge-based partition such as the broad phonetic classes? In order to answer this question the following experiment was conducted in each of the SWB-I partitions independently. First, for each file, the manual phonetic transcriptions of SWB-I [5] were used to align each feature vector with a broad phonetic class. The following set of phonetic classes was used: {liquids (LIQ), nasals (NAS), voiced/unvoiced fricatives (V/U-F), voiced/unvoiced stops (V/U-S), diphthongs (DIP) and back/center/front vowels (B/C/F-V)}. Then a probabilistic alignment of each feature vector with their corresponding UBM was performed. Only the top-1 scoring mixture was used for each feature vector. During this process, we kept track of the number of times each one of the 2048 UBM’s mixtures was used to score a frame with a particular phonetic class label. As a result, we obtained a probabilistic alignment of each UBM mixture with the aforementioned set of broad phonetic classes. As an

example, if a given mixture was used 80% of the time to score frames in nasal regions, the process would assign a 0.8 probability mass to that mixture with respect to nasals. Two important observations were made. First, every mixture had a non-zero probability mass for each broad phonetic class. Second and most important, the probability mass was not uniformly distributed and was highly concentrated on one or two phonetic classes for each mixture. Moreover, in order to establish a connection between the data-driven clusters and the broad phonetic classes we averaged the probabilistic assignments among all the mixtures in the same data-driven cluster. The top panel of Figure 3 shows the result of thresholding this averaged probabilistic alignment to keep approximately 90% of the probability mass. Each data-driven cluster gets aligned with at most 2 broad phonetic classes. After a close analysis of the resulting pairings between data-driven clusters and phonetic classes the authors believe that there is a good matching between the formant regions of the clusters and the canonical formant regions of the phonetic classes (see [7] for examples of these). The reader is encouraged to perform its own comparison. Due to the lack of space, Figure 3 only depicts the results for the partition P1 of SWB-I. However, the same

observations were made by analyzing the results of P2. This supports the generality of our results. Based on the experiments presented in this section we can claim that not only GSVs capture average patterns of spectral allocation of energy, but also a phonetic meaning can be attached to partitions of the GSV.

of the original GSV. The improved discriminative power of this representation seems to indicate that this less smooth GSV captures the most idiosyncratic characteristics of the speaker’s spectral allocation of energy. Further analysis of this claim will be conducted in future work by contrasting the GSVs of different speakers.

5. GSV space structure As a motivation for our interest in the structure of the IVS, we run the MIT-LL UBM-GMM recognition system [1] on SWB-I to quantify the effects of intersession compensation. The same protocol used in the Super-SID evaluation was followed [6]. We focused our attention on the 1conv-1conv recognition task. Each speaker model was trained with approximately 2.5 minutes of speech and tested against segments of similar duration. The total number of trials was 16280 from which around 10% were non-target trials. The solid line of Figure 4 shows the performance for the baseline system for which no intersession compensation was used. The methodology presented in [8] was used to compute the IVS and also to constrain the adaptation of the speaker models to the test files into the intersession subspace. The data in partition P1 of SWB-I was used for the IVS of P2 and vice versa. Table 1 shows the influence of the dimension of the IVS in the recognition performance in terms of Equal Error Rate (EER). The best results were obtained for a 32dimensional IVS. This number of dimensions is a little bit smaller than those reported for other databases with more sources of intersession variability [3]. The dashed line of Figure 4 shows an impressive improvement of more than 50% in the performance of the intersession variability compensated system with respect to the baseline. Table 1. Influence of the dimensionality of the intersession variability subspace in recognition performance. # Dim. IVS EER (%)

2 4.61

4 3.61

8 3.26

16 3.01

32 2.89

64 3.16

128 3.45

The great improvement obtained indicates that projecting away the components of the IVS from the speaker’s GSVs result in a much more discriminative representation. However, the exact composition of the sources of variability removed from the GSVs is still an unexplored area. In order to get a first look into this issue, we selected a female speaker from the partition P1 and obtained her GSV by MAP adaptation of the means of the corresponding UBM. Panel (c) of Figure 3 shows the result. It is interesting to observe how similar to the UBM’s GSV the speaker’s GSV remains after adaptation. A more surprising result is the one obtained by projecting the speaker’s GSV into a 32dimensional IVS (panel (d) in Figure 3). This projection resembles a smoothed version of the GSV in which the location of the formants becomes less defined. Therefore, the components of the projection into the IVS seem to mostly capture the variability in the spectral allocation of energy due to linguistic constraints. To this point, it is still not clear to the authors how the channel/handset variability of SWB-I is reflected in this representation. Further analysis will be carried out in the future to try to understand this issue. Finally, by projecting away the IVS components from the GSV (see bottom panel in Figure 3), the most discriminative representation of the speaker’s identity is obtained. A lot of the smoothness of the GSV is removed. The components of the compensated GSV seem much more irregular than those

Figure 4: DET curves for the GMM-UBM baseline system (19 MFCC + RASTA) with no intersession compensation and projecting away a 32 dimensional IVS (dashed line).

6. Conclusions A procedure to display GSVs was introduced in this paper. The potential benefits of this visualization tool were explored by using it as a visual aid to establish a probabilistic relation between a partition of the acoustic space and a set of broad phonetic classes. A meaningful relation was obtained that reveals useful information about the components of the GSV. Moreover, a qualitative analysis of the structure of the speaker information subspace and the intersession variability subspace provided a first look into the actual composition of the sources of variability present in the IVS.

7. References [1]

[2]

[3]

[4] [5]

[6]

[7]

[8]

D. A. Reynolds, T. F. Quatieri, and R. B. Dunn, “Speaker verification using adapted Gaussian mixture models,” Digital Signal Process., vol.10, pp. 181–202, 2000. W. M. Campbell, et al., “Support Vector Machines using GMM supervectors for speaker verification,” IEEE Signal Process. Letters, vol. 13, no. 5, pp. 308–311, May 2006. L. Burget, et al., “Analysis of feature extraction and channel compensation in a GMM speaker recognition system,” Trans. on Audio, Speech and Language Processing, vol. 15, no. 7, pp. 1987-1998, September 2007. P.Kenny, et al., “Factor analysis simplified,” in Proc. ICASSP, Philadelphia, PA, vol. 1, pp. 637–640, Mar. 2005. J. Godfrey, E. Holliman and J. McDaniel, “SWITCHBOARD: telephone speech corpus for research and development”, in Proc. ICASSP, San Francisco, California, pp. 517-520, 1992. D. A. Reynolds, et al., “The SuperSID project: exploiting highlevel information for high-accuracy speaker recognition”, in Proc. ICASSP, pp. 784–787, 2003. J. P. Olive, A. Greenwood and J. Coleman, “Acoustics of American English speech: a dynamic approach”, Springer, May, 1993. J. Deng, T. Fang Zheng and W. Wu, “Session variability subspace projection based model compensation for speaker verification”, in Proc. ICASSP, Honolulu, Hawaii, April 2007.

Intersession Variability in Speaker Recognition

Switchboard-I database (SWB-I) is used to study the ... subspace (IVS), computed from the SWB-I database, is ... under the name of intersession variability.

168KB Sizes 6 Downloads 264 Views

Recommend Documents

Speaker Recognition in Adverse Conditions
parametric density estimation methods are generally known to be superior when .... Application of a threshold to perform the verification decision creates to two ...

Speaker Recognition in Adverse Conditions
cation is performed in the presence of little or no interference. In general, these systems ..... ing the Gaussian kernel is that it does not have finite support. However, for ..... AT&T Bell Labs in Murray Hill, New Jersey, and also for the. Speech 

Speaker Recognition Final Report - GitHub
Telephone banking and telephone reservation services will develop ... The process to extract MFCC feature is demonstrated in Figure.1 .... of the network. ..... //publications.idiap.ch/downloads/papers/2012/Anjos_Bob_ACMMM12.pdf. [2] David ...

Speaker Recognition in Two-Wire Test Sessions
The authors would like to thank Oshry Ben-Harush, Itshak. Lapidot, and Hugo Guterman for providing automatic speaker diarization labels, and Reda Dehak, Najim Dehak, and David van Leeuwen for their assistance with the NAP-SVM system. 8. References. [

SPEAKER-TRAINED RECOGNITION USING ... - Vincent Vanhoucke
advantages of this approach include improved performance and portability of the ... tion rate of both clash and consistency testing has to be minimized, while ensuring that .... practical application using STR in a speaker-independent context,.

Speaker Recognition in Two-Wire Test Sessions - Semantic Scholar
techniques both in the frame domain and in the model domain. The proposed .... summed the two sides of the 4w conversations to get the corresponding 2w ...

Efficient Speaker Recognition Using Approximated ...
metric model (a GMM) to the target training data and computing the average .... using maximum a posteriori (MAP) adaptation with a universal background ...

Speaker Recognition in Encrypted Voice Streams
Alexi, Diana, Vitaly, Daniel, Elena, Gancho, Anna, Daniela for never giving up on me and supporting me in the past several months. Last (and least), I would like to thank my gym for the long opening hours, Trey Parker and Matt Stone for the philosoph

Fusion of heterogeneous speaker recognition systems in ... - CiteSeerX
Speech@FIT, Faculty of Information Technology Brno University of Tech- nology, Czech Republic. ... The STBU consortium was formed to learn and share the tech- nologies and available ...... as Associate Professor (Doc.) and Deputy Head of.

Speaker Recognition in Two-Wire Test Sessions - Semantic Scholar
system described in [7]. ... virtual 2w training method described in Section 5. ... 7.0. (0.0328). 3.2. Feature warping. Feature warping is the process of normalizing ...

Speaker Recognition in Encrypted Voice Streams
Transmitting voice communication over untrusted networks puts personal ... technical challenges in the design and analysis of the systems and programs ...... Administration of the President of the Russian Federation. ... Master's thesis, Cleve-.

Fusion of heterogeneous speaker recognition systems in ... - CiteSeerX
tium of 4 partners: Spescom DataVoice (South Africa), TNO .... eral frame selection criteria were applied: (a) frame energy must be more than than ..... UBM sources. S99–S03 ...... tant basic workhorses in speaker recognition, but alternative.

Speaker Recognition in Two-Wire Test Sessions - CiteSeerX
cheating experiment by replacing each 2w session with a concatenation of its two 4w sides (in the audio domain). For the GMM system, we received an EER of ...

SPEAKER-TRAINED RECOGNITION USING ... - Vincent Vanhoucke
approach has been evaluated on an over-the-telephone, voice-ac- tivated dialing task and ... ments over techniques based on context-independent phone mod-.

Automatic Speech and Speaker Recognition ... - Semantic Scholar
7 Large Margin Training of Continuous Density Hidden Markov Models ..... Dept. of Computer and Information Science, ... University of California at San Diego.

Automatic speaker recognition using dynamic Bayesian network ...
This paper presents a novel approach to automatic speaker recognition using dynamic Bayesian network (DBN). DBNs have a precise and well-understand ...

THE UMD-JHU 2011 SPEAKER RECOGNITION SYSTEM D Garcia ...
presence of reverberation and noise via the use of frequency domain perceptual .... best capture the speaker and the channel variabilities in supervector space.

Approaches to Speech Recognition based on Speaker ...
best speech recognition submissions in its Jan- ... ity such as telephone type and background noise. ... of a single vector to represent each phone in context,.

Speaker Recognition using Kernel-PCA and ...
Modeling in common speaker subspace (CSS). The purpose of the projection of the common-speaker subspace into Rm using the distance preserving ...

THE UMD-JHU 2011 SPEAKER RECOGNITION SYSTEM D Garcia ...
University of Maryland, College Park, MD, USA. 2 ..... (FL), FDLP-mel (FM) and Cortical (CC) features for the NIST SRE 2010 extended core data set. For the ...

Speaker Recognition using Kernel-PCA and ... - Semantic Scholar
[11] D. A. Reynolds, T. F. Quatieri and R. B. Dunn, "Speaker verification using adapted Gaussian mixture models,". Digital Signal Processing, Vol. 10, No.1-3, pp.