Speaker Verification Using Fisher Vector Yao Tian1 , Liang He1 , Zhi-yi Li1 , Wei-lan Wu2,3 , Wei-Qiang Zhang1 , Jia Liu1 1

National Laboratory for Information Science and Technology, Department of Electronic Engineering, Tsinghua University 2 State Key Laboratory of Transducer Technology, Institute of Electronics, Chinese Academy of Sciences 3 University of Chinese Academy of Sciences [email protected]

Abstract This paper introduces an approach based on Fisher vector feature representation for speaker verification. The Fisher vector is originated from Fisher Kernel and represents each utterance as a high-dimensional vector by encoding the derivatives of the loglikelihood of the UBM model with respect to it’s mean and variances. This representation captures the average first and second order differences between the utterance and each of the Gaussian centers of the UBM model. And the Fisher vector is further projected to a low-dimensional space using PPCA which is conducted in a similar way of factor analysis. We compare the proposed method with the state-of-art i-vector approach on the telephone-telephone condition of NIST SRE2010 female and male core task. The experimental results indicate that the proposed Fisher vector based method is competitive with i-vector. It can also provide complementary information to i-vector and the fusion of these two approach obtains a relative improvement of 11.8% and 14.7% in EER and 9.2% and 2.7% in minDCF for female and male than i-vector alone. Index Terms: speaker verification, i-vector, Fisher vector

1. Introduction The core task of speaker verification is to decide whether two utterances belong to the same person. The Gaussian Mixture Models-Universal Background Models(GMM-UBM)[1] lay the foundation of modeling speaker space and many approaches based on GMM-UBM framework has been proposed to improve the performance of speaker verification including Support Vector Machine(SVM)[2][3], Joint Factor Analysis(JFA)[4][5], total variability factor analysis(i-vector)[6] and etc. The classical JFA try to model the speaker and session variability by defining two distinct subspaces: the speaker space and the channel space. In contrast, the i-vector which originates from JFA aims at modeling both the speaker and channel space by generating a low-dimensional space named total variability space since the channel space estimated using JFA also contain information about speakers[6]. Then each utterance is represented as a compact, fixed-length vector called i-vector in this low-dimensional space. Intersession compensation techniques, such as Linear Discriminant Analysis(LDA) and Within Class Covariance Normalization(WCCN) are further carried out in the total variability space[6] as opposed to the high-dimensional GMM supervector space[4] to improve the performance of speaker verification. Recently, the Fisher vector representation[7] based on GMM has been successfully used in image classification[8][9] and

face verification[10]. The local descriptors extracted from large number of images are used to train a Gaussian Mixture Model firstly which is similar to the process of building UBM in speaker verification. Then each image is represented as a highdimensional vector called Fisher vector by encoding the derivatives of the log-likelihood of the GMM model with respect to it’s mean and variances using local descriptors[10]. This leads to the representation which captures the average first and second order differences between the features and each of the GMM centers. In this study, we propose a novel feature representation method by using Fisher vector for speaker verification. And we further project it to a low-dimensional space using PPCA[11][12] which is conducted in a similar way as the factor analysis in the i-vector model approach. There are three motivations for our work: Firstly, Fisher vector representation takes data’s higher-order(second-order) moments into consideration directly while i-vector doesn’t. Secondly, Fisher vector models each utterance in a different perspective compared with i-vector approach which may provide complementary information. Thirdly, kernel methods had been wildly discussed in speaker verification area[2][3], but most of them remained at the supervector-level and it was not quite convenient to do verification as the amount of data grows dramatically nowadays. The rest of this paper is organized as followed. in Section2, we give a review of the background of i-vector modeling, the approach of LDA, WCCN and the cosine distance scoring classifier. Section3 presents the process of extracting Fisher vector and dimension reduction method. Experimental setup and the results are showed in Section4. Finally, conclusions and future directions are discussed in Section5.

2. Theoretical Background 2.1. i-vector Inspired by the classical Joint Factor Analysis(JFA)[4][5] model which is based on speaker and channel factors separately, Dehak et al.[6] proposed the i-vector based speaker verification technique which defines a compact low-dimensional space called total variability space. In total variability space modeling, no distinction is made between the speaker effects and channel effects in GMM supervector space since experiments showed that the channel space estimated using JFA also contains information about speakers[6]. Given an utterance, the speaker-and-channel dependent G-

2.3. Classifier

MM supervector can be written as: M = m + Tω

(1)

where m is the speaker-and-channel independent component which can be taken to be the UBM supervector. T is a rectangular matrix of low rank representing the primary directions of the total variability space. ω is a random vector having a standard distribution N (0, I) and the i-vector is obtained as the MAP point estimate of ω. The process of training the the matrix T is the same as the training of eigenvoice in JFA except for that all the utterances from the same speaker is regarded as produced by different speakers. The i-vector modeling can be seen as a simple factor analysis, which allows us to represent an utterance by a compact, low-dimensional single factor. 2.2. Intersession Compensation After the extraction of i-vector, the intersession compensation can be carried out in the total variability space. In our experiments, we use linear discriminant analysis(LDA) and within class covariance normalization(WCCN) for intersession compensation. 1)Linear discriminant analysis: linear discriminant analysis(LDA)[13] is a technique for dimensionality reduction that is widely used in the field of pattern recognition. It tries to seek new orthogonal axes that maximizing between-class variance and minimizing intra-class variance. All the utterances from the same speaker are labeled as an individual class. The LDA optimization problem can be defined by maximizing the following Fisher criteria: J(v) =

v t Sb v v t Sw v

(2)

Where v is the projection direction. Sb and Sw are the betweenclass covariance and within-class covariance defined as: Sb =

S X

(¯ ωs − ω ¯ )(¯ ωs − ω ¯)

(3)

ns S X 1 X s (ωi − ω ¯ s )(ωis − ω ¯ s )t n s s=1 i=1

(4)

S is the number of speakers, ω ¯ s is the mean i-vectors for P of s s speaker s can be computed as ω ¯ s = (1/ns ) n i=1 ωi . ns is the number of i-vectors for speaker s. ω ¯ is the speaker population mean. The projection matrix V is obtained by stacking the top N −1 eigenvectors of the general matrix Sw Sb . 2)Within class covariance normalization:Within class covariance normalization(WCCN) tries to use the within-class covariance to normalize the cosine kernel function in order to compensate for intersession variability[14]. All utterances from the same speaker are labeled as an individual class. The withinclass covariance matrix is denoted as: W =

score(ωtarget , ωtest ) =

hωtarget , ωtest i kωtarget k kωtest k

ns S 1X 1 X s (ωi − ω ¯ s )(ωis − ω ¯ s )t S s=1 ns i=1

(5)

S is the number of speakers, ω ¯ s is the mean i-vectors for P of s s speaker s and is computed as ω ¯ s = (1/ns ) n i=1 ωi . ns is the number of i-vectors for speaker s. The feature mapping matrix L can be obtained through a Cholesky decomposition of the 0 matrix W −1 = LL .

(6)

3. Fisher Vector Fisher vector is derived from the Fisher Kernel introduced by Jaakkola and Haussler[7] and applied successfully to image classification[8][9] and face verification[10] lately. Let X = {xt , t = 1...T } be the set of frames of an utterance and the dimension of xt is denoted as d. They are modeled by UBM uλ with parameters λ = {ωk , µk , σk }k , where k is the mixture index, and ωk , µk , σk are the mixture weights, means, and diagonal covariances of the UBM. X can be described by the gradient vector[7]: GX λ =

1 ∇λ log uλ (X) T

(7)

The gradient of the log-likelihood describes the direction in which the data parameters should be modified. A natural kernel between two utterances X and Y on these gradients is[7] 0

−1 Y Kernel(X, Y ) = GX λ Fλ Gλ

(8)

Here Fλ is the Fisher information matrix of uλ and is denoted as: (9) Fλ = Ex∼uλ [∇λ log uλ (x)∇λ log uλ (x)0 ] With Cholesky decomposition, Fλ can be denoted as Fλ = 0 Lλ Lλ . Thus Kernel(X, Y ) can be rewritten as a dot-product between normalized vectors φ: X φX λ = Lλ Gλ

t

s=1

Sw =

Cosine distance scoring: In an i-vector based speaker verification system, cosine distance scoring[6] has been proven to be a fast and efficient way of measuring the similarity of a pair of i-vectors. The cosine distance between ωtarget and ωtest is:

(10)

Here φX λ is referred as Fisher vector which encodes the derivatives of the log-likelihood of the UBM model with respect with its parameters[7]. Similar to the supervector derived from KL divergence[3], the Fisher vector is also generated from a kernel but it can take data’s higher-order(second-order) moments into consideration. In this paper, we use Fisher vector as a novel feature representation method for speaker verification and only consider the derivatives with respect to the means and variances since the derivatives to weights doesn’t bring additional improvements in our experiments. They are computed as follows:   T xt − µk 1 X γt (k) (11) φX √ µ,k = T ωk t=1 σk φX σ,k =

  T X (xt − µk )2 1 √ γt (k) − 1 σk2 T 2ωk t=1

(12)

Here, γt (k) is the soft assignment weight of the t-th frame xt to the k-th Gaussian component ukλ and is computed as: γt (k) =

ωk ukλ (xt ) K P i=1

(13)

ωi uiλ (xt )

Fisher vector φX λ is then obtained by concatenating the derivaX X X X tives of each Gaussian as: φX λ = [φµ,1 , φσ,1 , ..., φµ,K , φσ,K ].

After the extraction of Fisher vector, we normalized each Fisher vector to unit length by L2 normalization which can be seen as a gaussianization step to satisfy the gaussian assumption in PPCA model. The dimension of Fisher vector is 2Kd which is usually very high(39936 for K=512 and d = 39). To further compare it’s performance with the i-vector approach, we use the Probabilistic Principle Component Analysis(PPCA) to project the Fisher vector to a low-dimensional space denoted T 0 . Actually, the i-vector modeling can be considered as a classical PPCA model[12]. In this study, the PPCA projection matrix is conducted with EM algorithm in a similar way as the i-vector approach. After the projection, the intersession compensation methods mentioned above are used as well. For convenience, the Fisher vector in the Experiments Section denotes the projected Fisher vector.

4. Experiments 4.1. Experimental setup The experiments were carried out on the telephone-telephone condition of the NIST SRE2010 female and male core task. The equal error rate(EER) and the minimum decision cost function(minDCF) were selected as metrics for evaluation[15]. 12 Mel Frequency Cepstral Coefficients(MFCC) together with log energy were extracted using a 20ms Hamming window and 10ms frame shift. This 13 dimensional feature vector was subjected to feature warping, using a 3s sliding window. Then Delta and delta-delta coefficients were calculated to produce final 39-dimensional feature vectors. The training data included Swb(Switchboard) II, Swb cellular, NIST SRE 2004, 2005, 2006 and 2008 corpora. The genderdependent UBM was trained with diagonal covariance and 512 mixture components. The dimension of i-vector and the Fisher vector were set to 400. The dimension of LDA was set to 200. The cosine distance scoring(CDS) was adopted as classifier to generate the verification score. The detailed usage of corpora to train the UBM, i-vector matrix, Fisher vector matrix, LDA, WCCN were listed in Table 1. Table 1: Corpora used to estimate the UBM, i-vector matrix T, Fisher vector matrix T’, LDA, WCCN. UBM Swb II Swb cellular NIST04 NIST05 NIST06 NIST08



T √ √ √ √ √ √

T’ √ √ √ √ √ √

LDA √ √ √ √ √ √

WCCN √ √ √ √

in minDCF compared with i-vector for female and male. The DET curves of i-vector system, Fisher vector system and fusion result are presented as Figure 1. In Table 3, the system performance of i-vector and Fisher vector using LDA and WCCN on female and male are compared. The fusion result is presented as well. It can be seen from Table 3 when LDA and WCCN is used, the i-vector approach outperforms the Fisher vector method. But the fusion of these two methods still show improvements over both EER and minDCF. the relative improvement of EER and minDCF is 11.8% and 9.2% for female, as well as 14.7% relative improvement in EER and 2.7% in minDCF for male. The DET curves of i-vector system, Fisher vector system and fusion result are presented as Figure 2. From the results above we can see that the Fisher vector approach is competitive with i-vector. It can also provide complementary information to i-vector and the fusion of these two systems further improve the performance of speaker verification. Another interesting observation is that that the Fisher vector approach is inferior to the i-vector approach while intersession compensation is applied. We thought that L2 normalization at the supervector-level helps gaussianizing Fisher vector while breaking some utterances’(short duration) structure. We’ll investigate this situation in our later work.

Table 2: EER(%) and minDCF of i-vector system, Fisher vector system and fusion result without intersession compensation on the NIST SRE2010 female and male tel-tel core condition. system i-vector Fisher vector score fusion

Data Set female male female male female male

EER(%) 7.12 7.08 6.76 6.82 6.20 6.00

minDCF 0.569 0.562 0.561 0.570 0.521 0.548

Table 3: EER(%) and minDCF of i-vector system, Fisher vector system and fusion result with intersession compensation(LDA, WCCN) on the NIST SRE2010 female and male tel-tel core condition. system i-vector Fisher vector score fusion

Data Set female male female male female male

EER(%) 4.48 3.41 4.79 3.69 3.95 2.91

minDCF 0.468 0.446 0.465 0.443 0.425 0.434

4.2. Experimental results In Table 2, we give the performance of the state-of-art ivector based system and our proposed Fisher vector based system without any intersession-compensation process on NIST SRE2010 female and male tel-tel core condition(condition 5). The score fusion result is also presented. From the result we can see that the EER(%) of Fisher vector approach outperforms i-vector. The fusion of these two approach show consistent improvement in both EER and minDCF where the relative improvement is 12.9% and 15.3 % in EER and 8.4% and 2.5%

5. Conclusions In this paper we propose a new approach based on Fisher vector feature representation to speaker verification. This feature represents each utterance in a different perspective compared with i-vector by using the derivatives of the log-likelihood of the UBM’s means and variances which takes both uterrances’ firstorder and second-order moments into consideration. Experimental results on NIST SRE 2010 female and male telephonetelephone core condition demonstrate the effectiveness of the

i−vector Fisher vector fusion i−vector+LDA+WCCN Fisher vector+LDA+WCCN fusion

80

60

40

40 Missing Rate (%)

Missing Rate (%)

60

ivector Fisher vector fusion ivector+LDA+WCCN Fisher vector+LDA+WCCN fusion

80

20 10 5

20 10 5

2

2

1 0.5

1 0.5

0.2 0.1

0.2 0.1

0.10.2 0.5 1 2

5 10 20 40 False Alarm Rate (%)

60

80

0.10.2 0.5 1

2

5 10 20 40 False Alarm Rate (%)

60

80

Figure 1: The DET curve of i-vector system, Fisher vector system and fusion result with and without intersession compensation(LDA,WCCN) on the NIST SRE2010 female tel-tel core condition.

Figure 2: The DET curve of i-vector system, Fisher vector system and fusion result with and without intersession compensation(LDA,WCCN) on the NIST SRE2010 male tel-tel core condition.

proposed method. The fusion result improve the state-of-art ivector approach with a relative 11.8% and 14.7% in EER and 9.2% and 2.7% in minDCF for female and male respectively. The fusion result outperforms the state-of-art i-vector approach. In future work, we would like to test the Fisher vector method on interview conditions.

for speaker verification,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 4, pp. 788– 798, 2011. Tommi Jaakkola, David Haussler, et al., “Exploiting generative models in discriminative classifiers,” Advances in neural information processing systems, pp. 487–493, 1999. Florent Perronnin and Christopher Dance, “Fisher kernels on visual vocabularies for image categorization,” in IEEE Conference on Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE, 2007, pp. 1–8. Florent Perronnin, Jorge S´anchez, and Thomas Mensink, “Improving the fisher kernel for large-scale image classification,” in Computer Vision–ECCV 2010, pp. 143–156. Springer, 2010. Karen Simonyan, Omkar M Parkhi, Andrea Vedaldi, and Andrew Zisserman, “Fisher vector faces in the wild,” in Proc. BMVC, 2013, vol. 1, p. 7. Michael E Tipping and Christopher M Bishop, “Probabilistic principal component analysis,” Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol. 61, no. 3, pp. 611–622, 1999. Yun Lei and John HL Hansen, “Speaker recognition using supervised probabilistic principal component analysis.,” in INTERSPEECH, 2010, pp. 382–385. Peter N. Belhumeur, Jo˜ao P Hespanha, and David Kriegman, “Eigenfaces vs. fisherfaces: Recognition using class specific linear projection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 711–720, 1997. Andrew O Hatch, Sachin S Kajarekar, and Andreas Stolcke, “Within-class covariance normalization for svmbased speaker recognition.,” in INTERSPEECH, 2006. “The nist year 2010 speaker recognition evaluation plan,” http://www.nist.gov/speech/tests/spk/2010/index.html, 2010.

6. Acknowledgements

[7]

[8]

This work was supported by the National Natural Science Foundation of China under Grant No. 61370034.

7. References [1] Douglas A Reynolds, Thomas F Quatieri, and Robert B Dunn, “Speaker verification using adapted gaussian mixture models,” Digital signal processing, vol. 10, no. 1, pp. 19–41, 2000. [2] Vincent Wan and Steve Renals, “Speaker verification using sequence discriminant support vector machines,” Speech and Audio Processing, IEEE Transactions on, vol. 13, no. 2, pp. 203–210, 2005. [3] William M Campbell, Douglas E Sturim, and Douglas A Reynolds, “Support vector machines using gmm supervectors for speaker verification,” Signal Processing Letters, IEEE, vol. 13, no. 5, pp. 308–311, 2006. [4] Patrick Kenny, Pierre Ouellet, Najim Dehak, Vishwa Gupta, and Pierre Dumouchel, “A study of interspeaker variability in speaker verification,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 16, no. 5, pp. 980–988, 2008. [5] Patrick Kenny, Gilles Boulianne, Pierre Ouellet, and Pierre Dumouchel, “Joint factor analysis versus eigenchannels in speaker recognition,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, no. 4, pp. 1435–1447, 2007. [6] Najim Dehak, Patrick Kenny, R´eda Dehak, Pierre Dumouchel, and Pierre Ouellet, “Front-end factor analysis

[9]

[10]

[11]

[12]

[13]

[14]

[15]

Speaker Verification Using Fisher Vector

Models-Universal Background Models(GMM-UBM)[1] lay the foundation of modeling speaker space and many approaches based on GMM-UBM framework has been proposed to improve the performance of speaker verification including Support Vec- tor Machine(SVM)[2][3], Joint Factor Analysis(JFA)[4][5], to- tal variability ...

208KB Sizes 1 Downloads 289 Views

Recommend Documents

speaker identification and verification using eigenvoices
approach, in which client and test speaker models are confined to a low-dimensional linear ... 100 client speakers for a high-security application, 60 seconds or more of ..... the development of more robust eigenspace training techniques. 5.

Speaker Verification Anti-Spoofing Using Linear ...
four major direct spoofing attack types against ASV systems. [11]. Among these ... training data. Therefore, SS and VC attacks are potential threats for falsifying ASV systems. For a detailed review and general information on spoofing attacks against

speaker identification and verification using eigenvoices
(805) 687-0110; fax: (805) 687-2625; email: kuhn, nguyen, [email protected]. 1. ABSTRACT. Gaussian Mixture Models (GMMs) have been successfully ap- plied to the tasks of speaker ID and verification when a large amount of enrolment data is av

Robust Speaker Verification with Principal Pitch Components
Abstract. We are presenting a new method that improves the accuracy of text dependent speaker verification systems. The new method exploits a set of novel speech features derived from a principal component analysis of pitch synchronous voiced speech

Speaker Verification via High-Level Feature Based ...
humans rely not only on the low-level acoustic information but also on ... Systems Engineering and Engineering Management, The Chinese University of Hong ...

Research on I-Vector Combination for Speaker ...
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China. Email: [email protected], {wqzhang, liuj}@tsinghua.edu.cn. Abstract—I-vector has been ... GMM are two stat-of-art technologies in SRE. Especially the i-

Multiple Background Models for Speaker Verification
Tsinghua National Laboratory for Information Science and Technology. Department ..... High Technology Development Program of China (863 Pro- gram) under ...

nearest neighbor vector based palmprint verification
nearest neighbor lines, and then a second level match is .... If 'a' is the mid-point of the center line segment, ... We call the line that connects the centerline.

End-to-End Text-Dependent Speaker Verification - Research at Google
for big data applications like ours that require highly accurate, easy-to-maintain systems with a small footprint. Index Terms: speaker verification, end-to-end ...

High-Level Speaker Verification via Articulatory-Feature ...
(m, p|k)+(1−βk)PCD b. (m, p|k), (2) where k = 1,...,G, PCD s. (m, p|k) is a model obtained from the target speaker utterance, and βk ∈ [0, 1] controls the con- tribution of the speaker utterance and the background model on the target speaker mo

Text-Independent Speaker Verification via State ...
phone HMMs as shown in Fig. 1. After that .... telephone male dataset for both training and testing. .... [1] D. A. Reynolds, T. F. Quatieri, and R. B. Dunn, “Speaker.

SPEAKER IDENTIFICATION IMPROVEMENT USING ...
Air Force Research Laboratory/IFEC,. 32 Brooks Rd. Rome NY 13441-4514 .... Fifth, the standard error for the percent correct is zero as compared with for all frames condition. Therefore, it can be concluded that using only usable speech improves the

SPEAKER-TRAINED RECOGNITION USING ... - Vincent Vanhoucke
advantages of this approach include improved performance and portability of the ... tion rate of both clash and consistency testing has to be minimized, while ensuring that .... practical application using STR in a speaker-independent context,.

Efficient Speaker Recognition Using Approximated ...
metric model (a GMM) to the target training data and computing the average .... using maximum a posteriori (MAP) adaptation with a universal background ...

SPEAKER-TRAINED RECOGNITION USING ... - Vincent Vanhoucke
approach has been evaluated on an over-the-telephone, voice-ac- tivated dialing task and ... ments over techniques based on context-independent phone mod-.

Automatic speaker recognition using dynamic Bayesian network ...
This paper presents a novel approach to automatic speaker recognition using dynamic Bayesian network (DBN). DBNs have a precise and well-understand ...

Agglomerative Hierarchical Speaker Clustering using ...
news and telephone conversations,” Proc. Fall 2004 Rich Tran- ... [3] Reynolds, D. A. and Rose, R. C., “Robust text-independent speaker identification using ...

beyond-reason-using-emotions-as-you-negotiate-fisher-roger ...
... the apps below to open or edit this item. pdf-1231\-beyond-reason-using-emotions-as-you-negoti ... ger-author-oct-01-2006-paperback-by-roger-fisher.pdf.

Online Signature Verification using PCA and Neural ...
vision-based ones include voice recognition and signature verification. Signature has been a ... electronic payments, access control, and so on. In this paper ...

Online Signature Verification using PCA and Neural Network - IJRIT
includes online banking transactions, electronic payments, access control, and so on. ... prevalence of credit cards and bank cheques has long been the target of ...

The Vn801 Chip Verification Environment Using Verilog ...
processor core and peripherals such as watchdog-timer, timer0, timer1, timer2, ccp1, ... program which converts assembly code to VN801 instructions. Instruction ...

Face Verification using Gabor Filtering and Adapted ...
2École Polytechnique Fédérale de Lausanne, Switzerland. 1. Abstract. ▷ Face authentication is difficult because of the high-variability of face images.