Subspace-constrained Supervector PLDA for Speaker Verification Daniel Garcia-Romero and Alan McCree Human Language Technology Center of Excellence, Johns Hopkins University, Baltimore, MD [email protected], [email protected]

Abstract In this paper, we consider speaker supervectors as observed variables and model them with a supervector probabilistic linear discriminant analysis model (SV-PLDA). By constraining the speaker and channel variability to lie in a common low-dimensional subspace, the model parameters and verification log likelihood ratios (LLR) can be computed in this lowdimensional subspace. Unlike the standard i-vector framework, SV-PLDA does not ignore the uncertainty arising from the variable length of a speech cut (observation noise). Moreover, the SV-PLDA model can be equivalently formulated in terms of an intermediate low-dimensional representation denoted as projected i-vectors (π-vectors). This intermediate representation facilitates the use of techniques that are important in practice such as length normalization and multi-cut enrollment averaging. We validate the proposed model on a subset of the NIST extended-SRE12 telephone dataset for which test segments of nominal durations of 300, 100, and 30 seconds are available. We show significant improvements over the standard i-vector system for the short-duration test cuts and also compare SVPLDA with recently proposed extensions of the i-vector framework that also include the observation noise. Index Terms: speaker recognition, subspace modeling, short duration

of i-vectors with uncertainty propagation [6], [7], is a two-step process. The first step, performs a “cleanup” through Wiener filtering [8], and then, the second step, acknowledges the inherent uncertainty of the cleanup process and reincorporates it into modeling and scoring. In this paper, we expand the previous work in [2]. We start by reformulating SV-BSC in terms of the more mainstream jargon of PLDA (hence, the name SV-PLDA). As a results of this, we obtain a new intermediate representation (denoted as πvector) that facilitates the use of practical techniques, such as: length normalization, and multi-cut enrollment averaging. We then present an EM algorithm, in terms of π-vectors, to learn the SV-PLDA parameters. Finally, we experimentally validate our model and compare it with the recently proposed extensions of the i-vector framework [6], [7].

2. Subspace-constrained SV-PLDA In this section, we define a supervector as an observed variable. Then, based on the additive observation noise model in [8], we prescribe a generative model of supervectors, and show that the verification likelihood ratios can be computed in a low-dimensional subspace. Finally, we propose a constrainedsubspace approach to learn the SV-PLDA parameters. 2.1. Supervector computation

1. Introduction The current state-of-the-art in speaker recognition is widely dominated by the use of i-vectors [1]. However, a drawback of the i-vector, is that it is derived as a point-estimate of a posterior mean, and therefore, ignores the observation noise due to the duration of the speech cut, and the phonetic variability (i.e., estimation uncertainty of the i-vector). Alternatively, the recently proposed framework of supervector Bayesian speaker comparison (SV-BSC) [2], keeps account of the observation noise throughout modeling and scoring. The effects of ignoring the observation noise uncertainty are negligible when the duration, of the enrollment and test cuts, is long and homogeneous (e.g., 300 seconds of nominal duration for the NIST SRE setups). Under these conditions, the standard PLDA i-vector systems [3], [4], [5], are a very good approximation to the SV-BSC model [2]. However, this approximation is inaccurate when dealing with speech cuts of arbitrary duration, since the observation noise is no longer negligible, compared to the speaker and channel variability. Recently, two equivalent extensions of the i-vector paradigm have been proposed to reincorporate the estimation uncertainty of the i-vector into the PLDA modeling and scoring [6], [7]. However, there is a big philosophical difference between these approaches and SV-BSC. In particular, SV-BSC, is a single-step model in supervector subspace, where the observation noise is directly manifested. On the other hand, PLDA

To compute a supervector, we assume that a Gaussian mixture model (GMM) with K components, denoted as universal background model (UBM), has been trained on a large collection of data representative of the task at hand. Then, given a speech cut, represented by a sequence of T frames, O = {o1 , . . . , oT }, with frame ot ∈ RD , we define P the data mean with respect to component k as µk = 1/γk t γkt ot , where the scalar γkt is the responsibility of component k for observation ot ; and γk is the soft count of mixture k. Centering the data mean, µk , with respect to the UBM mean, mk , we obtain the offset1 µ ¯ k = µk − mk . The supervector x ¯, for the speech cut O, is then defined as x ¯ = [µ ¯ T1 , . . . , µ ¯ TK ]T . In this way, we consider the supervector computation as a front-end that maps a speech cut O into a high-dimensional supervector. 2.2. Generative model of supervectors Given a collection of Ji supervectors from speaker i, Di = {¯ xi1 , . . . , x ¯iJi }, we prescribe a generative model of the form:     ¯ ¯  h   x ¯i1 F G ... 0  i  ¯i1 w i1  .  .   .  .. ..   .. (1)  ..  =  .. .  +  ..  . . .   ..  ¯ 0 ... G ¯ x ¯iJi ¯iJi F wiJi 1 We

define this offset to be 0 if no data is assigned to a component

with a common latent speaker variable hi ∼ N (0, I), i.i.d latent channel variables wij ∼ N (0, I), and the supervector ¯ ij ). All the ladependent observation noise ¯ij ∼ N (0, Σ tent variables and noise terms are assumed to be mutually in¯ ij = N−1 Σ ¯ o , where the block diagodependent. We define Σ ij nal matrix of soft counts, Nij , contains K blocks of the form ¯ o is a block diagonal matrix constructed from diag(γk I); and Σ the K covariance matrices of the UBM. Also, we define the ¯ = TF, and the channel subspace speaker subspace matrix as F ¯ = TG. Moreover, we assume that the subspace matrix as G matrix T ∈ RDK×PT , F ∈ RPT ×PF , G ∈ RPT ×PT , and rank(F) = PF ≤ rank(G) = PT . Therefore, the speaker subspace is constrained to be a subset of the channel subspace ¯ ⊆ range(G) ¯ = range(T)), and we refer to (i.e., range(F) this model as subspace-constrained SV-PLDA. However, unlike in the original PLDA model [9], the observation noise is not assumed i.i.d2 . Under these assumptions, the joint distribution of the supervectors Di is eV eT + Σ e i ), p(Di ) = N (Di ; 0, V

Ji Y

¯ i, G ¯G ¯T + Σ ¯ ij ). N (¯ xij ; Fh

(3)

j=1

2.3. Log likelihood ratio for speaker verification The goal of speaker verification is to determine whether a supervector x ¯t belongs to speaker i or not. In the SV-PLDA framework, this is equivalent to asking whether x ¯t was generated from the same latent speaker variable, hi , as Di or not. This corresponds to a model selection problem between two alternative generative models. Under the same-speaker hypothesis, Hs , the generative model assumes that hi = ht . Under the different-speaker hypothesis, Hd , the generative model assumes that hi and ht are independently drawn from a standard Gaussian. Since we are interested in a probabilistic answer, we compute a log likelihood ratio (LLR) between the two competing hypothesis: p(Di , x ¯t |Hs ) p(Di , x ¯t ) L(Di , x ¯t ) = log = log . p(Di , x ¯t |Hd ) p(Di )p(¯ xt )

(4)

In this way, the larger the LLR, the stronger the support in favor of the same-speaker hypothesis Hs . L(Di , x ¯t ) is nothing more than the log ratio of two Gaussian distributions with equal mean and different covariances. By exploiting the structure of the covariance matrix in (2), the LLR can be efficiently computed as: L(Di , x ¯t ) =

x ˆi =

Ji P

¯T Λ ¯ −1 x F ij ¯ij ,

j=1

¯T Λ ¯ −1 x ˆt = F ¯t , t x

Hi =

Ji P

¯T Λ ¯ −1 F ¯ F ij

j=1

(6)

¯T Λ ¯ −1 ¯ Ht = F t F,

¯• = G ¯G ¯T + Σ ¯ • , can be and the supervector dependent, Λ inverted using the matrix inversion lemma. 2.4. Equivalent low-dimensional formulation In this section, we show that the subspace-constrained SVPLDA model is equivalent to a generalized PLDA model of a low-dimensional representation denoted as π-vector3 . In particular, we show that the LLR in (4) can also be obtained from the following generative model of π-vectors xij = Fhi + Gwij + ij ,

(7)

(2)

e corresponds to the first matrix in the right hand side where, V e i is a block diagonal matrix with blocks, Σ ¯ ij , for of (1), and Σ j = 1, . . . , Ji . Also, given the latent speaker variable hi , the conditional distribution of Di factorizes into p(Di |hi ) =

where

|I + Hi + Ht | 1 − log 2 |I + Hi ||I + Ht | 1 + (ˆ xi + x ˆt )T (I + Hi + Ht )−1 (ˆ xi + x ˆt ) 2 1 T 1 T − x ˆ (I + Hi )−1 x ˆi − x ˆt (I + Ht )−1 x ˆt , 2 i 2 (5)

2 Although this is a big deviation from the original PLDA model, and the term generalized-PLDA would be more appropriate, we still refer to our model as PLDA.

where F, hi , G, and wij are the same as in section 2.2. The π-vector xij is computed as −1 T ¯ −1 ¯ −1 ¯ −1 ¯ij , ¯ij = Σij TT Σ xij = (TT Σ T Σij x ij T) ij x

(8)

and the projected observation noise ij ∼ N (0, Σij ). Note that the observation noise in π-vector space is also not i.i.d. Since, based on (7), the functional form of the LLR for πvectors is the same as (5), all that is needed to show the equivalence, is to express x ˆi , Hi , x ˆt , and Ht as functions of the π-vector space variables. Without loss of generality, we show this for the case of Ji = 1, using the generic subscript “• ”. Making use of the matrix inversion lemma: ¯ T (G ¯G ¯T + Σ ¯ • )−1 x x ˆ• = F ¯• ¯ −1 ¯TΣ ¯ T ¯ −1 ¯ −1 G ¯ −1 ¯ −1 ¯ x• = FT TT [Σ • ]¯ • − Σ• G(I + G Σ• G) −1 T −1 −1 T T ¯ −1 ¯ −1 x• G T Σ• ]¯ = FT [TT Σ • − Σ• G(I + G Σ• G) −1 T −1 −1 T = FT [Σ−1 G Σ−1 • ]x• • − Σ• G(I + G Σ• G)

= FT (GGT + Σ• )−1 x• , (9) where, in the 4th row, we have made use of the π-vector defini−1 T ¯ −1 ¯ −1 tion in (8), and the simple identity TT Σ • = Σ• Σ• T Σ• . Using the same mechanics as in (9), it is easy to see that: ¯ T (G ¯G ¯T +Σ ¯ • )−1 F ¯ = FT (GGT + Σ• )−1 F. (10) H• = F Hence, the subspace constrained SV-PLDA model in (1) can be equivalently understood in terms of a generative model of π-vectors. 2.5. Parameter learning In this section, we show how to learn the parameters of the subspace constrained SV-PLDA model. We first learn the subspace matrix T in an unsupervised way. Then, we use it to compute π-vectors, xij , and to obtain the observation noise covariances Σij . Finally, we learn F and G based on the model in (7). 3 The “p” in π-vector comes from “projected”, since, as shown in (8), it corresponds to the coordinates of a weighted projection.

3. Experiments

2.5.1. Unsupervised learning of constrain subspace Supervised learning in high-dimensional spaces is a very challenging problem (i.e., curse of dimesnionality) [10]. A common practice is to first perform unsupervised dimensionality reduction followed by supervised training [10]. In this way, the main objective of dimensionality reduction is to remove the unreliable dimensions for classification. For this reason, we propose to first learn the subspace matrix T in an unsupervised way, and then, constraint the speaker and channel variability to lie in the range of T. This opens the door for the use of large datasets for which speaker labels are not available. We learn a ML point estimate of T using the EM algorithm in [1]. 2.5.2. Supervised learning of speaker and channel variability After learning the matrix T, we compute the π-vectors and the observation noise covariances using (8). Once in π-vector space, we can obtain maximum likelihood point estimates of F and G using the EM algorithm. To do so, we first rewrite the model in (7) as xij = Fhi + Uij zij + 0ij ,

(11)

Uij UTij

where = Σij (i.e., Cholesky), zij is a standard Gaussian latent variable, and the noise term 0ij ∼ N (0, GGT ) is i.i.d. Note that this is just a notational device and does not affect the model, since the within-speaker variability GGT +Σij is still the same. However, it is convenient to use this notation so that we can treat the residual noise as i.i.d. Also, the EM recipe becomes the same as the one used in [6]. In particular, for the E-step, the posterior mean and covariances of the hidden variables of a speaker, (hi , {zij }), can be efficiently computed using the block Cholesky algorithm presented in section III.D of [11]. Then, the M-step results in X  X −1 F= xij hhTi i − Uij hxij hi i hhi hTi i , (12) ij

ij

We evaluate the performance of SV-PLDA on the NIST Speaker Recognition Evaluation (SRE12) extended clean-telephone recordings from male speakers (condition 2). This subset comprises 763 speaker models with an average of 8 enrollment cuts per speaker. The number of enrollment cuts was allowed to vary per speaker. We only use the telephone cuts for enrollment, which results in 131 out of the 763 speakers only have 1 enrollment cut. We report results for multi-cut, and single-cut enrollment. For the single-cut case, the cut is picked randomly. The nominal duration of the test segments was either 300, 100, or 30 seconds; which facilitates the analysis of verification performance as a function of the test cut duration. The experiments in this paper only use the actual model involved in the trial to produce a score (i.e., no use of Bayes rule). We report performance in terms of three calibration-insensitive metrics: equal error rate (EER), and the two normalized minimum Detection Cost Functions defined by NIST for the SRE08 (mDCF08) and SRE10 (mDCF10) evaluations [13]. We used 40-dimensional MFCCs (20 base + deltas) with short-time mean and variance normalization. A 2048-mixture gender-independent UBM was used to obtain the supervectors. It was trained using all the telephone and microphone data from the male and female speakers in the enrollment pool (i.e., around 35K cuts from the SRE06, SRE08, and SRE10 databases). The same data was also used to to learn a 600dimensional subspace matrix T. The parameters of the baseline Gaussian PLDA i-vector model and the SV-PLDA model were learned only on the male subset (around 25K cuts), which contained synthetically corrupted versions of the data with babble and hvac noise at 15 dB and 6 dB SNR. We used a 400 dimensional speaker subspace for all the models. 3.2. Practical implementations

0ij

= xij − Fhi − Uij zij , 1 X 0 0T hij ij i, GGT = J ij

and, using the identity

3.1. Dataset and experimental setup

(13)

where, J is the total number of π-vectors, and the operator h i represents the moments with respect to the posterior distribution of the latent variables (with the previous values of Fold and Gold ). After each M-step, we perform a minimum divergence step [12] to accelerate the convergence.

In order to produce state-of-the-art results with Gaussian PLDA, it is necessary to apply length-normalization (LN) to the ivectors [5]. This non-linear transformation is a two-step process that involves whitening of the i-vectors and projection into the unit sphere. An equivalent process can be applied to the πvectors. However, a new issue arises with respect to the transformation of the observation noise covariance. This issue is also present in the recently proposed i-vector extensions [6], [7]. We use the following transformation for our experiments :

2.6. Relation to i-vector systems with uncertainty The difference in perspective, between SV-PLDA and i-vector systems with uncertainty [6], [7], is mainly manifested in two interrelated practical differences. In particular, comparing an i-vector [1] −1 T ¯ −1 ¯ −1 ˙ ij TT Σ ¯ −1 x˙ ij = (I + TT Σ T Σij x ¯ij = Σ ¯ij , ij T) ij x (14) with its π-vector counterpart, xij , in (8), we can see that an i-vector corresponds to a shrunk version of a π-vector. This shrinkage can be understood in terms of the “cleanup” performed by Wiener filtering [8]. Consequently, the observation ˙ ij , is smaller than the noise propagated to the i-vector space, Σ π-vector counterpart, Σij , in (8). We quantify the implications of these differences, in verification performance, in our experi¯ ij is negmental results. Finally, when the observation noise Σ ligible, both approaches are equivalent.

Σij ←

WΣij WT , kWxij k2

(15)

where W is the whitening transformation of the LN. This transformation was motivated in [7] as an approximation to a firstorder Taylor expansion of the non-linear LN. Another practical issue arises in the multi-cut enrollment scenario. In particular, the conditional independence assumption indicated in (3) is not very consistent with the type of enrollment data provided in the NIST SRE (i.e., where multiple recordings from the same channel or same phonetic content are provided). A common practical solution, in the standard Gaussian PLDA systems, is to average all the enrollment i-vectors of each speaker and compute LLRs pretending that there is only one enrollment cut. We explore the implications of this approach for the SV-PLDA system in section 3.3.2.

Table 1: Verification performance of three systems, as a function of test cut duration for single-cut and multi-cut enrollment. Baseline i-vector mDCF10 mDCF08 EER(%)

Uncertainty i-vector SV-PLDA (pi-vector) mDCF10 mDCF08 EER(%) mDCF10 mDCF08 EER(%) Multi-cut enrollment Pool 0.377 0.154 3.90 0.308 0.103 2.69 0.379 0.150 3.33 300 sec 0.134 0.039 1.44 0.136 0.042 1.27 0.163 0.041 1.08 100 sec 0.212 0.072 1.92 0.193 0.067 1.67 0.205 0.060 1.51 30 sec 0.629 0.219 5.58 0.440 0.172 4.66 0.470 0.156 4.01 Pool 0.252 0.080 2.24 0.251 0.080 1.96 0.245 0.078 1.94 300 sec 0.114 0.024 0.62 0.124 0.026 0.60 0.125 0.024 0.59 100 sec 0.160 0.052 1.49 0.167 1.31 0.163 0.051 1.29 Baseline i-vector Uncertainty0.053 i-vector SV-PLDA (pi-vector) SRE12 C2-ext 30 sec mDCF10 0.450 mDCF08 0.157 EER(%) 4.13 mDCF10 0.450 mDCF08 0.155 EER(%) 3.95 mDCF10 0.422 mDCF08 0.148 EER(%) 3.79 male only clean Single-cut enrollment Multi-cut enrollment PoolPool 0.3770.588 0.1540.269 3.90 6.31 0.3080.518 0.1030.212 2.69 4.98 0.3790.539 0.1500.225 3.33 5.29 300 sec 0.1340.384 0.0390.132 1.44 2.93 0.1360.371 0.0420.119 1.27 2.7 300 sec 0.1630.397 0.0410.127 1.08 2.69 100 sec 0.496 0.191 4.5 0.449 0.17 3.97 0.451 0.161 3.66 100 sec 0.212 0.072 1.92 0.193 0.067 1.67 0.205 0.060 1.51 30 sec 0.825 0.393 9.13 0.69 0.333 8.14 0.661 0.307 6.64 30 sec 0.629 0.219 5.58 0.440 0.172 4.66 0.470 0.156 4.01 Pool 0.495 0.17 3.47 0.487 0.167 3.56 0.483 0.158 3.37 Pool 0.252 0.080 2.24 0.251 0.080 1.96 0.245 0.078 1.94 300 sec 0.312 0.065 1.08 0.334 0.07 1.51 0.341 0.07 1.45 300 sec 0.114 0.024 0.62 0.124 0.026 0.60 0.125 0.024 0.59 100 sec 0.444 0.113 2.36 0.436 0.116 2.31 0.424 0.11 2.14 100 sec 0.160 0.052 1.49 0.167 0.053 1.31 0.163 0.051 1.29 30 sec 0.718 0.318 6.75 0.671 0.294 6.31 0.676 0.286 6.29 30 sec 0.450 0.157 4.13 0.450 0.155 3.95 0.422 0.148 3.79 ! Single-cut enrollment Pool 0.588 0.269 6.31 0.518 0.212 4.98 0.539 0.225 5.29 ! 3.3. Results the raw scores seem to0.397 be better aligned across2.69 durations. How300 sec 0.384 0.132 2.93 0.371 0.119 2.7 0.127 ever, this is no longer true when LN is used. ! 100 sec 0.496 0.191 0.17 3.97 0.451 0.161 3.66 In this section, we experimentally validate the4.5 proposed 0.449 SVthe three systems conditions. This 30 sec 0.825 0.393 9.13 0.69 0.333Applying 8.14LN helps0.661 0.307 in all6.64 PLDA! model for the single-cut and multi-cut enrollment sceis quite remarkable since LN was 0.158 initially introduced to alleviPool 0.495 0.17 3.47 0.487 0.167 3.56 0.483 3.37 narios.! In both cases, we analyze the performance as a funcate the dataset [14]. Also,1.45 LN reduces the 300 sec 0.312 0.065 1.08 0.334 0.07 1.51 shift problem 0.341 [5], 0.07 tion of! test100 cutsec duration,0.444 and also 0.113 show the effects of lengthgap, with respect to the baseline,0.11 of SV-PLDA 2.36 0.436 0.116 2.31 0.424 2.14and i-vectors normalization. with uncertainty. However, the benefits of LN 30 sec 0.718 0.318 6.75 0.671 0.294 6.31 0.676 0.286 6.29get smaller as ! No LN LN

LN

No LN

LN

No LN

LN

No LN

SRE12 C2-ext male only clean

LN

No LN

the duration gets shorter. Additionally, for the 30 second case, ! SRE12 C2-ext Baseline i-vector Extended i-vector SV-PLDA (pi-vector) 3.3.1. Single-cut enrollment SV-PLDA without LN outperforms the baseline with LN. To the mDCF08 EER(%) mDCF10 mDCF08 EER(%) ! male only clean mDCF10 mDCF08 EER(%) mDCF10 best of our knowledge, this is the first time a result like this has Table 1 shows the results of0.588 the baseline Gaussian 6.31 PLDA sys-0.518 been Pool 0.269 0.212 4.98 0.539 0.225 5.29 reported. ! tem, the i-vector system uncertainty, and the proposed 300 sec with 0.384 0.132 2.93 SV-0.371 0.119 2.7 0.397 0.127 2.69 ! system,100 PLDA as asec function0.496 of the test 0.191 cut duration 4.5 (as well as0.449 3.3.2. 0.17Multi-cut 3.97 0.451 0.161 3.66 enrollment Single-cut enrollment mDCF10The mDCF08 EER(%) 0.69 pooling the scores across durations). single-cut9.13 enrollment 30 sec 0.825 0.393 0.333 8.14 0.661 0.307 6.64 the trends observed for the single-cut enrollment are also Enroll and testThe0.495 0.539 are 0.17 0.225with3.47 is shown at the bottom. results shown or5.29 without0.487 All0.167 Pool 3.56 0.483 0.158 3.37 No LN true for the multi-cut case. Note that all the multi-cut results in Test 0.544 0.228 5.32 test LN. As expected, the performance shorter 300 sec 0.312 degrades 0.065 with 1.08 0.334 0.07 1.51 0.341 0.07 1.45 Table 1 are based on the practical approach (described in secEnroll andperformance test 0.483 0.158 3.37 cuts. Looking at the 100 sec 0.444 without 0.113LN (which 2.36directly0.436 tion0.116 2.31 the 0.424 0.11 2.14 2 shows LN 3.2) of averaging enrollment π-vectors. Table Test 0.486 corresponds to the theoretical derivation of0.158 SV-PLDA),3.34 we can 30 sec 0.718 0.318 6.75 0.671 the0.294 6.31across 0.676 0.286other approaches. 6.29 results (pooled durations) with It observe ! ! that the relative improvement of SV-PLDA over the is quite obvious that the assumption in (3), which is known in baseline gets larger as the test cut duration gets shorter; which the community as scoring ”by-the-book” is not consistent with is Pooled!results!for!SV.PLDA!single.cut!enrollment!including!the!observation!noise!in!both! consistent with the theory. While this is true for the three ! the NIST enrollment data. Moreover, we also tested the alternametrics, the EER seems to benefit the most, with a 27% relative enrollment!and!test,!or!only!in!test.! tive of directly summing all the supervector sufficient statistics improvement at the nominal duration of 30 seconds. A simi! to produce a single π-vector; which assumes that all the enrolllar trend is observed for the i-vector system with uncertainty; Oracle-pooling no LN mDCF10 mDCF08 EER(%) ment data for a speaker comes from the same channel. This whose performance is very similarly to the SV-PLDA model. Uncertainty i-vector 4.83 alternative is not as good as averaging the π-vectors. However, for the 30 second case,0.504 SV-PLDA0.208 is noticeably better. SV-PLDA (pi-vector) 0.503 0.199 4.31 An important observation is that the i-vector system with uncer! seems to pool better than the SV-PLDA system. That is, tainty 4. Conclusions

Oracle!pooling!across!test!durations!for!i.vector!with!uncertainty!and!SV.PLDA!(single.cut! In this paper, we presented a generative model of supervectors that keeps account of the observation noise throughout enrollment).! Table 2: Comparison of three scoring approaches for SV-PLDA modeling and scoring. Under a subspace constrain assumpsystem ! with multi-cut enrollment (see section 3.3.2 for details). Multi-cut enrollment “By-the-book” No LN Average pi-vec Sum all SV “By-the-book” LN Average pi-vec Sum all SV

mDCF10 0.427 0.379 0.375 0.384 0.245 0.369

mDCF08 0.223 0.150 0.150 0.204 0.078 0.158

tion, we showed an equivalent formulation in terms of a lowdimensional representation (denoted as π-vector). We then introduced an EM algorithm, in terms of π-vectors, to learn the SV-PLDA parameters. Finally, we validated our model on a subset of the NIST SRE12 dataset, and showed significant improvements over the standard i-vector system for the shortduration test cuts. We also compared our model with recently proposed extensions of the i-vector framework.

EER(%) 6.24 3.33 3.32 5.54 1.94 3.31

! Comparison!of!three!scoring!approaches!for!SV.PLDA!system!with!multi.cut!enrollment!(see! text!for!details).!! ! !

N

SRE12 C2-ext male only clean Pool

Baseline i-vector

Extended i-vector

SV-PLDA (pi-vector)

mDCF10

mDCF08

EER(%)

mDCF10

mDCF08

EER(%)

mDCF10

mDCF08

EER(%)

0.588

0.269

6.31

0.518

0.212

4.98

0.539

0.225

5.29

5. References [1] N. Dehak, P. Kenny, R. Dehak, P. Dumouchel, and P. Ouellet, “Front-end factor analysis for speaker verification,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 4, pp. 788 –798, may 2011. [2] B. J. Borgstrom and A. McCree, “Supervector Bayesian speaker comparison,” in International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Vancouver, Canada, May 2013 (accepted). [3] P. Kenny, “Bayesian speaker verification with heavy-tailed priors,” in Odyssey: The Speaker and Language Recognition Workshop, Brno, Czech Republic, 2010. [4] P. Matejka, O. Glembek, F. Castaldo, M. J. Alam, P. Kenny, L. Burget, and J. Cernocky, “Full-covariance UBM and HeavyTailed PLDA in i-vector speaker verification,” in International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Prague, Czech Republic, 2011. [5] D. Garcia-Romero and C. Y. Espy-Wilson, “Analysis of i-vector length normalization in speaker recognition systems,” in Interspeech, Florence, Italy, August 2011, pp. 249–252. [6] P. Kenny, T. Stafylakis, P. Ouellet, M. J. Alam, and P. Dumouchel, “PLDA for speaker verification with utterances of arbitrary duration,” in International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Vancouver, Canada, May 2013 (accepted). [7] S. Cumani, O. Plchot, and P. Laface, “Probabilistic linear discriminant analysis of i–vector posterior distributions,” in International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Vancouver, Canada, May 2013 (accepted). [8] A. McCree, D. Sturim, and D. Reynolds, “A new perspective on GMM subspace compensation based on PPCA and Wiener filtering,” in Interspeech, 2011, pp. 145–148. [9] S. J. D. Prince and J. H. Elder, “Probabilistic linear discriminant analysis for inferences about identity,” in IEEE International Conference on Computer Vision (ICCV), Rio de Janeiro, 2007, pp. 1 – 8. [10] X. Jiang, “Linear subspace learning-based dimensionality reduction,” IEEE Signal Processing Magazine, vol. 28, no. 2, pp. 16– 26, March 2011. [11] P. Kenny, “Joint factor analysis of speaker and session variability : Theory and algorithms,” Montreal, Tech. Rep., 2005. [12] N. Brummer, “EM for probabilistic LDA,” (Available at https:// sites.google.com/site/nikobrummer/), February 2010. [13] “The NIST year 2010 Speaker Recognition Evaluation plan.” (Available at http://www.itl.nist.gov/iad/mig/tests/sre/2010/ NIST SRE10 evalplan.r6.pdf), 2010. [14] C. Vaquero, “Dataset shift in PLDA based speaker verification,” in Odyssey: The Speaker and Language Recognition Workshop, 2012.

Subspace-constrained Supervector PLDA for Speaker ...

speaker subspace matrix as ¯F = TF, and the channel subspace matrix as ¯G ... 2Although this is a big deviation from the original PLDA model, and the term ...

317KB Sizes 0 Downloads 176 Views

Recommend Documents

Bayesian PLDA
Aug 13, 2010 - There is one path between M and π, which is blocked when the head-to-tail node θt is observed, so that: P(M|θt,π) = P(M|θt). (7). Note that when the head-to-head node Dt, which is on the same path, is also observed, this node is n

Best Speaker for Echo Dot - AUDIOSOUNDCENTRAL.COM.pdf ...
Get the Best Picks for the Best. Amazon Echo DOT Bluetooth. Speakers. Page 5 of 21. Best Speaker for Echo Dot - AUDIOSOUNDCENTRAL.COM.pdf.

A Speaker Count System for Telephone Conversations
The applications of this research include three-way call detection and speaker tracking, and could be extended to speaker change-point detection and indexing.

Best Speaker for Echo Dot KW's- AUDIOSOUNDCENTRAL.COM ...
Page 4 of 37. Best Speaker for Echo Dot KW's- AUDIOSOUNDCENTRAL.COM - Drive Links.pdf. Best Speaker for Echo Dot KW's- AUDIOSOUNDCENTRAL.

A Speaker Count System for Telephone Conversations
3, which shows the intra- and inter-speaker Mahalanobis distance distributions (1000 comparisons for each class) for 5- segment based models, suggests some ...

Best Speaker for Echo Dot - AUDIOSOUNDCENTRAL.COM.pdf ...
What Does Echo Dot Do ... The built-in. speaker does work but it can be hard to hear ... COM.pdf. Best Speaker for Echo Dot - AUDIOSOUNDCENTRAL.COM.pdf.

SPEAKER SELECTION TRAINING FOR LARGE ...
adaptation methods such as MAP and MLLR [1] do not work well. ... may be a good index to select the reference speakers from many ... phone (or phone class).

Cheap Portable Mini Bluetooth Speaker Car Music Center Speaker ...
Cheap Portable Mini Bluetooth Speaker Car Music Cen ... mputer Speakers Free Shipping & Wholesale Price.pdf. Cheap Portable Mini Bluetooth Speaker Car ...

Cheap E104 Hand Wireless Bluetooth Speaker Loud Speaker ...
Cheap E104 Hand Wireless Bluetooth Speaker Loud Sp ... etooth Spinner Free Shipping & Wholesale Price.pdf. Cheap E104 Hand Wireless Bluetooth Speaker ...

Cheap Ribbon Tweeter,Speaker Piezo Tweeter,Speaker Driver ...
Cheap Ribbon Tweeter,Speaker Piezo Tweeter,Speaker Driver⁄ Free Shipping & Wholesale Price.pdf. Cheap Ribbon Tweeter,Speaker Piezo Tweeter,Speaker ...

speaker 2_CV_Lemack.pdf
“Expression of rat tropoelastin in the transgenic mouse bladder: histologic and physiologic effects,” Annual Meeting of the American. Urological Association, May ...

Cheap Ribbon Tweeter,Piezo Tweeter Speaker,Speaker Driver ...
Cheap Ribbon Tweeter,Piezo Tweeter Speaker,Speaker Driver⁄ Free Shipping & Wholesale Price.pdf. Cheap Ribbon Tweeter,Piezo Tweeter Speaker,Speaker ...

Cheap Mifa F6 Nfc Bluetooth Speaker Outdoor Wireless Speaker ...
Cheap Mifa F6 Nfc Bluetooth Speaker Outdoor Wireles ... 6S,Ipad,Samsung Free Shipping & Wholesale Price.pdf. Cheap Mifa F6 Nfc Bluetooth Speaker ...

Speaker/Commencement.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item.

Speaker Biographies
workers are properly classified as employees under the law. A lifelong resident of New Jersey, Fred .... Environment & Design Center at The New School focusing on climate and environmental justice research. Ana has a PhD in Urban ...... ACLU Grover D

january speaker
their loved ones. I also show them that 15 years after open heart surgery, I am fully functional and able to live a full life. Visiting heart patients after they leave the.

Speaker Bios.pdf
Bachelor of Science degree, he obtained a Master. of Science degree and a PhD in Entomology (gold. medal award) from Kochi University, Japan in 1995.

Uniform Multilingual Multi-Speaker Acoustic Model for Statistical ...
single-speaker system from small amounts of data for that lan- ... In recent years, statistical parametric speech synthesis has seen .... analysis components.

Uniform Multilingual Multi-Speaker Acoustic Model for Statistical ...
Uniform Multilingual Multi-Speaker Acoustic Model for Statistical Parametric ... training data. This type of acoustic models, the multilingual multi-speaker (MLMS) models, were proposed in [12, 13, 14]. These approaches utilize a large input feature