INTERSPEECH 2005

Statistical Properties of the Warped Discrete Cosine Transform Cepstrum compared with MFCC R. Muralishankar, Abhijeet Sangwan and Douglas O’Shaughnessy INRS-EMT (Telecommunications), University of Quebec, Montreal, Canada. Department of Electrical and Computer Engineering, Concordia University, Montreal, Canada. [email protected], a [email protected], [email protected] Abstract

the clean and noisy vowel clusters formed by WDCTC and MFCC features and present the average separability of the vowel classes.

In this paper, we continue our investigation of the warped discrete cosine transform cepstrum (WDCTC), which was earlier introduced as a new speech processing feature [1]. Here, we study the statistical properties of the WDCTC and compare them with the mel-frequency cepstral coefficients (MFCC). We report some interesting properties of the WDCTC when compared to the MFCC: its statistical distribution is more Gaussian-like with lower variance, it obtains better vowel cluster separability, it forms tighter vowel clusters and generates better codebooks. Further, we employ the WDCTC and MFCC features in a 5-vowel recognition task using Vector Quantization (VQ) and 1-Nearest Neighbour (1-NN) as classifiers. In our experiments, the WDCTC consistently outperforms the MFCC.

2. WDCTC Algorithm The new WDCTC algorithm is outlined below. Consider a finite duration, real sequence x(n), defined for 0 ≤ n ≤ N − 1 and zero elsewhere. Taking an N -point Warped discrete cosine transform (WDCT, [3]) of the above sequence, we have XW DCT (k) defined for 0 ≤ k ≤ N − 1. We can write XW DCT (k) as XW DCT (k) = exp(ξ(k)) |XW DCT (k)|

(1)

where

jπ (sgn(XW DCT (k)) − 1) 2 and ‘sgn’ is the sign of the WDCT coefficients. Taking natural logarithm on both sides of eq. 1, we obtain the WDCTC of x(n) as ξ(k) =

1. Introduction We recently introduced the warped discrete cosine transform cepstrum (WDCTC) as a new speech processing feature and demonstrated its better performance than the melfrequency cepstral coefficients (MFCC) in a vowel recognition and speaker-identification task [1]. The WDCTC has shown good promise as a speech processing feature and we are encouraged to further investigate the WDCTC feature and its statistical properties. A large volume of training data is required to build speaker-independent speech recognition systems. One technique of reducing the data size is clustering the data and choosing a reasonable number of representative feature vectors to form codebooks [2]. Hence, codebook techniques are very relevant and practical to speech recognition systems. We form WDCTC and MFCC codebooks using a k-means clustering algorithm and compare the codebook statistics for clean and noisy vowels using the coefficient of variance and overlap ratio (defined later). Our experiment demonstrates that the WDCTC codebooks represent the underlying vowel data better than MFCC. In order to compare the classification capability of the features, the WDCTC and MFCC are employed in a 5vowel recognition task. Vector quantization (VQ) and 1nearest neighbor (1-NN, [2]) are used as classifiers and their recognition performance is reported. We also investigate

x (n) = Re{IDCT (ξ(k) + ln |XW DCT (k)|)}.

(2)

Here, an inverse discrete cosine transform (IDCT) [4] is used to get the WDCTC sequence and it is denoted as x (n).

3. Vowel Recognition Task Vowel recognition experiments are conducted on the TIMIT database. We select 5 vowels /aa/,/eh/,/iy/,/ow/ and /uw/ for our experiments. The vowel segments are extracted from continuous speech using the train and test sets of the TIMIT database (Dialect Region: North Midland) to form the train and test sets for our experiments. The number of speakers in our train and tests sets were 72 and 26, respectively. Each vowel segment is sampled at 16 kHz. The duration of each frame of speech is 16 ms, with an overlap of 8 ms between successive frames. Each frame of speech is preemphasized with a factor equal to 0.98 and is Hamming windowed. Eighteen-dimensional feature vectors (MFCC and WDCTC) are obtained for each frame. For obtaining MFCC, the Mel-scale is simulated using a set of 18 triangular filters. For the WDCTC, we choose a warping parameter in the WDCT, for a given fs , that closely follows the

1

September, 4-8, Lisbon, Portugal

INTERSPEECH 2005

Vowel Class Separability in Car noise

3

8

x 10

5

3

Foc

1

0.5 0

0.5

(a)

0

1.5

1

0.5

0

(b)

0

5

10 bins

bins

9 3

WDCTC 15 10 SNR (dB)

25

20

3

Clean

1.5 1

1

0.5

0

0.5

(c)

0

0.2

0.4

0.6

10.5

0.8

0

1

(d)

3

x 10

5

4

x 10

12

10

8

6 bins

4

2

0

bins

4

20

5

2

1.5

Vowel Class Separability in Babble noise

x 10

15

2.5

2

Foc

5

5

2.5

MFCC 0

x 10

Foc

8

7 −5

5

10 Foc

3

9.5

3

Foc

Separability measure (10 )

1.5

1

10

5

2

1.5

11

x 10

2.5

2

7 Foc

Separability measure (103)

2.5

2

1

9

(e)

0

0

5

10 15 SNR (dB)

20

25

0.2

0.4

0.8

0.6

1

1.2

(f)

1.4

0

0

2

WDCTC 0.014 0.100

MFCC 0.265 1.010

χ2 ndf

WDCTC 0.011 0.063

8

6

10

12

14

bins

3.1.2. Separability of Vowel Clusters We compute the separability between two vowel classes as:

Table 1: Gaussian fit for MFCC and WDCTC features. Average Variance

4

Figure 2: Histograms of the within-class distance averaged over all vowel classes for clean vowels: (a) WDCTC (b) MFCC, and vowels corrupted by -5 dB SNR: (c) WDCTC, Babble noise (d) MFCC, Babble noise (e) WDCTC, Car noise (f) MFCC, Car noise. Foc = Frequency of occurrence.

Clean

Figure 1: Average separability of vowel classes using WDCTC and MFCC features in varying SNR for: (a) Car noise (b) Babble noise.

Clean Noisy

0

bins

8.5

8 −5

2 1

1 −1 )Sij ) D(i, j) = tr( (Sii−1 + Sjj 2

MFCC 0.033 0.048

(3)

where D(i, j) is the separability and Sij is the cross-covariance between the ith and the j th classes. Sii is the covariance of the ith class and ‘tr’ denotes the trace of a matrix. A similar separability measure is used to compare the MFCC with the linear discriminant analysis (LDA) applied to logspectral or cepstral feature vector [7]. Finally, the average class separability is calculated as:

mel-scale [5]. We incorporate this WDCT in the discrete cosine transform cepstrum (DCTC, [6]) to obtain WDCTC. The first 18 WDCTC coefficients, excluding the gain term, form the feature vector. 3.1. Statistical Analysis of the WDCTC and MFCC

Dave =

The statistical nature of the WDCTC and MFCC features is presented prior to their recognition performance.

 1 D(i, j) m(m − 1) i j

(4)

where m is the total number of vowel classes and Dave is the average separability across all the classes. The average separability of the vowel classes for WDCTC and MFCC is shown in Fig. 1. It is observed from Fig. 1 that the WDCTC vowel classes are consistently more separable than MFCC, under varying SNR conditions.

3.1.1. Statistical Distribution of the WDCTC and MFCC We compare the closeness of the statistical distribution of the WDCTC and MFCC to a Gaussian distribution. In order to obtain the statistical distribution of the features, 1000 samples each of 5 vowels from 72 speakers are extracted from the TIMIT database. The variance and χ2 normalized by the number of degrees of feedom of the Gaussian fit averaged across C1 to C18 for the WDCTC and MFCC features are shown in Table 1 for clean and noisy vowels. From Table 1, it is observed that the WDCTC coefficients fit the Gaussian distribution more closely than MFCC in clean vowels. Another important observation is the lower variance of the WDCTC coefficients as compared to the MFCC coefficients in clean and noisy vowels (vowels corrupted with Car noise) at a -5 dB signal-to-noise ratio (SNR). Low variance is a desirable property for estimation or classification, which makes the WDCTC more attractive.

3.1.3. Within-Vowel Cluster Distance Statistics Within-class distance is the distance between two feature vectors of the same class and is given by:  a  b  (5) dab ij = fi − fj th where dab and j th feature ij is the distance between the i a b vectors, fi and fj of the classes a and b respectively. The distance of a feature vector for a given class a from its cluster center is given by: a a daa i = fi − E[fi ]

2

(6)

INTERSPEECH 2005

300

250

200

200

Foc

Foc

300

250

150

150

100

100

50

50

(a)

0 0.1

0.3

0.2

0.4

0.5 bins

0 0.6

0.7

0.8

0.8

3

0.9

0.7 2.5 (b)

2

0

6 bins

4

8

0.6

12

10

200

200

150 100

0.5

ρ − intra−codebook, MFCC

150

ρ − cluster−codebook, MFCC

1.5

κ − MFCC

100

50

50

(c)

0 0.1

0.2

0.5

0.4

0.3

0

0.6

0

1

4

3

2

5

6

ρ − cluster−codebook, WDCTC

7

800

300

600

0.3

κ − WDCTC

1

bins

400

0.4

ρ − intra−codebook, WDCTC

(d)

bins

κ, ρ

300 250

κ, ρ

300

250

Foc

Foc

2

0.2

Foc

Foc

0.5 200

100

400

0.1

200 (e)

0

0

0.1

0.2

0.3

0.4 bins

0.5

0.6

0.7

(a)

(f)

0.8

0

0

2

4

6 bins

8

10

0

12

20 15 Codebook length

25

(b)

30

0 5

10

20 15 Codebook length

25

30

where ρad is the coefficient of variance of the intra-codebook distance for class a and std[.] is the standard deviation. A low value of ρ indicates formation of tighter codebooks where members of the same class are closely packed. In order to compare the separation of the codebooks in the two feature spaces, a new term called the overlap ratio, κ is introduced. It is defined as the ratio of the intra to intercodebook distance, i.e.,

where E[.] is the expectation operator. We term this distance as the feature-centroid distance (for a given class a). The statistical distributions of the within-class and featurecentroid distances are given in Figs. 2 and 3. It is seen that the statistical distributions of both the distances in clean and noisy vowels resemble a Gaussian distribution more closely for the WDCTC as compared to MFCC. Further, it is observed from Fig. 2 that the within-class distance for the WDCTC has lower variance than MFCC, indicating a tighter cluster formation for the WDCTC. This is a useful property for constructing a codebook, where the WDCTC needs a smaller codebook length than MFCC for good data representation.

κab d =

E[daa ij ] E[dab ij ]

(8)

where κab d is the overlap-ratio between vowel classes a and b. A good feature should have a low ρ and κ. We define the cluster-codebook distance as the distance between the feature vectors of a given class and its codebook. Finally, in order to measure the closeness of a codebook to its cluster, we compute the coefficient of variance of the clustercodebook distance. Figure 4 shows the codebook statistics for clean and noisy vowels. The coefficient of variance of the intra-codebook and cluster-codebook distance is higher for MFCC than the WDCTC for clean and noisy vowels, which indicates that the WDCTC codebooks are tighter and represent vowel data more closely. Particularly, it is noted that for small-length codebooks (eg., 8-length) the difference between the coefficient of variance of clustercodebook distance of the WDCTC and MFCC is very high in clean and noisy vowels. This agrees with our hypothesis that the WDCTC indeed represents data with smaller codebook length. The overlap-ratio of the MFCC codebook is higher than with WDCTC in clean and noisy vowels, which is an indicator of the separability of the codebooks of different vowel classes.

3.1.4. Codebook Statistics We have observed that the WDCTC exhibits low variance and tighter cluster formation. Hence, we expect that for a given length the WDCTC must form better codebooks than MFCC and represent underlying data more closely. We set up an experiment to verify our hypothesis. We use the kmeans clustering algorithm to form 8, 16 and 32-length vector codebooks and compute codebook statistics. We compute the intra-codebook distance, which is defined as the average distance between the vectors belonging to the same codebook using eq. 5, where now fia is a codebook vector for class a. Similarly, the average distance between the codebook vectors of different classes is computed and termed as the inter-codebook distance. Further, the coefficient of variance, ρ, of the intra-codebook distance is computed as: std[daa ij ] aa E[dij ]

10

Figure 4: Coefficient of variance of Intra-codebook distance, Coefficient of variance of cluster-codebook distance and overlap-ratio for WDCTC and MFCC codebooks using (a) noisy vowels corrupted by -5 dB SNR, Babble noise (b) clean vowels.

Figure 3: Histograms of the feature-centroid distance averaged over all vowel classes for clean vowels: (a) WDCTC (b) MFCC, and vowels corrupted by -5 dB SNR: (c) WDCTC, Babble noise (d) MFCC, Babble noise (e) WDCTC, Car noise (f) MFCC, Car noise. Foc = Frequency of occurrence.

ρad =

5

(7)

3

INTERSPEECH 2005

VQ Performance in Babble Noise

80

70

70 Recognition %

Recognition %

VQ Performance in Car Noise 80

60

50

40

30 20 −5

60 50 40 30

(a)

0

5 10 SNR (dB)

15

20 −5

20

(b)

0

15

20

CTC consistently outperforms MFCC.

20

The WDCTC is compared to MFCC using VQ and 1-NN classifiers. We choose basic learning algorithms to minimize their influence in the classification results and maximize the feature’s capability. The results show that the WDCTC is consistently better than MFCC. Interesting facts about the WDCTC are presented, such as good vowel class separability, low variance, better codebook representation and robustness to noise. It is also shown that the WDCTC feature and its statistics fit a Gaussian distribution more closely than MFCC. This is useful information as the ‘state of the art’ HMM classifiers inherently assume a Gaussian distribution for speech features. The next logical step to this study is to test the WDCTC feature using more complex machine learning algorithms such as HMM and SVMs. We are currently working in this direction.

4. Conclusion

80

70 Recognition %

70

60

50

60

50

40

40

30 (c) 20 −5

10 5 SNR (dB)

1−NN Performance in Babble Noise

1−NN Performance in Car Noise 80

Recognition %

Table 2: Comparison of clean vowel average recognition performance of MFCC and WDCTC features. VQ model 1-NN model Feature Train Test Train Test MFCC 90.86 67.83 99.01 71.51 WDCTC 90.97 69.48 99.75 73.00

30

0

5 10 SNR (dB)

15

20

20 −5

WDCTC MFCC

(d)

0

10 5 SNR (dB)

15

Figure 5: Noisy vowels’ average recognition performances of MFCC and WDCTC. (a) Using VQ model in presence of Car noise. (b) Using VQ model in presence of Babble noise. (c) Using 1-NN model in presence of Car noise. (d) Using 1-NN model in presence of Babble noise. 3.2. Vowel Recognition using VQ and 1-NN

5. References

We have presented statistical properties of the WDCTC and MFCC in the previous subsection. These properties are exploited by VQ and 1-NN [2] classifiers for the vowel recognition task. VQ and 1-NN classifiers will tap the useful codebook properties. Further, these sets of classifiers are basic and have little influence in the final recognition performance. In this manner, we are able to limit the influence of the classifier and rely heavily on the features themselves in the final recognition performance. ‘State of the art’ classifiers like the hidden Markov model (HMM) and support vector machines (SVM) are beyond the scope of this paper. Each vowel is modeled using a 32-length VQ codebook, consisting of code vectors of MFCC and WDCTC features. The codebooks are trained using the k-means clustering algorithm, employing a Euclidean distance measure. Vowels are identified by evaluating the distortion between the features of the test vowel sample and the vowel codebooks. Car noise and babble noise are added to clean vowels. SNR is varied from -5 dB to 20 dB and the noisy vowel recognition performances using MFCC and WDCTC features are obtained. The vowel recognition performance of MFCC and WDCTC features for clean speech using VQ and the 1-NN model is shown in Table 2. Figure 5 shows the noisy vowel recognition performances of MFCC and WDCTC. Figure 5(a) and (b) show the recognition accuracy in percentage and the comparative performances in the presence of car noise and babble noise for the VQ model. Similarly, Figs. 5(c) and (d) show the results for the 1-NN model. From Table 2 and Fig. 5, we can see that the WD-

[1] R. Muralishankar, A. Sangwan and D. O’Shaughnessy, “Warped Discrete Cosine Transform Cepstrum: A new feature for speech processing,” accepted EUSIPCO 2005. [2] Vishwa N. Gupta, Matthew Lennig, and Paul Mermelstein, “Decision rules for speaker-independent isolated word recognition,” ICASSP ’84, vol. 9, pp. 336-339, 1984. [3] N. I. Cho and S. K. Mitra, “Warped discrete cosine transform and its application in image compression,” IEEE Trans. Circuits Syst. Video Technol., vol. 10, pp. 1364-1373, Dec. 2000. [4] S. A. Matrucci, “Symmetric convolution and the discrete sine and cosine transforms,” IEEE Transactions on Signal Processing, vol. 42, pp. 1038-1051, 1994. [5] J. O. Smith III and J. S. Abel, “Bark and ERB Bilinear Transforms,” IEEE Trans. Speech, Audio Processing, vol. 7, pp. 697-708, June 1999. [6] R. Muralishankar and A. G. Ramakrishnan, “Pseudo complex cepstrum using discrete cosine transform,” accepted, International Journal of Speech Technology. [7] T. Eisele, R. Haeb-Umbach and D. Langmann, “A comparative study of linear feature transformation technique for automatic speech recognition,” Proc. ICSLP96, pp. 252-255, Philadelphia, PA, Oct. 1996.

4

Statistical Properties of the Warped Discrete Cosine ...

of the TIMIT database (Dialect Region: North Midland) to form the train and test ... SNR (dB). Vowel Class Separability in Babble noise. WDCTC. MFCC. Figure 1: ...

148KB Sizes 1 Downloads 216 Views

Recommend Documents

Warped Discrete Cosine Transform Cepstrum
MFCC filter bank is composed of triangular filters spaced on a logarithmic scale. ..... Region: North Midland) to form the train and test set for our experiments.

Image Compression Using the Discrete Cosine Transform
NASA Ames Research Center. Abstract. The discrete ... The discrete cosine transform of a list of n real numbers s(x), x = 0, ..., n-1, is the list of length n given by:.

Image Compression and the Discrete Cosine Transform ...
We are now ready to perform the Discrete Cosine Transform, which is accomplished by matrix multiplication. D : TMT r. 5. In Equation (5) matrix M is first multiplied on the left by the DCT matrix T from the previous section; this transforms the rows.

two dimensional discrete cosine transform using bit ...
Discrete orthogonal transforms (DOTs), which are used frequently in many applications including image and speech processing, have evolved quite rapidly over the last three decades. Typically, these applications require enormous computing power. Howev

Statistical Properties of Generalized Discrepancies
Here, an inequality of Kiefer and. Wolfowitz (see [KW58]) (indeed a multivariate extension of the Dvoretzky-Kiefer-. Wolfowitz or DKW inequality) implies the existence of positive constants c and c such that: (5.4). P. {. D∗. ∞ (Pn,u) ≥ m. √

statistical properties of generalized discrepancies ...
vestigation of the asymptotic properties of D2 (Pn) and have derived two alternative statistics, strictly linked to this one, that can be used to test statistically the efficiency of numerical integration procedures. Recently, other authors have inve

A statistical examination of the evolution and properties ...
Sep 22, 2009 - FreeBSD ports. ○ 12,010 ports. ○ 6,556 written in C. ○ 1,109,697 files. ○ 447,612 in C. AnsiC. C++. Shell. Perl. Java. PHP. Python. Lisp.

DISPERSIVE PROPERTIES FOR DISCRETE SCHR ...
Let us now consider the following system of difference equations. (1.7) .... Using the well-known results of Keel and Tao [7] we obtain the following Strichartz-like.

Warped Mixture Models - GitHub
We call the proposed model the infinite warped mixture model. (iWMM). ... 3. Latent space p(x). Observed space p(y) f(x). →. Figure 1.2: A draw from a .... An elegant way to construct a GP-LVM having a more structured latent density p(x) is.

Statistical Properties of Generalized Discrepancies and ...
Liang et al. ([11]) have started an investigation of the asymp- totic properties of D2 (Pn) and have derived two alternative statistics, strictly linked to this one, that can be used to test statisti- cally the efficiency of numerical integration pro

Statistical Verification of Probabilistic Properties with ...
probabilistic model checking, PRISM [17], relies on iterative methods to verify properties with unbounded until. Each iteration involves a matrix–vector multi- plication, which in the worst case is O(n2), but often O(n) (for sparse models), where n

Statistical Properties and Memory of Excursions in ...
The correlation method applied to the return intervals shows a weak correlation for both groups with changes as the threshold q increases. The DFA analysis.

Statistical Properties and Memory of Excursions in ...
statistics of return intervals between long excursions above ... The DFA analysis confirms the presence ... behavior (shuffled records) is observed in original data.

Statistical Properties and Memory of Excursions in ...
cial indexes [13], x-ray solar flares [14], climate records. [15], or even long heartbeat ..... invariance in the nonstationarity of human heat rate. Phys. Rev Lett 2001 ...

PDF Online Discrete Random Signals and Statistical ...
Consistent treatment of both real and complex cases throughout. Over 300 written problems and ... A disk of data for use with the computer assignments is ...

Exploring Statistical Properties for Semantic Annotation ...
kth most common word. Proof by ... tures,” J. Machine Learning Research, vol. 3, 2003. [6] David M. Blei, ... Statistical Learning in Computer Vision, 2004.

DP Mixture of Warped Correlated GPs for Individualized ...
a set of PPA update equations requires the covariance of the observation point with all existing data. 5 Experiment and Results. Model training and validation are conducted using 29 and 43 sets of heat strain data measured from laboratory and field r

Graphing Sine and Cosine Notes.pdf
Graphing Sine and Cosine Notes.pdf. Graphing Sine and Cosine Notes.pdf. Open. Extract. Open with. Sign In. Main menu.

Properties of Water
electron presence. Electron density model of H2O. 1. How many hydrogen atoms are in a molecule of water? 2. How many oxygen atoms are in a molecule of ...