USABLE SPEECH DETECTION USING A CONTEXT DEPENDENT GAUSSIAN MIXTURE MODEL CLASSIFIER Robert E. Yantorno, Brett Y. Smolenski, Ananth N. Iyer, Jashmin K. Shah Temple University/ECE Dept. 12th & Norris Streets, Philadelphia, PA 19122-6077, USA [email protected], [email protected], [email protected], [email protected] http://www.temple.edu/speech_lab

ABSTRACT Speech that is corrupted by nonstationary interference, but contains segments that are still usable for applications such as speaker identification or speech recognition, is referred to as “usable” speech. A common example of nonstationary interference occurs when there is more than one person talking at the same time, which is known as co-channel speech. In general the above speech processing applications do not work in cochannel environments; however, they can work on the extracted usable segments. Unfortunately, currently available usable speech measures only detect about 75% of the total available usable speech. The first reason for this high error stems from the fact that no single feature can accurately identify all the usable speech characteristics. This situation can be resolved by using a Gaussian Mixture Model (GMM) based classifier to combine several usable speech features. A second source of error stems from the fact that the current usable speech measures treat each frame of co-channel data independently of the decisions made on adjacent frames. This situation can be resolved when a Hidden Markov Model (HMM) is used to incorporate any context dependent information in adjacent frames. Using this approach we were able to obtain 84% detection of usable speech with a 16% false alarm rate.

1. INTRODUCTION In the field of audio restoration, such as the removing of clicks in gramophone recordings, it is common practice to first detect and remove the damaged portions of the signal and then interpolate the removed sections [1]. This approach has the advantage of only processing the damaged portions of the signal. It is interesting to note that this approach to audio restoration is commonly used when dealing with nonstationary interference, but has not yet been applied to the nonstationary interference in cochannel speech. In addition, even for speech having little interference there exists several common situations like coughing, yawning, laughing, etc., that lie outside the training sets of typical speech processing applications. Further, most speech processing systems use feature vectors derived from physical models of the speech production process, such as the Linear Prediction Coefficients (LPC), which assume an all-pole model of the vocal tract [2]. However, for some classes of phonemes, such as nasals, it is known that the underlying production mechanisms are best described using pole-zero models [3]. It is unlikely that the above mentioned kinds of data would be very useful when input to a speech processing application, and hence, one would not want to

processes these segments. The situation is similar to when the statistician identifies and removes outliers or when one chooses not to process low energy and unvoiced speech frames The traditional approach to processing highly corrupted speech has been to enhance the speech while attenuating the interference [4]. However, recently a novel approach to cochannel speech processing has been proposed [5]. Like the audio restoration approach to click removal, the portions of the cochannel speech that are highly corrupted are first detected and removed. Within a co-channel utterance, where both speakers are contributing the same overall energy, there exist several segments of speech where one of the speakers is 15 dB or more above the other speaker [5]. It has been shown that when the target speaker is at least 15 dB greater than the interfering speaker, 80% reliable identification of the target speaker can be obtained [6]. Hence, these segments with a high Target-to-Interferer Ratio (TIR) may be considered usable with respect to speaker identification. The TIR was computed by taking the value, in dB, of the ratio of signal power to interferer power. Since for speaker identification it is not necessary to make a decision on every frame of data, the system can be implemented in a co-channel environment by extracting out and processing only the usable segments. Fortunately, current research has shown that about 35% of a cochannel utterance is usable speech [6]. Recent advances in co-channel speech processing have produced several usable speech measures, which yield some indication of the TIR [7] [8] [9] [10]. Such measures are necessary to determine usability in an operational environment, since a priori knowledge of the TIR would not be available. Unfortunately, currently available usable speech measures only detect about 75% of the total available usable speech. One reason for this high detection error stems from the fact that the measures treat each frame of co-channel data independently of the decisions made on adjacent frames. Another source of error stems from the fact that no single usable speech measure is capable of identifying all of the characteristics of usable speech [11]. It is the goal of this research to increase the performance of usable speech identification by combining several measures and making the classification process context dependant.

The system in Figure 1 (below) illustrates the approach taken. The features used in this research were Linear Prediction Coefficients (LPC) along with a linear discriminant (LD) based feature derived from the LPC residual. The features were then orthogonalized, to make them independent, and passed through a GMM classifier. Previously proposed approaches used the much less sophisticated classification techniques of nonlinear estimation and Quadratic Discriminant Analysis (QDA) on only two usable speech measures, which did not make use of contextual information [11] [12]. The decisions of the GMM are

then passed trough a Maximum Likelihood Sequence Detector to determine the most probable sequence of usable and unusable states given the output of the output of the GMM.

2.2 Gaussian Mixture Model Formally a random vector x that is described by a mixture of M Gaussians has a probability density function of the form:

f ( x) = Co-Channel Frames LPC Analysis

GMM Classifier

A(z)

LD Analysis

1 A( z )



MLSE

Usable/ Unusable Frames

Residual

Figure 1: Block diagram of context dependent usable speech classifier. Using a GMM type classifier is used, the desirable features are those that have a distribution well modeled by a mixture of Gaussians. Although the actual distribution of the LPC’s of a speaker for a particular phoneme may not be Gaussian distributed, the estimate of them is [13]. When one includes the estimates of the LPC’s across several phonemes, a mixture of Gaussians should result regardless of the orthogonalization stage. In addition, Linear Discriminant Analysis (LDA) is used on the LPC residual to yield an additional novel feature that incorporates any remaining useful information. Further research using additional nonlinear features having the above desirable properties is currently ongoing. 2. BACKGROUND 2.1 Linear Discriminant Analysis Linear Discriminant Analysis (LDA) was used in an attempt to capture all the remaining information left in the LPC residual using one additional feature. The goal of linear discriminant analysis is to use a linear transformation to project the set of raw testing data vectors onto a vector space of lower dimension such that some metric of class discrimination is maximized [14]. The metric most often used is the ratio of the between class scatter (variance) to the within class scatter:

trace{S w−1 S b }

λi N ( µ i , Σ i )

(3)

where the N(µ i, i) are multivariate Gaussian distributions having mean vector µ i and covariance matrix I [16]. The λi sum to one and indicate the relative weight of each Gaussian component in the mixture. It can be shown that any distribution can be approximated with arbitrary precision using a mixture of enough Gaussians [14]. To obtain the parameters λi, µ i, and i, the Expectation Maximization (EM) algorithm was used, which is an iterative implementation of maximum likelihood estimation using incomplete information about the underlying probability distributions [16]. Eight mixture components (M=64) were used, since this amount produced the lowest detection error. In general, each covariance matrix in the mixture contains N2 elements (where N is the dimension of the feature vector) that need to be estimated. If the features are chosen such that they are independent, than all the off-diagonal elements of the covariance matrices will be zero [15]. Hence, one would like to use independent features, since only N parameters would need to be estimated. 2.3 Hidden Markov Model In order to make the classifier context dependent, it would be helpful to use a statistical model that exploits as much a priori information about the TIR as possible [14]. One challenge regarding this is that the segmental TIR process is a nonstationary process. To model the non-stationary aspects of the TIR, the following HMM is proposed, Figure 2 (below).

1-p Un -

p

q

Usable Usable

1-q

(1)

The result of this minimization for the two class (usable or unusable) problem is the following linear transformation (matrix equation):

yˆ = ( µ1 − µ 2 ) T S w−1 x

M i =1

(2)

where the µ are the two mean vectors of the two class’s data vectors and x is the data test vector [14]. The mean vectors and within-class-scatter matrix were estimated using the sample mean and sample variance of the training vectors. This transformation produces the 1-dimensional feature yˆ from the LPC-residual data frames, which for this research where 80 samples (10msec frame at 8kHz sampling rate) in length. Hence, the transformation is from ℜ80 to ℜ. Since the feature generated by this approach is a linear combination of a large number of independent identically distributed random variables, the feature’s probability distribution will be highly Gaussian regardless of the distribution of x [15]. Exploring other metrics as well as nonlinear transformations is currently ongoing.

Figure 2: State diagram of the HMM process of co-channel speech frames. We say the model is ‘hidden’ because one cannot observe the actual states, just the statistical characteristics of the signal for a particular state [17]. For the usable state, one person is talking with little interference. For the unusable state, both talkers are contributing about equal energy. Hence, the transition probabilities of this process are related to the statistics of the silent portions of speech. Each state corresponds to a 40ms frame of the co-channel signal and, hence, the signal is quasi-stationary in this time interval [2]. The state transition matrix T of this process is:

T=

p 1− p 1− q q

(4)

where p is the probability of the next frame being usable given that the current frame is usable, and q is the probability of the

next frame being unusable given that the current frame is unusable. These probabilities were estimated using the measure’s training data. One can notice that this model will only make use of dependence between adjacent frames. Fortunately, current research has shown that little dependence exists between anything but adjacent frames [5]. Using the state transition matrix in conjunction with the celebrated Forward-Backward algorithm it is possible to determine the maximum likelihood sequence of states given the output of the GMM classifier [17]. 3. METHODS For this research 10 single speaker utterances, 5 male and 5 female, were taken from the TIMIT speech corpus. These 10 utterances were used to form a co-channel speech database of 45 co-channel files (10 choose 2 = 45). The files were down sampled to 8kHz and the longer file in each pair was truncated to make both files the same length. The files were then combined at 0 dB overall TIR to form the co-channel utterance. To control the variability and eliminate any bias between the dialect regions, only one dialect region was used (region 1 of the TIMIT data base). It should be noted that in an operational environment it is highly unlikely that two speakers would be talking over each other during the entire utterance. In addition, each utterance would not have exactly the same length or have the same energy. The reason for using this approach was to capture the worst possible scenario, with respect to both speakers, that one could expect in a co-channel environment. Once the co-channel utterance was formed it was broken down into 40 ms frames with no overlap, since it has been demonstrated that speaker identification reliability has little dependence on overlap [6]. For each frame, the values for the features, TIR, signal energy, and spectral flatness were computed. Signal energy and spectral flatness were necessary in order to exclude silence and unvoiced frames, since usable speech measures would not be used with these frames. Usable speech measures are designed to work with only voiced speech, since unvoiced frames provide little information useful for speaker identification [6]. Training data was used for obtaining the parameters of the GMM classifier and MLSE detector. Once these models were obtained, it was possible to use testing data to classify what frames of the co-channel speech were usable (|TIR| > 15dB) and those frames that were unusable (|TIR| < 15dB). The absolute value is necessary, since usable speech can come from both speakers. Half of the 45 co-channel speech files (22) were randomly selected to train the system. The remaining half (23 cochannel files) were used for testing. 4. RESULTS Figure 3 (below) shows the classification results for the APPC measure alone, quadratic discriminant analysis (QDA) classifier using APPC and SAPVR-residual measures as features, context independent GMM classifier using the LPC-based features, and the context dependent GMM classifier. Since the minimum probability of error criterion is used in determining the decision boundary surface of the classifier, the percent of Misses (%Misses = 100% - %Hits) equals the percentage of false alarms. However, one can easily choose to weight the false alarms differently than the misses and obtain a different decision boundary surface. The context dependent GMM classifier was able to obtain 84% detection of usable speech with a 16% false

alarm rate. This amounts to a 38% reduction in total detection error over the APPC measure alone.

90% 80% 70% 60% 50% 40% 30% 20% 10% 0%

Hits FA's

APPC

QDA

GMM

CDGMM

Figure 3: Percent hits and False Alarms (FA) for APPC measure, QDA classification using APPC and SAPVR-residual measures, context independent GMM classifier using LPC-based features, and context dependent (CD) GMM classifier. 5. FURTHER RESEARCH More usable features need to be developed. Some current candidates include using pole-zero parameters, sinusoidal model parameters, as well as, nonlinear features of the vocal tract such as those derived from the Teager energy operator [18]. Further, parameters derived from the glottis such as the Liljencrantz-Fant model of the glottal flow derivative, may give an indication of the quality of the speech [19]. In addition, the use of other classification techniques such as Support Vector Machines (SVM) has yet to be explored. The current approach to usable speech segmentation is to partition the signal into short fixed length frames with no overlap. Segmentation is always necessary, since the speech signal is nonstationary [2]. However, a more intelligent approach to segmenting the speech signal would be to identify the stationary regions in the speech signal and process those entire segments. Iterative feature extraction and sequential usable speech detection should improve the resolution capabilities of the classifier as well as make it context dependent by default. Also, usable speech can be defined with respect to the intended application, as opposed to the TIR value, by studying what types of frames work with the system. In addition to improving speaker identification systems, several other applications of usable speech are currently under development including a speaker count and speaker separation system [20]. ACKNOWLEDGEMENT This effort was sponsored by the Air Force Research Laboratory, Air Force Material Command, and USAF, under agreement number F30602-02-2-0501. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright annotation thereon. DISCLAIMER The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily the official policies or endorsements, either expressed or implied, of the Air Force Research Laboratory, or the U.S. Government.

[1]

[2] [3]

[4] [5]

[6]

[7]

[8]

[9]

[10]

[11]

[12]

[13] [14] [15]

[16]

6. REFERENCES S. J. Godsill and P. J. W. Rayner, Digital Audio Restoration: A Statistical Model Based Approach, New York: Springer, 1998. L. R. Rabiner and R. W. Schafer, Digital Processing of Speech Signals, Englewood Cliffs, NJ: Prentice-Hall, 1978. D. O'Shaughnessy, Speech Communications: Human and Machine, New York: Institute of Electrical and Electronics Engineers, 2000. J. S. Lim, ed., Speech Enhancement, Englewood Cliffs, NJ: Prentice-Hall, 1983. R. E. Yantorno, “Co-Channel Speech Study”, Final Report for Summer Research Faculty, Sponsored by AFRL/IF Laboratory, Rome, NY. 1999. J. Lovekin, R. E. Yantorno, D. S. Benincasa, S. J. Wenndt, and M. Huggins, "Developing Usable Speech Criteria for Speaker Identification", ICASSP 2001, pp. 421-424, May 2001. K. R. Krishnamachari, R. E. Yantorno, D. S. Benincasa, and S. J. Wenndt, “Spectral Autocorrelation Ratio as a Usability Measure of Speech Segments Under Co-channel Conditions”, IEEE International Symposium on Intelligent. Signal Processing and Communication Systems, November 2000. J. Lovekin, K. R. Krishnamachari, R. E. Yantorno, D. S. Benincasa, and S. J. Wenndt, “Adjacent Pitch Period Comparison (APPC) as a Usability Measure of Speech Segments Under Co-channel Conditions”, IEEE International Symposium on Intelligent Signal Processing and Communication Systems, November 2001. A. R. Kizhanatham, R. E. Yantorno, S. J. Wenndt, “Cochannel Speech Detection Approaches Using Cyclostationarity or Wavelet Transform”, 4th IASTED International Conference on Signal and Image Processing, July 2002. N. Chandra, R. E. Yantorno, D. S. Benincasa, and S. J. Wenndt, “Usable Speech Detection Using the Modified Spectral Autocorrelation Peak-to-Valley Ratio Using the LPC residual”, 4th IASTED International Conference on Signal and Image Processing, July 2002. B. Y. Smolenski, R. E. Yantorno, and S. J. Wenndt, “Fusion of Co-Channel Speech Measures Using Independent Components and Nonlinear Estimation”, IEEE International Symposium on Intelligent Signal Processing and Communication Systems, November 2002. B. Y. Smolenski, R. E. Yantorno, and S. J. Wenndt, “Fusion of Usable Speech Measures Quadratic Discriminant Analysis”, IEEE International Symposium on Intelligent Signal Processing and Communication Systems, December 2003. S. M. Kay, Fundamentals of Statistical Signal Processing, Englewood Cliffs, NJ: Prentice-Hall, 1998. S. Theodoridis and K. Koutroumbas, Pattern Recognition, San Diego, CA: Academic Press, 1999. H. Stark and J. W. Woods, Probability, Random processes, and Estimation Theory for Engineers, Englewood Cliffs, NJ: Prentice-Hall, 1994. G. J. McLachlan and K. E. Basford, Mixture Models: Inference and Applications to Clustering, New York, NY: M. Dekker, 1988.

[17] X. D. Huang, Y. Ariki, and J. A. Mervyn, Hidden Markov Models for Speech Recognition, Edinburgh: Edinburgh University Press, 1990. [18] T. F. Quatieri, Discrete-time Speech Signal Processing: Principles and Practice, Upper Saddle River, NJ: PrenticeHall, 2002. [19] D. G. Childers, Speech Processing and Synthesis Toolboxes, New York: John Wiley, 2000. [20] B. Y. Smolenski, R. E. Yantorno, D. S. Benincasa, and S. I. Wenndt, “Co-Channel Speaker Segment Separation”, ICASSP, May 2002.

USABLE SPEECH DETECTION USING A CONTEXT ...

ABSTRACT. Speech that is corrupted by nonstationary interference, but contains segments that are still usable for applications such as speaker identification or speech recognition, is referred to as. “usable” speech. A common example of nonstationary interference occurs when there is more than one person talking at.

42KB Sizes 2 Downloads 167 Views

Recommend Documents

Structural Usable Speech Measure Using LPC Residual
of speaker identification systems under degraded condi- tions and use of usable .... false alarm if a frame is identified as usable by the mea- sure, but unusable ...

Structural Usable Speech Measure Using LPC Residual
E-mail: [email protected], [email protected], [email protected] ... also be extended to other applications such as automatic speech recognition ...

Integrating Visual Context and Object Detection within a ...
other promising papers on visual context for object detection [10,8,11,9], we define the .... The LabelMe image database [13] is often used for such pur- poses.

A framework for visual-context-aware object detection ...
destrian detection in urban images using a state-of-the-art pedes- ... nation of this derived context priors with a state-of-the-art object detection ..... For illustration.

User Simulations for context-sensitive speech ...
ing n-best automatic speech recognition ... In such systems, complex repair strategies are re- quired when ..... Recall, F-measure per class and total Accuracy of.

Keylogger Detection Using a Decoy Keyboard
OPC server account, which includes the IP address of the server computer, the name of the .... external IP addresses. In the latter .... RAND CORP (1980). 35.

A Generalized Data Detection Scheme Using ... - Semantic Scholar
Oct 18, 2009 - We evaluated the performance of the proposed method by retrieving a real data series from a perpendicular magnetic recording channel, and obtained a bit-error rate of approximately 10 3. For projective geometry–low-density parity-che

Human eye sclera detection and tracking using a ...
Keywords: Human eye detection; Eye sclera motion tracking; Time-adaptive SOM; TASOM; .... rectly interact with the eye tracking system when it is in oper- ation ...

A study of OFDM signal detection using ... - Semantic Scholar
use signatures intentionally embedded in the SS sig- ..... embed signature on them. This method is ..... structure, channel coding and modulation for digital ter-.

Fast Pedestrian Detection Using a Cascade of Boosted ...
on pedestrian detection using state-of-the-art locally extracted fea- tures (e.g. ... meaningful features if there is a large variation in object's ap- pearance .... The final strong classifier can be .... simple nonpedestrian patterns in the early s

A Survey on Brain Tumour Detection Using Data Mining Algorithm
Abstract — MRI image segmentation is one of the fundamental issues of digital image, in this paper, we shall discuss various techniques for brain tumor detection and shall elaborate and compare all of them. There will be some mathematical morpholog

A Review on Segmented Blur Image using Edge Detection
Image segmentation is an active topic of research for last many years. Edge detection is one of the most important applications for image segmentation.

A Saliency Detection Model Using Low-Level Features ...
The authors are with the School of Computer Engineering, Nanyang ... maps: 1) global non-linear normalization (being simple and suit- ...... Nevrez İmamoğlu received the B.E. degree in Com- ... a research associate for 1 year during 2010–2011 ...

A Framework for Malware Detection Using Ensemble Clustering and ...
Static Analysis module, that provides static information, such as files, antivirus reports, PE structure, file entropy, Packer Signature, and strings. 2. Dynamic Analysis module, which extracts the program behavior, by using a DLL, that will be added

Network Anomaly Detection Using a Commute ...
Anomaly detection in the context of computer network is finding unusual and ... (e.g. Distributed Denial of Service - DDoS) to unusual network traffic (e.g. flash ...

User Simulations for context-sensitive speech ...
1 Introduction. A crucial problem in the design of spoken dia- logue systems (SDS) is to decide for incoming recognition hypotheses whether a system should.

Detection-Based ASR in the Automatic Speech Attribute ...
School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA, USA1. Department of Computer ... As a result, a wide body of expert knowledge in ... Specifically, we present methods of detector design in the Auto-.

Voicing Features for Robust Speech Detection
to greatly varying SNR and channel characteristics and 3) per- formance ..... [3] S. Basu, “A linked-hmm model for robust voicing and speech detection,” in. Proc.

The Usability of Ambiguity Detection Methods for Context-Free ...
Problem: Context-free grammars can be ambiguous ... Overview. 1. Ambiguity in Context-Free Grammars. 2. .... Architecture and Software Technology, 2001.

Image-Based Localization Using Context - Semantic Scholar
[1] Michael Donoser and Dieter Schmalstieg. Discriminative feature-to-point matching in image-based localization. [2] Ben Glocker, Jamie Shotton, Antonio Criminisi, and Shahram. Izadi. Real-time rgb-d camera relocalization via randomized ferns for ke

Lead Detection Using a Pineapple Bioelectrode
fer and mass transport) to diverse applications particu- larly in biosensing and .... apple peeling to the carbon powder is an important factor on the current peak ...

A Framework for Malware Detection Using Ensemble Clustering and ...
information or gain access to the private computer systems without the owner's informed consent (e.g., viruses, backdoors, spyware, Trojans. , and worms). Now a day's malware writers ..... [6] Garfinkel T, Rosenblum M. A Virtual Machine Introspection

A New Shot Change Detection Method Using ...
Department of Electronic Engineering, Shanghai Jiao Tong University ..... Consumer Electronics, vol. 49, pp. ... Circuits and Systems for Video Technology, vol.