2003 2003 ISPACS Awaji Awaji Island Island

2003 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS 2003) Awaji Island, Japan, December 7-10, 2003

C2-3

Structural Usable Speech Measure Using LPC Residual Ananth N. Iyer, Melinda Gleiter, Brett Y. Smolenski and Robert E. Yantorno Speech Processing Laboratory, Temple University 12th & Norris Streets, Philadelphia PA 19122-6077 USA Tel: 215-204-6984 E-mail: [email protected], [email protected], [email protected], [email protected] Abstract: In an operational environment speech is degraded by many kinds of interferences. The operation of many speech processing techniques are plagued by such interferences. Usable speech extraction is a novel concept of processing degraded speech data. The idea of usable speech is to identify and extract portions of degraded speech that are considered useful for various speech processing systems. The performance reduction of speaker identification systems under degraded conditions and use of usable speech concept to improve the performance has been demonstrated in previous work. A new usable speech measure, based on the structure of Linear Predictive Coding (LPC) residual is developed to identify usable speech frames. It is shown that the method has 72% success in identifying the usable frames with 28% false rate.

1. Introduction Speaker identification system performance is plagued by many kinds of situations in an operational environment. Of particular interest is when two or more people are talking at the same time over the same channel, known as co-channel speech. Traditional methods of co-channel speech processing have been to enhance the prominent speaker (target), suppress the interfering speaker speech or both [1]. However, previous studies on co-channel speech have shown that it is desirable to process only portions of the co-channel speech which are minimally degraded [2]. Such portions of speech considered usable for speaker identification are referred to as “usable speech”. The concept of usable speech can also be extended to other applications such as automatic speech recognition [3]. Recently, a usable speech extraction system was proposed to classify co-channel speech frames into usable speech and unusable speech for speaker identification [4]. The speech segments can be declared usable based upon a Target-to-Interferer energy ratio (TIR) threshold. Studies indicate that by considering only usable speech segments for speaker identification, there is a considerable increase in the accuracy of speaker identification [5]. The block diagram of the usable speech extraction system and speaker identification is shown in Figure 1. The segment separator block [6] is employed to distinguish frames of speech data belonging to each

236

speaker. Several usable speech measures are being developed to identify usable speech frames [7] [8] [9] [10] [11]. These measures are used in a fusion/desicion system to reliably identify usable speech frames [12] [13] [14] by harnessing the complementary information provided by the usable speech measures. Speech

-

Usable Speech Extractor

-

Segment Separator

Speaker - Identification

SID

-

Figure 1. Block diagram showing the application of usable speech detection Hence to achieve higher identification accuracy, one needs to fuse many measures providing complimentary information. A new usability measure for spotting the usable speech segments from a co-channel utterance is proposed here. This measure is based on the fact that there is a significant change in the excitations of speech belonging to the two classes in co-channel speech, i.e., usable speech and unusable speech.

2. Background 2. 1 Linear Prediction The composite spectrum effects of radiation and vocal tract is assumed to be represented by a time-varying digital filter whose steady-state system function is of the form G S(z) Pp (1) = H(z) = U (z) 1 − k=1 αk z −k where αk are the digital filter coefficients. This all-pole system is excited by the voice source to produce speech data. A linear predictor with predictor coefficients, αk is defined as a system whose output is sˆ(n) =

p X

αk s(n − k)

(2)

k=1

The all-pole linear predictor models a signal sn by a linear combination of its past values. The LPC analysis is performed to determine the predictor coefficients αk , which represents the vocal tract model of the speaker, directly from the speech signal.

2. 2 Minimum Description Length Under co-channel conditions one would expect the model order of the LPC analysis needed to be twice that used for a single speaker speech. Hence the model order for LPC analysis is dynamically selected for each frame of speech. Principle of Minimum Description Length (MDL) is employed to choose the model order that gives the shortest description of the speech frame data [15]. The model order i is determined by using the equation given below. The value of i is incremented and the the order which leads the the lowest MDL is chosen as the order for LPC analysis. Mi = N log10 (Ps )2 + i log10 (N)

(3)

where N is the speech frame size and Ps is the residual power.

residual of the unusable frame. The LPC residual shows no presence of definite pulses and is unstructured. The property of usable speech residual having a good periodic structure and unusable speech having no definite structure was used to develop a usable speech measure. Speech Data 2000 1000 0 −1000

0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.015 0.02 0.025 Time (Seconds)

0.03

0.035

0.04

LPC Residual 1500 1000 500 0

2. 3 LPC Residual of Co-channel Speech

−500

The LPC residual is spectrally flattened speech, where the vocal tract shape is removed by creating an inverse filter using the LPC coefficients. Figure 2, below shows a typical usable frame, which is voiced and having a high TIR value, exhibiting good periodic structure (upper panel). The bottom panel of the figure is the LPC residual of the speech frame. It can be seen that the periodic structure of usable speech is represented as pulses with constant frequency in the LPC residual of speech [16]. Speech Data 4000 2000 0 −2000 0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.015 0.02 0.025 Time (Seconds)

0.03

0.035

0.04

LPC Residual

2000 1000 0 −1000 0

0.005

0.01

Figure 2. Usable speech frame (top panel) and the corresponding LPC residual of speech frame (lower panel) showing good structure Figure 3 shows, an example of an unusable (mixed speech data from two speakers) speech frame (upper panel). The frame of speech shown is voiced and has low TIR value. It is seen that the speech data has no defined periodic pattern. The lower panel shows the LPC

237

0

0.005

0.01

Figure 3. Unusable speech frame (top panel) and the corresponding LPC residual of speech frame (lower panel) showing no presence of structure

3. Algorithm Description The presence of equidistant pulses in the residual suggested the possibility of measuring the distances between the pulses and use the distance to measure the periodic structure of speech frame. The various steps to obtain the distances is explained below 1. Processing of data was performed on a frame-byframe basis. A frame of length 32ms equal to 240 samples for speech data sampled at 8kHz, from the co-channel utterance was used. The frame was determined to be - voiced, unvoiced or silence. Unvoiced and silent frames were rejected from further processing. Voiced/unvoiced decision was done using the Spectral Flatness Method [17] and silence decision was done using an energy threshold. 2. LPC model order was determined based on minimum description length. 3. The LPC coefficients were used to create an inverse filter and the speech frame was filtered using the inverse filter. The output of the inverse filter is the LPC residual. 4. An intelligent peak-picking algorithm was designed to determine the prominent peaks from the residual frame. The peak-picker was implemented based on a sorting method and selecting the peaks which satisfies three conditions - a) the magnitude of the peak should be at least 40% of the highest peak; b) The distance between the peaks picked should be greater than 30 time points

(3.7ms); and c) the distance between the peaks picked should be smaller that 100 time points (12.5ms). If any of the peaks picked did not satisfy these conditions, it was dropped. The peak-picking algorithm was run iteratively until only peaks satisfying the above conditions were picked. 5. The distances between the peaks picked in the previous step were determined. These distances represent the pitch period of the speech frame. Hence the distances in male speech would nearly be twice that of a female speech. The peak distances are normalized such that the distances for both male speech and female speech lie in the same range. The normalized distances is the product of the individual distances between the peaks and the number of peaks in the frame. For example, female speech with pitch 200 Hz or 40 points and 7 peaks would result in a value of 280, and male speech of pitch 100 Hz or 80 points and 3 peaks would result in a value of 240. 6. The average of the normalized peak distances in each frame was computed and used as a usable speech measure. A threshold was determined based on the constructed probability density function.

4. Experimental Results

1.5

1

0.5

0

The LPC residual peak based usable speech measure developed above was tested on co-channel data obtained from the TIMIT database. To determine the optimum threshold for the proposed usable speech measure, the probability density function of the measure values was generated and is shown in Figure 4. Usable Speech Unusable Speech

−0.5

−1

−1.5 0

0.5

1

1.5 2 Time Points

2.5

3 4

x 10

Figure 5. Usable speech detected by TIR and usable speech measure. TIR usable speech (black), TIR unusable speech (gray), Usable speech measure detected frames (black box).

0.12

0.1 Probability

TIR < –20dB [5] and the black bars represent the measure values for unusable speech frames. From the probability density function, a threshold value of 100 was considered optimum. To determine the effectiveness of the proposed measure in spotting usable speech frames, the percent correct identifications (hits) and percent incorrect identifications (false alarms) was determined. The measure is said to have a hit if, the measure as well as TIR identify a frame as usable. The measure is said to have a false alarm if a frame is identified as usable by the measure, but unusable based on TIR. The proposed measure has resulted in 72% correct decisions or hits and 30% incorrect decisions or false alarms. These values were obtained on a set of 861 co-channel utterances. These utterances were created using 21 male speech and 21 female speech files.

0.08

0.06

0.04

0.02

0 0

50

100 150 Measure Value

200

Figure 4. Probability density function - The mean value of normalized peak distances in each frame. Black bars represent unusable data and gray bars represents usable data. For the Figure 4 above, the gray bars in the probability density function represent measure value for the usable speech frames, i.e., frames with TIR > 20dB or

238

The rectangles above the speech in Figure 5 indicates detection of usable speech, a value of +1 indicates a hit in identifying usable speech and a value of –1 indicates a false alarm. Portions of speech having a TIR value of 20dB or above are shown in gray. The portions of the co-channel utterance in black unusable speech frames. A closer look at the false alarms in the usable speech measure, reveals that many frames lie in a transition region - from voiced to unvoiced/silence. This suggests that the false alarm rate can be reduced if such transition frames are identified and eliminated from processing. The identified usable speech frames were extracted and tested in a speaker identification system, which uses LPC-Cepstrum features and a Vector Quantization scheme to perform the classification. The system was tested using the extracted usable speech frames and the

resulted in 82% correct identification compared to 45% correct detection, with co-channel speech. [5]

5. Summary The purpose of this paper was to identify the usable portions of co-channel speech in the context of speaker identification. It was found that the structural changes in the LPC residual of speech can be used as an effective method to identify usable speech segments. The proposed usable speech measure detected 72% of the usable portions from co-channel data. Further improvements in this algorithm are possible, to make the performance more robust. One possible scenario would be to identify and remove outlier frames such as those lying in transition regions. It is also expected that the performance of usable speech extraction system would improve with the inclusion of this new measure. Further, the LPC coefficients can be potentially used to develop a model based usable speech measure.

[6]

[7]

[8]

Acknowledgements The Air Force Research Laboratory, Air Force Material Command, and USAF sponsored this effort, under agreement number F30602-02-2-0501. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright annotation thereon. Further we wish to extend thanks to Nithya Sundaram for providing information about the principle of minimum description length for model order selection.

[9]

[10]

Disclaimer The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of Air Force Research Laboratory, or the U.S. Government.

[11]

References [1] J. A. Naylor and S. F. Boll, “Techniques for suppression of an interfereing talker in co-channel speech,” Proc. IEEE ICASSP, pp. 205–208, 1987. [2] R. E. Yantorno, “Co-channel speech study, final report for summer research faculty program,” tech. rep., Air Force Office of Scientific Research, Speech Processing Lab, Rome Labs, New York, 1999. [3] W. T. Hicks, B. Y. Smolenski, and R. E. Yantorno, “Testing for intelligibility of corrupted speech with an automated speech recognition system,” IIIS Systemics, Cybernetics and Informatics, 2003. [4] R. E. Yantorno, “Co-channel speech and speaker identification study,” tech. rep., Air Force Office of

239

[12]

[13]

[14]

Scientific Research, Speech Processing Lab, Rome labs, New York, 1998. J. M. Lovekin, R. E. Yantorno, K. R. Krishnamachari, D. Benincasa, and S. J. Wenndt, “Developing usable speech criteria for speaker identification,” IEEE, International Conference on Acousitcs and Signal Processing, pp. 424–427, May 2001. B. Y. Smolenski, R. E. Yantorno, D. S. Benincasa, and S. J. Wenndt, “Co-channel speaker segment separation,” IEEE, International Conference on Acousitcs and Signal Processing, May 2002. K. R. Krishnamachari and R. E. Yantorno, “Spectral autocorrelation ratio as a usability measure of speech segments under co-channel conditions.,” IEEE International Symposium on Intelligent Signal Processing and Communication Systems, Nov 2000. J. M. Lovekin, K. R. Krishnamachari, and R. E. Yantorno, “Adjacent pitch period comparison (appc) as a usability measure of speech segments under co-channel conditions,” IEEE International Symposium on Intelligent Signal Processing and Communication Systems, Nov 2001. N. Chandra and R. E. Yantorno, “Usable speech detection using modified spectral autocorrelation peak to valley ratio using the lpc residual,” 4th IASTED International Conference Signal and Image Processing, pp. 146–150, 2002. A. R. Kizhanatham, R. E. Yantorno, and B. Y. Smolenski, “Peak difference autocorrelation of wavelet transform (pdawt) algorithm based usable speech measure.,” IIIS Systemics, Cybernetics and Informatics, 2003. N. Sundaram, A. N. Iyer, B. Y. Smolenski, and R. E. Yantorno, “Usable speech detection using linear predictive analysis - a model-based approach,” IEEE International Symposium on Intelligent Signal Processing and Communication Systems, ISPACS 2003, 2003. B. Y. Smolenski and R. E. Yantorno, “Fusion of cochannel speech measures using independent components and nonlinear estimation.,” IEEE International Symposium on Intelligent Signal Processing and Communication Systems, ISPACS, 2002. B. Y. Smolenski and R. E. Yantorno, “Fusion of usable speech measures using quadratic discriminant analysis.,” IEEE, International Symposium On Intelligent Signal Processing and Communication Systems (submitted), 2003. J. K. Shah, B. Y. Smolenski, and R. E. Yantorno, “Decision level fusion of usable speech measures using consensus theory,” IEEE International Sympo-

sium on Intelligent Signal Processing and Communication Systems, ISPACS 2003, 2003. [15] M. H. Hansen and B. Yu, “Model selection and the principle of minimum description length,” Journal of the American Statistical Association, vol. 96, no. 454, pp. 746–774, 2001. [16] M. Sharma, Y. Kogan, R. Ramachandran, and Y. Li, “Speaker count determination, final technical report,” Tech. Rep. AFRL-IF-RS-TR-1999-57, TNetix, Inc., Air Force Research Laboratory, Rome, NY, April 1999. [17] D. S. Benincasa, “Voicing state determination of co-channel speech,” IEEE, International Conference on Acousitcs and Signal Processing, vol. II, pp. 1021–1024, 1998.

240

Structural Usable Speech Measure Using LPC Residual

E-mail: [email protected], [email protected], [email protected] ... also be extended to other applications such as automatic speech recognition ...

331KB Sizes 2 Downloads 171 Views

Recommend Documents

Structural Usable Speech Measure Using LPC Residual
of speaker identification systems under degraded condi- tions and use of usable .... false alarm if a frame is identified as usable by the mea- sure, but unusable ...

USABLE SPEECH DETECTION USING A CONTEXT ...
ABSTRACT. Speech that is corrupted by nonstationary interference, but contains segments that are still usable for applications such as speaker identification or speech recognition, is referred to as. “usable” speech. A common example of nonstatio

LPC Checklist.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. LPC Checklist.

Residual Deterrence
School of Economics, and at the University of Edinburgh for helpful discussions. ... drink driving, lead to reductions in offending that extend past the end of the ...

IC_55.Dysarthric Speech Recognition Using Kullback-Leibler ...
IC_55.Dysarthric Speech Recognition Using Kullback-Leibler Divergence-based Hidden Markov Model.pdf. IC_55.Dysarthric Speech Recognition Using Kullback-Leibler Divergence-based Hidden Markov Model.pdf. Open. Extract. Open with. Sign In. Main menu.

using glycol to effectively control distortion and residual ...
quenching and as a result the full benefit of this technology is not currently being ..... cool the furnace down to room temperature, remove the parts from the ...

Speech Recognition Using FPGA Technology
Figure 1: Two-line I2C bus protocol for the Wolfson ... Speech recognition is becoming increasingly popular and can be found in luxury cars, mobile phones,.

Speech Recognition Using FPGA Technology
Department of Electrical Computer and Software Engineering ..... FSM is created to implement the bus interface between the FPGA and the Wolfson. Note that ...

LPC corrected NLJAN2017.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. LPC corrected ...

LPC - Reactivate Inactive License.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. LPC - Reactivate Inactive License.pdf. LPC - Reactivate Inactive License.pdf. Open. Extract. Open with. Sign

LPC - Reinstate Expired License.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. LPC - Reinstate ...Missing:

AITP - LPC-format (1).pdf
Mk; ehs; Kw;gfy; / gpw;gfy; mYtyfj;jpd; nghWg;Gfis xg;gilj;jhh;. 5. Nkw;gb muR Copahpd; Cjpaq;fs; Kjypatw;wpypUe;J nra;ag;gl Ntz;ba gpbj;jq;fspd; tptuk;. gpd;gf;fk ...

Using a Sensitivity Measure to Improve Training ...
Engineering, Hohai University, Nanjing 210098, China (email: [email protected]). In our study, a new learning algorithm based on the MRII algorithm is developed. We introduce a sensitivity of. Adalines, which is defined as the probability of an Adalin

Can we measure influenza vaccine effectiveness using ...
Spanish Influenza Sentinel Surveillance System, 2008-9. Instituto. De Salud. Carlos III. C. Savulescu, S. de Mateo, M. Valenciano,. A. Larrauri, and cycEVA study ...

Designing Usable Web Forms - Research at Google
May 1, 2014 - 3Dept. of Computer Science ... guidelines to improve interactive online forms when .... age, level of education, computer knowledge, web.

Residual Deterrence
ual deterrence occurs when reductions in offending follow a phase of active .... evaluation of “top antitrust authorities” focuses on successful prosecutions. ... similar features to ours in the so-called “bad news” case, where the worker inc

Residual Deterrence
to Environmental Protection Agency fines against other nearby mills (Shimshack and. Ward, 2005); reductions ... offense is only worthwhile for a firm if the regulator is not inspecting, while inspecting is only worthwhile for the ..... to have data o

Quantifying Residual Finiteness
Aug 3, 2015 - Then FL. H(n) ≼ FS. G(n). Proof. As any homomorphism of G to Q restricts to a ..... Mathematics Department, Queen Mary College, Lon-.

Isolated Tamil Word Speech Recognition System Using ...
Speech is one of the powerful tools for communication. The desire of researchers was that the machine should understand the speech of the human beings for the machine to function or to give text output of the speech. In this paper, an overview of Tam

audio-visual speech integration using coupled hidden ...
Email: {asubram, sabrig, epatter, jgowdy}@clemson.edu. Abstract. In recent years ... visual information for the purposes of Automatic Speech. Recognition (ASR) ...

Fully Automated Non-Native Speech Recognition Using ...
tion for the first system (in terms of spoken language phones) and a time-aligned .... ing speech recognition in mobile, open and noisy environments. Actually, the ...

Multilingual Non-Native Speech Recognition using ...
cept that associates sequences of native language (NL) phones to spoken language (SL) phones. Phonetic confusion rules are auto- matically extracted from a ...