Introduction Theory Examples Conclusion

Measuring, refining and calibrating speaker and language information extracted from speech Niko Brümmer

Ph.D. Defence, 18 October 2010 Dept. E&E Engineering. Univ. Stellenbosch Promoter: Prof. J.A. du Preez

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Outline

1

Introduction

2

Theory

3

Examples

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Y

We can now rewrite the integral (for the case i = N ) as: Z Γ(N )C ∗ (p|θN ) dη N Z ∞ Z y1 Z ∞ Z ∞ ··· I(x) dxN −1 dxN −2 · · · dx1 = −∞ −∞ Z−∞ Z −∞ ∞ Z y2 Z ∞ ∞ + ··· I(x) dxN −1 dxN −2 · · · dx1 −∞ −∞ Zy1∞ Z−∞ Z Z ∞ y3 ∞ + ··· I(x) dxN −1 dxN −2 · · · dx1 y1

y1



y2

−∞

−∞

y2

+··· Z ∞Z +

Z



y3

···

Z

yN −1

−∞

I(x) dxN −1 dxN −2 · · · dx1

where PN −1

N

I(x) =

Γ(N − 1) Y ηk = Γ(N − 1)  ηN k=1

Niko Brümmer

Ph.D. Defence

1+

e k=1 PN −1 k=1

xk

exk

N −1

Introduction Theory Examples Conclusion

Theorem 1

Lemmas 2&3

Lemma 4

Theorem 3

Lemmas 5 & 6

Lemma 1 Theorem 2

Theorem 4

PAV algorithm

Theorem 5: PAV-LLR algorithm Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

Introduction 1

Introduction Subject definition Speaker and language recognition Representing uncertainty Goals

2

Theory

3

Examples

4

Conclusion Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

The basic problem

How to evaluate the goodness of a certain class of automatic pattern recognizers. Automatic pattern recognizers are not infallible. They make errors. If one wants to design, improve, sell, buy, or use a pattern recognizer, then it is important to have some understanding of these errors. In short, the need exists to evaluate the goodness of pattern recognizers.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

The pattern recognizers

We discuss two kinds of pattern recognizers that automatically extract information from speech: Automatic speaker recognition Automatic spoken language recognition Although this work is grounded in the literature and presented in the terminology of these two fields, it is more generally applicable also to other pattern recognition problems.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

The canonical speaker recognizer

input

pattern recognizer

recognized class

Class 1: The segments are of the same speaker.

speech segment 1 speaker recognizer speech segment 2

Class 2: The segments are of two different speakers.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

Language recognizer

input

pattern recognizer

recognized class

Afrikaans

English speech segment

language recognizer Xhosa

Zulu

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

The question is simple

The questions asked of speaker and language recognizers are very simple: Is it the same speaker or not? Which of these four languages is it? There is a small discrete number of possibilities. Compare this to speech recognition, where the question, “What was said?”, is much more complex, because there are very many possibilities.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

The answer is not

The answers to these questions are however complicated by the fact that with current technology, they cannot be answered with certainty. An honest (and therefore more useful) pattern recognizer should reflect the degree of uncertainty in its answer.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

How to reflect uncertainty existing state of the art

For example, a speaker recognizer could output: Hard discrete decisions: class 1 or class 2. This is a poor solution, with no indication of uncertainty. Scores: more positive scores favour the same-speaker hypothesis, more negative scores favour the different-speaker hypothesis. This is a good solution and is still part of the current state of the art. But the score is uncalibrated: it is up to the user to exercise the recognizer in order to learn how to interpret the score magnitude as an indication of uncertainty. Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

How to reflect uncertainty proposed here

or the speaker recognizer could output various forms of calibrated scores: Probability distribution: P(class 1|speech,prior),

P(class 2|speech,prior)

Likelihood distribution: L(class 1|speech),

L(class 2|speech)

Likelihood and posterior probability are closely related, but the likelihood (our preferred format), is prior-independent and more useful because it allows user-supplied priors. Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

Goals

The main goals of this work were: 1

To evaluate the goodness of speaker and language recognizers (or other similar pattern recognizers) that provide outputs in class likelihood format.

2

Given that we can measure goodness of pattern recognizers, how can we improve them?

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Subject definition Speaker and language recognition Representing uncertainty Goals

Relevance This work builds on the series of NIST Speaker Recognition Evaluations and NIST Language Recognition Evaluations, which have been a major driving force for research in these fields for more than a decade. In the period 2000 to 2010, the author has participated in 7 speaker recognition and 3 language recognition evaluations. The practical algorithms developed in this work have been made available in a MATLAB toolkit, the FoCal Toolkit, which has been used by many other researchers, especially for their work in the NIST evaluations.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Theory 1

Introduction

2

Theory Why likelihoods? Evaluation Calibration Discriminative training

3

Examples

4

Conclusion Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Why likelihoods?

Likelihoods convey the information extracted from the speech by the recognizer, which the user can employ to make optimal Bayes decisions. The information in the likelihoods is application independent. The Bayes decision framework allows the user to apply the likelihoods to a wide range of different applications.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

input: speech

Why likelihoods? Evaluation Calibration Discriminative training

recognizer likelihoods: α P (speech | class 1) α P (speech | class 2)

prior: P (class 1 | other info) P (class 2 | other info)

Bayes’ rule

recognizer

user makes Bayes decisions

posterior: P (class 1 | speech, other info) P (class 2 | speech, other info)

cost: C (decision j | class i ) i = 1… N, j = 1…K

choose decision with optimal expected outcome

Niko Brümmer

Ph.D. Defence

decision

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Cost Example A language recognition application

prior, cost speech in unknown language

language recognizer

likelihoods

Bayes decision agent

Koos John Nelson Jacob

Afrikaans 0 1 1 1

English 0 0 0 0

Xhosa 2 2 0 1

Zulu 2 2 1 0

Cost of assigning a speech segment to an agent. Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

input: speech

Why likelihoods? Evaluation Calibration Discriminative training

recognizer likelihoods: α P (speech | class 1) α P (speech | class 2)

prior: P (class 1 | other info) P (class 2 | other info)

applicationindependent recognizer

user Bayes’ rule

posterior: P (class 1 | speech, other info) P (class 2 | speech, other info)

applicationdependent parameters cost: C (decision j | class i ) i = 1… N, j = 1…K

choose decision with optimal expected outcome

Niko Brümmer

Ph.D. Defence

decision

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

application-dependent recognizer input: speech

user

recognizer likelihoods: α P (speech | class 1) α P (speech | class 2)

prior: P (class 1 | other info) P (class 2 | other info)

Bayes’ rule posterior: P (class 1 | speech, other info) P (class 2 | speech, other info)

cost: C (decision j | class i ) i = 1… N, j = 1…K

choose decision with optimal expected outcome

Niko Brümmer

Ph.D. Defence

decision

Introduction Theory Examples Conclusion

input: speech

Why likelihoods? Evaluation Calibration Discriminative training

recognizer likelihoods: α P (speech | class 1) α P (speech | class 2)

prior: P (class 1 | other info) P (class 2 | other info)

Bayes’ rule posterior: P (class 1 | speech, other info) P (class 2 | speech, other info)

applicationdependent recognizer user

cost: C (decision j | class i ) i = 1… N, j = 1…K

choose decision with optimal expected outcome

Niko Brümmer

Ph.D. Defence

decision

Introduction Theory Examples Conclusion

input: speech

Why likelihoods? Evaluation Calibration Discriminative training

recognizer likelihoods: α P (speech | class 1) α P (speech | class 2)

prior: P (class 1 | other info) P (class 2 | other info)

applicationindependent recognizer

user Bayes’ rule

posterior: P (class 1 | speech, other info) P (class 2 | speech, other info)

applicationdependent parameters cost: C (decision j | class i ) i = 1… N, j = 1…K

choose decision with optimal expected outcome

Niko Brümmer

Ph.D. Defence

decision

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Theory Evaluation

1 2

Introduction Theory Why likelihoods? Evaluation Calibration Discriminative training

3

Examples

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Recognizer evaluation To evaluate the goodness of speaker or language recognizers, we follow the paradigm provided by the NIST Speaker Recognition Evaluations and the NIST Language Recognition Evaluations: The recognizer under evaluation is exercised on a large database of speech inputs, producing an output in likelihood form for each input. The evaluator compares the recognizer’s likelihoods with the true class labels to produce a summary of the goodness of the recognizer.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Evaluation via supervised evaluation database

evaluation database: speech input for trial 1 speech input for trial 2 … …

recognizer

likelihoods for trial 1 likelihoods for trial 2 … … evaluation database: true class label 1 true class label 2 … …

evaluator

criterion of goodness of recognizer (as evaluated on this database)

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Supervised evaluation of likelihoods

How does the evaluator judge the goodness of the recognizer’s likelihoods? The evaluator is not given ‘true’ likelihoods to compare against. The evaluator has only the true class labels.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Evaluation by Bayes decision

How does the evaluator judge the goodness of the recognizer’s likelihoods? Apply the likelihoods for their designed purpose: use them to make Bayes decisions. Evaluate the goodness (cost) of those decisions.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database input 1 input 2 … input t …

black box recognizer

fixed evaluation parameters

? decision t

true class of trial t

prior

cost

Σ P(class t ) x C( decision t | class t ) t

expected cost

existing NIST recipe evaluation criterion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database input 1 input 2 … input t …

recognizer

true class of trial t

uncalibrated scores

fixed evaluation parameters

variety of ad hoc solutions decision t

prior

cost

Σ P(class t ) x C( decision t | class t ) t

expected cost

existing NIST recipe evaluation criterion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database input 1 input 2 … input t …

recognizer calibrated scores: likelihoods

true class of trial t

variable evaluation parameters Bayes decision

decision t

prior

cost

Σ P(class t ) x C( decision t | class t ) t

expected cost

proposed evaluator evaluation criterion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database input 1 input 2 … input t …

recognizer calibrated scores: likelihoods

true class of trial t

evaluation parameters Bayes decision

decision t

prior

cost

Σ P(class t ) x C( decision t | class t ) t

expected cost

proposed: Bayes cost evaluator evaluation criterion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database

recognizer

input 1 input 2 … input t …

likelihoods

application-dependent evaluation parameters: prior, cost

Bayes cost evaluator

application-dependent evaluation criterion

Niko Brümmer

Ph.D. Defence

true class of trial t

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

evaluation database input 1 input 2 … input t …

recognizer likelihoods



prior( λ ) cost( λ )

Bayes cost evaluator



application-spanning evaluation criterion

Niko Brümmer

Ph.D. Defence

true class of trial t

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Solving the integral



prior( λ ) cost( λ )

Bayes cost evaluator



With appropriate choices of the parametrizations, prior(λ) and cost(λ), this integral can be solved analytically. This solution forms a family of strictly proper scoring rules, a tool from statistics literature for the evaluation of the goodness of for example probabilistic weather forecasts.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Y

Why likelihoods? Evaluation Calibration Discriminative training

We can now rewrite the integral (for the case i = N ) as: Z Γ(N )C ∗ (p|θN ) dη N Z ∞ Z y1 Z ∞ Z ∞ ··· I(x) dxN −1 dxN −2 · · · dx1 = −∞ −∞ Z−∞ Z −∞ ∞ Z y2 Z ∞ ∞ + ··· I(x) dxN −1 dxN −2 · · · dx1 −∞ −∞ Zy1∞ Z−∞ Z Z ∞ y3 ∞ + ··· I(x) dxN −1 dxN −2 · · · dx1 y1

y1



y2

−∞

−∞

y2

+··· Z ∞Z +

Z



y3

···

Z

yN −1

−∞

I(x) dxN −1 dxN −2 · · · dx1

where PN −1

N

I(x) =

Γ(N − 1) Y ηk = Γ(N − 1)  ηN k=1

Niko Brümmer

Ph.D. Defence

1+

e k=1 PN −1 k=1

xk

exk

N −1

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Bayes cost evaluation Summary

The recognizer’s likelihoods are evaluated by how good the decisions are that those likelihoods can make. This gives a family of practical evaluation recipes, which can be used in two ways: 1

Evaluation over a range of different application parameters (prior, cost), which can be displayed graphically.

2

Or integrated, to provide an application-spanning, scalar measure of goodness.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Theory Calibration

1 2

Introduction Theory Why likelihoods? Evaluation Calibration Discriminative training

3

Examples

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Calibration Example of a calibration problem

Tue 68

Wed 72

Thu 64

Fri 63

Sat 64

Predicted1 maximum temperature for the next 5 days. Can this be an accurate prediction?

1

http://www.weathersa.co.za Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Calibration Example of a calibration problem

◦F ◦C

= 59 (◦ F − 32)

Tue 68 20

Wed 72 22

Thu 64 18

Fri 63 17

Sat 64 18

Predicted maximum temperature for the next 5 days, now re-calibrated, so that we can understand it.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Calibration

Probabilities and likelihoods can also be badly calibrated in the same way as the above temperature prediction: all the information is there, but not in the format we expect it. In this work we propose ways to: Measure the degree of miscalibration of a pattern recognizer. Re-calibrate the recognizer to improve calibration. All of this is still based on Bayes decisions.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Theorem 1

Lemmas 2&3

Why likelihoods? Evaluation Calibration Discriminative training

Lemma 4

Theorem 3

Lemmas 5 & 6

Lemma 1 Theorem 2

Theorem 4

PAV algorithm

Theorem 5: PAV-LLR algorithm Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Theory Discriminative training

1 2

Introduction Theory Why likelihoods? Evaluation Calibration Discriminative training

3

Examples

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Why likelihoods? Evaluation Calibration Discriminative training

Discriminative training FoCal Toolkit

A scalar evaluation criterion (such as we formed with the above integral) is the most important ingredient needed for discriminative training. In particular, one of the members of this family, the logarithmic proper scoring rule, has many desirable properties as a discriminative training objective function. The FoCal Toolkit (the practical embodiment of this work, which is used by many other researchers) uses the logarithmic scoring rule as discriminative training criterion to optimize calibration and fusion transformations of the scores of speaker and language recognizers.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

Examples

1

Introduction

2

Theory

3

Examples Speaker Recognition Language Recognition

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

Speaker Recognition Examples Calibration analysis

Below we show two examples of our evaluation methods as applied to two different speaker recognition systems in the NIST 2010 Speaker Recognition Evaluation. 1

The first example shows a recognizer with good calibration over a wide range of different operating points.

2

The second example shows a recognizer which could have been good, but bad calibration spoils the applicability of this recognizer at most operating points.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

BUT PLDA i−vector condition 2 new DCF point dev misses dev false−alarms dev act DCF eval misses eval false−alarms eval min DCF eval act DCF eval DR30

1

normalized DCF

0.8

0.6

0.4

0.2

0 −10

−9

−8

−7

−6

Niko Brümmer

−5 logit Ptar

−4

Ph.D. Defence

−3

−2

−1

0

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

BUT i−vector full−cov condition 2 new DCF point dev misses dev false−alarms dev act DCF eval misses eval false−alarms eval min DCF eval act DCF eval DR30

1

normalized DCF

0.8

0.6

0.4

0.2

0 −10

−9

−8

−7

−6

Niko Brümmer

−5 logit Ptar

−4

Ph.D. Defence

−3

−2

−1

0

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

Speaker Recognition Example Discriminatively trained fusion

Below is an example of a discriminatively trained fusion of multiple speaker recognition subsystems in the NIST 2006 Speaker Recognition Evaluation. The fusion shows a dramatic improvement in accuracy.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

DET1: 1conv4w-1conv4w 40

STBU sub-systems and fusion, NIST SRE 2006

30

Mis s proba bility (in %)

20

10

sub-systems

5 2 1

fusion

0.5 0.2 0.1 0.1 0.2

0.5

1 2 5 10 20 Fa ls e Ala rm proba bility (in %)

Niko Brümmer

Ph.D. Defence

30 40

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

Language Recognition Example Discriminatively trained calibration

The next slide shows an example of four different language recognizers, submitted by other researchers to the NIST 2007 Language Recognition Evaluations. The blue bars show the error-rates of the originally submitted systems. The red bars show the improved accuracy of the the same systems, re-calibrated by using the FoCal Toolkit.

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Speaker Recognition Language Recognition

3 original re-calibrated

2.5

CAVG %

2

1.5

1

0.5

0

1

2

3 recognizers

Niko Brümmer

Ph.D. Defence

4

Introduction Theory Examples Conclusion

Conclusion

1

Introduction

2

Theory

3

Examples

4

Conclusion

Niko Brümmer

Ph.D. Defence

Introduction Theory Examples Conclusion

Conclusion

In summary: Pattern recognizers with likelihood outputs are good for making Bayes decisions. Bayes decisions are good for evaluating such pattern recognizers. Such evaluation is good for building better recognizers. The dissertation and associated software can be downloaded from http://niko.brummer.googlepages.com.

Niko Brümmer

Ph.D. Defence

Measuring, refining and calibrating speaker and ...

40. False Alarm probability (in %). M is s p ro b a b ility (in. %. ) DET1: 1conv4w-1conv4w sub-systems fusion. STBU sub-systems and fusion,. NIST SRE 2006.

416KB Sizes 1 Downloads 131 Views

Recommend Documents

Measuring, refining and calibrating speaker and ...
George Doddington again, whose crazy idea to shift the operating point by two orders of ..... If one wants to design, improve, sell, buy, or use a pattern recognizer, then it is important ..... 5available online at http://niko.brummer.googlepages.com

Testing and Measuring Instruments
United Systems & Automation is a business enterprise deals in testing, measuring instruments and automation products and it is one of the fastest-growing Automation company in Mohali Punjab. Having built a large clientele in the domestic market, our

PETROLEUM REFINING AND PETROCHEMICAL TECH.pdf
PETROLEUM REFINING AND PETROCHEMICAL TECH.pdf. PETROLEUM REFINING AND PETROCHEMICAL TECH.pdf. Open. Extract. Open with. Sign In.

Efficient Speaker Identification and Retrieval - Semantic Scholar
Department of Computer Science, Bar-Ilan University, Israel. 2. School of Electrical .... computed using the top-N speedup technique [3] (N=5) and divided by the ...

Efficient Speaker Identification and Retrieval
(a GMM) to the target training data and computing the average log-likelihood of the ... In this paper we aim to (a) improve the time and storage efficiency of the ...

Star Petroleum Refining
Feb 14, 2018 - Room 1907-1909 Tow er A , N o. 100 Zunyi Road,. Shanghai, PRC 200051. Shenzhen. Room 24D1, 24/F, A Unit, Zhen Ye Building,2014 ...

Efficient Speaker Identification and Retrieval - Semantic Scholar
identification framework and for efficient speaker retrieval. In ..... Phase two: rescoring using GMM-simulation (top-1). 0.05. 0.1. 0.2. 0.5. 1. 2. 5. 10. 20. 40. 2. 5. 10.

Automatic Speech and Speaker Recognition ... - Semantic Scholar
7 Large Margin Training of Continuous Density Hidden Markov Models ..... Dept. of Computer and Information Science, ... University of California at San Diego.

speaker identification and verification using eigenvoices
approach, in which client and test speaker models are confined to a low-dimensional linear ... 100 client speakers for a high-security application, 60 seconds or more of ..... the development of more robust eigenspace training techniques. 5.

TOWARDS SPEAKER AND ENVIRONMENTAL ...
for improving the performance of speech recognition system under adverse ... on HEQ by imposing constraints on the type of histogram transfor- mation that can be ..... The same phone (spoken language) may be mis-pronounced in a dif-.

Read PDF Petroleum Refining: Technology and ...
Book sinopsis. Addresses changes in refinery feedstocks, product distribution, and processing requirements resulting from federal and state legislation. Offering ...

Measuring and Predicting Software Productivity
Jun 14, 2010 - The environments (all management information systems) can be characterized as follows: • Environment 1: ..... definition of a function point. Hence, when measuring function point productivity one ... from the application management s