INTERSPEECH 2010

Signal Interaction and the Devil Function John R. Hershey, Peder A. Olsen, Steven J. Rennie IBM T. J. Watson Research Center {jrhershe,pederao,sjrennie}@us.ibm.com

Abstract

is equivalent to adding log gain to the mixture. There is also a symmetry property due to the cosine: v(x, n, θ) = v(x, n, 2π−θ). We can use this relation to define an interaction function,

It is common in signal processing to model signals in the log power spectrum domain. In this domain, when multiple signals are present, they combine in a nonlinear way. If the phases of the signals are independent, then we can analyze the interaction in terms of a probability density we call the “devil function,” after its treacherous form. This paper derives an analytical expression for the devil function, and discusses its properties with respect to model-based signal enhancement. Exact inference in this problem requires integrals involving the devil function that are intractable. Previous methods have used approximations to derive closed-form solutions. However it is unknown how these approximations differ from the true interaction function in terms of performance. We propose Monte-Carlo methods for approximating the required integrals. Tests are conducted on a speech separation and recognition problem to compare these methods with past approximations.

def

p(y|x, n, θ) = δ (y − v(x, n, θ)) ,

(3)

where δ(.) is the Dirac delta function. The surface defined by v(x, n, θ) = y, for fixed y, we call the unicorn function after its single-pronged shape (see Figure 1). To perform inference

1. Introduction Signals are often analyzed and classified using models of their log power spectra. In the context of noise, or any other interfering signal, such model-based classifiers must compensate for the noise in some way. Model-based noise compensation requires a signal interaction model, which describes the effect of adding two signals on the resulting acoustic features. Traditionally the influence of phase has either been ignored through the use of approximate interaction models, or has been diminished by averaging, especially when working in the log spectrum domain. We describe and illustrate the signal interaction model, which we call the “devil function.” Exact inference using this function is difficult because the required integrals are intractable. Even efficient approximate inference can be elusive. Just how important it is to accurately model this signal interaction is an empirical question. To address this question we perform experiments using accelerated Monte Carlo simulations that can approximate inference arbitrarily well if enough samples are used, and compare the results to simpler approximations on a speech separation and recognition task.

Figure 1: The unicorn function, defined by v(x, n, θ) = y, for y = 0. Axes (counter clockwise from top left) are n, x and θ.

in this model, given priors on x and n, we have to compute Z p(y) = p(x)p(n)p(θ)δ (y − v(x, n, θ)) dθ dx dn Z = p(x)p(n)p(y|x, n)dx dn , (4) R where p(y|x, n) = p(θ)δ (y − v(x, n, θ)) dθ . It is also useful to compute the posterior expected value. Z p(x)p(n)p(y|x, n) E(x|y) = x dx dn (5) p(y)

2. Interaction models

An empirical histogram of p(y|x, n) is shown in Figure 2, for mixed speech signals, along with the histogram, p(x, n), of the original signals prior to mixing. Define the inverse function of v(x, n, θ) that yields θ as a function of y in the range [ymin = v(x, n, θ = π), ymax = v(x, n, θ = 0)], for fixed x and n:  y  e − ex − en def Θ(y|x, n) = arccos ∈ [0, π]. (6) x+n 2e 2

The relationship between log power spectra x, n, with phases θx , θn , and the log power of their mixture y is x

n

y = log(|e 2 +iθx + e 2 +iθn |2 ) x

n

= log(e + e + 2e

x+n 2

cos(θ)),

(1) (2)

def

where θ = |θx − θn | is the phase difference. def For convenience we define the function v(x, n, θ) = log(ex + x+n n e + 2e 2 cos(θ)). Note that v obeys a shift property, v(x + s, n + s, θ) = v(x, n, θ) + s, so adding log gain to both sources

Copyright © 2010 ISCA

Θ(y|x, n) is a monotonic function ranging in value from π at ymin down to 0 at ymax . Since θ is uniformly distributed, and considering the symmetry of v over the ranges θ ∈ [0, π], and

334

26- 30 September 2010, Makuhari, Chiba, Japan

log of the expected value in the power domain E(ey ) = ex +en . def

p(y|x, n) ≈ N (y | log(ex + en ), ψ) = plogsum (y|x, n), (11) where ψ is a variance designed to compensate for the effects x of phase. The log-sum approximation can   be written log(e +

en ) = max(x, n) + log 1 + e−|x−n| , which motivates the max approximation: def

p(y|x, n) ≈ δ (y − max(x, n)) = pmax (y|x, n),

where δ(.) is the Dirac delta function. It was shown in [5] that if the phase difference is uniformly distributed, then the expected value of the log power of the sum of the signals is Eθ (y|x, n) = max(x, n). The max model was first used in [6] for noise adaptation, and in [7], to compute state likelihoods. Other approximate interaction functions have been proposed to handle the phase term for log averages over power spectrum bins, such as [8].

Figure 2: The empirical histogram P (y|x, n), obtained from 120 mixtures of utterances from two speakers. Also shown are the max and log-sum approximations θ ∈ [π, 2π], the cumulative distribution function (cdf) of y ∈ [ymin , ymax ] is P (y ≤ y|x, n) =2P (θ ≤ Θ(y, x, n))

4. Naive Monte Carlo

(7)

Θ(y, x, n) , (8) =1 − π where the factor of 2 accounts for the symmetry in θ. The probability density function (pdf) p(y|x, n) can be found by taking the derivative with respect to y ∈ [ymin , ymax ]. ∂ P (y ≤ y|x, n) ∂y ey−c , = q π 1 − 14 (ey−c − ez − e−z )2

p(y|x, n) =

(12)

We can instead estimate p(y) directly using Monte Carlo methods. It is tempting to directly use the devil function, but the fact that it goes to infinity in some places and is zero over much of the space is a potential pitfall. Instead, here we use the unicorn function, replacing the delta function with a Gaussian with variance ψ. Taking samples {xi , ni , θi } from a proposal distribution h(x, n, θ), the standard importance sampling approximation [9] is 1 X p(xi )p(ni )p(θi )N (y|v(xi , ni , θi ), ψ) , (13) p(y) ≈ N i h(xi , ni , θi )

(9) (10)

where c = (x + n)/2 and z = (x − n)/2. Note that the allowable region for y, x, and n is such that y = v(x, n, θ) for some θ. Outside the allowable region, p(y|x, n) = 0. Along the edges of this region, defined by θ ∈ {0, π, 2π}, p(y|x, n) = ∞. These singularities make the distribution somewhat tricky to work with. Because of these hazards, we call p(y|x, n) the devil function (see Figure 3). Like the unicorn function, the devil function obeys a shift property, p(y−s|x, n) = p(y|x + s, n + s). A similar distribution can be derived in the amplitude domain, and is known in the wireless communications literature as the two-wave envelope pdf [1].

and similarly for the expected value. When sampling from the prior distribution h(x, n, θ) = p(x)p(n)p(θ) we can cancel out P these priors, so that p(y) ≈ N1 i N (y|v(xi , ni , θi ), ψ) This sampling scheme can suffer from the fact that the samples may be far from the true posterior, and far from the set of allowable points, x, n given the observation. In that case, the estimated posterior mean E(x|y) may be inaccurate.

5. Sampling the devil function Let us consider the devil function more carefully. We have Z p(y) = px (x)pn (n)p(θ)δ (y − v(x, n, θ)) dx dn dθ (14) We introduce the change of variables used in (10), via a transformation matrix,        x−n  1 1 −1 z x 2 , A= (15) =A = x+n c n 2 1 1 2 and a new function, u(z, θ) = log ez + e−z + 2 cos θ def



= log (2 cosh(z) + 2 cos θ) , Figure 3: The devil function, p(y|x, n). Horizontal axes are x and n with density on the vertical axis.

(16) (17)

so that v(x, n, θ) = c + u(z, θ). Then Z px (c + z)pn (c − z)δ (y − c − u(z, θ)) 1 p(y) = dc dz dθ 2π | det A| Z 1 = px (y − u(z, θ) + z) pn (y − u(z, θ) − z) dz dθ π Z 1 = px (˜ x) pn (˜ n) dz dθ , (18) π

3. Approximate interaction functions The devil function is traditionally avoided via various approximations. The log-sum approximation, used in [2, 3, 4], uses the

335

compute ni = Φ−1 n (pi ), zi = (y − ni )/2. The sampling distribution in z is then  πx 2Lx≤y (xi ) : n = y h(z) = (23) πn 2Ln≤y (ni ) : x = y.

(a) Prior density

Note that instead of πx and πn derived from the posteriors, which depend on the prior models for x and n we can choose a uniform sampling distribution over the two segments, in order to be independent of the models. This allows the samples to be shared across all combinations of x and n distributions in a state-dependent model.

(b) Posterior density

Figure 4: Max model: a) the prior normal density p(x, n), with max interaction shown in red. b) the posterior p(x, n|y) .

6.2. Algonquin importance sampling def

def

where x ˜ = y − u(z, θ) + z and n ˜ = y − u(z, θ) − z. Note that u(z, θ) ≈ |z| when |z| is large, and c + |z| = max(x, n); that is, the max approximation is accurate for large |z|. We can then approximate p(y) by sampling θi ∼ U (0, 2π) and zi ∼ h(z), and computing x ˜i = y − u(zi , θi ) + zi and n ˜i = y − u(zi , θi ) − zi .

In Algonquin, [10, 11] the log-sum interaction function is used. To handle the intractable likelihood and expected value integrals, the log-sum function is linearized, which yields a Gaussian approximation to the posterior. This posterior in turn gives a better linearization point and the process is iterated. The Gaussian posterior,          x x ηx φxx φxn p y ≈N , , (24) n n ηn φxn φnn

2 X px (y − u(zi , θi ) + zi ) pn (y − u(zi , θi ) − zi ) N i h(zi ) X px (˜ xi ) pn (˜ ni ) 2 = (19) N i h(zi )

p(y) ≈

can be used to obtain a sampling distribution in z: h(z) = N (z|(ηx − ηn )/2, (φxx + φnn − 2φxn )/4)

(25)

The x ˜i and n ˜ i we have thus derived can be seen as samples from the devil function multiplied by h(z).

6. Refined importance sampling The question remains of what proposal distribution h(z) to use. We could again start with the priors, this time projecting points such that they satisfy the constraints imposed by y. However, such samples still may not fall predominantly in the region of the posterior, which is necessary for a good approximation. An easy option is to sample from a wide range or z values, but this may be expensive in the number of samples. A better choice is to sample z values from an approximation to the posterior given the observation and the prior, such as provided by inference in either the log-sum or the max model. We call this approach ”refined importance sampling,” because it makes an educated guess about the region of interest.

Figure 5: Integration using control variates: a) (solid) the function to be integrated, b) (dashed) the control variate approximation, c) (blue) the known integral of the approximation, and d) (green) the remaining area to be estimated. When the two signals x and n are conditioned on state variables in a larger model, we typically have to estimate the likelihood under all combinations of states. This presents one drawback of Algonquin-based importance sampling: it requires independent samples under all combinations of states.

6.1. Max model importance sampling

7. Control variates

In the max model, the posterior can be computed analytically in a single step, which makes it an attractive choice. Here we use the posterior as a sampling distribution. The max model likelihood function is piece-wise linear and thus admits a closed form solution for the posterior, p(x, n|y). The likelihood is pmax (y) =px (y)Φn (y) + pn (y)Φx (y),

(20)

def

using px (y) = p(x = y) for random R yvariable x, and the cumulative distribution function Φx (y) = −∞ p(xt )dx. The posterior distribution is p(x)p(n)δ (y − max(x, n)) p(y) = πx Lx≤y (x)δ(y − n) + πn Ln≤y (n)δ(y − x)

pmax (x, n|y) ≈

(21) (22)

with πx = pn (y)Φx (y)/p(y), πn = 1 − πx , and Lx≤y (x) = px (x)1x≤y /Φx (y) is p(x) truncated to [−∞, y] and re-normalized. To sample we first take ui ∼ U (0, 1). If ui < πx , we sample pi ∼ U (0, Φx (x)) and compute xi = Φ−1 x (pi ), zi = (xi − y)/2. Otherwise, we sample pi ∼ U (0, Φn (n)), and

336

The method of control variates [9] gives us another means to improve the sampling efficiency. In this method we use an integrable approximation to the function of interest, and use sampling to compute the difference between the exact function and its approximation, as shown in Figure 5. This results in a faster convergence to the true integral. Let us consider, for example, using the max model as a control variate for sampling from the devil function. The max model likelihood integral, given in (20) can be transformed according to (15), noting that max(x, n) = c + |z|: Z pmax (y) = px (x)pn (n)δ (y − max(x, n)) dx dn (26) Z px (c + z)pn (c − z)δ (y − c − |z|) dc dz = | det A| Z = 2 px (y − |z| + z) pn (y − |z| − z) dz = px (y)Φn (y) + pn (y)Φx (y). Emax (x|y) = πx y + πn Emax (x|x < y).

(27) (28)

Algorithm Naive Monte Carlo Devil Function MC Max Control Variate

10 30.8 33.3 24.6

Number of Samples 20 40 80 28.8 27.6 27.1 28.2 26.2 25.2 24.3 24.0 22.7

of 28.8% in [13]. The speech recognition task is a relatively complicated test of these methods, and may be dominated by other factors such as the constraints of the task grammar. Further work is underway to directly measure the convergence of the likelihood and expected value estimates in isolation to further validate the proposed Monte Carlo methods. It should be noted that the Monte Carlo methods are quite slow in comparison with the max model and log-sum methods: the devil function is so far mainly of theoretical value. It is also still unknown whether the integrals were approximated well enough to give a clear comparison with the max model and log-sum methods. However, early indications are that there is room for improvement to these models.

160 25.7 23.4 22.8

Table 1: Percent word error rate (WER) on a speech separation task using different Monte Carlo methods to estimate likelihoods and expected values, as a function of the number of samples used per Gaussian combination. Baseline for the max model is 23.9% and for Algonquin is 24.9%. For Gaussian px (x), with mean µx , and variance σx2 , it can be shown that the expected value, Emax (x|x < y), is µx − σx2 px (y)/Φx (y) [12]. Defining x0i = y − |zi | + zi , and n0i = y − |zi | − zi , we can thus use the max model as a control variate for the devil function:

9. References [1] G. Durgin, T. Rappaport, and D. De Wolf, “New analytical models and probability density functions for fading in wireless communications,” IEEE Transactions on Communications, vol. 50, no. 6, pp. 1005–1015, 2002.

def

p(y) ≈ pcvmax (y) = α (px (y)Φn (y) + pn (y)Φx (y))   . 2 X h(zi ), px (˜ xi ) pn (˜ ni ) − αpx x0i pn n0i + N i

[2] S. Boll, “Suppression of acousic noise in speech using spectral subtraction,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 27, pp. 114–120, 1979.

where α is a free parameter that can be chosen to reduce the variance of the estimator, or to keep the estimate positive . The latter term corrects for inaccuracies in the max model by measuring the difference in likelihood between samples from the devil function, x ˜i , n ˜ i , and the projections x0i , n0i of those samples onto the max function. The expected value is given by

[3] P. Moreno, B. Raj, and R. Stern, “A vector Taylor series approach for environment-independent speech recognition,” in ICASSP, 1996. [4] B. Frey, T. Kristjansson, L. Deng, and A. Acero, “Algonquin - learning dynamic noise models from noisy speech for robust speech recognition,” NIPS, pp. 1165–1171, 2001.

def

E(x|y) ≈ Ecvmax (x|y) = αEmax (x|y) (29)   0 0 0 X x ˜i px (˜ xi ) pn (˜ ni ) − αxi px (xi ) pn (ni ) 2 . + N i h(zi )pcvmax (y)

[5] M. Radfar, R. Dansereau, and A. Sayadiyan, “Nonlinear minimum mean square error estimator for mixture-maximisation approximation,” Electronics Letters, vol. 42, no. 12, pp. 724–725, 2006.

Similarly Algonquin and many other methods may be used as control variates.

[6] A. N´adas, D. Nahamoo, and M. A. Picheny, “Speech recognition using noise-adaptive prototypes,” IEEE Trans. Acoust., Speech, Signal Processing, vol. 37, pp. 1495–1503, 1989.

8. Experiments Experiments were conducted on a speech separation and recognition task [13] using the model-based approach described in [14]. The model consists of a hidden Markov model for each speaker with Gaussian observation densities in the log spectrum domain. The front-end of the system in [14] was modified to eliminate spectral averaging that would alter the interaction function. Inference involves computing the likelihoods (4) and expected values (5). The likelihood values are computed for each of 319 frequency bins, for every combination of 256 acoustic Gaussians for the pair of speakers, x and n. Band quantization is used to speed up this computation as described in [14]. The 2-D Viterbi algorithm is used to estimate state sequences through the task grammar. The time-domain signal is then reconstructed using the expected values of the hidden signals x and n, conditioned on the state sequences, and the result is fed to a conventional recognizer. Results for different Monte Carlo algorithms and sample numbers are presented in Table 1 for the 0 dB signal-to-noise ratio (SNR) condition, which contains 600 two-speaker mixtures of utterances from a six-word grammar. The naive Monte Carlo algorithm, which sampled from the prior Gaussians fared the worst, whereas the devil-function sampling, using a Gaussian importance-sampling distribution derived from the mean and variance speech model, worked better. The control variate method, using the max model and the same Gaussian importance sampling distribution, worked best, and the 22.7% WER compares favorably with the Algonquin baseline of 24.9% and the max model baseline of 23.9%. Although the improvements relative to baseline are not dramatic, all of the methods outperform the human listener scores

337

[7] A. Varga and R. Moore, “Hidden Markov model decomposition of speech and noise,” ICASSP, pp. 845–848, 1990. [8] L. Deng, J. Droppo, and A. Acero, “Enhancement of log Mel power spectra of speech using a phase-sensitive model of the acoustic environment and sequential estimation of the corrupting noise,” IEEE Transactions on Speech and Audio Processing, vol. 12, no. 2, pp. 133–143, 2004. [9] R. Rubinstein, Simulation and the Monte Carlo Method. 1981.

Wiley,

[10] B. Frey, L. Deng, A. Acero, and T. Kristjansson, “Algonquin: Iterating laplace’s method to remove multiple types of acoustic distortion for robust speech recognition,” in Eurospeech, September 2001. [11] T. Kristjansson, J. Hershey, and H. Attias, “Single microphone source separation using high resolution signal reconstruction,” ICASSP, 2004. [12] J. K. Patel and C. B. Read, Handbook of the Normal Distribution. New York: Marcle Dekker, 1982. [13] M. Cooke, J. Hershey, and S. Rennie, “Monaural speech separation and recognition challenge,” Computer Speech & Language, vol. 24, no. 1, pp. 1–15, 2010. [14] J. Hershey, S. Rennie, P. Olsen, and T. Kristjansson, “Superhuman multi-talker speech recognition: A graphical modeling approach,” Computer Speech & Language, vol. 24, no. 1, pp. 45–66, 2010.

Signal Interaction and the Devil Function

(b) Posterior density. Figure 4: Max ... max interaction shown in red. b) the posterior p(x, n|y) . ..... [8] L. Deng, J. Droppo, and A. Acero, “Enhancement of log Mel.

2MB Sizes 5 Downloads 201 Views

Recommend Documents

Signal Interaction and the Devil Function
in terms of a probability density we call the “devil function,” after its treacherous form. This paper derives an analytical ex- pression for the devil function, and ...

MHS-5200A-Function-Signal-Generator.pdf
Page 2 of 25. 1. Unpacking. When you get a new MHS-5200A Series dual-channel DDS signal generator, it. is recommended that you follow these steps to ...

Multimodal Signal Processing and Interaction for a Driving ... - CiteSeerX
In this paper we focus on the software design of a multimodal driving simulator ..... take into account velocity characteristics of the blinks are re- ported to have ...

Multimodal Signal Processing and Interaction for a ...
attention and fatigue state is based on video data (e.g., facial ex- pression, head ... ment analysis – ICARE – Interaction modality – OpenInterface. – Software ..... elementary components are defined: Device components and Interaction ...

CHAPTER 17 Signal Sequence Function in the ...
had been previously analyzed for their capacity to support LamB export in viva. (Emr and Silhavy, 1982, 1983), Briggs and Gierasch (1984) demonstrated a clear.

Structure and function of mucosal immun function ...
beneath the epithelium directly in contact with M cells. Dendritic cells in ... genitourinary tract, and the breast during lactation (common mucosal immune system).

Executive Function and Medial Frontal Lobe Function ...
events, risk management, shifting focus, flexibility, inhibition ... word denotes--an incongruent trial (i)--the resulting conflict regarding which action plan to execute ...

The Devil and Tom Walker questions.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. The Devil and ...

Women Read the Romance: The Interaction of Text and Context
to have sold 168 million romances throughout the world in the single year of .... about the best romances to buy and those to avoid. When I ...... Lentricchia, After the New Criticism (Chicago: University of Chicago Press, 1980). 7. Two good ...

Genetic and Economic Interaction in the Formation of Health: The ...
Intro GxEcon Empirics Structural Conclusion. Research question Motivation. Outline. 1. Introduction. Research question. Motivation. 2. Gene x Econ. 3. Empirical Findings. ALSPAC Data. Results. Productivity Effect. Preferences. Robustness Checks. Repl

+( Free Download The Devil and Webster pdf free ...
+( Free Download The Devil and Webster pdf free download books ebooks ID: ... expertly plotted thriller moves along at a slow burn, building up to shocking.

Watch The Devil and Miss Jones (1941) Full Movie Online.pdf ...
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.

The Devil and Tom Walker newspaper.pdf
... visually attractive with captions for photos and drawings? • Did your group use their time wisely? Page 1 of 1. The Devil and Tom Walker newspaper.pdf.

pdf-49\tale-of-the-devil-the-biography-of-devil-anse-hatfield-by-dr ...
Page 1 of 9. TALE OF THE DEVIL: THE BIOGRAPHY OF. DEVIL ANSE HATFIELD BY DR. COLEMAN. C. HATFIELD AND ROBERT SPENCE. DOWNLOAD EBOOK : TALE OF THE DEVIL: THE BIOGRAPHY OF DEVIL. ANSE HATFIELD BY DR. COLEMAN C. HATFIELD AND ROBERT SPENCE PDF. Page 1 of

Women Read the Romance: The Interaction of Text and Context
We use information technology and tools to increase productivity and facilitate new forms ... publishing houses currently issue from two to six romantic novels.