Estimating Local Optimums in EM Algorithm over Gaussian Mixture Model Zhenjie Zhang

Bing Tian Dai Anthony K.H. Tung Department of Computer Science School of Computing National Univ. of Singapore {zhenjie, daibingt atung}@comp.nus.edu.sg May 1, 2008

Abstract EM algorithm is a very popular iteration-based method to estimate the parameters of Gaussian Mixture Model from a large observation set. However, in most cases, EM algorithm is not guaranteed to converge to the global optimum. Instead, it stops at some local optimums, which can be much worse than the global optimum. Therefore, it is usually required to run multiple procedures of EM algorithm with different initial configurations and return the best solution. To improve the efficiency of this scheme, we propose a new method which can estimate an upper bound on the logarithm likelihood of the local optimum, based on the current configuration after the latest EM iteration. This is accomplished by first deriving some region bounding the possible locations of local optimum, followed by some upper bound estimation on the maximum likelihood. With this estimation, we can terminate an EM algorithm procedure if the estimated local optimum is definitely worse than the best solution seen so far. Extensive experiments show that our method can effectively and efficiently accelerate conventional multiple restart EM algorithm.

1

Introduction

Gaussian Mixture Model (GMM) (McLachlan & Peel, 2000) is a powerful tool in unsupervised learning to model unlabelled data in a multi-dimensional space. However, given an observation data set, estimating the parameters of the underlying Gaussian Mixture Model of the data is not a trivial task, especially when the dimensionality or the number of components is large. Usually, this model estimation problem is transformed to a new problem, which try to find parameters maximizing the likelihood probability on the observations from the Gaussian distributions. In the past decades, EM algorithm (Dempster et al., 1977) has become the most widely method used in the problem of learning Gaussian Mixture Model (Ueda et al., 2000; Jordan & Xu, 1995; McLachlan & Krishnan, 1996). Although EM algorithm can converge in finite iterations, there is no guarantee on the convergence to global optimum. Instead, it usually stops at some local optimum, which can be arbitrarily worse than the global optimum. Although there have been extensive studies on how to avoid bad local optimums, it is still required to run EM algorithm with different random initial configurations and the best local optimum is returned as final result. This leads to a great waste of computation resource since most of the calculations do not have any contribution to the final result.

1

In this paper, we propose a fast stopping method to overcome the problem of trapping into bad local optimums. Given any current configuration after an EM iteration, our method can estimate an upper bound on the final likelihood of the local optimum current configuration is leading to. Therefore, if the estimated local optimum is definitely not better than the best local optimum achieved in previous runs, current procedure can be terminated immediately. To facilitate such local optimum estimation, we first prove that a region in the parameter space can definitely cover the unknown local optimum. If a region covers the current configuration and any configuration on the boundary of the region gives lower likelihood than the current one does, we can show that the local optimum is “trapped” in the region; and we call such region as a maximal region. In this paper, we adopt a special type of maximal region, which can be computed efficiently. Since the best likelihood of any configuration in a maximal region can be estimated in relatively short time, it can be decided immediately on whether current procedure still has potential to achieve a better local optimum. In our experiments, such method is shown to greatly improve the efficiency of original EM algorithm for GMM, on both synthetic and real data sets. The rest of the paper is organized as follows. We first introduce the definitions and related works on Gaussian Mixture Model and EM algorithm in Section 2 . Section 3 proves the local trapping property of EM algorithm on GMM; and Section 4 presents our study on maximal region of local optimum. We propose our algorithm on estimating the likelihood of a local optimum in Section 5 . Section 6 shows some experimental result. Finally, section 7 concludes this paper.

2

Model and Related Works

In this section, we review the basic models of Gaussian Mixture Model, EM algorithm, and some acceleration method proposed for a special type of Gaussian Mixture Model (K-Means Algorithm).

2.1

Gaussian Mixture Model

In GMM model (McLachlan & Peel, 2000), there exist k underlying components {ω1 , ω2 , . . . , ωk } in a d-dimensional data set. Each component follows some Gaussian distribution in the space. The parameters of the component ωj include Θj = {µj , Σj , πj }, in which µj = (µj [1], . . . , µj [d]) is the center of the Gaussian distribution, Σj is the covariance matrix of the distribution and πj is the probability of the component ωj . Based on the parameters, the probability of a point coming from component ωj appearing at x = (x[1], . . . , x[d]) can be represented by Pr(x|Θj ) =

½ ¾ 1/2 |Σ−1 1 j | T −1 exp − (x − µ ) Σ (x − µ ) j j j 2 (2π)d/2

Thus, given the component parameter set Θ = {Θ1 , Θ2 , . . . , Θk } but without any component information on an observation point x, the probability of observing x is estimated by Pr(x|Θ) =

k X

Pr(x|Θj )πj

j=1

The problem of learning GMM is estimating the parameter set Θ of the k component to maximize the likelihood of a set of observations D = {x1 , x2 , . . . , xn }, which is represented by Pr(D|Θ) =

n Y

i=1

2

Pr(xi |Θ)

(1)

Based on the parameters of the GMM model, the posterior probability of xi from component ωj (or the weight of xi in component j), τij , can be calculated as follows. Pr(xi |Θj )πj τij = Pk l=1 Pr(xi |Θl )πl

(2)

To simplify the notations, we use Φ to denote the set of all τij for any pair of i, j, and use Ψ(Θ) to denote the corresponding Φ based on current configuration Θ. For ease of analysis, the original optimization problem on equation (1), is usually transformed to an equal maximization problem on the following variable, called log likelihood.

L(Θ, Φ) =

n X k X

τij

i=1 j=1

Ã

ln |Σ−1 (xi − µj )T Σ−1 πj j | j (xi − µj ) ln + − τij 2 2

!

(3)

L is actually a function over Θ and Φ, the latter of which is usually optimized according to Θ. Thus, the problem of learning GMM is finding an optimal parameter set Θ∗ which can maximize the function L(Θ∗ , Ψ(Θ∗ )).

2.2

EM Algorithm

EM algorithm (Dempster et al., 1977) is a widely used technique for probabilistic parameter estimation. To estimate Θ = {Θ1 , . . . , Θk }, it starts with a randomly chosen initial parameter configuration Θ0 . Then, it keeps invoking iterations to recompute Θt+1 based on Θt . Every iteration consists of two steps, E-step and M-step. In E-step, the algorithm computes the expected value of τij for each pair of i and j based on Θt = {Θt1 , . . . , Θtk } and equation (2). In M-step, the algorithm finds a new group of parameters Θt+1 to maximize L based on Φt = {τijt } and {x1 , x2 , . . . , xn }. The details of the update process for µj , Σj and πj are listed below. µt+1 j Σt+1 j

=

Pn

Pn t τij xi = Pi=1 n t i=1 τij

− µt+1 )T (xi − µt+1 j ) Pn j t i=1 τij Pn t τij = i=1 n

t i=1 τij (xi

πjt+1

(4)

(5) (6)

The iteration process stops only when ΘN = ΘN −1 after N iterations. We note that both E-step and M-step always improve the objective function, i.e. L(Θt , Φt ) ≥ L(Θt , Φt−1 ) ≥ L(Θt−1 , Φt−1 ). Based on this property, EM-algorithm will definitely converge to some local optimum.

2.3

K-Means Algorithm and Its Acceleration

K-Means algorithm can be considered as a special problem of GMM learning with several constraints (Duda et al., 2001). First, the covariance matrix for each component must be identity matrix. Second, the posterior probability τij can only be 0 or 1. Therefore, in E-step of the algorithm, each point is assigned to the closest center under Euclidean distance; whereas in M-step, the set of geometric center of each cluster is used to replace the old set. With the problem simplification from GMM to K-Means, there have been many methods proposed to accelerate the multiple restart EM algorithm for K-Means. In (Kanungo et al., 2002), for

3

Table 1: Table of Notations Notation n d k ωj Θj µj Σj πj Θ xi τij Φ Ψ(Θ) S L(Θ, Φ) ∆

Description number of points in data dimensionality of data space number of components component j parameter set of ωj center of ωj covariance matrix of ωj probability of ωj configuration of all components ith point in the data posterior probability Pr(ωj |xi ) the set of all τij the optimal Φ with Θ solutions space for configurations objective log likelihood function a parameter for a maximal region

example, Kanungo et al. applied indexing technique to achieve a much more efficient implementation of E-step. In (Elkan, 2003), Elkan accelerated both E-step and M-step by employing triangle inequality of Euclidean distance to reduce the time for distance computations. In (Zhang et al., 2006), Zhang et al. introduced a lower bound estimation on the k-means local optimums to efficiently cut the procedures not leading to good solutions. However, all these methods proposed for k-means algorithm cannot be directly extended to the general GMM. As far as we know, our paper is the first study on acceleration of the multiple restart EM algorithm with robustness guarantee. To improve the readability of the paper, we summarize all notations in Table 1.

3

Local Trapping Property

In this section, we prove the local trapping property of EM algorithm on GMM. To derive the analysis, we first define a solution space S, containing (d2 + d + 1)k dimensions where d is the dimensionality of the original data space. Any configuration Θ, either valid or invalid, can be represented by a point in S. Without loss of generality, we use Θ to denote the configuration as well as the corresponding point in solution space S. The rest of the section will be spent to prove the following theorem. Theorem 1 Given a closed region R in the solution space S covering current configuration Θt , EM algorithm converges to a local optimum in R if every configuration Θ on the boundary of R has L(Θ, Ψ(Θ)) < L(Θt , Φt ) Given two configurations Θt and Θt+1 across one EM iteration, we define a path between Θ and Θt+1 in S as follows. This path consists of two parts, called P 1 and P 2 respectively. # # # t t P 1 starts at Θt and ends at Θ# , where Θ# = {Θ# j }. Here Θj = {µj , Σj , πj }, and Σj = P t P t t # t t T i τij (xi − µj )(xi − µj ) / i τij . An intermediate configuration between Θ and Θ is defined α α α t as Θ , in which µtj and πjt remain the same, while Σj in Θ is ((1 − α)(Σ )−1 + α(Σ# )−1 )−1 . t

4

When α increases from 0 to 1, we can move from Θt to Θ# in the solutions space S. The second part of the path starts at Θ# and ends at Θt+1 . Any intermediate configuration Θβ = {Θβj }, P t β β β T P t β t 1−β (π t+1 )β . where µβj = (1 − β)µtj + βµt+1 i τij (xi − µj )(xi − µj ) / i τij , and πj = (πj ) j , Σj = j Similarly, a continuous movement from Θ# to Θt+1 can be made by increasing β from 0 to 1. The following lemmas prove that any intermediate configuration on the path is a better solution than Θt . Lemma 1 Given any intermediate configuration Θα between Θt and Θ# , we have L(Θα , Ψ(Θα )) ≥ L(Θt , Φt ). Proof: By the optimality property of Ψ(Θα ), we have L(Θα , Ψ(Θα )) ≥ L(Θα , Φt ). P t P t t t T t t t Since Σ# i τij (xi − µj )(xi − µj ) / i τij is the optimal choice for Σj if τij , µj and πj are j = fixed, we also have L(Θ# , Φt ) ≥ L(Θt , Φt ). By the definition of Θα and the property of Θ# above, the following equations can be easily derived. L(Θα , Φt ) = (1 − α)L(Θt , Φt ) + αL(Θ# , Φt ) ≥ L(Θt , Φt ) Therefore, it is straightforward to reach the conclusion that L(Θα , Ψ(Θα )) ≥ L(Θt , Φt ).

2

Lemma 2 Given any intermediate configuration Θβ between Θ# and Θt+1 , we have L(Θβ , Ψ(Θα )) ≥ L(Θt , Φt ). Proof: Again, the basic inequality L(Θβ , Ψ(Θβ )) ≥ L(Θβ , Θt ) holds. Based on this, we can prove the lemma by showing L(Θβ , Φt ) ≥ L(Θ# , Φt ), since L(Θ# , Φt ) ≥ L(Θt , Φt ). P P t P t β β T P t If Σβj = j i τij (xi − i τij , a very interesting result is that i τij (xi − µj )(xi − µj ) / µβj )T (Σβj )−1 (xi − µβj ) remains constant for any β, as is proved below. In the following proof, tr(A) denotes the trace of a matrix A. XX j

=

i

XX j

i

=

XX

=

X

j

i

tr

=

³ ´ τijt tr (xi − µβj )(xi − µβj )T (Σβj )−1

ÃÃ X

tr(I ·

τijt (xi − µβj )(xi − µβj )T

XX j

X

!

(Σβj )−1

!

τijt )

i

j

=

³ ´ τijt tr (xi − µβj )T (Σβj )−1 (xi − µβj )

i

j

X

τijt (xi − µβj )T (Σβj )−1 (xi − µβj )

τijt d

i

= n·d

(7) 5

Therefore, for any Θβ , we only need to consider the sum

i

By the definition of πjβ , since πjβ = (πjt )1−β (πjt+1 )β , we have Then, n X

τijt ln

i=1

Therefore,

t j τij

P P i

ln

πjβ t τij

πjβ τijt



= (1 − β)

n X

τijt ln

i=1

t j τij

P P i

ln

πjt t τij

=

n X

´

n X πjt+1 πjt t τ ln + β ij τijt τijt

(8)

i=1

, since Σβj ,

On the other hand, based on the definition of

Σβj

³

β β t t j τij ln(πj /τij ) − ln(|Σj |)/2 . ln πjβ = (1 − β) ln πjt + β ln πjt+1 .

P P

Pn

t i=1 τij

we have

ln

πjt+1 t τij



Pn

t i=1 τij

πt

ln τ tj . ij

t τi,j (xi − µβj )(xi − µβj )T

i=1

=

n X

t τi,j (xi − µtj − β(µt+1 − µtj ))(xi − µtj − β(µt+1 − µtj ))T j j

i=1

=

n X

t τi,j (xi



µtj )(xi



µtj )T

i=1

−β =

t τi,j (xi − µtj ))(µt+1 − µtj )T j

i=1

n X i=1

n X

− β(

n X

n X t+1 t t t τi,j (µj − µj )( (xi − µtj ))T + β 2 (µt+1 − µtj )(µt+1 − µtj )T τi,j j j

t τi,j (xi

i=1



i=1

µtj )(xi



µtj )T

n X t − µtj )T + (β − 2β)( − µtj )(µt+1 τi,j )(µt+1 j j 2

i=1

β Since β 2 −2β ≤ 0 for any β between 0 and 1, ln |Σβj | ≤ ln |Σ# j |. And thus, we have − ln |Σj |/2 ≥

− ln |Σ# j |/2. Combing the results above, we reach the conclusion that L(Θβ , Φt ) ≥ L(Θ# , Φt ), leading to the proof of the lemma. 2 Proof for Theorem 1 Proof: We prove the theorem by contradiction. If R satisfies the boundary condition but EM algorithm converges to some configuration out of R in S, there is at least one pair of configurations {Θs , Θs+1 } that Θs is in R but Θs+1 is not. By setting up the path {Θα } ∩ {Θβ } between Θs and Θs+1 as defined above, we know there is at least one Θα (Θβ ) that Θα (Θβ ) is exactly on the boundary of R. By Lemma 1 (Lemma 2), we know L(Θα , Ψ(Θα )) ≥ L(Θs , Φs ) (L(Θβ , Ψ(Θβ )) ≥ L(Θs , Φs )). On the other hand, any Θα or Θβ is better than Θt by the definition of R. This leads to the contradiction, since L(Θs , Φs ) > L(Θt , Φt ). 2

4

Maximal Region

Based on Theorem 1, we define the concept of Maximal Region in GMM as follows. Given the current configuration Θt , a region R in S is the maximal region for Θt , if (1) R covers Θt , and (2) any boundary configuration Θ of R has L(Θ, Ψ(Θ)) < L(Θt , Φt ), by Theorem 1, EM algorithm converges to some local optimum in R. Given the current configuration Θt , there are infinite number of valid maximal regions in the solution space, most of which are hard to verify and manipulate. To facilitate efficient computation, 6

we further propose a special class of maximal regions. Given Θt and a positive real value ∆ < 1, we define a closed region R(Θt , ∆) ⊆ S as the union of any configuration Θ, each θj = {µj , Σj , πj } of t which satisfies all of the conditions below: (1) (1 − ∆)πjt ≤ πj ≤ (1 + ∆)πjt ; (2) −∆ ≤ tr(Σ−1 j (Σj ) − I) ≤ ∆; and (3) (µj − µtj )T (Σtj )−1 (µj − µtj ) ≤ ∆2 ; where tr(M ) denotes the trace of the matrix M and I is the identity matrix of dimension d. In the rest of the section, we will derive some sufficient condition for R(Θt , ∆) being a valid maximal region. Theorem 2 Any configuration Θ on the boundary of R(Θt , ∆) has L(Θ, Φt−1 ) ≤ L(Θt , Φt−1 ) − n minj πjt ∆2 /6. Proof: Given any R(Θt , ∆), any configuration on the boundary must satisfy one of the following t conditions for at least one j (1 ≤ j ≤ k): (1) (1 − ∆)πjt = πj ; (2) πjt = (1 + ∆)πjt ; (3) |tr(Σ−1 j (Σj ) − t I)| = ∆; and (4) (µj − µtj )T (Σj )−1 (µj − µtj ) = ∆2 . If Θ satisfies condition (1) for some component l, L(Θ, Φt−1 ) is maximized if µtj and Σtj remain unchanged for all j, while πj = bound.

1−(1−∆)πlt t πj 1−πlt

for all j 6= l. Therefore, we have the following upper

L(Θ, Φt−1 ) − L(Θt , Φt−1 )     n n t t X X X X π π π π j j t−1 t−1 τ t−1 ln l + τ t−1 ln l + ≤ τij ln t−1  − τij ln t−1  il il t−1 t−1 τ τ τ τ ij ij il il i=1 i=1 j6=l j6=l   n X X πj t−1 τ t−1 ln πl + τij ln t  = il t πl πj i=1 j6=l

1 − (1 − ∆)πlt 1 − πlt µ ¶ ∆πlt t t = nπl ln(1 − ∆) + n(1 − πl ) ln 1 + 1 − πlt µ ¶ ∆πlt ∆2 ≤ nπlt −∆ − + n(1 − πlt ) 2 1 − πlt

≤ nπlt ln(1 − ∆) + n(1 − πlt ) ln

= −

nπlt ∆2 2

The second inequality from the bottom is achieved by applying Taylor expansion on ln(1 − ∆). By iterating l with all k components, we have L(Θ, Φt−1 ) ≤ L(Θt , Φt−1 ) − minj nπlt ∆2 /2. If Θ satisfies condition (2) for some component l, L(Θ, Φt−1 ) can be maximized similarly. We have

7

L(Θ, Φt−1 ) − L(Θt , Φt−1 )   n X X πj t−1 τ t−1 ln πl + τij ln t  ≤ il t π π j l i=1 j6=l

1 − (1 + ∆)πlt 1 − πlt µ ¶ ∆πlt nπlt ln(1 + ∆) + n(1 − πlt ) ln 1 − 1 − πlt µ ¶ ∆2 ∆πlt ∆3 nπlt ∆ − + n(1 + πlt ) + 2 3 1 − πlt ¶ µ 2 ∆3 ∆ − −nπlt 2 3 t 2 nπ ∆ − l 6

≤ nπlt ln(1 + ∆) + n(1 − πlt ) ln = ≤ = ≤

Again, the third inequality from the bottom is due to Taylor expansion of ln(1 + ∆). The last inequality is because ∆3 ≤ ∆2 for any 0 ≤ ∆ ≤ 1. If Θ satisfies condition (3) for some component l, L is maximized if all other parameters remain the same. Thus, L(Θ, Φt−1 ) − L(Θt , Φt−1 ) n X ¢ τilt−1 ¡ ln |(Σj )−1 | − (xi − µtl )T (Σl )−1 (xi − µtl ) − ≤ 2 i=1

= = = = ≤ =

n X ¢ τilt−1 ¡ ln |(Σtj )−1 | − (xi − µtl )T (Σtl )−1 (xi − µtl ) 2 i=1 ¶ µ n X τilt−1 |Σ−1 t −1 −1 t t T l | ln − (xi − µl ) ((Σl ) − (Σl ) )(xi − µl ) 2 |(Σtl )−1 | i=1 µ µµ Pn ¶ ¶¶ t−1 t t T ¡ −1 ¢ nπlt −1 t t −1 i=1 τil (xi − µl )(xi − µl ) ln |Σl Σl | − tr Σ − (Σ ) l l 2 nπlt ¡¡ −1 ¢ ¢¢ nπlt ¡ t ln |Σ−1 Σl − (Σtl )−1 Σtl l Σl | − tr 2 ¡ ¢¢ nπlt ¡ ¡ ¡ −1 t ¢¢ t tr log Σl Σl − tr Σ−1 l Σl − I 2 µ ¶ t 2 tr(Σ−1 nπlt l Σl − I) − 2 2 t 2 nπ ∆ − l 4

The fourth equality is derived by the definitions of Σtl and πlt . And the second inequality from bottom is due to the taylor expansion on the logarithm matrix. P τil (xi −P µl )(xi −µl )T . Finally, if Θ satisfies condition (4) for some component l, L is maximized if Σl = τil P t−1 P −1 t−1 t In this case, i τl (xi − µj )T Σj (xi − µj ) = i τl (xi − µtj )T (Σj )−1 (xi − µtj ) = nπj d. Thus, the only difference on the log likelihood function L stems from the change on the determinant of the covariance matrix.

8

L(Θ, Φt−1 ) − L(Θt , Φt−1 ) n X ¢ τil ¡ t −1 ln |Σ−1 | ≤ l | − ln |(Σl ) 2 i=1 = = ≤ ≤

n X ¢ τil ¡ − ln |Σl | + ln |Σtl | 2 i=1

n X ¢ τil ¡ − ln |Σtl + (µl − µtl )(µl − µtl )T )| + ln |Σtl | 2 i=1

n X ¢ ¢ ¡ τil ¡ − ln |Σtl | + |(µl − µtl )(µl − µtl )T | + ln |Σtl | 2 i=1

n X ¡ ¢ ¢ τil ¡ − ln |Σtl | + ∆2 |Σtl | + ln |Σtl | 2 i=1

= − ≤



n X τil ln(1 + ∆2 )

i=1 nπlt ∆2

2

2

The fourth inequality applies the property of positive definition matrices that |A+B| > |A|+|B| (Lutkepohl, 1996). n minj π t ∆2

j In all of the four cases, the reduction on the likelihood function L is at least − . This 6 completes the proof of the theorem. 2 t 2 Last theorem implies that Θ will reduce the log likelihood function by at least n minj πj ∆ /6 if Φ remains Φt−1 . The following question is how much we can increase the likelihood if we use the optimal Ψ(Θ) instead of Φt−1 . t

Lemma 3 Given Θ ∈ R(Θt , ∆), P r(xi |θj )πj is no larger than (1+∆) where M (x, Θj ) is ³

ln(1 + ∆) − (1 − ∆) max

|(Σj )−1 |1/2 (2π)d/2

exp{M (x, Θj )}πjt ,

o´2 np ((x − µtj )T (Σtj )−1 (x − µtj )) − ∆, 0 /2

Proof: By the definition on the range of πj , we have

πj ≤ (1 + ∆)πjt

(9)

By the definition of Σj , the following inequalities derive the change between ln |Σj | and ln |Σtj | ln |Σj | − ln |Σtj | = tr ln Σj − tr ln Σtj = tr ln(Σj (Σtj )−1 ) ¡¡ ¢ ¢ = tr ln Σj (Σtj )−1 − I + I ≤ ln(1 + ∆)

9

(10)

At last, we analyze the change on (xi − µj )T (Σj )−1 (xi − µj ). First,

q xT (Σtj )−1 x can be

regarded as the Mahalanobis distance with respect to matrix Σtj . Since Mahalanobis distance is a metric distance, the triangle inequality implies that q (xi − µj )T (Σtj )−1 (xi − µj ) q q ≥ (xi − µtj )T (Σtj )−1 (xi − µtj ) − (µtj − µj )T (Σtj )−1 (µtj − µj ) q ≥ (xi − µtj )T (Σtj )−1 (xi − µtj ) − ∆ (11) o´2 ³ np ((x − µtj )T (Σtj )−1 (x − µtj )) − ∆, 0 . On Therefore, (xi − µj )T (Σtj )−1 (xi − µj ) ≥ max

the other hand, the increase by using Σj instead of Σtj can be bounded as follows.

= = = ≤

(xi − µj )T (Σtj )−1 (xi − µj ) − (xi − µj )T Σj (xi − µj ) ³ ´ (xi − µj )T (Σtj )−1 − Σ−1 (xi − µj ) j ³ ³ ´´ tr (xi − µj )(xi − µj )T (Σtj )−1 − Σ−1 j ³ ³ ´´ t tr (xi − µj )(xi − µj )T (Σtj )−1 I − Σ−1 Σ j j ³ ´ ¡ ¢ t tr (xi − µj )(xi − µj )T (Σtj )−1 tr I − Σ−1 Σ j j

= (xi − µj )T (Σtj )−1 (xi − µj )∆

(12)

Combining equation (11) and (12), we have o´2 ³ np ((x − µtj )T (Σtj )−1 (x − µtj )) − ∆, 0 −(xi − µj )T Σj (xi − µj ) ≤ −(1 − ∆) max

The summarization of the inequalities lead to the conclusion of the lemma.

2

t |(Σj )−1 |1/2 Lemma 4 Given Θ ∈ R(Θt , ∆), P r(xi |θj )πj is no smaller than (1−∆) (2π) exp{N (x, Θj )}πjt , d/2 where N (x, Θj ) is µ ³p ´2 ¶ t T t −1 t ((x − µj ) (Σj ) (x − µj )) + ∆ /2 ln(1 − ∆) − (1 + ∆)

.

Proof: The proving strategy of the lemma is similar to Lemma 3. First, πj ≥ (1 − ∆)πjt . Second, ln |Σj | ≥ ln |Σtj | + ln(1 − ∆). Third, −(xi − µj )T Σj (xi − µj ) ≥ ³p ´2 −(1 + ∆) ((x − µtj )T (Σtj )−1 (x − µtj )) + ∆ This completes the proof. 2 Lemma 5 Given a region R(Θt , ∆) as defined above, an upper bound, Uij , on τij ∈ Ψ(Θ) for any Θ ∈ R(Θt , ∆) can be calculated in constant time.

10

Proof: For any configuration Θ on the boundary of R(Θt , ∆), the optimal value of τij can be calculated by equation (2). By Lemma 3 and Lemma 4, we can compute maxΘ Pr(xi |ωj )πj and minΘ Pr(xi |ωj )πj . Therefore, max Pr(xi |ωj )πj τij ≤ Uij = P Θ l minΘ Pr(xi |ωl )πj

min Pr(xi |ωj )πj τij ≥ Lij = P Θ l maxΘ Pr(xi |ωl )πj

The calculations can be finished in constant time with the two sums pre-computed. 2 By Lemma 5, the increase upper bound from L(Θ, Φt−1 ) to L(Θ, Ψ(Θ)) can be calculated by the following equation. L(Θ, Ψ(Θ)) − L(Θ, Φt−1 ) XX ≤ ln Uij max Pr(xi |ωj )πj − Θ XX ln τij max Pr(xi |ωj )πj Θ

(13)

The following theorem gives a sufficient condition on a maximal region R(Θt , ∆) for some positive value ∆. Theorem 3 R(Θt , ∆) is a maximal region for Θt if ln L(Θt , Φt ) − L(Θt , Φt−1 )

P P

Pi Pj i

Uij max

j τij max

Θ Pr(xi |ωj )πj − n min π t ∆2 /6 < j Θ Pr(xi |ωj )πj

Proof: By the definition on the log likelihood function L, we have P P i j Uij Pr(xi |ωj )πj t t−1 L(Θ, Ψ(Θ)) − L(Θ , Φ ) ≤ ln P P i j τij Pr(xi |ωj )πj

It is not hard to verify that the derivative of L(Θ, Ψ(Θ)) − L(Θt , Φt−1 ) to Pr(xi |ωj )πj is always positive. Therefore, the equation above can be maximized if we employ the maximum value of Pr(xi |ωj )πj . Based on the analysis above, we know that P P i j Uij maxΘ Pr(xi |ωj )πj t t−1 L(Θ, Ψ(Θ)) − L(Θ , Φ ) ≤ ln P P i j τij maxΘ Pr(xi |ωj )πj

By Theorem 2, L(Θ, Φt−1 )−L(Θt , Φt−1 ) ≤ n minj πjt ∆2 /6. Therefore, by Theorem 1, L(Θ, Ψ(Θ)) < L(Θt , Φt ) if the condition of the theorem is satisfied. 2 ∗ t For any local optimum Θ in the maximal region R(Θ , ∆), the following theorem upper bound the likelihood function L(Θ∗ , Ψ(Θ∗ )). Theorem 4 Given a valid maximal region R(Θt , ∆), if EM algorithm converges to local optimum Θ∗ , L(Θ∗ , Ψ(Θ∗ )) ≤ L(Θt , Φt ) + n min πjt ∆2 /6. P P Proof: Since i j (Uij − τijt ) maxΘ Pr(ωj |xi ) < n min πjt ∆2 /6 by Theorem 3 and L(Θ, Ψ(Θ)) − P P L(Θ, Φt ) ≤ i j (Uij − τijt ) maxΘ Pr(ωj |xi ), we have L(Θ, Ψ(Θ)) − L(Θt , Φt ) ≤ L(Θ, Ψ(Θ)) − L(Θ, Φt ) ≤ n min πjt ∆2 /6. 2 11

Algorithm 1 New Iteration(Data Set D, Current Θt−1 , current Φt−1 , component number k, sample number m, current best result L∗ ) 1: Compute new Θt by M-Step. 2: Compute new Φt by E-Step. 3: if L(Θt , Φt ) < L∗ then ½ r ¾ t t 6(L∗ −L(Θ ,Φ )) 4: Let ∆ = min 1, n min πjt

5:

6: 7: 8: 9: 10: 11: 12: 13: 14: 15: 16: 17: 18: 19: 20: 21:

5

S=0 X=0 for each xi do for each dimension j do Get lij = maxΘ Pr(xi |θj )πj by Lemma 3. Get sij = minΘ Pr(xi |θj )πj by Lemma 4. Get Uij by Lemma 5. S+ = Uij ∗ lij X+ = τijt−1 ∗ lij end for end for if ln S − ln X − n min πj ∆2 /6 < L(Θt , Φt ) − L(Θt , Φt−1 ) then Stop the current procedure of EM algorithm. end if else Return (Θt , Φt ) end if

Algorithm

Theorem 3 provides an easy way to verify whether R(Θt , ∆) is a valid maximal region. On the other hand, Theorem 4 implies that a smaller ∆ can lead to tighter bound on the likelihood function L. However, it is not necessary to get the tightest bound on local optimum in our algorithm, since the goal of our algorithm is estimating whether the ¾ current configuration can lead to better solution. ½ r t t ∗ Instead, we set ∆ as min 1, 6(L −L(Θ ,Φ )) , where L∗ is the best result we have seen so far. n min πjt

This ∆ is the maximal one of all ∆ values, which are able to prune the current EM procedure by Theorem 4 The details of the algorithm are summarized in Algo 1. In this algorithm, conventional M-step and E-step are invoked first. If the current configuration is better than the best solution we have seen before, there is no need to test the upper bound of the local optimum. Otherwise, the value of ∆ is set according to min πjt , L∗ and L(ΘT , Φt ). For each point and each component, lij , sij and Uij are collected according to Lemma 3, Lemma 4 and Lemma 5 respectively. With the information collected from each point, the condition of Theorem 1 can be tested. If this condition is satisfied, we can assert that current local optimum can never be better than L∗ , leading to the termination of the current procedure.

12

4.5 4 3.5 3 2.5 2 1.5 1 10

15

20 25 30 Dimensionality

(a) CPU Time

35

40

65 60 55 50 45 40 35 30 25 20 15 10

1.2

AEM OEM

Ratio of AEM Time to OEM Time

AEM OEM

5

Average Number of Iteration

Average Time until Convergence

5.5

D=10 D=20 D=30 D=40

1.1 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2

10

15

20 25 30 Dimensionality

35

(b) Number of Iterations

40

0

10

20

30

40 50 60 70 Number of Runs

80

90 100

(c) CPU Time Ratio

Figure 1: Performance comparison with varying dimensionality

6

Experiments

In this section, we report the experimental results on the comparison of our accelerated EM algorithm (AEM) and the conventional EM algorithm (OEM). We note that in our implementation, either AEM or OEM will be stopped if it does not converge after 100 iterations. We employ both synthetic and real data sets in our empirical studies. The synthetic data sets are generated in a d-dimensional unit cube. There are k components in the space. Each component follows some Gaussian distribution. The center, size and covariance matrix of each component are randomly generated independently. Two real data sets are also tested, including Cloud and Spam, both of which are available on UCI Machine Learning Repository. The Cloud data set consists of 1024 points in 10-dimensional space, while Spam data set has 4601 points in 58 dimensions. Both of the real data set are normalized before being used in our experiments. Two performance measurements are recorded in our experiments, including CPU time and number of iterations. We also report the ratio of CPU time between original algorithm to the new accelerated algorithm with respect to the number of restart time from 1 to 100. An algorithm is supposed to be better if it spends less CPU time and invokes less time of iterations. All of the experiments are compiled and run on a Fedora Core 6 linux machine with 3.0 GHz Processor, 1GB of memory and GCC 4.1.2. In the experiments on the data sets, we test the performances of the algorithms with varying dimensionality D, number of components K, and the number of points in the data S. The default setting of our experiments is D = 20, K = 20, and S = 100K.

6.1

Results on Synthetic Data

In Figure 1(a) and Figure 1(b), we present the experimental result by varying the dimensionality from 10 to 40. The results show that AEM is much more efficient than OEM. On data set with low dimensionality, AEM is almost two times faster than OEM, both on the CPU time and the number of iterations. Even on high dimensional data, AEM is much faster than OEM. From Figure 1(c), we can see that our method can quickly set up the advantage over original EM algorithm even when there are very few restart time. The ratio is stable after 30 runs even on high dimensional space. The results of our experiments on varying component number are summarized in Figure 2(a) and Figure 2(b). From the figures, we can see the performance advantage of AEM is stable, with the increase of component number. The CPU time and number of iterations on AEM is only about half of those of OEM. In terms of the CPU time ratio on varying restart time, Figure 2(c) implies that AEM’s advantage converges more quickly when the number of components is larger. As is shown in Figure 3(a), Figure 3(b) and Figure 3(c), AEM has much better performance 13

35

10 8 6 4 2 0

1.1

AEM OEM

Ratio of AEM Time to OEM Time

AEM OEM

Average Number of Iteration

Average Time until Convergence

12

30 25 20 15 10

10

15

20 25 30 Number of Clusters

35

40

0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1

10

(a) CPU Time

K=10 K=20 K=30 K=40

1 0.9

15

20 25 30 Number of Clusters

35

40

0

(b) Number of Iterations

10

20

30

40 50 60 70 Number of Runs

80

90 100

(c) CPU Time Ratio

Figure 2: Performance comparison with varying component number 60

8 7 6 5 4 3 2 1 0

1.1

AEM OEM

55

Ratio of AEM Time to OEM Time

AEM OEM

9

Average Number of Iteration

Average Time until Convergence

10

50 45 40 35 30 25 20 15 10

40

60

80

100 120 140 Data Size (K)

160

180

200

(a) CPU Time

E=50 E=100 E=150 E=200

1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2

40

60

80

100 120 140 Data Size (K)

160

180

(b) Number of Iterations

200

0

10

20

30

40 50 60 70 Number of Runs

80

90 100

(c) CPU Time Ratio

Figure 3: Performance comparison with varying data size than OEM when we increase the data size from 50K to 200K. AEM can detect those worse local optimums much earlier, if there are more data available. The number of iterations invoked by AEM is almost the same, even when the data has been doubled. The ratio of CPU time is more stable when the data size is larger.

6.2

Results on Real Data

On Spam data set, AEM also show great advantage over OEM, on CPU time (Figure 4(a)) and on the number of iterations (Figure 4(b)). AEM is more efficient than OEM by one magnitude, independent to the number of components K. Also, AEM gains the advantage quickly even when there are very running time (Figure 4(c)). However, the experiments on Cloud data set show quite different results than the pervious results, where AEM has very limited advantage. We believe the difference on the results stems from normalization problem of the cloud data, which leads to some good local optimum hard to stop in most cases. In Figure 6, we present the log likelihood of the best solution found by AEM and OEM, when given the same CPU time. Since AEM is faster than OEM, it can discover better solution on two real data sets, Spam and Cloud.

7

Conclusion

In this paper, we propose a new acceleration method for multiple restart EM algorithm over Gaussian Mixture Model. We derive an upper bound on the local optimum of the likelihood function in

14

80

50 40 30 20 10 0

AEM OEM

70

Ratio of AEM Time to OEM Time

AEM OEM

Average Number of Iteration

Average Time until Convergence

60

60 50 40 30 20 10 0

1

10 Number of Clusters

100

1 0.8 0.6 0.4 0.2 0

1

(a) CPU Time

K=2 K=4 K=8 K=16

1.2

10 Number of Clusters

100

0

(b) Number of Iterations

10

20

30

40 50 60 70 Number of Runs

80

90 100

(c) CPU Time Ratio

Figure 4: Performance comparison with varying component number on Spam data

2.5 2 1.5 1 0.5 0

1.1

AEM OEM

100

Ratio of AEM Time to OEM Time

AEM OEM

3

Average Number of Iteration

Average Time until Convergence

3.5

80 60 40 20 0

1

10 Number of Clusters

(a) CPU Time

100

K=2 K=4 K=8 K=16 K=32

1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1

1

10 Number of Clusters

(b) Number of Iterations

100

0

10

20

30

40 50 60 70 Number of Runs

80

90 100

(c) CPU Time Ratio

Figure 5: Performance comparison with varying component number on Cloud data the solution space. This upper bound computation turns out to be both efficient and effective in pruning un-promising procedures of EM algorithm.

References Dempster, A. P., Laird, N. M., & Robin, D. B. (1977). Maximum likelihood from incomplete data via the em algorithm (with discussion). Journal of Royal Statistical Society B, 39, 1–38. Duda, R. O., Hart, P. E., & Stork, D. G. (2001). Pattern classification (2nd ed.). Wiley-Interscience. Elkan, C. (2003). Using the triangle inequality to accelerate k-means. ICML (pp. 147–153). Jordan, M. I., & Xu, L. (1995). Convergence results for the em approach to mixtures of experts architectures. Neural Networks, 8, 1409–1431. Kanungo, T., Mount, D., Netanyahu, N., Piatko, C., Silverman, R., & Wu, A. (2002). An efficient k-means clustering algorithm: analysis and implementation. Lutkepohl, H. (1996). Handbook of matrices. John Wiley & Sons Ltd. McLachlan, G., & Krishnan, T. (1996). The em algorithm and extensions. Wiley-Interscience. McLachlan, G., & Peel, D. (2000). Finite mixture models. Wiley-Interscience. Ueda, N., Nakano, R., Ghahramani, Z., & Hinton, G. E. (2000). Smem algorithm for mixture models. Neural Computation, 12, 2109–2128. 15

-125000

-19560 -19580

-130000

-19600 Likelihood

Likelihood

-135000 -140000 -145000

-19620 -19640 -19660 -19680

-150000

-19700 -155000

-19720

AEM OEM

-160000 0

200

400 600 CPU Time

800

AEM OEM

-19740 1000

0

(a) on Spam data

5

10

15 20 CPU Time

25

30

(b) on Cloud data

Figure 6: Likelihood comparison with fixed CPU time Zhang, Z., Dai, B. T., & Tung, A. K. (2006). On the lower bound of local optimums in k-means algorithm. ICDM.

16

Estimating Local Optimums in EM Algorithm over ...

May 1, 2008 - School of Computing. National Univ. ..... On the other hand, any Θα or Θβ is better than Θt by the definition of R. This leads ..... The Cloud data set consists of .... Convergence results for the em approach to mixtures of experts.

243KB Sizes 0 Downloads 139 Views

Recommend Documents

On the Lower Bound of Local Optimums in K-Means ...
Then R(M, ∆) is a maximum region if f(∆) > 0. ... Theorem 3 Given a positive ∆ satisfying f(∆) > 0, if k- ..... grams were compiled with gcc 3.4.3 in Linux system.

Polony Identification Using the EM Algorithm Based on ...
Wei Li∗, Paul M. Ruegger†, James Borneman† and Tao Jiang∗. ∗Department of ..... stochastic linear system with the em algorithm and its application to.

A Global–Local Approach for Estimating the Internet's ...
applications (antivirus software, intrusion detection systems etc.). This paper introduces PROTOS (PROactive ... eration and management of all things) poses several security is- sues. The reason is that the attack surface .... analyse them to identif

Estimating Video Quality over ADSL2+ under ...
pure best-effort Internet service, are starting to play a damaging role when it comes ... video content, two Linux boxes were used at the extremes of the DSL line.

A local mobility agent selection algorithm for mobile ...
Email: {yxu, hlee, vriz}@i2r.a-star.edu.sg. Abstract— The Mobile IP ... dent service. No matter where a host resides physically, the network should be able to identify it correctly with transpar- ent support for communications above network layer.

Fast Markov Blanket Discovery Algorithm Via Local Learning within ...
is further increased through collecting full statistics within a single data pass. We conclude that IPC-MB plus AD-tree appears a very attractive solution in very large applications. Keywords: Markov blanket, local learning, feature selection, single

An algorithm for improving Non-Local Means operators ...
number of an invertible matrix X (with respect to the spectral norm), that is ...... Cutoff (ω). PSNR Gain (dB) barbara boat clown couple couple_2 hill house lake lena .... dynamic range of 0 to 0.2, corresponding to blue (dark) to red (bright) colo

A Robust Algorithm for Characterizing Anisotropic Local ...
755 College Road East, Princeton, NJ 08540, USA. 2. INRIA Rhône- ...... lung walls and near rib structures. All the 14 ... Academic Press, San Diego, 1990. 5.

Local Learning Algorithm for Markov Blanket Discovery
Ecole Polytechnique de Montreal,. C.P.6079, Succ. Centre-ville, Montreal, Quebec, Canada ... show that (i) BFMB significantly outperforms IAMB in measures of data ...... Spirtes, P., Glymour, C.: An algorithm for Fast Recovery of Sparse Casual ...

A Robust Algorithm for Local Obstacle Avoidance
Jun 3, 2010 - Engineering, India. Index Terms— Gaussian ... Library along with Microsoft Visual Studio for programming. The robot is equipped with ...

On Codes over Local Frobenius Rings: Generator ...
Jul 30, 2014 - of order 16 for illustration. ... It is well known, see [7], that the class of finite rings for which it makes ... codes is the class of finite Frobenius rings.

Golden Bear baseball rolls over Evergreen - Bryan Times: Local ...
Golden Bear baseball rolls over Evergreen - Bryan Times: Local Sports.pdf. Golden Bear baseball rolls over Evergreen - Bryan Times: Local Sports.pdf. Open.

Living Over the Store Architecture and Local Urban Life - Howard ...
There was a problem previewing this document. Retrying. ... Living Over the Store Architecture and Local Urban Life - Howard Davis.pdf. Living Over the Store ...

Estimating parameters in stochastic compartmental ...
Because a full analytic treat- ment of a dynamical ..... Attention is restricted to an SIR model, since for this case analytic solutions to the likelihood calculations ...

Estimating parameters in stochastic compartmental ...
The field of stochastic modelling of biological and ecological systems ..... techniques to estimate model parameters from data sets simulated from the model with.

Experiments in Separating Computational Algorithm ...
DUAL devout target "/dev/tty"; // declaration : the 'target' variable is external ..... if ( DEVICE == file){ ofstream to; to.open(Ch,ios::ate); to.seekp(index);.

DC_CD_VPD-Measles-Algorithm-to-Determine-Susceptibility-in ...
Page 1 of 1. General. Population. Born BEFORE 1957. No exclusion. Monitor for S/S. If no history of disease, consider titer or. single dose of vaccine. Born in or AFTER 1957. No vaccine doses. Consult with CDPHE (Meghan, Emily or Amanda). for 21-day

Expectation-Maximization Algorithm in Image ...
Mar 15, 2007 - Specification. Estimate. Maximize. Gaussian. E-Step. M-Step. Results. Reference. Expectation-Maximization Algorithm in Image. Segmentation.

Unnatural Monopolies in Local Telephone
We use information technology and tools to increase productivity and facilitate new .... cross-sectional time series data with enough degrees of freedom to yield ...

Unnatural Monopolies in Local Telephone
We use information technology and tools to increase productivity and ... The divestiture of the American Telephone and Telegraph Company (AT&T) on January ... AT&T, an important public policy question is whether to permit entry and ...