On The Eigenvectors of p-Laplacian Dijun Luo, Heng Huang, Chris Ding, and Feiping Nie Department of Computer Science and Engineering, University of Texas, Arlington, Texas, USA

Abstract. Spectral analysis approaches have been actively studied in machine learning and data mining areas, due to their generality, efficiency, and rich theoretical foundations. As a natural non-linear generalization of Graph Laplacian, p-Laplacian has recently been proposed, which interpolates between a relaxation of normalized cut and the Cheeger cut. However, the relaxation can only be applied to two-class cases. In this paper, we propose full eigenvector analysis of p-Laplacian and obtain a natural global embedding for multi-class clustering problems, instead of using greedy search strategy used by previous researchers. An efficient gradient descend optimization approach is introduced to obtain the p-Laplacian embedding space, which is guaranteed to converge to feasible local solutions. Empirical results suggest that the greedy search method often fails in many real-world applications with non-trivial data structures, but our approach consistently gets robust clustering results. Visualization results also indicate our embedding space preserves the local smooth manifold structures existing in real-world data.

1

Introduction

Graph-based methods, such as spectral embedding [1, 2], spectral clustering [3, 1, 2], and semi-supervised learning [4–6], have recently received much attention from the machine learning community. Due to their generality, efficiency, and rich theoretical foundations [7, 1, 4, 8–10], these methods have been widely explored and applied into various machine learning related research areas, including computer vision [3, 11, 12], data mining [13], speech recognition [14], social networking [15], bioinformatics [16], and even commercial usage [17, 18]. Recently, as a nonlinear generalization of the standard graph Laplacian, graph p-Laplacian starts to attract attentions from machine learning community, such as B¨ uhler et. al. [19] proved the relationship between graph p-Laplacian and Cheeger cuts. Meanwhile, discrete p-Laplacian has also been well studied in mathematics community and solid properties have been investigated by previous work [20–22]. B¨ uhler [19] provided a rigorous proof of the approximation of the second eigenvector of p-Laplacian to the Cheeger cut. Unlike other graph-based approximation/relaxation techniques (e.g. [23]), the approximation to the optimal Cheeger cut is guaranteed to be arbitrarily exact. This discovery starts, theoretically and practically, a direction to graph cut based applications. Unfortunately, the p-Laplacian eigenvector problem leads to an untractable optimization, which

2

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

was solved (see [19]) by a somewhat complicated way. Moreover, they only solved for the second eigenvector and provided a direct approach to solve two-class clustering problems. For multi-class problems, they employed hierarchical strategy, which often leads to poor clustering quality in real world data with complicated structures due to its intrinsically greedy property. Putting the nice theoretical foundations of p-Laplacian and its difficulties together, one might immediately raise a question: can we obtain a full eigenvector space of p-Laplacian, similar to other regular spectral techniques, and easily derive a complete clustering analysis using p-Laplacian? To solve this question, in this paper, we investigate the whole eigenvector space of p-Laplacian and provide (1) an approximation of the whole eigenvectors which lead to a tractable optimization problems, (2) a proof to show that our approximation is very close to the true eigenvector solutions of p-Laplacian, and (3) an efficient algorithm to solve the resulting optimization problems, which is guaranteed to converge to feasible solutions. After introducing several important research results from mathematics community, we further explore the new properties of the full eigenvector space of p-Laplacian. Our main theoretical contributions are summarized in Theorems 2 and 3. Through our theoretical analysis and practical algorithm, the p-Laplacian based clustering method can naturally and optimally find the cluster structures in multi-class problems. Empirical studies in real world data sets reveal that greedy search often fails in complicated structured data, and our approach consistently obtains high clustering qualities. Visualizations of images data also demonstrate that our approach extracts the intrinsic smooth manifold reserved in the embedding space.

2

Discrete p-Laplacian and Eigenvector Analysis

Given a set of similarity measurements, the data can be represented as a weighted, undirected graph G = (V, E), where the vertices in V denote the data points and positive edge weights W encode thePsimilarity of pairs of data points. We denote the degree of node i ∈ V by di = j wij . Given function f : V → R, the p-Laplacian operator is defined as follows: X (∆W wij φp (fi − fj ), (1) p f )i = j

where φp (x) = |x|p−1 sign(x). Note that φ2 (x) = x, which becomes the standard graph Laplacian. In general, the p-Laplacian is a nonlinear operator. The eigenvector of p-Laplacian is defined as following. Definition 1. f : V → R is an eigenvector of p-Laplacian ∆W p , if there exists a real number λ, such that (∆W p f )i = λφp (fi ), i ∈ V. λ is called as eigenvalue of

∆W p

associated with eigenvector f .

(2)

p-Laplacian Embedding

3

One can easily verify that when p = 2, the operator ∆W p becomes the regular graph Laplacian ∆W = L = D −W , where D is a diagonal matrix with Dii = di , 2 and the eigenvectors of ∆W become the eigenvectors of L. The eigenvector of p p-Laplacian is also called p-eigenfunction. 2.1

Properties of Eigenvalues of p-Laplacian

Proposition 1. [24] If W represents a connected graph, and if λ is an eigenvalue of ∆W p , then λ ≤ 2p−1 max di . i∈V

This indicates that the eigenvalues of p-Laplacian are bounded by the largest volume. It is easy to check that for connected bipartite regular graph, the equality is achieved. 2.2

Properties of Eigenvectors of p-Laplacian

Starting from previous research results on p-Laplacian, we will introduce and prove our main theoretical contributions in Theorems 2 and 3. The eigenvectors of p-Laplacian have the following properties. Theorem 1. [19]. f is an eigenvector of p-Laplacian ∆W p , if and only if f is a critical point of the following function P p ij wij |fi − fj | , (3) Fp (f ) = p 2kf kp P where kf kpp = i |fi |p . The above theorem provides an equivalent statement of eigenvector and eigenvalue of p-Laplacian. It also serves as the foundation of analysis of eigenvector. Notice that Fp (αf ) = Fp (f ) which indicates the following property of p-Laplacian: Corollary 1. If f is an eigenvector of ∆W p associated with eigenvalue λ, then for any α 6= 0, αf is also an eigenvector of ∆W p associated with eigenvalue λ. W W Notice that ∆W p is not a linear operator, i.e. ∆p (αf ) 6= α∆p f , if p 6= 2. However, Corollary 1 shows that the linear transformation of a single eigenvector W remains an eigenvector of the p-Laplacian. Also note that ∆W p f = ∆p (f + d) W for any constant vector d. Thus ∆p is translation invariant, and we have

Corollary 2. c1 is an eigenvector of ∆W p for c 6= 0, associated with eigenvalue 0, where 1 is a column vector with all elements 1 and proper size. In the supplement (Lemma 3.2) of [19], authors also provided the following property of the non-trivial eigenvector of p-Laplacian.

4

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

Proposition 2. If f is a non-trivial eigenvector of ∆W p , then X φp (fi ) = 0.

(4)

i

The non-trivial eigenvectors mean those eigenvectors associated with non-zero eigenvalues. Inspired by the above properties of eigenvectors of p-Laplacian, we propose the following new theoretical analysis on eigenvectors of p-Laplacian. Definition 2. We call f 6= 0 and g 6= 0 as p-orthogonal if the following condition holds X φp (fi )φp (gi ) = 0. (5) i

As one of the main results in this paper, the following property of the full eigenvectors of p-Laplacian is proposed, Theorem 2. If f and g are two eigenvectors of p-Laplacian ∆W p associated with different eigenvalues λf and λg , and W is symmetric, and p ≥ 1, then f and g are p-orthogonal up to the second order Taylor expansion. Proof: By definitions, we have (∆W f )i = λf φ(fi ),

(6)

(∆W g )i

(7)

= λg φ(gi ).

Multiplying φp (gi ) and φp (fi ) on both sides of Eq. (6) and Eq. (7), respectively, we have (∆W f )i φ(gi ) = λf φ(fi )φ(gi ),

(8)

(∆W g )i φ(fi )

(9)

= λg φ(gi )φ(fi ).

By summing over i and taking the difference of both sides of Eq. (8) and Eq. (9), we get X X£ ¤ (λf − λg ) φp (fi )φp (gi ) = (∆W f )i φp (gi ) − (∆W g)i φp (fi ) . i

i

Notice that for any p > 1, a, b ∈ R, φp (a)φp (b) =|a|p−1 sign(a)|b|p−1 sign(b) =|a|p−1 |b|p−1 sign(a)sign(b) = |ab|p−1 sign(ab) = φp (ab). We have

X£ ¤ (∆W f )i φp (gi ) − (∆W g)i φp (fi ) i

=

X

wij (φp (fi − fj )φp (gi ) − φp (gi − gj )φp (fi ))

ij

=

X ij

wij (φp (fi gi − fj gi ) − φp (gi fi − gj fi ))

p-Laplacian Embedding

5

Since any constant vector c1 is a valid eigenvector of p-Laplacian, we write φp (x) as φp (x) = φp (c) + φ0p (c)(x − c) + o2 . And notice that wij = wji , then the above equation becomes X wij (φp (c) + φ0p (c)(fi gi − fj gi − c)) ij

− ≈

X ij

+

X

wji (φp (c) + φ0p (c)(fj gj − fj gi − c)) + o2

ij

wij φ0p (c)(fi gi − fj gj ) + X

X

wji φ0p (c)(fj gi − fj gi )

ij

wij (φp (c) +

cφ0p (c)

− φp (c) − cφ0p (c))

ij

=0 where o2 is the sum of second order Taylor expansion terms at constant c which is ignored. This leads to X (λf − λg ) φp (fi )φp (gi ) ≈ 0. i

Since λf 6= λg , we have

X

φp (fi )φp (gi ) ≈ 0.

i

If p = 2, the second order term of Taylor expansion is 0, then the approximation is exact. This property of p-Laplacian is significant different from the those in existing literacy, in the sense that it explores the relationship of the full eigenvectors space. Theorem 3. If f ∗1 , f ∗2 , · · · , f ∗n are n eigenvectors of operator ∆W p , associated with unique eigenvalues λ∗1 , λ∗2 , · · · , λ∗n , then f ∗1 , f ∗2 , · · · , f ∗n are local solution of the following optimization problem X min J(F) = Fp (f k ), (10) F

s.t.

X

k

φp (fik )φp (fil ) = 0, ∀k 6= l.

i

¡ ¢ where F = f 1 , f 2 , · · · , f n . Proof: We do the derivative of J(F) w.r.t f k : P

wij |f k −f k |p

∂ ij 2kf kikpp j ∂Fp (f k ) ∂J(F) = = ∂f k ∂f k ∂f k # " P k k p 1 ij wij |fi − fj | W k k = k p ∆p (f ) − φp (f ) . kf kp kf k kpp

(11)

6

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

From Theorem 3.1 in [19], P λ∗k =

ij

wij |fi∗k − fj∗k |p kf ∗k kpp

,

and by definition, ∗k ∗ ∗k ∆W p (f ) − λk φp (f ),

thus we have, ∂J(F) = 0, ∂f ∗k and according to Theorem 2, the constraints in Eq. (16) are satisfied. Thus f ∗k , k = 1, 2, · · · , n are local solutions for Eq. (10). On the other hand, one can show the following relationship between the Cheeger cut and the second eigenvector of p-Laplacian when K = 2. Definition 3. Given a undirected graph W and a partition of the nodes {C1 , C2 , · · · , CK }, the Cheeger cut of the graph is CC =

K X Cut(Ck , C¯k ) , min1≤l≤K |Cl |

(12)

k=1

where X

Cut(A, B) =

Wij ,

(13)

i∈A,j∈B

and C¯k is the complement of Ck , k = 1, 2, · · · , K. Proposition 3. Denote by CCc∗ the Cheeger cut value obtained by thresholding ∗ the second eigenvector of ∆W p , and CC is the global optimal value of Eq. (12) with K = 2, then the following holds CC ∗ ≤ CCc∗ ≤ p(max di ) i∈V

p−1 p

1

(CC ∗ ) p .

(14)

This property of the second eigenvector of ∆W p indicates that when p → 1, CCc∗ → CC ∗ . Notice that this approximation can be achieved arbitrarily accurate, which is different from other relaxation-based spectral clustering approximation. Thus it opens a total new direction of spectral clustering. However, this relationship holds only in the case of K = 2. In previous research, a greedy search strategy is applied to obtain Cheeger cut results for multiclass clustering [19]. In the algorithm, they first split data in to two parts and recursively dichotomize the data until a desired number of clusters is achieved. In our study, we discover that in many real world data sets, this greedy search strategy isn’t efficient and effective. This limitation inspires us to explore the whole eigenvectors of p-Laplacian to obtain better solution of Cheeger cuts.

p-Laplacian Embedding

3

7

Solving Complete Eigenfunctions for p-Laplacian

In previous section, we derive a single optimization problem for full eigenvectors of p-Laplacian. However, the optimization problem remains intractable. In this section, we propose an approximation algorithm to obtain full eigenvectors of p-Laplacian. We also provide a proof to show how good our approximation is. 3.1

Orthogonal p-Laplacian

Instead of solving Eq. (10), we solve the following problem: XX min Jo (F) = wij |fik − fjk |p , F

k

(15)

ij

s.t. F T F = I, kf k kpp = 1, k = 1, 2, · · · , n 3.2

(16)

The Approximation Evaluation

Here we show that the approximation is tight. By introducing Lagrangian multiplier, we obtain, X X ¡ ¢ k T L= QW ξk kf k kpp − 1 , (17) p (f ) − TrF FΛ − k

k

where QW p (f ) =

X

wij |fi − fj |p .

ij

Taking the derivative of L w.r.t. f k and set it to be zeros, we have, X p wij φp (fik − fjk ) − λk fik − pξk φp (fik ) = 0, i = 1, 2, · · · , n,

(18)

j

which leads to λk = or

i h k k p ∆W p (f ) − ξk φf

i

fik

,

h i k k p ∆W p (f )/ξk − φf

λk i . (19) = ξk fik h i k k Denote ηi = ∆W p (f )/ξk − φf , from [24], we know that ηi is a constant w.r.t. i

i. Notice that Eq. (19) holds for all i, thus ηi ≈ 0, indicating that comparing with ξk , λk can be ignored. Thus, Eq. (18) becomes

8

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

p

X

wij φp (fik − fjk ) − pξk φp (fik ) = 0, i = 1, 2, · · · , n,

j

and by definition, f k is an eigenvector of ∆W p associate with eigenvalue ξk .

4

p-Laplacian Embedding

Since Fp (f ) = Fp (αf ) for α 6= 0, we can always scale f without any change. Thus, we propose the following p-Laplacian Embedding problems. min JE (F) = F

X

P ij

wij |fik − fjk |p kf k kpp

k

,

(20)

s.t. F T F = I. 4.1

(21)

Optimization

The gradient of JE w.r.t. fik can be written as,   k ∂JE φ (f ) 1 X p i  wij φp (fik − fjk ) − . = k p k kp kf kp kf ∂fik p j

(22)

If we simply use the gradient descend approach, the solution f k might not be orthogonal. Thus we modify the gradient as following to enforce the orthogonality, ∂JE ∂JE ← −F ∂F ∂F

µ

∂JE ∂F

¶T F.

We summarize the p-Laplacian Embedding algorithm as following, Here the parameter α is the step length, which is set to be P |Fik | P . α = 0.01 ik |G ik | ik One can easily see that if F T F = I, then using the a simple gradient descend approach can guarantee to give a feasible solution. More explicitly, we have the following results. Theorem 4. The solution obtained from Algorithm 1 satisfies the constraint in Eq. (21).

p-Laplacian Embedding

9

Input: Pairwise graph similarity W , number of embedding dimension K Output: Embedding space F Compute L = D − W , where D is a diagonal matrix with Dii = di . Compute eigenvector decomposition of L: L = U SU T , Initialize F ← U (:, 1 : K) while not converged do ³ ´T E E E G ← ∂J − F ∂J F , where ∂J is computed using Eq. (22) ∂F ∂F ∂F F ← F − αG. end

Algorithm 1: p-Laplacian Embedding

Proof: Since Laplacian L is symmetric, we have F T F = I for initialization, and ¡ ¢T GT F t + F t G " Ã µ ¶T !T µ ¶T # ¡ ¢ ∂JE ∂J ∂J ∂JE T E E −F F Ft + Ft −F F = ∂F ∂F ∂F ∂F µ ¶T µ ¶T ¡ t ¢T ∂JE ¡ ¢T ∂JE ∂JE ∂JE t = F − F − Ft + Ft ∂F ∂F ∂F ∂F =0, By Algorithm 1 we have,

F t+1 = F t − αG,

Thus ¡

F t+1

¢T

¡ ¢T ¡ t ¢ F t+1 = F t − αG F − αG h ¡ ¢T ¡ ¢T i = F t F t − α GT F t + F t G = I.

This technique is a special case of Natural Gradient, which can be found in [25]. Since JE (F) is bounded: JE (F) ≥ 0, our algorithm also has the following obvious property, Theorem 5. Algorithm 1 is guaranteed to converge.

5

Experimental Results

In this section, we will evaluate the efficiency of our proposed p-Laplacian Embedding algorithm. We use 8 benchmark data sets to demonstrate the results: AT&T, MNIST, PIE, UMIST, YALEB, ECOLI, GLASS, and DERMATOLOGY.

10

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

5.1

Data Set Descriptions

In the AT&T database 1 , there are ten different images of each of 40 distinct subjects. For some subjects, the images were taken at different times, varying the lighting, facial expression (open/close eyes, smiling/no-smiling) and facial details (glasses/no glasses). All images were taken against a dark homogeneous back-ground with the subjects in an upright, frontal, position (with tolerance for some side movement). MNIST hand- written digits data set consists of 60,000 training and 10,000 test digits [26]. The MNIST data set can be downloaded from website 2 with 10 classes, from digit “0” to “9”. In the MNIST data set, each image is centered (according to the center of mass of the pixel intensities) on a 28x28 grid. We select 15 images for each digit in our experiment. UMIST faces is for multi-view face recognition, which is challenging in computer vision because the variations between the images of the same face in viewing direction are almost always larger than image variations in face identity. This data set contains 20 persons with 18 images for each. All these images of UMIST database are cropped and resized into 28 × 23 images. Due to the multiview characteristics, the images shall lie in a smooth manifold. We further use this data set to visually test our embedding smoothness. CMU PIE (Face Pose, Illumination, and Expression) face database contains 68 subjects with 41,368 face images as a whole. Preprocessing to locate the faces was applied. Original images were normalized (in scale and orientation) such that the two eyes were aligned at the same position. Then, the facial areas were cropped into the final images for matching. The size of each cropped image is 64 × 64 pixels, with 256 grey levels per pixel. No further preprocessing is done. In our experiment, we randomly pick 10 different combinations of pose, face expression, and illumination condition. Finally we have 68 × 10 = 680 images. Another images benchmark used in our experiment is the combination of extended and original Yale database B [27]. These two databases contain single light source images of 38 subjects (10 subjects in original database and 28 subjects in extended one) under 576 viewing conditions (9 poses x 64 illumination conditions). Thus, for each subject, we got 576 images under different lighting conditions. The facial areas were cropped into the final images for matching [27], including: 1) preprocessing to locate the faces was applied; 2) original images were normalized (in scale and orientation) such that the two eyes were aligned at the same position. The size of each cropped image in our experiments is 192×168 pixels, with 256 gray levels per pixel. We randomly pick up 20 images for each person and also sub-sample the images down to 48 × 42. To visually test the quality of the embedding space, we pickup the images such that they are taken in different illumination conditions. All other three data sets (ECOLI, GLASS, and DERMATOLOGY) are selected from UCI Repository [28]. The detailed information of the 8 selected data sets can be found in Table 1. 1 2

http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html http://yann.lecun.com/exdb/mnist/

p-Laplacian Embedding

11

Table 1. Detailed information of data sets used in our experiments. Data set AT&T MNIST PIE UMIST YALEB ECOLI GLASS DERMATOLOGY

#samples 400 150 680 360 1984 336 214 366

#Attribute 644 784 1024 644 2016 343 9 34

#class 40 10 68 20 31 8 6 6

For all the data sets used in our experiments, we directly use the original space without any preprocessing. More specifically, for images data sets, we use the raw gray level values as features. 5.2

Experimental Settings

We construct the pairwise similarity of data points as follows. ³ ´ ( kx −x k2 exp − iri rjj , xi , xj are neighbors Wij = otherwise 0,

(23)

where ri and rj are the average distances of K-nearest neighbors of data point i and j, respectively. K is set to 10 in all our experiments, which is the same as in [19]. By neighbors here we mean xi is a K-nearest neighbors of xj or xj is a K-nearest neighbors of xi . For our method (Cheeger cut Embedding or CCE), we first obtain the embedding space using Algorithm 1. Then a standard K-means algorithm is applied to further determine the clustering assignments. For visualization, we use the second and third eigenvectors as the x-axis and y-axis, respectively. For direct comparison and succinct presentation, we only compare our results with greedy search Cheeger cut algorithm [19] in terms of three clustering quality measurements. We download their codes and directly use them with default settings. For both methods, we set p = 1.2, which is suggested in previous research [19]. 5.3

Measurements

We measure 3 metrics in our experiments: the objective of Eq. (3), the Cheeger cut defined in Eq. (12), and clustering accuracy. Clustering accuracy (ACC) is defined as: Pn δ(li , map(ci )) , (24) ACC = i=1 n

12

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

where li is the true class label and ci is the obtained cluster label of xi , δ(x, y) is the delta function, and map(·) is the best mapping function. Note δ(x, y) = 1, if x = y; δ(x, y) = 0, otherwise. The mapping function map(·) matches the true class label and the obtained cluster label and the best mapping is solved by Kuhn-Munkres algorithm. A larger ACC indicates a better performance. And Lower Objective of Eq. (3) or lower Cheeger cuts suggests better clustering quality. 5.4

Evaluation Results

Embedding Results We use 4 data sets (AT&T, MNIST, UMIST, YALEB) to visualize the embedding results obtained by our method. For each data set, we select samples in four different clusters. We use the second and third eigenvector as x-axis and y-axis, respectively. The embedding results are shown in Figure 1 (a) – (d). For AT&T data, the four persons are well separated. For MNIST data, the four digits are separated in most of the images. Three images (“3”, “2”, and “0” as highlighted in Figure 1(b)) are visually different from other images from the same group. The embedding result also show that these three images are far way from the other objects in the same group, This result indicates that our embedding space reserves the visual characteristics. For UMIST and YALEB data, since the images from the same group are took under different face expression or illumination conditions, they are arranged in a smooth manifold. This structure also remains in our embedding space, see Figures 1(c) and 1(d). Clustering Analysis on Confusion Matrices We select 10 groups for AT&T, MNIST, PIE, UMIST, and YALEB, 6 for GLASS and DERMATOLOGY, and 8 for ECOLI. We compare the greedy search Cheeger cut [19] (GSCC) to our method (CCE). The confusion matrices are show in Figure. 2. In AT&T, MNIST, and ECOLI, our method obviously outperforms GSCC, because the diagonals of our confusion matrices are much stronger than those in GSCC results. Clustering Quality Analysis We use three criteria mentioned above to measure the quality of clustering results. We compare our method with greedy search Cheeger cut in various experimental settings. For AT&T, MNIST, PIE, and UMIST, we choose k = 2, 3, 4, 5, 6, 8, 10, where k is the number of clusters. Typically, the larger k is, the more difficult the clustering task is, and the lower clustering accuracy should be. For ECOLI, GLASS, YALEB, and DERMATOLOGY we set k = 2, 3, 4, 5, 6, 8, k = 2, 3, 4, 5, 6, 7, k = 2, 4, 5, 6, 8, 10 and k = 2, 3, 4, 5, 6, respectively. We set these numbers of k according to the size of the original data sets and also for convenient presentation. Results are shown in Table 2. Notice that for greedy search, if k > 2, there is no way to calculate the objective function values defined in Eq. (3). In Table 2, when the data set is simple (i.e. k is small), the accuracy of the two methods is close to each other. However, if the data is complex (i.e. when k is large), our method has much better clustering results than greedy search. For

p-Laplacian Embedding

(a) AT&T

(c) UMIST

13

(b) MNIST

(d) YALEB

Fig. 1. Embedding results on four image data sets using the second and third eigenvector of p-Laplacian as x-axis and y-axis, respectively, where p = 1.2. Different colors indicate different groups according to ground truth. In (b) the highlighted are images which are visually far away from other images in the same group.

example, in AT&T, when k = 10, our approach remains high (78%) in clustering accuracy, while greedy search only achieves 38%. Also we can see that when k is large, our algorithm obtains both objective and Cheeger cut much lower than greedy search. One should notice that the setting of MNIST data used in our experiment is different from the one used in previous research [19].

6

Conclusions

Spectral data analysis is important in machine learning and data mining areas. Unlike other relaxation-based approximation techniques, the solution obtained by p-Laplacian can approximate the global solution arbitrarily tightly. Meanwhile, Cheeger cut favors the solutions which are more balanced. This paper is the first one to offer a full eigenvector analysis of p-Laplacian. We provide an efficient gradient descend approach to solve the full eigenvector problem. Moreover,

1 2 3 4 5 6 7 8 9 10 Prediction

0

4 2 1 2 3 4 5 6 7 8 9 10 Prediction

0

10

5

1 2 3 4 5 6 7 8 9 10 Prediction

6 4 2 0

1 2 3 4 5 6 7 8 9 10 Prediction

1 2 3 4 5 6 7 8 9 10

4

2

1 2 3 4 5 6 7 8 9 10 Prediction

0

1 2 3 4 5 6 7 8 9 10

10

5

10 5 1 2 3 4 5 6 7 8 Prediction

0

20 15 10

20

3

15

4

10

5

0

5

1 2 3 4 5 6 7 8 9 10 Prediction

0

8 6 4 2

(g) YALEB

0

20

2 15

3

10

4 5

2

3 4 5 Prediction

0

6

5 1

2

3 4 5 Prediction

0

6

(f) GLASS 10

1 2 3 4 5 6 7 8 9 10 Prediction

5

6 1

30

12

1 2 3 4 5 6 7 8 9 10

10

1

6

0

Ground Truth

5

Ground Truth

Ground Truth

10

15

1 2 3 4 5 6 7 8 9 10 Prediction

2

(e) ECOLI 1 2 3 4 5 6 7 8 9 10

20

1 2 3 4 5 6 7 8 9 10

25

1

5 1 2 3 4 5 6 7 8 Prediction

0

(d) UMIST 25

1 2 3 4 5 6 7 8

0

1 2 3 4 5 6 7 8 9 10 Prediction

Ground Truth

15

Ground Truth

Ground Truth

20

5

1 2 3 4 5 6 7 8 9 10 Prediction

15

(c) PIE 1 2 3 4 5 6 7 8

10

(b) MNIST 6

Ground Truth

1 2 3 4 5 6 7 8 9 10

Ground Truth

Ground Truth

(a) AT&T 8

0

15

1 2 3 4 5 6 7 8 9 10

Ground Truth

2

6

1 2 3 4 5 6 7 8 9 10

Ground Truth

4

8

30

1

1

2

2

20 3 4 10 5 6

Ground Truth

6

10

1 2 3 4 5 6 7 8 9 10

Ground Truth

8

Ground Truth

10

1 2 3 4 5 6 7 8 9 10

Ground Truth

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

Ground Truth

14

20 3 4 10 5 6

1

2

3 4 5 Prediction

6

0

1

2

3 4 5 Prediction

6

0

(h) DERMATOLOGY

Fig. 2. Comparison of confusion matrices of GSCC (left in each panel) and our CCE (right in each panel) on 8 data sets. Each column of the matrix represents the instances in a predicted class, while each row represents the instances in an actual class.

we give out new analysis of the properties of eigenvectors of p-Laplacian. Empirical studies show that our algorithm is much more robust in real world data sets clustering than the previous greedy search p-Laplacian spectral clustering. Therefore, both theoretical and practical results proposed by this paper provide a promising direction to machine learning community and related applications.

References 1. Belkin, M., Niyogi, .: Laplacian eigenmaps and spectral techniques for embedding and clustering. In: NIPS 14, MIT Press (2001) 585–591 2. Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. In Dietterich, T.G., Becker, S., Ghahramani, Z., eds.: NIPS, MIT Press (2001) 585–591 3. Shi, J., Malik, J.: Normalized cuts and image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence 22 (2000) 888–905 4. Zhou, D., B., O., Lal, T.N., Weston, J., Sch¨ olkopf, B.: Learning with local and global consistency. NIPS 16 (2003) 321–328

p-Laplacian Embedding

15

Table 2. Clustering quality comparison of greedy search Cheeger cut and our method. Obj is the objective function value defined in Eq. (3), CC is the Cheeger cut objective defined in Eq. (12), and Acc is the clustering accuracy defined in Eq. (24). For greedy search, objective in Eq. (3) is not provided when k > 2.

AT&T

Greedy Search

k

Obj

CC

2 3 4 5 6 8 10

31.2 -

29.1 124.9 385.9 1092.5 3034.6 5045.6 8012.7

PIE

Acc

Greedy Search Obj

CC

Acc

2 3 4 5 6 8 10

38.4 -

31.0 144.0 514.3 2224.7 3059.9 5362.7 7927.2

60.0 50.0 43.0 34.0 28.0 24.0 22.0

Greedy Search

k

Obj

2 3 4 5 6 8

70.7 65.7 189.5 529.0 738.1 - 1445.6 - 16048.2

YALEB

CC

CC

Acc

100.0 31.2 29.1 100.0 80.0 187.4 91.3 100.0 60.0 333.2 176.4 100.0 50.0 500.7 306.6 90.0 45.0 673.3 436.8 73.0 36.0 1134.9 862.6 80.0 38.0 1712.8 1519.7 78.0

k

ECOLI

Our Method Obj

Acc 97.0 73.0 72.0 57.0 61.0 58.0

Greedy Search

k

Obj

CC

Acc

2 4 6 8 10

73.5 -

68.6 425.7 1642.4 3953.5 4911.5

50.0 33.0 27.0 27.0 24.0

Our Method Obj

CC

38.4 38.4 189.1 112.8 324.3 179.6 477.1 336.6 673.6 489.7 1146.1 985.9 1707.4 1776.3 CC

70.7 67.8 318.3 180.6 458.4 306.0 790.0 566.7 1083.5 993.1 1969.5 1736.0 CC

73.5 68.6 740.1 405.0 2300.6 1534.5 3044.4 2109.8 4690.6 3368.1

2 3 4 5 6 8 10

46.0 42.9 100.0 46.0 42.9 100.0 182.9 56.0 268.1 132.3 98.0 534.0 47.0 459.7 252.9 97.0 - 1129.9 45.0 680.7 402.8 92.0 - 6356.0 40.0 923.2 582.7 89.0 - 10785.6 33.0 1608.9 1110.4 86.0 - 16555.5 35.0 2461.2 1731.2 85.0

UMIST

CC

Acc

Our Method

Obj

Greedy Search

k

Obj

CC

Acc

65.0 67.0 60.0 58.0 62.0 45.0 45.0

2 3 4 5 6 8 10

86.2 -

80.8 269.1 732.3 973.5 1888.4 8454.9 4204.1

57.0 42.0 42.0 42.0 33.0 29.0 34.0

GLASS

Greedy Search

Acc

k

Obj

98.0 89.0 84.0 77.0 80.0 79.0

2 3 4 5 6 7

71.4 85.3 259.5 821.3 - 7659.5 - 12160.9 - 12047.2

Our Method Obj

Greedy Search

k

Acc

Our Method Obj

MNIST

DERMA

CC

k

Obj

CC

50.0 39.0 34.0 31.0 28.0

2 3 4 5 6

74.1 -

69.1 275.0 493.8 1120.5 2637.5

CC

Acc

Our Method Obj

CC

86.2 80.4 388.4 193.2 669.6 399.8 1026.1 639.1 1426.0 898.2 2374.8 1939.8 3793.1 2673.7

Acc 57.0 58.0 62.0 57.0 60.0 55.0 54.0

Our Method

Acc

Obj

CC

Acc

63.0 39.0 48.0 43.0 38.0 37.0

71.4 386.4 617.9 862.7 1253.7 1253.7

111.3 198.8 389.6 498.9 814.5 816.3

63.0 66.0 71.0 57.0 62.0 62.0

Greedy Search

Acc

Obj

Acc

Our Method Obj

CC

Acc

100.0 74.1 69.1 100.0 78.0 426.7 192.0 100.0 81.0 770.9 403.3 95.0 77.0 1206.6 661.2 96.0 45.0 1638.6 1115.1 96.0

5. Kulis, B., Basu, S., Dhillon, I.S., Mooney, R.J.: Semi-supervised graph clustering: a kernel approach. Machine Learning 74 (2009) 1–22

16

Dijun Luo, Heng Huang, Chris Ding, Feiping Nie

6. Belkin, Matveeva, Niyogi: Regularization and semi-supervised learning on large graphs. In: COLT: Proceedings of the Workshop on Computational Learning Theory, Morgan Kaufmann Publishers. (2004) 7. Chung, F.: Spectral graph theory. AMS (1997) 8. Hein, Audibert, von Luxburg: From graphs to manifolds – weak and strong pointwise consistency of graph laplacians. In Auer, P., Meir, R., eds.: Proc. of the 18th Conf. on Learning Theory (COLT), Springer. (2005) 486–500 9. Robles-Kelly, Hancock: A riemannian approach to graph embedding. PATREC: Pattern Recognition, Pergamon Press 40 (2007) 10. Guattery, Miller: On the quality of spectral separators. SIJMAA: SIAM Journal on Matrix Analysis and Applications 19 (1998) 11. Jain, V., 0002, H.Z.: A spectral approach to shape-based retrieval of articulated 3D models. Computer-Aided Design 39 (2007) 398–407 12. Chen, G., Lerman, G.: Spectral curvature clustering (SCC). International Journal of Computer Vision 81 (2009) 317–330 13. Jin, R., Ding, C.H.Q., Kang, F.: A probabilistic approach for optimizing spectral clustering. (2005) 14. Bach, F.R., Jordan, M.I.: Learning spectral clustering, with application to speech separation. Journal of Machine Learning Research 7 (2006) 1963–2001 15. White, S., Smyth, P.: A spectral clustering approach to finding communities in graph. In: SDM. (2005) 16. Liu, Y., Eyal, E., Bahar, I.: Analysis of correlated mutations in HIV-1 protease using spectral clustering. Bioinformatics 24 (2008) 1243–1250 17. Anastasakos, T., Hillard, D., Kshetramade, S., Raghavan, H.: A collaborative filtering approach to ad recommendation using the query-ad click graph. In Cheung, D.W.L., Song, I.Y., Chu, W.W., Hu, X., Lin, J.J., eds.: CIKM, ACM (2009) 1927–1930 18. Cheng, H., Tan, P.N., Sticklen, J., Punch, W.F.: Recommendation via query centered random walk on K-partite graph. In: ICDM, IEEE Computer Society (2007) 457–462 19. B¨ uhler, T., Hein, M.: Spectral clustering based on the graph p -laplacian. In: ICML. Volume 382., ACM (2009) 81–88 20. Amghibech, S.: Eigenvalues of the discrete p-laplacian for graphs. Ars Comb 67 (2003) 283 – 302 21. Allegretto, W., Huang, Y.X.: A picone’s identity for the p-laplacian and applications. Nonlinear Anal. 32 (1998) 819–830 22. Bouchala, J.: Resonance problems for p-laplacian. Mathematics and Computers in Simulation 61 (2003) 599–604 23. Ding, C.H.Q., He, X.: On the equivalence of nonnegative matrix factorization and spectral clustering. In: SDM. (2005) 24. Amghibech, S.: Bounds for the largest p-laplacian eigenvalue for graphs. Discrete Mathematics 306 (2006) 2762–2771 25. Amari, S.: Natural gradient works efficiently in learning. Neural Computation 10 (1998) 251–276 26. Cun, Y.L.L., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proceedings of IEEE 86 (1998) 2278–2324 27. Georghiades, A., Belhumeur, P., Kriegman, D.: From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intelligence 23 (2001) 643–660 28. Asuncion, A., Newman, D.: UCI machine learning repository (2007)

On The Eigenvectors of p-Laplacian

working [15], bioinformatics [16], and even commercial usage [17,18]. Recently, ..... In the AT&T database 1, there are ten different images of each of 40 distinct.

1MB Sizes 0 Downloads 194 Views

Recommend Documents

Efficiently Training A Better Visual Detector With Sparse Eigenvectors
Acknowledgments. NICTA is funded by the Australian Government as represented by the ... pages 97–105, San Francisco, CA, USA, 1999. [14] J. Leskovec.

Overview of comments received on Guideline on the conduct of ...
Jan 19, 2017 - 30 Churchill Place ○ Canary Wharf ○ London E14 5EU ○ United Kingdom. An agency of ... Send a question via our website www.ema.europa.eu/contact. © European ... clinical studies according to Good Clinical Practice. (GCP), Good ..

Overview of comments received on Guideline on the conduct of ...
Jan 19, 2017 - Telephone +44 (0)20 3660 6000 Facsimile +44 (0)20 3660 5555. Send a question via ... industrial and commercial property, the applicant shall ...

Corrigendum on the Schedule of the Training-Workshop on Records ...
There was a problem previewing this document. Retrying... Download ... Corrigendum on the Schedule of the Training-Workshop on Records Management.pdf.

On the Value of Variables
Call-by-value and call-by-need λ-calculi are defined using the distinguished ... (or imperative extensions of Plotkin's calculus [3]) employ a notion of practical ..... and theoretical values it would evaluate exactly in the same way as for CBN.

On the Value of Variables
rewriting rules at top level, and then taking their closure by evaluation contexts. A peculiar aspect of the LSC is that contexts are also used to define the rules at top level. Such a use of contexts is how locality on proof nets (the graphical lang

Note on the Voice of the Customer
on the ambient room light and reflections, the colors that the software designer chooses, the ... 10. Classes have a balance between theory and real-world application. ... A wide selection of companies and industries that recruit at the business.

Report Of The Commission On The National
Report of the Commission on the National and the Colonial Questions. Comrades, I shall ... by some big imperialist power, have become greatly dependent on that power by virtue of peace .... but more and more data will gradually accumulate.

Decision of the Management Board - On the appointment of the ...
Mar 16, 2017 - Lenihan as the Agency's Accounting Officer (EMA/MB/124276/2012). Done at London, 16 March 2017. [Signature on file]. Christa Wirthumer- ...

Detailed guidance on the electronic submission of information on ...
Apr 19, 2017 - basis in line with the requirements and business processes described in this ...... to support data analytics and business intelligence activities; ...

On the Effect of Bias Estimation on Coverage Accuracy in ...
Jan 18, 2017 - The pivotal work was done by Hall (1992b), and has been relied upon since. ... error optimal bandwidths and a fully data-driven direct plug-in.

On the Impact of Kernel Approximation on Learning ... - CiteSeerX
The size of modern day learning problems found in com- puter vision, natural ... tion 2 introduces the problem of kernel stability and gives a kernel stability ...

On the Effect of Bias Estimation on Coverage Accuracy in ...
Jan 18, 2017 - degree local polynomial regression, we show that, as with point estimation, coverage error adapts .... collected in a lengthy online supplement.

Detailed guidance on the electronic submission of information on ...
Apr 19, 2017 - marketing authorisation holders to the European Medicines Agency in accordance with Article .... Pharmacovigilance enquiry email (AP.7) .

Concept paper on revision of Guidelines on the clinical investigation ...
Jul 21, 2016 - ... to regulatory decisions e.g. potency labelling and monitoring of patient .... the clinical trial concept taking into account the limits in availability ...

Study on the influence of dryland technologies on ...
Abstract : A field experiment was conducted during the North East monsoon season ... Keywords: Sowing time, Land management, Seed hardening and Maize ...

On the Inconsistency of Certain Axioms on Solution ...
of California at ... At each of a player's information sets, a strategy for that player specifies ... Historically, this encouraged the view that while information is lost in.

On the Impact of Kernel Approximation on ... - Research at Google
termine the degree of approximation that can be tolerated in the estimation of the kernel matrix. Our analysis is general and applies to arbitrary approximations of ...

ON THE NODAL LINES OF EISENSTEIN SERIES ON ...
where z ∈ H2 and ξ ∈ ∂H2 belongs to the domain of discontinuity of Γ that is .... several standard properties of Busemann functions which can be checked ..... proof, but the price to pay to obtain a result at this level of generality is that

On the Representation of Preference Orders on ...
*Department of Economics, College of Business, Florida Atlantic University, ... throughout the analysis: for x, y * X we say x $ y iff xi $ yi for all i * M; x>y iff x $ y iff ...

Divergent position on a CVMP opinion on the establishment of ...
Reading summary reports and the assessment report that was prepared for the current procedure, the following can be summarised: •. We do not have accessible any relevant scientific information concerning the behaviour and fate of fluazuron in milk,

Divergent position on a CVMP opinion on the establishment of ...
The CVMP has estimated the safe level based on a mathematic calculation only (even if the conclusion is that exposure does not represent a sizable part of the ...

22.12.14 - Report on the training on Marketing of Agricultural .pdf ...
22.12.14 - Report on the training on Marketing of Agricultural .pdf. 22.12.14 - Report on the training on Marketing of Agricultural .pdf. Open. Extract. Open with.

On the Inconsistency of Certain Axioms on Solution ...
moving from the extensive form to the normal form, all strategically relevant information is preserved. It is worth remarking that if (Al) is weakened to require only that two extensive form games have the same solution if their agent-normal forms (s