Discriminative High Order SVD: Adaptive Tensor Subspace Selection for Image Classification, Clustering, and Retrieval Dijun Luo, Heng Huang, Chris Ding The University of Texas at Arlington 701 S Nedderman Drive, Arlington, Texas, USA [email protected], {heng,chqding}@uta.edu

Abstract

the projective feature vector space. Although PCA/SVD and LDA work well for data dimension reduction of two-dimensional arrays, they traditionally express the input data as vectors and it is not natural to apply them into higher dimensional data, known as high-order data. For high-order data we can use Tucker decomposition [16]. After adding orthogonality constraints to tensor decomposition, the High-Order Singular Value Decomposition (HOSVD) [11] has been widely used in computer vision applications [18]. Several other tensor based methods have been also proposed, e.g. Two Dimensional PCA (2DPCA) [21], Generalized Low Rank Approximation of Matrices (GLRAM) [23], Two Dimensional Singular Value Decomposition (2DSVD) [6], rank-R tensors [19], dynamic and streaming tensor factorization [14]. Substantial efforts have also been devoted to improve the vector-based classical LDA. Ye et al. [24] introduced two dimensional LDA (2DLDA) with an iterative solution algorithm, while paper [12] solved the ambiguity problem in 2DLDA. Both of them consider the projection of the data onto a space which is the tensor product of two vector spaces. Yan et al. [20] used tensor representation in discriminative analysis. In previous research, HOSVD usually is used for data compression and classification. In paper [9], the authors pointed out HOSVD simultaneously does subspace selection and tensor clustering. In computer vision applications, more discriminative subspaces usually are desired to be selected to provide better clustering, classification, or retrieval tasks. In this paper, we introduce discriminative analysis into tensor decomposition to propose a novel Discriminative HOSVD that achieves better clustering/classification/retrieval results. First, an Adaptive 2DLDA method is proposed to find the more discriminative subspaces U and V . In our Adaptive 2DLDA method, the K-means clustering and 2DLDA algorithms are iteratively used to find class labels and to obtain subspaces U and V unsupervisedly. After that, we calculate matrix W with improved clustering capability based on our theoretical proof. We emphasize that the Discriminative HOSVD is still an

Tensor based dimensionality reduction has recently attracted attention from computer vision and pattern recognition communities for both feature extraction and data compression. As an unsupervised method, High-Order Singular Value Decomposition (HOSVD) searches for low-rank subspaces such that the low-rank approximation error is minimized. The data projection on subspace can be used as features for different computer vision tasks such as image clustering, classification, and retrieval. However, without the class labels, the discriminative power of selected features is restricted. In this paper, we propose a new unsupervised high-order tensor decomposition approach which employs the strength of discriminative analysis and K-means clustering to adaptively select subspaces that improve the clustering, classification, and retrieval capabilities of HOSVD. We provide both theoretical analysis to guarantee that our new method generates more discriminative subspaces and empirical studies on several public computer vision data sets to show the consistent improvement of image clustering, classification, and content-based image retrieval results over existing methods.

1. Introduction Subspace learning methods such as Principal Component Analysis (PCA), Singular Value Decomposition (SVD), and Linear Discriminant Analysis (LDA) were popular as tools for the analysis of two-dimensional arrays of data in a wide variety of computer vision applications, such as previous research in papers [10, 15, 2, 17]. As subspace analysis approaches, PCA and SVD employ unsupervised ways to learn the low-dimensional structure of high-dimensional data, but LDA is a supervised method and seeks for a set of vectors that maximized Fisher Discriminant Criterion. LDA simultaneously minimizes the withinclass scatter while maximizing the between-class scatter in 1

unsupervised learning method. Therefore, it can relieve human effort of classifying large amount data points. In our extensive experiments on four datasets, our Discriminative HOSVD outperforms the standard unsupervised learning methods (PCA, HOSVD, and GLRAM) in clustering, classification, and retrieval accuracy. Although many studies on tensor factorization appeared in computer vision areas, to our knowledge, this paper is the first one to propose and prove the feasibility of discriminative analysis, a supervised learning idea, in unsupervised tensor factorization. Our novel Adaptive 2DLDA and Discriminative HOSVD approaches have broad applications in high dimensional image clustering, classification, and image retrieval. In previous research, there are several papers ever explored the relations between unsupervised dimension reduction and unsupervised learning [3, 25, 4, 5]. But all previous methods only work for vector based input data, not for high dimensional data (e.g. images and videos). The main contributions of this paper are the following: (a) We show that 2DLDA and tensor clustering are optimizing the same objective function, i.e., they both minimize the within-class scatter and maximize the betweenclass scatter. (b) Based on the above theoretical analysis, we show that the objective function for our proposed Adaptive 2DLDA provides a natural generalization which combines the strength of 2DLDA and K-means tensor clustering to select more discriminative subspaces U and V that provide better tensor classification results. (c) A new Discriminative HOSVD is proposed to improve the clustering capability of standard HOSVD, i.e. a better clustering structure exists in W . We also prove Discriminative HOSVD is equivalent to tensor clustering with subspaces selected by Adaptive 2DLDA.

2. Related Work In computer vision areas, HOSVD has been widely used for dimensionality reduction and data compression. In previous work, researchers showed the HOSVD can do simultaneous subspace selection and K-means clustering [9], i.e., the clustering structure exists in matrix W of HOSVD in (2). In this section, we will introduce the related previous research work, including HOSVD and tensor clustering.

2.1. High Order SVD HOSVD searches low-rank matrices and a core tensor to approximate high order tensors. Given a 3D tensor X = {X1 , · · · , Xn3 } where each Xi is a 2D matrix of size n1 ×

n2 , the standard HOSVD factorizations is to solve: min

U,V,W,S

J1 = ||X − U ⊗1 V ⊗2 W ⊗3 S||2

(1)

s.t. U T U = I, V T V = I, W T W = I where U, V, W are 2D matrices and S is a 3D core tensor: S ∈
2.2. Tensor Clustering Given a three dimensional tensor X, or, equivalently a set of two dimensional images X1 , X2 , · · · , Xn , the Kmeans tensor clustering minimizes min

{Ck }

n X

min ||X` −Ck ||2 =

1≤k≤K

`=1

K X X

k=1 `∈Ck

||X` −Ck ||2 (2)

where Ck is the centroid tensor of cluster Ck . After running HOSVD on tensor X, we obtain U and V . We can get: Yi = U T Xi V. (3) Using the distance relationship, the tensor clustering can be equivalently done in {Y` }: min JKmeans (Y ) =

{Ck }

K X X

k=1 `∈Ck

||Y` − CkY ||2

(4)

where CkY = U T Ck V .

3. Discriminative High Order SVD We propose a new tensor factorization method, Discriminative HOSVD, which adaptively selects the subspaces U and V in an unsupervised way so as to improve the clustering capability of W , and the clustering and classification capability of embedding space projection by U , V . In order to select more discriminative subspaces U and V , we first develop the Adaptive 2DLDA (A2DLDA) method in §3.2, which select subspaces by explicitly combining 2DLDA and K-means clustering in a coherent way. After selecting discriminative subspaces U and V , we can solve matrix W for Discriminative HOSVD. Compared to the standard HOSVD, our Discriminative HOSVD can find better subspaces for further supervised and unsupervised learning tasks.

3.1. Discriminative Subspace Selection Suppose performing HOSVD on tensor X and obtaining U , V . Our goal is to find U and V such that the clustering structure of data become more apparent in the subspace as (3). Thus, after the subspace projection, the distances of data points within the same class should be minimized and the distances of data points between classes should be maximized. The intuition here is to introduce LDA-like idea into

unsupervised subspace selection. We should notice the data points here have high dimensions and are not vectors used in LDA. One way to achieve the goal is to maximize the between class scatter: maxU,V Tr SbY , and minimize the within class Y scatter: minU,V Tr Sw , or another alternative way is to use objective function as: max U,V

Tr SbY , Y Tr Sw

(5)

where scatter matrices in projection subspaces of (3) as SbY =

n X j=1

Y Sw =

nj CjY − C Y

n X X



 Y

Y i − Cj

j=1 i∈Πj

T

,

 Y T

,

CjY − C Y Y i − Cj

(6)

3.2. Adaptive 2DLDA (7)

where Yi is the subspace projection result of the data point Xi by U and V as (3), CjY is the centroid of cluster j, nj is the total number of data points in cluster j, and C Y is the centroid of the whole dataset. We use (5) as the objective function to select the discriminative subspaces U and V . The advantage of using (5) is to be revealed by Lemma 1. Lemma 1 Tensor clustering in (4) is equivalent to minY imize the within-class scatter matrix Sw or maximize the between-class scatter matrix SbY in projection subspaces. Proof: Y Tr Sw

=

=



Tr 

n X X



j=1 i∈Πj

n X X

Tr



LDA is widely used to select the subspace which has the maximal discriminant power. However, LDA is a supervised learning method which requires we know the class label for each data point as prior knowledge. Since 2DLDA and K-means (tensor) clustering both Y minimize Sw and maximize SbY , we propose the following main algorithmic approach to optimize the objective function (5) and select discriminative subspaces U and V in an unsupervised way: (A1) Tensor clustering in (4) provides class labels for data points; (A2) With these labels, (5) can be solved as 2DLDA. Iteratively repeating these two steps is the main approach of our new Adaptive 2DLDA framework as following.

U T (Xi − Cj ) V V T (Xi − Cj )T U 

U T (Xi − Cj ) V V T (Xi − Cj )T U



The principle of Adaptive 2DLDA is to do tensor clustering and 2DLDA iteratively. Once we have the K-means clustering results, we obtain the clustering labels for each data point. Thus, we can perform 2DLDA on the original data space to derive more discriminative subspaces according to the clustering labels. With the discriminative subspace, we can find better clustering labels. Till converge, we will get discriminative subspaces U and V . We use 3D Tensor decomposition as example to show how our Adaptive 2DLDA framework works. They can be easily generalized to higher dimension tensors. The goal of our Adaptive 2DLDA method is to optimize (5) in an unsupervised way. Because Tr SbY = Tr U T SbV U = Tr V T SbU V and Y V U Tr Sw = Tr U T Sw U = Tr V T Sw V,

(5) can be written as:

j=1 i∈Πj

=

n X

=

n X X

X

max

(8)

where

j=1 i∈Πj

kYi − CjY k2

j=1 i∈Πj

=

U,V

kU T Xi V − U T Cj V k2

Tr U T SbV U Tr V T SbU V or max , VU UV U,V Tr V T Sw Tr U T Sw

SbV

=

JKmeans (Y )

Therefore, K-means (tensor) clustering minimizes the Y in projection subspaces, or within-class scatter matrix Sw maximizes the between-class scatter matrix SbY since the total scatter matrix is a constant. u – From Lemma 1, we know LDA like (actually it is 2DLDA and please see §3.2) objective function (5) has very similar properties as K-means (tensor) clustering: minimizY ing within-class scatter Sw and/or maximizing betweenY class scatter Sb .

V Sw

=

n X

j=1 n X

nj (Cj − C)V V T (Cj − C)T X

j=1 Xi ∈Πj

SbU

=

n X

=

n X X

j=1

U Sw

(Xi − Cj )V V T (Xi − Cj )T

nj (Cj − C)T U U T (Cj − C)

j=1 Xi ∈Πj

(Xi − Cj )T U U T (Xi − Cj ).

The labels for {Xi } can be found by running K-means clustering on {Yi }. Using these labels, similar to standard LDA

3.3. Discriminative HOSVD Algorithm by

W

(9)

where Hkk0 =

X

Xijk Xi0 j 0 k0 (U U T )ii0 (V V T )jj 0 .

DHOSVD Iteration 1

0 −0.05 −0.1 −0.1

Second Component

0.1 0.05 0 −0.05 −0.1 −0.2

0 First Component

0.2

0.05 0 −0.05 −0.1 −0.2

0 0.1 First Component DHOSVD Iteration 3

Second Component

0.05

DHOSVD Iteration 2

0.1

0.15 0.1 0.05 0 −0.05 −0.2

0 First Component

0.2

0.3 0.2 0.1 0 −0.1 −0.1

0 0.2 First Component DHOSVD Iteration 4 Second Component

Second Component

0.1

0 0.1 First Component DHOSVD Iteration 5

0.1 0.05 0 −0.05 −0.1 −0.2

0 First Component

0.2

(a) 40 images are projected onto subspaces

(b) Images used in (a)

After getting U and V from § 3.1, we can obtain W [11] max Tr(W T HW ) s.t. W T W = I,

HOSVD Second Component

Algorithm 1 Adaptive 2DLDA Algorithm Input X = {X1 , · · · , Xn3 }, parameters k1 , k2 , iteration number t Initialize: Compute U and V as eigenvectors of PN PN F = i=1 Xi XiT and G = i=1 XiT Xi , Yi = U T Xi V, i = 1, 2, ..., n3 . Do a) Run K-means on {Yi } to obtain class label for each Xi b) Seek discriminative subspaces U and V using class labels of {Xi } and repeating t iterations: V −1 V ) Sb i) Calculate U as first k1 eigenvectors of (Sw U −1 U ) Sb ii) Calculate V as first k2 eigenvectors of (Sw Until convergence Output Cluster indicator Q as class labels, and discriminative subspaces U, V .

Algorithm 2 Discriminative HOSVD Algorithm Input X = {X1 , · · · , Xn3 }, Parameters k1 , k2 , k3 Do 1. Set (U, V ) as the results of Adaptive 2DLDA (see §3.1). 2. Solve W using (10) where H is calculated using U and V from the results of Adaptive 2DLDA. 3. S = U T ⊗1 V T ⊗2 W T ⊗3 X. Output U, V, W, S.

Second Component

solutions, (8) can be solved as a generalized eigenvalue problem. U can be calculated as the first k1 eigenvectors V −1 V (Sw ) Sb , and V can be calculated as the first k2 eigenU −1 U ) Sb . vectors of (Sw We summarize the Adaptive 2DLDA algorithm in Alg. 1.

(10)

Figure 1. Comparison of Discriminative HOSVD and HOSVD. Both methods are performed on 40 randomly selected images for four people from Yale dataset. The two coordinates are (uT1 Xi v1 , uT1 Xi v2 ), where u1 , u2 are the first two columns of U and v1 , v2 are first two columns of V . Images from the same subject use the same color.

ii0 jj 0

The solution W of (9) including eigenvectors of matrix H. These U , V , and W are the final results of Discriminative HOSVD. Our Discriminative HOSVD algorithm is described in Alg. 2. The subspaces U and V are adaptively selected. With the discriminative U and V , the clustering capability of matrix W is also improved accordingly. We theoretically prove that in Theorem 1 in next section.

All four subjects in HOSVD result mix to each other and the within-class distances are large. The clustering result is poor. But DHOSVD can unsupervisedly select good discriminative subspaces for these subjects to improve the clustering results. After two iterations, the within-class distances are already small. After five iterations, the betweenclass distances are increased substantially. Therefore, results shown in Fig. 1 demonstrate the discriminative power of DHOSVD subspaces.

Fig. 1 illustrates the other example to demonstrate the discriminative power of DHOSVD by clustering application. The real world Yale image dataset is used (detailed experimental settings can be found in § 4). We randomly select 4 subjects with 40 images totally. (U , V ) are obtained from HOSVD and DHOSVD. In each figure, the data points are plotted using two components (there is no difference for two or three components and we randomly select this number): uT1 Xi v1 and uT1 Xi v2 , where u1 , u2 are the first two columns of U and v1 , v2 are first two columns of V .

3.5. Analysis of Discriminative HOSVD

3.4. Demonstration of Discriminative HOSVD

In Discriminative HOSVD, U and V are calculated by Adaptive 2DLDA algorithm and improve the clustering capability of data points after subspace projection. In following Theorem 1, we will prove Discriminative HOSVD is equivalent to tensor clustering with subspaces U and V , and the clustering structure exists in matrix W . Because U and V improve the tensor clustering, the clustering capability of matrix W is also improved.

Theorem 1 Discriminative HOSVD is equivalent to tensor clustering of (5) with U and V selected by Adaptive 2DLDA and the clustering structure exists in matrix W . Proof: In Lemma 1, we have proved Y Tr(Sw )=

n X X

j=1 i∈Πj

n3 X

kYi − CjY k2

4. Experimental Results

n X 1 X kYi k − = Tr(YiT Yi0 ) n k 0 j=1 i=1 2

i,i ∈Cj

where CjY is centroid the jth cluster in the U, V subspace. We introduce the clustering indicator matrix Q = (q1 , · · · , qK ),

qTi qj = δij .

(11)

where nj

z }| { √ (12) qj = (0, · · · , 0, 1, · · · , 1, 0, · · · , 0)T / nj P n3 kYi k2 is constant, It is obvious that QT Q = I. Since i=1 Y minimizing Tr(Sw ) is equivalent to solve max

n X 1 n j=1 j

X

Tr(YiT Yi0 ) = Tr QT XQ

i,i0 ∈CjY



(13)

where Xii0

= Tr(YiT Yi ) = Tr(V T XiT U U T Xi0 V ),

(14)

and T

(QQ )``0 =



0 1/nj

if if

Y` or Y`0 6∈ CjY Y` and Y`0 ∈ CjY

(15)

Obviously, (14) is the same as (10). As a result, (13) s.t. QT Q = I is equivalent to the objective function to solve matrix W in HOSVD decomposition, also in our Discriminate HOSVD: max Tr(W T HW ) s.t. W T W = I. W

clustering results from W in DHOSVD is better than standard one in HOSVD. We have proved that in Theorem 1 and will also demonstrate that by empirical experiments in §4. In §4, the tensor clustering results based on U and V will also be further discussed. If we use U and V in clustering, DHOSVD has the same results as Adaptive 2DLDA. When we use DHOSVD, we usually emphasize W ; when we use Adaptive 2DLDA, we emphasize U and V .

(16)

Thus, solution W in (16) is exact the same as the solution Q in (13). u –

3.6. DHOSVD vs Adaptive 2DLDA Although in [9], the authors proved the clustering capability of HOSVD. In fact, they only focus on matrix W at that time. Here we clarify that U and V can be also used for clustering via embedding space projection as (3). Especially after selecting discriminative subspaces U and V , the

In order to evaluate the discriminative power of our method, we perform clustering and classification experiments on four commonly used datasets: AT&T [13], Binary Alphadigits (BinAlpha) [1], UMIST [8], and YaleB [7]. We also validate the quality of the subspaces using a contentbased image retrieval problem on the People Playing Musical Instrument (PPMI) data set [22]. In the benchmark AT&T face database, images from 40 distinct persons were taken at different times, different facial expression, and facial details. We use the whole AT&T dataset and resize all images (400 images from 40 persons, each of which has 10 images in similar illumination conditions) from 112 × 92 to 28 × 23. The second dataset is Binary Alphadigits corpus. This is a small vocabulary task which is fairly challenging. The corpus has a vocabulary of 36 characters: 26 letters (from ‘A’ to ‘Z’) and 10 digits (from ‘0’ to ‘9’). For each word, we have 39 images (20 × 16). In our experiments, we use the whole dataset and the original size. UMIST dataset includes multi-view images of human faces and consists of 564 images of 20 individuals (mixed race, gender, and appearance). Each individual is shown in a range of poses from profile to frontal views. The image resolution is 220 × 220 and we resize them to 28×23. YaleB face images database used in our experiment is the combination of extended and original Yale database B [7]. We resize the images from 192 × 168 to 24 × 21. Because there is a set of images which are corrupted during the image acquisition [7], we have 31 subjects in total without any corrupted images. We randomly select ten illumination conditions for all 31 subjects to create the experimental data set with 310 images. The PPMI dataset contains images of humans interacting with twelve different musical instruments, bassoon, cello, clarinet, erhu, flute, French horn, guitar, harp, recorder, saxophone, trumpet, and violin. In order to compare the interpretability and discriminability of DHOSVD and HOSVD, we first use all four data sets to evaluate the clustering structures existing in matrix W of DHOSVD and HOSVD. After that, we use two data sets to show the performance of discriminative subspaces U and V selected by A2DLDA. Both clustering and classification applications are used. As we discussed in §3.6, DHOSVD is the same as A2DLDA if

we only use U and V to do projection and clustering.

0.9

Our goal of proposing Discriminative HOSVD is to improve the clustering capability of HOSVD. Thus, we compare the clustering performance (clustering accuracy and normalized mutual information) between our DHOSVD and standard HOSVD. Four standard image data sets are employed. Both HOSVD and DHOSVD are performed on data set to get matrix W and K-means clustering method is utilized to find the clustering structure in W . We use two standard metrics to measure the quality of clustering results: clustering accuracy and normalized mutual information. The clustering accuracy is defined by the ratio of the number of images that are clustered into the default subject clusters and the number of total images. Because clustering is an unsupervised method, the subject label is not identical to the cluster label. For each cluster in clustering results, we find the largest group of images from the same subject and use their label for this cluster. If two clusters are marked with the same label, for the cluster whose largest group of images has a smaller number, we change its label to the second largest group of images. We iteratively label the clusters, till convergence. The Normalized Mutual Information (NMI) is calculated by: M I(C, C 0 ) , NMI = max(H(C), H(C 0 )) where C is a set of clusters obtained from the true labels and C 0 is a set of clusters obtained from the clustering algorithm. M I(C, C 0 ) is the mutual information metric, and H(C) and H(C 0 ) are the entropies of C and C 0 , respectively. NMI is between 0 and 1. A larger NMI value indicates a better performance. The clustering results using DHOSVD and HOSVD are compared in Fig. 2. Our new DHOSVD provides better clustering capability than standard HOSVD in all four data sets. Thus, DHOSVD decomposition results are more discriminative.

4.2. Clustering Performance Evaluation of Adaptive 2DLDA In § 3.6, we discussed there are two ways to compare the clustering results of DHOSVD and HOSVD. In Fig. 2, the clustering results using W of both DHOSVD and HOSVD have been shown. Now we emphasize U and V to compare the clustering results. In this way, DHOSVD has the same results as Adaptive 2DLDA. In Fig. 3, we compare Adaptive 2DLDA to four other unsupervised methods: HOSVD + K-means, GLRAM [23] + K-means, PCA + K-means, and K-means. Our method has

0.9 HOSVD DHOSVD

0.8

0.7

0.7

0.6

0.6

0.5

0.5

NMI

4.1. Clustering Capability Enhancement in Matrix W of DHOSVD

Clustering Accuracy

0.8

0.4

0.4

0.3

0.3

0.2

0.2

0.1 0

HOSVD DHOSVD

0.1

UMIST

AT&T

BinAlpha

YaleB

0

UMIST

AT&T

BinAlpha

YaleB

Figure 2. Clustering accuracy (left) and Normalized Mutual Information (NMI, right) comparison between Discriminative HOSVD and HOSVD. The x-axis shows four data sets and the values on yaxis are the clustering accuracy or NMI. More details can be found in §4.1

better clustering results than others in terms of both clustering accuracy and normalized mutual information. For Adaptive 2DLDA, HOSVD, and GLRAM, we first project image Xi into their U and V subspaces as (3). We then run K-means on {Yi }. For PCA+K-means, we always set the reduced dimension Kpca = k1 k2 and perform K-means on the Kpca principle components subspace. For K-means, we do K-means on the original space treating an image as a one line vector. The clustering accuracy is calculated by the way described in §4.1. Since K-means clustering is used in all 5 methods with different subspaces, we are unable to start K-means clustering in all methods with the same centroid. In order to generate more reliable comparison results, we start K-means using the same random partition of data points for all methods in each round. For each method, we run 30 rounds. In each round, 30 times K-means are performed and the result with minimum K-means clustering objective function error is selected (totally 900 K-means are computed for each method). We use the average results of all 30 rounds as the final results which are show in Fig. 3.

4.3. Classification Performance Evaluation of A2DLDA Since Adaptive 2DLDA selects more discriminative subspaces, we are interested whether it will help classification. We compare subspaces selected by Adaptive 2DLDA to subspaces selected by PCA, GLRAM, and HOSVD, and the original space. In all cases, we use nearest neighbor classifier (1-NN) and 5-fold cross validation scheme. For A2DLDA, GLRAM, HOSVD, the distance between 2-D objects can be interpreted as a bilinear metric: d(Xi , Xj ) = kXi − Xj kU,V = kU T Xi V − U T Xj V k (17) We can use this metric to perform 1-NN classification. For PCA, the distance is the Euclidean distance of the principle component subspace. For the original space, d(i, j) =

5. Conclusions

0.78 0.91 0.9

A2DLDA K−means PCA+K−means GLRAM+K−means HOSVD+K−means

0.74

0.72

A2DLDA K−means PCA+K−means GLRAM+K−means HOSVD+K−means

0.89 0.88 NMI

Accuracy

0.76

0.87 0.86 0.85

0.7

0.84

0.68

2

4

6 # iteration

8

0.83

10

2

4

6 # iteration

8

10

(a) AT&T 0.46 0.61

0.45

0.6 0.59

0.43

0.58

A2DLDA K−means PCA+K−means GLRAM+K−means HOSVD+K−means

0.42 0.41 0.4

NMI

Accuracy

0.44

0.56 0.55

0.39 0.38

A2DLDA K−means PCA+K−means GLRAM+K−means HOSVD+K−means

0.57

0.54

2

4

6 # iteration

8

0.53

10

2

4

6 # iteration

8

10

(b) BinAlpha

Figure 3. Clustering accuracy (left) and normalized mutual information (NMI, right) comparison on AT&T (a) and BinAlpha (b) datasets. The compared methods include K-means, PCA+Kmeans, GLRAM+K-means, HOSVD+K-means, and A2DLDA.

0.985

0.725 0.72 A2DLDA Original PCA GLRAM HOSVD

0.715 Accuracy

Accuracy

0.98

0.975

0.71

References

A2DLDA Original PCA GLRAM HOSVD

0.705 0.7 0.695 0.69

0.97 2

4

6 # iteration

(a) AT&T

8

In this paper, we present a new tensor approach using adaptive subspace learning, in which high order objects are discriminatively separatively. We first proved K-means tensor clustering and 2DLDA use the same objective functions to minimize the within-class scatter or/and maximize the between-class scatter. Based on our theoretical analysis, we proposed an adaptive 2DLDA framework to iteratively do K-means clustering and select discriminative subspaces. Combining the strength of K-means clustering and 2DLDA, we unsupervisedly selected more discriminative subspaces U and V . Moreover, we proved Discriminative HOSVD is equivalent to tensor clustering with subspaces U and V selected by Adaptive 2DLDA. Because U and V are more discriminative and improve the tensor clustering, the clustering capability of matrix W in DHOSVD is enhanced. Experimental results on clustering, classification, and image retrieval indicate that the quality of the subspaces selected by our methods is much higher than all the other methods. Acknowledgment This research is partially supported by NSF-CCF-0830780, NSF-DMS-0915228, NSF-CCF0917274.

10

0.685

2

4

6 # iteration

8

10

(b) BinAlpha

[1] Algoval system. http://algoval.essex.ac.uk/ also can be downloaded from http://www.cs.toronto.edu/ roweis/data.html. 5

Figure 4. Classification accuracy comparison on original space and subspace of Adaptive 2DLDA, PCA, GLRAM, and HOSVD.

[2] P. Belhumeur, J. Hespanha, and D. Kriengman. Eigenfaces vs fisherfaces: recognition using class specific linear projection. IEEE Trans. Pattern Analysis and Machine Intelligence, 19(7):711–720, 1997. 1

kXi − Xj k, where Xi is the original image. The average classification accuracy of the 5-fold cross validation is plotted in Fig 4.

[3] C. Ding and X. He. K-means clustering via principal component analysis. Proc. of Int’l Conf. Machine Learning, pages 225–232, 2004. 2

The classification results shown in Fig 4 indicate A2DLDA outperforms all other methods.

[4] C. Ding, X. He, H. Zha, and H. Simon. Adaptive dimension reduction for clustering high dimensional data. ICDM, pages 147–154, 2002. 2

4.4. Content-Based Image Retrieval

[5] C. Ding and T. Li. Adaptive dimension reduction using discriminant analysis and k-means clustering. Proc. Int’l Conf. on Machine Learning (ICML), pages 521–528, 2007. 2

Since A2DLDA learns a bilinear space metric, we can also utilize this metric in image retrieval. We use 4 metrics: Euclidean, d(Xi , Xj ) = kxi − p xj k, Mahalanobis metric: d(Xi , Xj ) = kxi −xj kP −1 = (xi − xj )T P −1 (xi − xj ), where P is the covariance matrix, PCA: kxi − xj kU = kU xi − U xj k, and our method using (17). Here xi and xj are the vectorization of Xi , Xj , respectively. We show an query example and the top 10 retrieved images for each method in Figure 5. The query image is a woman playing violin. One can visually conclude that our method retrieval much more reasonable images.

[6] C. Ding and J. Ye. Two-dimensional singular value decomposition (2dsvd) for 2d maps and images. SIAM Int’l Conf. Data Mining, pages 32–43, 2005. 1 [7] A. Georghiades, P. Belhumeur, and D. Kriegman. From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intelligence, 23(6):643– 660, 2001. 5

Our Method PCA Mahalanobis Euclidean

Figure 5. One query example for the four distance matric measurements: Euclidean , Mahalanobis, PCA, and our method. For each method, the first image is the query image and the following 10 images are the top 10 retrieved images using the corresponding method. Our method makes two errors in the top 10 retrieval images: the 4-th and 10-th in which the person is playing guitar. This error is reasonable since the shapes of violin and guitar are similar.

[8] D. B. Graham and N. M. Allinson. Face recognition: From theory to applications. NATO ASI Series F, Computer and Systems Sciences, 163:446–456, 1998. 5 [9] H. Huang, C. Ding, D. Luo, and T. Li. Simultaneous tensor subspace selection and clustering: The equivalence of high order svd and k-means clustering. The 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 327– 335, 2008. 1, 2, 5 [10] Kirby and Sirovich. Application of the kl procedure for the characterization of human faces. IEEE Trans. Pattern Anal. Machine Intell., 12:103–108, 1990. 1 [11] L. D. Lathauwer, B. D. Moor, and J. Vandewalle. On the best rank-1 and rank-(r1, r2, . . . , rn) approximation of higher-order tensors. SIAM J. Matrix Anal. Appl., 21:1324–1342, 2000. 1, 4 [12] D. Luo, C. Ding, and H. Huang. Symmetric two dimensional linear discriminant analysis (2dlda). 2009. 1 [13] F. Samaria and A. Harter. Parameterisation of a stochastic model for human face identification. Proceedings of 2nd IEEE Workshop on Applications of Computer Vision, 1994. 5 [14] J. Sun, D. Tao, and C. Faloutsos. Beyond streams and graphs: Dynamic tensor analysis. Proceedings of ACM KDD, pages 374–383, 2006. 1 [15] D. Swets and J. Weng. Using discriminant eigenfeatures for image retrieval. IEEE Trans. Pattern Analysis and Machine Intelligence, 18(8):831–836, 1996. 1 [16] L. Tucker. Some mathematical notes on three-mode factor analysis. Psychometrika, 31(3):279–311, 1966. 1 [17] M. A. Turk and A. P. Pentland. Face recognition using eigenfaces. CVPR, 1991. 1

[18] M. Vasilescu and D. Terzopoulos. Multilinear analysis of image ensembles: Tensorfaces. European Conf. on Computer Vision, pages 447–460, 2002. 1 [19] H. Wang and N. Ahuja. A tensor approximation approach to dimensionality reduction. International Journal of Computer Vision, pages 1573–1405, 2007. 1 [20] S. Yan, D. Xu, Q. Yang, L. Zhang, X. Tang, and H.-J. Zhang. Discriminant analysis with tensor representation. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), pages 526–532, 2005. 1 [21] J. Yang, D. Zhang, A. F. Frangi, and J. Yang. Twodimensional pca: A new approach to appearancebased face representation and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 26(1), 2004. 1 [22] B. Yao and L. Fei-Fei. Grouplet: A structured image representation for recognizing human and object interactions. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, USA, June 2010. 5 [23] J. Ye. Generalized low rank approximations of matrices. International Conference on Machine Learning, 2004. 1, 6 [24] J. Ye, R. Janardan, and Q. Li. Two-dimensional linear discriminant analysis. Advances in Neural Information Processing Systems (NIPS 2004), 17:1569–1576, 2004. 1 [25] H. Zha, C. Ding, M. Gu, X. He, and H. Simon. Spectral relaxation for k-means clustering. Neural Information Processing Systems, 14:1057–1064, 2001. 2

Discriminative High Order SVD: Adaptive Tensor ...

for Image Classification, Clustering, and Retrieval ... sets to show the consistent improvement of image cluster- ...... http://www.cs.toronto.edu/ roweis/data.html. 5.

448KB Sizes 1 Downloads 107 Views

Recommend Documents

A High Order Periodic Adaptive Learning Compensator ...
IEEE Trans. on Ind. Electron., vol. 51, pp. 526–536, 2004. ... 2006, pp. 457–462. [14] K. L. Moore and YangQuan Chen, “A separative high-order framework.

A High Order Periodic Adaptive Learning Compensator ...
trical and Computer Engineering department, Utah State University, Logan, UT ... ¶[email protected]; Dept. of Automation Science and Engineering, South.

Dual-high-order Periodic Adaptive Learning ...
stability of the equilibrium points ex(t), ev(t) and ea(t), as t → ∞ (s → ∞), with the initial condition, exi(t) = evi(t) = eai(t)=0, as (t − ∑i j=1 Pk+1−j) ≤ 0, where.

Fractional Order Adaptive Compensation for ... - Semantic Scholar
ing the FO-AC is much smaller than that using the IO-AC. Furthermore, although the ... IEEE Trans. on Ind. Electron., 51:526 – 536, 2004. D. Y. Xue, C. N. Zhao, ...

Adaptive Algorithms Versus Higher Order ... - Semantic Scholar
sponse of these channels blindly except that the input exci- tation is non-Gaussian, with the low calculation cost, com- pared with the adaptive algorithms exploiting the informa- tion of input and output for the impulse response channel estimation.

Fractional Order Adaptive Compensation for ... - Semantic Scholar
1. J. µ + B1)Vd(s). −. µs1−ν. J vd(s)+(. µ. Js. + 1)vd(0). (36). Denote that ν = p q. , sν = s p q , ..... minimization. IEEE Trans. on Ind. Electron., 51:526 – 536, 2004.

High Order to Trade Ratio (OTR). - NSE
4 days ago - Sub: High Order to Trade Ratio (OTR). This has reference to SEBI Circular No. SEBI/HO/MRD/DP/CIR/P/2018/62 dated April 09, 2018.

High-order synchronization, transitions, and ...
May 7, 2008 - Department of Physics, Lancaster University, Lancaster LA1 4YB, United Kingdom .... “constant force,” thanks to which a rotator has the features of a self-sustained oscillator) and the first harmonic. By means of a translation in th

High Order to Trade Ratio (OTR). - NSE
4 days ago - not be permitted to place any orders for the first 15 minutes on the next trading ... securities in SME, ETF and securities with designated market.

diffusion tensor imaging
T2-EP image, and this facilitates visualization of the data by providing ..... FIGURE 27.8. Comparison of diffusion tensor imaging in the brainstem (medulla oblongata) of ... ant for, or has spoken on behalf of the following companies within the ...

Diffusion tensor imaging
... that MR imaging may be a powerful tool to assess developmental changes that predict .... ki А Dav р. Ю2 r ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi k2. 1 ю k2. 2 ю k2. 3 q. ¼ ..... imaging of the brain in v

Adaptive Control for a Discrete-time First-order ...
Adaptive Control for a Discrete-time First-order Nonlinear System with ... of non-parametric part is characterized by a Lipschitz constant. L, and the nonlinearity of ...

Adaptive penalized quantile regression for high ...
For example, there is no guarantee that K-fold cross-validation would provide a choice of ln with a proper rate. Third, their statistical properties are still uncharted ...

Adaptive finite elements with high aspect ratio for ...
of grid points adaptive isotropic finite elements [12,13] have been used. Further reduction of ...... The initial solid grain is a circle of. Fig. 15. Dendritic growth of ...

Adaptive finite elements with high aspect ratio for ...
Institut d'Analyse et Calcul Scientifique, Ecole Polytechnique Fйdйrale de Lausanne, 1015 Lausanne, ..... degrees of freedom, the triangles may have large aspect ..... computation of solidification microstructures using dynamic data structures ...

Adaptive Logarithmic Mapping For Displaying High ...
Categories and Subject Descriptors (according to ACM CCS): I.3.3 [Image Processing and Computer Vision]: Image. Representation. 1. Introduction. Mainstream ...

An Adaptive Protocol Stack for High-Dependability based on ... - EWSN
In Wiselib 802.15.4, pack- ets are limited to 116Bytes and as a result, it may include a maximum of 37 neighbors. If we need to operate on a larger neighborhood we can use the Wiselib Fragmenting Radio and transmit beacons larger than a single messag

Adaptive Finite Elements with High Aspect Ratio for ... - Springer Link
An adaptive phase field model for the solidification of binary alloys in two space dimensions is .... c kρsφ + ρl(1 − φ). ( ρv + (k − 1)ρsφvs. )) − div. (. D(φ)∇c + ˜D(c, φ)∇φ. ) = 0, (8) where we have set .... ena during solidif

Adaptive Logarithmic Mapping For Displaying High ...
Our adaptive logarithmic mapping technique is capable of producing perceptually tuned images with ... Mainstream imaging and rendering software are now ad-.

Energy momentum tensor
In [Joot()], the electrodynamic energy density U and momentum flux density ... Additionally the energy and momentum flux densities are components of.

IMPROVED DISCRIMINATIVE TRAINING ...
Techniques for improving lattice-based Maximum Mu- ... 2. MMIE OBJECTIVE FUNCTION. MMIE training was proposed in [1] as an alternative to .... This stage.

In search of an SVD and QRcp Based Optimization ... - CiteSeerX
Therefore, an optimum design of neural network is needed towards real-time ... (corresponding author to provide phone: +91-3222-283556/1470; fax: +91-. 3222-255303 .... or not in a Digital Signal Processor based system. We use that.