1218

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

Non-negative Patch Alignment Framework Naiyang Guan, Dacheng Tao, Member, IEEE, Zhigang Luo, and Bo Yuan

Abstract— In this paper, we present a non-negative patch alignment framework (NPAF) to unify popular non-negative matrix factorization (NMF) related dimension reduction algorithms. It offers a new viewpoint to better understand the common property of different NMF algorithms. Although multiplicative update rule (MUR) can solve NPAF and is easy to implement, it converges slowly. Thus, we propose a fast gradient descent (FGD) to overcome the aforementioned problem. FGD uses the Newton method to search the optimal step size, and thus converges faster than MUR. Experiments on synthetic and real-world datasets confirm the efficiency of FGD compared with MUR for optimizing NPAF. Based on NPAF, we develop non-negative discriminative locality alignment (NDLA). Experiments on face image and handwritten datasets suggest the effectiveness of NDLA in classification tasks and its robustness to image occlusions, compared with representative NMF-related dimension reduction algorithms. Index Terms— Image occlusion, non-negative matrix factorization, patch alignment framework.

I. I NTRODUCTION

D

IMENSION reduction plays an important role in pattern recognition. However, the learned bases by traditional algorithms, e.g., principal component analysis (PCA) [1] and Fisher’s linear discriminant analysis (FLDA) [2] are inconsistent with the psychological evidence of parts-based representation in human brain. In practice, they cannot perform robustly on noised data. Recently, non-negative matrix factorization (NMF) [3] was proposed as a new dimension reduction algorithm. Since the learned bases by NMF have psychological and physiological intuition of combing parts to form a whole in human brain, it has been widely applied to pattern recognition and other applications. In recent years, different NMF-related dimension reduction algorithms [4]–[6] have been proposed for practical applications. By adding penalties to NMF, Li et al. [6] presented the local NMF (LNMF) that learns spatially localized parts-based representation for visual patterns and applied to face recognition. To encode the data geometric structure in

Manuscript received September 12, 2010; revised January 5, 2011; accepted May 12, 2011. Date of publication June 30, 2011; date of current version August 3, 2011. This work was supported in part by the National Natural Science Foundation of China under Grant 91024030/G03. N. Guan and Z. Luo are with the School of Computer Science, National University of Defense Technology, Changsha 410073 China (e-mail: ny [email protected]; [email protected]). D. Tao is with the Center for Quantum Computation and Intelligent Systems, Faculty of Engineering and Information Technology, University of Technology Sydney, New South Wales 2007, Australia (e-mail: [email protected]). B. Yuan is with the Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai 200240 China (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TNN.2011.2157359

a nearest neighbor graph, Cai et al. [5] proposed the graph regularized NMF (GNMF), which was successfully applied to image clustering. Since LNMF and GNMF completely ignore the discriminative information, they cannot perform well in classification tasks. To incorporate the discriminative information, Zafeiriou et al. [4] developed the discriminant NMF (DNMF), which combines Fisher’s criterion with NMF for frontal face verification. However, the aforementioned NMF-related algorithms were designed according to specific intuitions and developed on the basis of the knowledge of field experts for their own purposes. Therefore, a general framework is helpful to better understand the common properties and intrinsic differences of these algorithms. In this paper, we propose the non-negative patch alignment framework (NPAF) to unify various NMF-related dimension reduction algorithms. It builds patches for each sample, forms one local coordinate for such patch, and aligns all the local coordinates to form the global coordinate of all the samples in the nonnegative subspace. To solve NPAF, we develope a multiplicative update rule (MUR). It is easy to implement and can be applied to solving existing NMF-related dimension reduction algorithms. By introducing a new auxiliary function, we prove that the MUR converges. However, the MUR converges slowly and thus it is difficult to apply the algorithm in practice. We thus propose the fast gradient descent (FGD) to overcome the slow convergence problem. FGD uses a Newton method to search the optimal step size for one factor with another fixed in each iteration round. Preliminary experiments on synthetic and real-world datasets confirm the efficiency of FGD. In general, the NPAF shows that: 1) various NMF related algorithms are intrinsically different in patches that they build; 2) all the algorithms share almost the same whole alignment procedure; and 3) their solutions can be obtained by using the MUR or FGD according to different alignment matrices. As an application of the framework, we show a new NMF-related dimension reduction algorithm, termed “non-negative discriminative locality alignment” (NDLA). In particular, we build the within-class local patch to preserve the data local geometric structure in learning the subspace as well as the between-class local patch on which the margins between different classes are maximized, and thus improve its performance in classification tasks. Finally, we apply NDLA to face recognition on several popular face image datasets under different partial occlusions. By using the 5 × 2 cv F-test [7], we statistically compare the performance of NDLA with three NMF-related algorithms and three other traditional dimension reduction algorithms. The experimental results suggest the effectiveness of NDLA in classification tasks and its robustness to image occlusion, compared with NMF and its variants.

1045–9227/$26.00 © 2011 IEEE

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1219

Base 1

z

II. NPAF Patch alignment framework [8] unifies popular dimension reduction algorithms, e.g., FLDA [2] and its extensions [9], [10], locally linear embedding (LLE) [11], ISOMap [12], and locality preserving projections (LPP) [13]. It contains two stages: part optimization and whole alignment, and can be applied to designing dimension reduction algorithms with specific objectives. Part optimization: For a given sample vi in a dataset, based on the labeling information, we can categorize the other samples in this dataset into two groups: 1) samples sharing the same class label with vi ; and 2) samples taking different labels with vi . Each sample vi associates with a patch v i , vi 1 , . . . , vi k1 , vi1 , . . . , vik2 ], wherein vi , vi 1 , . . . , vi k1 , Vi = [ i.e., the k1 nearest samples of vi , are from the same class as vi , and vi1 , . . . , vik2 , i.e., the other k2 nearest samples of vi , are from different classes against vi . For each patch Vi , the corresponding low-dimensional representation is denoted by Hi = [hi , hi 1 , . . . , hi k1 , hi1 , . . . , hik2 ]. In this local patch, specific statistical properties, e.g., discrimination and local geometry, can be encoded. For example, the discriminative locality alignment (DLA) [14] encodes the discriminative information over the local patch Hi by keeping the distances between hi and its k1 nearest samples (from the same class as hi ) as small as possible and the distances between hi and its k2 nearest samples (from different classes against hi ) as large as possible. The part optimization over the patch Hi is min Hi tr (Hi L i HiT ), where tr (·) is the trace operator, and L i varies with different dimension reduction algorithms to encode the discriminative information and the local geometry of the patch. Whole alignment: Each patch Hi has its own coordinate system and all Hi s can be unified together as a whole one by assuming that the coordinate of the i th patch Hi is selected from the global coordinate H = [h1 , h2 , . . . , hl ], i.e., Hi = H Si , where Si ∈ Rl×(k1 +k2 +1) is the selection matrix and an entry is defined by  1, if p = Fi (q) (Si ) pq = 0, otherwise where Fi = {i, i 1 , . . . , i k1 , i 1 , ¡, i k2 } denotes indices for the i th patch built by vi and its related samples. The alignment strategy [15] is adopted to build the global coordinate for all patches

y

y

min H

i=1

tr (Hi L i HiT ) = min H

l 

tr (H Si L i SiT H T )

i=1

= min tr (H L H T ) H

(1)

 where L = li=1 Si L i SiT . For linearization, H = W T V is usually considered, where W is the projection matrix. We can impose different constraints, e.g., or W T W = I , to uniquely determine H . Under this constraint and H = W T V , the solution of (1) can be obtained by using the conventional Lagrangian multiplier method [16] or the generalized eigenvalue decomposition.

Base 2

0 local patch

x

0

x (a)

(b)

Fig. 1. Example of NPAF. (a) Nonnegative samples in the 3-D space. (b) Embedded 2-D representation.

A. Definition of NPAF In this paper, we expect both the bases and the coordinate to be nonnegative, which is inspired by the intuition that the whole is formed by combining parts, and thus yields parts-based representation. Such nonnegativity constraints on both the bases and the coordinates are desirable in many applications where the underlying components have a physical interpretation, e.g., face components in face recognition experiments. By incorporating the nonnegativity constraints on both the bases and the coordinate for manifold learning based dimension reduction, we arrive at the definition of NPAF. Given n nonnegative samples in R m that are arranged in matrix V ∈ R m×n , NPAF projects them to Rr , wherein r is the reduced dimension. Fig. 1 gives an example of NPAF. The 3-D nonnegative samples [see Fig. 1(a)] are projected onto the 2-D space [see Fig. 1(b)]. The local coordinates are aligned with the global coordinate that is spanned by Base 1 and Base 2. By incorporating the nonnegativity constraint, we obtain the objective of NPAF min

W ≥0,H ≥0

γ tr (H L H T ) + D(V, W H ) 2

(2)

where W ∈ R m×r signifies the bases vectors, H ∈ Rr×n refers to the coordinate, and D(V, W H ) is the error between samples V and its approximation W H . The error can be measured by the Kullback–Leibler divergence (KLD)  v i j log K L(V, W H ) = ij

l 

global coordinate

 vi j − v i j + (W H )i j . (W H )i j

It can be replaced by the Frobenius matrix norm [17]. We use γ as the tradeoff parameter and (1/2) over γ to simplify subsequent derivations. In the following section, we will show that different algorithms build different patches, and thus have different alignment matrices L.

B. MUR for NPAF Note that (2) is nonconvex on both W and H , so it is unrealistic to find the global optimal. Considering its second

1220

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

derivatives with respect to wab and h i j

Equations (4) and (5) lead to the following MUR: 

γ (H L − )i j + l vl j wli w h  k lk kj hi j ← hi j + γ (H L )i j + k wki

 v al h bl

l w kh l k a k . wab ← wab h k bk

 2v al h 2bl ∂2 F  = 2 ( k wak h kl )2 ∂wab l

 2vl j wli2 ∂2 F  = + γljj ( k wlk h kj )2 ∂h 2i j l

where F(W, H ) = (γ /2)tr (H L H T ) + K L(V, W H ) is the 2 ) > 0 and objective function. We can find that (∂ 2 F/∂wab 2 2 (∂ F/∂h i j ) > 0 if l j j ≥ 0. Thus, (2) is convex with respect to wab or h i j , and an iterative optimization can be used to learn W and H . To solve the constrained optimization problem (2), we introduce Lagrangian multipliers ϕ ∈ Rm×r and φ ∈ Rr×n for constraints W ≥ 0 and H ≥ 0, respectively. The Lagrangian function is L=

γ tr (H L H T ) + K L(V, W H ) + tr (ϕ T W ) + tr (φ T H ). 2

Then the problem (2) is equal to min W,H L. To minimize L, we obtain its partial derivatives with respect to wab and h i j  v al h bl  ∂L  =− + h bk + ϕab , ∂wab k wak h kl l k  vl j wli  ∂L  = γ (H L)i j − + wik + φi j . ∂h i j k wlk h kj l

k

By using the KKT conditions [16], i.e., ϕab wab = 0 and φi j h i j = 0, we obtain the following equalities:  γ (H L)i j −

 l

 −

 l

  vl j wli  + wik h i j = 0, k wlk h kj k   + h bk wab = 0.

v al h bl  k wak h kl

k

By separating L into two parts, i.e., L = some algebra, (3) is equivalent to  +

γ (H L )i j +

 k

+ =



vl j wli  k wlk h kj

l  l

L+ −

L −,

and using



 wik h 2i j = γ (H L −)i j 

v al h bl  k wak h kl

h 2i j ,

 

(4)

 2 h bk wab

k

 2 wab

(7)

We can iteratively update W and H until the objective value does not change. Since V can be approximated by W H columnwise, i.e., v j ≈ W h j , we can project a sample x from the original high-dimensional space to the low-dimensional space, i.e., y = W † x, wherein the projection matrix W † = (W T W )−1 W T is the pseudoinverse of W . By introducing a new auxiliary function, we prove the convergence of MUR. Definition 1 (Auxiliary Function): The function G(x, x  ) is an auxiliary function for F(x), if G(x, x  ) ≥ F(x) and G(x  , x  ) = F(x  ). Proposition 1: If G(x, x  ) is an auxiliary function of F(x), then F(x) is nonincreasing under the update x = arg min G(x, x  ). x

Proof: F(x) ≤ G(x, x  ) ≤ G(x  , x  ) = F(x  ) We prove that the update (6) with W fixed and the update (7) with H fixed do not increase the objective function F(W, H ) by introducing two auxiliary functions and using Proposition 1 in the following proposition. Proposition 2: The objective function F(W, H ) is nonincreasing by using the update (6) with W fixed and using the update (7) with H fixed. The detailed proof is given in [17], wherein an inequality, shown in Proposition 3, plays an important role. r×n and Proposition 3: For any positive matrices H ∈ R+ r×n n×n  , H ∈ R+ , and symmetric nonnegative matrix A ∈ R+ the following inequality holds:  (H  A)i j h 2i j

(3)

(6)

i, j

h i j

≥ tr (H AH T ) 

hi j . h i j i, j The left-hand side inequality comes from [18]. The detailed proof is given in [17]. The time complexity of MUR is No.o f i ter ati on × O(mnr + n 2r ), wherein No.o f i ter ati on is the iteration number and O(mnr + n 2r ) is the time complexity of one iteration round of MUR. ≥ tr (H  AH T ) + 2

(H  A)i j h i j log

C. FGD for NPAF (5)

where both L + and L − are nonnegative symmetric matrices. It is easy to separate the matrix L, e.g., we can separate L into − the positive and negative parts as L + i j = (|L i j | + L i j )/2, L i j = + − (|L i j | − L i j )/2, where both L and L are nonnegative symmetric matrices because L itself is a symmetric matrix.

Although MUR can optimize NPAF, it converges slowly. That is because MUR is actually a first-order method. Some methods have applied a Newton-type method to optimize the Frobenius norm-based NMF [19], but it is unsuitable for solving NPAF (2) because there is no closed-form representation for the Hessian matrix. In our previous work [20], we proposed an FGD method to accelerate MUR. Here we name it as “old FGD” (or OFGD for short). Particularly,

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1221

OFGD uses the Newton method at each iteration round to search the optimal step size along the direction of scaled negative gradient. However, OFGD searches a single step size for the whole factor (W or H ), so it will make the step size close to 1 for some problems (see Fig. 3), i.e., OFGD almost degenerates to MUR for these problems. Here, we propose a new efficient FGD method to optimize NPAF. It assigns a step size for each column of the matrix factor and searches the optimal step size vector in each iteration round. Since the objective of FGD is convex, we apply the multivariate Newton method to optimize its objective function. Although the inverse of Hessian matrix in the multivariate Newton method is time consuming, we introduce the special structure of the Hessian matrix to efficiently approximate its inverse, and thus FGD efficiently searches the optimal step size vector without increasing the time cost compared with OFGD. Therefore, FGD rapidly reduces the objective function at each iteration round. Without loss of generality, we take the procedure of updating H as an example, and W can be updated with a similar procedure by setting γ = 0. The overall optimization is alternatively updating H with W fixed, and updating W with H fixed. At the kth iteration round, the scaled negative gradient can be written in matrix form as   T V T + − + W E + γ Hk L − γ Hk L (8) ∇ = −η ⊗ −W W Hk where E ∈ Rm×n is the matrix whose entries are all 1, ⊗ is the elementwise product operator, and η the scaling factor. By setting η = Hk /(γ Hk L + + W T E) and substituting into (8), we have ∇ = Hk ⊗

γ Hk L − + W T γ Hk L + +

V W Hk WT E

− Hk .

(9)

In this paper, we set a step size θ j for each h j , j = 1, . . . , n, i.e., the j th column of Hk . To make H nonnegative, θ j should  j ≥ 0, θ j > 0}, be selected in the range of D j = {θ j |h j + θ j ∇  j is the j th column of ∇ = [∇  1, . . . , ∇  n ]. Then the where ∇  where new value of H should be Hk+1 = Hk + ∇ × di ag(θ), θ = [θ1 , . . . , θn ]T . We use the multivariate Newton method to search the optimal step size θ. The problem can be written as min φ(θ) = F(W, Hk+1 ). θ

(10)

T  = tr (L Hk+1 μ(θ) Hk+1 ) = tr (L HkT Hk )

+ 2θ1 tr (L HkT ∇ 1 ) + · · · + 2θn tr (L HkT ∇ n )  T θi θ j tr (L∇ i ∇ j ). + i, j

Then the first-order and second-order sub-derivatives of  with respect to θj are μ(θ)  ∂μ T = 2tr (L HkT ∇ j ) + 2 θi tr (L∇ i ∇ j ) (12) ∂θ j i

and

∂ 2μ T = 2tr (L∇ i ∇ j ). ∂θ j ∂θi

(13)

Equations (12) and (13) can be rewritten in matrix forms as  ∂μ = 2tr ((Hk L)T ∇ j ) + 2 θi L i j (∇ T ∇)i j ∂θ j i  j +2 = 2bTj ∇ θi L i j (∇ T ∇)i j i

and

∂ 2μ T = 2tr (L∇ i ∇ j ) = 2L i j (∇ T ∇)i j ∂θ j ∂θi

(14)

where b j is the j t h column of Hk L = B  [b1 , . . . , bn ]. By combining (11), (13), and (14) together, we can obtain the Hessian matrix of φ(θ) as H essi an(φ) = A + H sn

(15)

where H sn = γ L ⊗(∇ T ∇), and A is a diagonal matrix whose j th entry is defined by ajj =

 j )2 vl j (W ∇ l

 l

 j )l )2 ((W h j )l + θ j (W ∇

> 0.

(16)

Since ∇ T ∇ is positive semidefinite, according to the Schur product theorem [21], H sn is positive semidefinite if L is positive semidefinite. Recall that A is positive definite, and thus the Hessian matrix H essi an(φ) is positive definite. Equation (10) is convex and thus the multivariate Newton method can be applied to search the optimal step size. The Newton update rule is given by θ(t +1) = θ(t ) − H essi an(φ)−1∇φ (θ(t ) )

According to (2), we obtain the first-order and second-order sub-derivative of φ(θ) with respect to θj , which are   j )l vl j (W ∇ ∂φ γ ∂μ   j )l − = + (W ∇  j )l ∂θ j 2 ∂θ j (W h j )l + θ j (W ∇ l

and

 j , . . . , 0 ], j = 1, . . . , n. With such a trick, we rewrite [0, . . . , ∇  μ(θ) into a polynomial with respect to θ j , j = 1, . . . , n as

l

⎧ ⎨γ ∂ 2φ 2 = ⎩γ ∂θ j ∂θi

  j )2 v l j (W ∇ ∂2μ l + l  j )l )2 ∂θ 2j ((W h j )l +θ j (W ∇ ∂2μ 2 ∂θ j ∂θi , i = j

(11)

T ). To calculate (11), we rewrite where μ(θ) = tr (Hk+1 L Hk+1 Hk+1 as Hk+1 = Hk + θ1 ∇ 1 + · · · + θn ∇ n , where ∇ j =

(17)

where ∇φ (θ(t )) is the sub-gradient of φ at θ(t ). Note that the matrix inverse operation in (17) with time complexity O(n 3 ) is time consuming. Since the Hessian of φ(θ) is composed of H sn, which is independent of θ and A whose inverse can be efficiently calculated, we introduce an approximation scheme by using the Sherman–Morrison– Woodbury formula [22] to obtain H essi an(φ)−1 whose complexity is O( p3 ) with p n H essi an(φ)−1 = (A + H sn)−1 = (A + U V T )−1 



≈ (A + U  V 

−A−1 U (

T

 −1

)−1 = A−1 +V

T



A−1U )−1 V

T

A−1

1222

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

where H sn = U V T is the singular value decomposition (SVD) of H sn which can be calculated in the previous   T iteration round, and U  V is an approximation of U V T , which can be obtained by selecting first p the most  p the 2 / n δ 2 ≤ 95%, δ important components such that i=1 i i=1 i wherein δi2 is the i th largest singular value. The update (17) stops and obtains the optimal step size θ∗ when |θ(t +1) − θ(t )| ≤ tol, wherein t ≥ 0 is the iteration counter which is empirically small (usually t < 10), and tol is a predefined small value, e.g., 10−4 used in all experiments. The initial step size is set as θ(0) = 1 due to 1 ∈ D j , in which case FGD degenerates to Hk+1 = Hk ⊗

γ Hk L − + W T γ Hk

L+

+

V W Hk WT E

m×n m×r r×n Input: V ∈ R+ , L, Wk ∈ R+ , Hk ∈ R+ r×n Output: Hk+1 ∈ R+

1: ∇ = Hk ⊗

V − Wk Hk +γ Hk L WkT E+γ Hk L +

WkT

U U T

− Hk .

⊗ (∇ T ∇).   T

2: SVD: =γL 3: Select p components: U  U ≈ U U T . 4: Initialize θ(0) = 1 , t = 0. repeat 5.1: Calculate A−1 according to (16). 5.2: Calculate Hessian inverse: H essi an(φ)−1 = A−1 − A−1U  (  T

where 0 < λ < 1 is used to ensure that Hk+1 is not too close to the boundary. Note that boundary for the j th column is  j )i |(∇  j )i < 0}, where xi signifies sup(D j ) = max{(h j )i /(∇ the i th entry of x , and sup(D j ) > 1, because 1 ∈ D j . Thus set the parameter λ as λ = 0.99 × min j {(sup(D j ) − 1)/(θ ∗j − 1)|θ ∗j > sup(D j )}. By using (19), we obtain the new step size, and then H is updated by (20)

The FGD procedure is summarized in Algorithm 1. Since the objective function (2) is nonincreasing under the update rule (18), which is proved in Lemma 1, it is also nonincreasing under the update rule (20), which is shown in Proposition 4. Therefore, according to Proposition 1, the FGD for NPAF converges. Lemma 1: The objective function (2) is nonincreasing under the rule of (18). Proposition 4: The objective function (2) is nonincreasing under the rule of (20). The proofs of Lemma 1 and Proposition 4 are given in [17]. By setting the parameter γ = 0 and considering V ≈ W H , the same procedure can be applied to optimize W . By updating H and W alternatively, we can optimize a particular implementation of NPAF, e.g., NDLA. FGD is efficient although it searches the optimal step size for each column separately, because it can be implemented with the matrix form. The main time cost of Algorithm 1 is on Statements 1, 2, and 5, whose complexities are O(mnr + n 2r ), O(n 3 ), and O(mn + p3 ), respectively. Thus the total complexity of Algorithm 1 is O(mnr +n 2 r +n 3 )+k × O(mn+ p3 ), where k is the iteration number of the Newton method in FGD (see Statement 5 in Algorithm 1). Since the iteration number k is small, usually k r , p n, the time cost of one iteration round of FGD is comparable to that of one iteration round of MUR O(mnr + n 2r ), especially when mr ≤ n 2 . However, FGD converges much faster than MUR, and thus the

−1

+ U  A−1U  )−1 U  A−1 .

(18)

where Hk+1 is obviously nonnegative matrix. The final step size is chosen such that the new factor value, i.e., Hk+1 is as close as possible to the minimum along the ∇ direction without exceeding the positive quadrant. Formally, we have θk+1 = λθ∗ + (1 − λ)1 (19)

Hk+1 = Hk + ∇ × di ag(θk+1 ).

Algorithm 1 FGD optimization for NPAF

T

5.3: Update θ(t +1): θ(t +1) = θ(t ) − H essi an(φ)−1∇φ (θ(t )). 5.4: t = t + 1. until Stopping criteria is met (h )  j )i < 0}, 1 ≤ j ≤ n. 6: sup(D j ) = max{  j i |(∇ (∇ j )i sup(D j )−1 (t +1) |θ j (t+1) θ −1

7: λ = 0.99 × min j {

> sup(D j )}.

j

8: θk+1 = λθ(t +1) + (1 − λ)1 . 9: Hk+1 = Hk + ∇ × di ag(θk+1 ).

overall time cost of FGD is much smaller than that of MUR. We will empirically show the efficiency of FGD by comparing it with MUR. Actually, both MUR and FGD cannot guarantee convergence to a local minimum because it is difficult to prove the existence of stationary point, whereas such stationarity is a necessary condition for converging to a local minimum. To overcome this deficiency, Lin [23] modified the multiplicative algorithms to guarantee stationarity. This is theoretically significant. We really appreciate this contribution but this modification brings in additional computations and cannot boost the performance of subsequent classification. Thus, we do not introduce this modification to our methods, and experimental results show that the performance of our methods is quite acceptable. D. Related Works This section compares NPAF with NGE and GNMF, and compares FGD with OFGD and PG. NDLA versus NGE: Recently, Yang et al. [24] proposed a unified framework termed “nonnegative graph embedding” (NGE) to understand NMF-related dimension reduction algorithms. NGE is intrinsically different from NPAF because it understands NMF-related algorithms based on different perspectives. In particular: 1) NGE reveals the underlying differences between algorithms by the design of their intrinsic and penalty graphs and their types of embeddings, and 2) NGE is optimized by MUR that is inefficient because the matrix inverse operator at each iteration round is time consuming. Although another MUR-based algorithm was further proposed

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1223

in [25] for optimizing NGE, it converges slowly compared to the proposed FGD. NDLA versus GNMF: Cai et al. [5] expected GNMF as a general framework that stems from both NMF and graph Laplacian. Although other information, e.g., label of data [26], can be used to construct the graph in GNMF, NPAF is different from GNMF in the following two aspects: 1) NPAF understands NMF-related algorithms on the viewpoint of patch alignment framework, whereas GNMF is based on the graph Laplacian, and 2) the proposed FGD converges much faster than MUR and can also solve all NMF-related algorithms under NPAF. We also show that NPAF includes GNMF as a special case. FGD versus OFGD: FGD is different from OFGD [20] and much faster than OFGD [17]. FGD versus PG: Recently, projected gradient (PG) [27] has been introduced to optimize NMF. FGD is different from PG in terms of the following two aspects: 1) PG cannot optimize NPAF because the projection operator may bring many zeros in the matrix factors and thus makes KLD in (2) ill defined, and 2) in each iteration round, PG minimizes one matrix factor (W or H ) with another one fixed, but FGD searches one step along the rescaled gradient direction with optimal step size.



where U = W T W , and U = H H T . In LNMF, three regularizations are incorporatedwith the bases and the coefficient matrix: 1) minimizing i = j Ui j makes  the bases approxthe imately orthogonal; 2) minimizing i Uii suppresses   overdecomposition of the bases W ; and 3) maximizing i Uii encourages retaining   components withimportant information. Considering i Uii = tr (H H T ) and i, j Ui j = tr (W eW T ), (23) is equivalent to     (24) min D(V, W H ) − βtr H H T + αtr W eW T W ≥0,H ≥0

where e is square matrix whose elements are all 1. To solve LNMF defined in (23), we optimize   (25) min D(V, W H ) − βtr H H T H ≥0

with fixed W and optimize min D(V, W H ) + αtr (W eW T )

W ≥0

with fixed H iteratively. This procedure can be understood under NPAF. In particular, (25) is equivalent to min D(V, W H ) + βtr (H (−I )H T ) H ≥0

III. U NIFYING NMF-R ELATED D IMENSION R EDUCTION A LGORITHMS This section unifies the various NMF-related dimension reduction algorithms under NPAF. By using NPAF, we can exploit their common property and their intrinsic differences. A. NMF From the perspective of dimension reduction, NMF [3] learns to represent original samples V = [ v 1 , . . . , vn ] as a m×r linear combination of low-dimensional bases W ∈ R+ (r r×n m), and the coefficient matrix H ∈ R+ , V ≈ W H , and both W and H are nonnegative. The objective of NMF is min

W ≥0,H ≥0

D(V, W H )

(21)

where D(V, W H ) is the distance between samples and their respective approximations. Note that (21) is equivalent to (2) by setting γ = 0 in (2). Equation (21) can be optimized by using MUR and FGD (see Section II) with γ = 0, and the MUR for (21) is a relaxed version of the algorithm in [3] V HT WT V , W ← W ⊗ . (22) WT W H W H HT MUR for (21) converges slowly compared to (22), but it is directly derived from the NPAF, i.e., NMF can be treated as a special case of NPAF. H←H⊗

B. LNMF To learn spatially localized parts-based representation, Li et al. [6] proposed the LNMF    Ui j − β Uii (23) min D(V, W H ) + α W ≥0,H ≥0

i, j

i

(26)

(27)

where I is identity matrix. Since (27) is equivalent to (2) with L replaced by −I , (25) can be unified by NPAF, and the patch built for any sample in LNMF degenerates to itself. It ignores the geometric structure information. Since −I is a diagonal matrix, (25) can be solved by the proposed MUR (see Section II-C) as  β H + W T WVH . H←H⊗ WT E Note that the parameter β should be set to a sufficiently small value to guarantee the convexity of the problem in (27). Similarly, (26) is equivalent to (2) with L replaced by e and thus (26) of LNMF can be unified by NPAF. The patch is built by itself and the remaining bases. Since e is symmetric, (26) can be solved by the proposed MUR (see Section II-C) as  V T WH H W ←W⊗ . αW e + E H T In summary, LNMF can be unified in NPAF by building two parallel patches on the columns of H and W , respectively. C. GNMF To encode the data geometric structure, Cai et al. [5] proposed the GNMF min

W ≥0,H ≥0

D(V, W H ) + λtr (H L H T )

(28)

where L is the graph Laplacian matrix. In GNMF, the data geometry is encoded in an adjacent graph, where each vertex corresponding to a sample and the weight between vertex vi and vertex v j is defined as  v j ) or v j ∈ Nk ( vi ) 1, if vi ∈ Nk ( Si j = 0, otherwise

1224

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

where Nk ( v i ) signifies the set of k nearest neighbors of vi . Then L is written as L = T − S, where T is a diagonal matrixwhose diagonal entries are column sums of S, i.e., Tii = j S j i . According to [5], rewrite the objective function (28) as  min D(V, W H ) + λ ||hi − h j ||2 Si j W ≥0,H ≥0

= =

min

W ≥0,H ≥0

min

W ≥0,H ≥0

i, j

D(V, W H ) + λ

k  i

D(V, W H ) + λ



||hi − hi j ||2

j =1

tr (Hi L i HiT )

(29)

i

where hi j , j = 1, . . . , k, are the low-dimensional coordinate of k connected samples of the given sample vi in the graph,  T   k −1 kT  −1 k , where and L i = di ag(1 k )[−1 k Ik ] = Ik −1 k diag(1 k ) Ik ∈ Rk×k is an identity matrix and 1 k = [1, . . . , 1]T ∈ Rk denotes a k-dimensional vector whose entries are all 1. This notation is used in the following. Equation (29) serves as the whole alignment for all the patches. For each sample vi , we have the part optimization: min Hi tr (Hi L i HiT ), and thus the patch for vi is built by itself and its k nearest neighbors in Nk ( v i ). With the alignment strategy in [15], we can construct the alignment matrix L g which is equal to the graph Laplacian matrix L. Therefore, GNMF can be unified by NPAF. Finally, (28) can be solved by MUR and FGD, because L is a symmetric and positive semidefinite matrix. However, both GNMF and LNMF ignore the discriminative information, and thus cannot perform well in classification tasks. D. DNMF Zafeiriou et al. [4] proposed the DNMF, which combines the Fisher’s criterion with NMF min

W ≥0,H ≥0

D(V, W H ) + γ SW − δS B

(30)

where SW and S B are the with-in class and between-class scatter, respectively. To unify DNMF into NPAF, we rewrite (30) as min

W ≥0,H ≥0

which is equivalent to (2) with L replaced by L W −(cδ/γ )L B . Therefore, DNMF can be unified by NPAF, and (32) can be optimized by MUR and FGD because L W and L B are both symmetric and positive semidefinite matrices. Note that the parameter (δ/γ ) should be set to a sufficiently small value to guarantee the convexity of the problem in (32). From (31), we can see that two patches should be built for vi in DNMF. One is built by vi itself and the rest in the same class Ci , the other is built by the centroid vim of Ci and the rest centroids of different classes. Since the patches it builds are global and include all samples, DNMF cannot encode the local geometry in learning.

D(V, W H ) + γ tr (H L W H T ) − δtr (H L B H T )

where L W and L B are alignment matrices for min H SW and max H S B . The alignment matrices can be obtained by using the alignment strategy in [15], wherein L iW and L iB are   (Ni − 1)2 −(Ni − 1)1 TNi −1 1 W Li = 2 Ni −(Ni − 1)1 Ni −1 1 Ni −1 1 TNi −1   T Ni (C − 1)2 −(C − 1)1 C−1 B (31) Li = 2 T C −(C − 1)1 C−1 1 C−1 1 C−1 where Ni is the number of samples that is in the same class with vi , and C is the class number. Then (30) becomes   δ B T W (32) min D(V, W H ) + γ tr H (L − L )H W ≥0,H ≥0 γ

IV. NDLA It has been experimentally proved that DLA [14] is an effective method for visual recognition. Therefore, we introduce the underlying strategy used in DLA to NPAF and obtain the NDLA that preserves the data local geometric structure and the discriminative information. Given a dataset V = [ v 1 , . . . , vn ], where vi ∈ Rm . For a sample vi , according to its label information, we divide the whole dataset into two parts V s and V d , where V s is composed of samples in the same class as vi , and V d is composed of samples in the different classes against vi . Then we build two types of local patch: 1) with-in class patch, v i , v1w , . . . , vkw1 ], containing denoted by the matrix Viw = [ itself and its k1 nearest neighbors in V s , and 2) betweenv i , v1b , . . . , vkb2 ], class patch, denoted by the matrix Vib = [ containing vi and its k2 nearest neighbors in V d . To preserve the data local geometric structure, we expect the samples in the same class to be as close as possible in the low-dimensional space, and thus obtain the part optimization onthe within-class patch min Hiw tr (Hiw L iw Hiw T ), where k1 w w T i j =1 (1i ) j −(1i ) Lw = . The set of indices, for vi , on −1 w diag(1 w i i ) the within-class patch is Fiw = {i, i 1 , . . . , i k1 }. To make samples in different classes separable, we obtain the part optimization on the between-class patch  k2  b ) j −(1 b )T ( 1 T b i b i i j =1 i . min H b tr (Hi L b Hi ), where L b = i −1 bi diag(1 bi ) The set of indices, for vi , on the between-class patch is Fib = {i, i 1 , . . . , i k2 }. By using the whole alignment (see Section II-A), we come up with the following two objective functions: min

l 

H

max H

    tr Hi L iw HiT = min tr H L w H T

i=1 l  i=1

H

    tr Hi L ib HiT = max tr H L b H T H

(33)

(34)

  T T where L w = li=1 Swi L iw Swi and L b = li=1 Sbi L ib Sbi are the alignment matrices of within-class patch and between-class patch, respectively, and Swi ∈ R n×(k1 +1) and Sbi ∈ R n×(k2 +1) are selection matrices for the with-in class patch and the

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1225

TABLE I C OMPARISON OF NMF-FGD WITH NMF-MUR AND NMF-OFGD ON R ANDOM D ENSE M ATRICES (b): 128 × 16 × 32 MUR OFGD FGD 0.409 0.409 0.409 383 55 48 1.060 0.343 0.343

between-class patch of the sample vi , and their entries are   1, if p = Fiw (q) 1, if j = Fib (k) (Swi ) pq = . , (Sbi ) j k = 0, otherwise 0, otherwise Noth that both L w and L b are symmetric and positive semidefinite. By combining (33) and (34), we arrive at −1

H

Since L b is unnecessarily positive definite, it could be noninvertible. According to [28], we consider adding a tiny perturbation to the diagonal of the alignment matrix, i.e., L˜ b = L b + ζ I , to make it invertible. It has been shown that the solution obtained by the perturbed alignment matrix is consistent with the original one as long as ζ is fixed to a small number, so here we empirically set ζ as 10−4 tr (L b ). In the rest of this paper, L b implies the perturbed matrix L˜ b . Through above analysis, we have the objective γ tr (H L H T ) + K L(V, W H ) W ≥0,H ≥0 2 −1/2

1

(35)

−1/2

where L = (L b )T L w L b . It is clear that L j j ≥ 0 due to L w being positive semidefinite, and thus (35) is convex with respect to either wab or h i j . Equation (35) can be solved by MUR (see Section II-C) −1/2 −1/2 with L + and L − replaced by D = (L b )T Dw L b and −1/2 −1/2 S = (L b )T Sw L b , which are obtained by separating L w into two parts L w = Dw − Sw . In order to penalize the nonsmoothness of H , we add a tiny perturbation to the diagonal of D, i.e., D˜ = D + ζ I , to impose the Tiknohov regularization over H , here we empirically set ζ = 10−4 . According to Section II-D, the MUR for NDLA, which we called NDLA-MUR, converges because both D and S are nonnegative symmetric matrices, which is presented in Proposition 5. However as mentioned in Section II, NDLAMUR converges slowly, thus we solve (35) with FGD, which we called NDLA-FGD. Proposition 5: Both D and S are nonnegative symmetric matrices. The proof is given in [17]. Both NDLA-MUR and NDLA-FGD stop at the k + 1th iteration if the objective function satisfies |F(Wk , Hk ) − F(W∗ , H∗ )| ≤τ |F(W1 , H1 ) − F(W∗ , H∗ )| where k ≥ 0 is the iteration counter, τ signifies the tolerance of precision which is usually a small value, e.g., τ = 10−4 , and (W1 , H1 ) is the initial point, (W∗ , H∗ ) signifies the final solution. In practice, it is impossible to know (W∗ , H∗ ) previously, so we use (Wk+1 , Hk+1 ) instead.

(d): 2048 × 128 × 256 MUR OFGD FGD 0.530 0.530 0.530 895 173 105 500.498 251.504 102.711 1

NMF-MUR NMF-OFGD NMF-FGD

0.9 0.8 0.7 0.6

NDLA-MUR NDLA-OFGD NDLA-FGD

0.8 0.6 0.4 0.2

0.5 0.4

−1

min tr (H (L b2 )T L w L b2 H T ).

min

(c): 2,048 × 32 × 256 MUR OFGD FGD 0.846 0.846 0.846 759 138 86 198.838 96.205 58.625

f(k)/f(1)

(a): 128 × 8 × 32 MUR OFGD FGD 0.643 0.643 0.643 411 160 83 0.889 0.811 0.546

f(k)/f(1)

Problem Algorithm f k / f1 N o.o f iteration CPU seconds

200

400 600 iteration (a)

800

0

200

400 600 iteration (b)

800 1000

Fig. 2. Objective values versus iteration numbers for MUR, OFGD, and FGD when solving (a) NMF and (b) NDLA.

The time complexity of one iteration round of NDLAMUR and NDLA-FGD is O(mnr + n 2r ) and O(mnr + n 2r + n 3 ) + k × O(mn + p3 ), respectively (Section II-C and II-D). We implement both two algorithms in MATLAB and replace Statements 2 and 3 with function ‘svds’, which further reduces the time cost of NDLA-FGD. Note that another time cost of NDLA is spent on the process of constructing the within-class and between-class patches, which includes two main parts: 1) calculating the Euclidean distance between every two data samples for constructing L w and L b , whose complexity is O(n 2 m), and 2) the SVD of L˜ b and matrix multiplication for constructing D and S, whose complexity is O(n 3 ). V. E XPERIMENTS In this section, we study the computational efficiency of MUR and FGD, and evaluate the proposed NDLA on several real-life datasets. Section V-A studies the efficiency of MUR and FGD for optimizing NPAF based on NMF and NDLA. Section V-B evaluates the effectiveness and robustness of NDLA by comparing with six representative algorithms, which are PCA [1], FLDA [2], DLA [14], NMF [3], LNMF [6], and DNMF [4], under different partial occlusions on two popular face image datasets, i.e., Oracle Research Laboratory (ORL) [29] and University of Manchester Institute of Science and Technology (UMIST) [30], and the popular handwritten dataset Mixing National Institute of Standard and Technology (MNIST) [31]. A. Study of MUR Versus FGD We compare FGD with MUR and OFGD in terms of efficiency by applying them to NMF and NDLA. The naming convention is given by NMF-MUR/NMF-OFGD/NMF-FGD and NDLA-MUR/NDLA-OFGD/NDLA-FGD. All six methods were conducted on their respective evaluation datasets with the same initial point. Fig. 2 shows the objective values versus iteration numbers for MUR, OFGD, and FGD when

1226

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

TABLE II C OMPARISON OF NDLA-FGD WITH NDLA-MUR, NDLA-OFGD ON UMIST AND MNIST D ATASETS

14 12

(a): UMIST (r = 50) MUR OFGD FGD 0.131 0.131 0.131 655 166 98 175.594 134.301 110.807

OFGD FGD

10 8 6 4 2 0

20 40 60 80 100 120 iteration (a)

Step sizing for optimizing H

Step sizing for optimizing W

Problem Algorithm f k / f1 N o.o f iteration CPU seconds

14 12

(b): UMIST (r = 200) MUR OFGD FGD 0.041 0.041 0.041 1000 168 121 480.389 192.224 162.225

OFGD FGD

10 8 6 4 2 0

20 40 60 80 100 120 iteration (b)

Fig. 3. Step size for optimizing (a) W and (b) H versus iteration number on the MNIST dataset with the subspace dimension set to 300. For FGD, the average step sizes for rows of W and columns of H are given.

solving NMF and NDLA on a 2048 × 256-dimension matrix and 1600 × 300-dimension matrix, respectively. The subspace dimensions for NMF and NDLA are 128 and 200, respectively. Fig. 2 shows that both OFGD and FGD reduce the objective function much faster than MUR in each iteration round. By comparing the OFGD and FGD curves in Fig. 2, we can see that FGD reduces the objective function faster than OFGD in each iteration round. To further evaluate NMF-FGD by comparing with NMFMUR and NMF-OFGD, Table I shows their objective values, iteration numbers, and CPU seconds on 128 × 32-dimension random dense matrix, and 2048×256-dimension random dense matrix, respectively. Table I shows that NMF-FGD uses less iteration number and CPU seconds to optimize the objective function than NMF-MUR. By comparing columns (c) with (d) of Table I, we can see that NMF-FGD becomes more efficient than NMF-MUR with the increasing of subspace dimension r . That is because, in one iteration round, the time complexity of FGD is t × O(mn + p3 ) + O(mnr ), wherein t is the iteration number of Algori thm 1 and t × O(mn + p3 ) does not increase as fast as the increasing of r , especially when m and n are large. This means the time complexity of FGD is comparable to that of MUR O(mnr +n 2 r ) in one iteration round with large m, n, and r , but it converges much faster than MUR. From Table I, we can see that NMF-FGD uses fewer iterations and CPU seconds to optimize the objective function than NMFOFGD. It means that FGD reduces the objective function faster than OFGD without increasing the time overhead. Table II compares NDLA-FGD with NDLA-MUR and NDLA-OFGD in term of efficiency on two real-life datasets, which are the face image dataset UMIST [30] and the handwritten dataset MNIST [31]. We selected 300 images in UMIST dataset which were taken from 20 subjects and reshaped each image to a vector in R 1600 . We set the subspace dimension as 50 and 200, respectively. The MNIST handwrit-

(c): MNIST (r = 100) MUR OFGD FGD 0.162 0.162 0.010 1000 579 32 237.308 281.831 35.864

(d): MNIST (r = 300) MUR OFGD FGD 0.055 0.055 0.003 278 135 33 138.981 200.898 50.793

ten dataset includes 3000 handwritten digits 0−9 written by different subjects, and each handwritten digit is stretched into a 28 × 28 image. We randomly selected 50 images for each integer 0 − 9 and reshaped them into a 784 × 500 matrix for NDLA training, and set the subspace dimension as 100 and 300, respectively. Without loss of the generality, for both datasets, we set k1 = 2 and k2 = 15 in NDLA. Experimental results are given in Table II. Based on Table II, we observe that FGD uses fewer iterations and CPU seconds to optimize the objective function than MUR and OFGD. Although OFGD converges much faster than MUR on UMIST dataset [see columns (a) and (b)], it fails to accelerate MUR on the MNIST dataset [see columns (c) and (d)]. This is because the MNIST dataset contains many zeros, which causes the single optimal step size in OFGD to easily exceed the positive quadrant and makes the OFGD degenerate to MUR in this case. However, FGD overcomes this deficiency according to columns (c) and (d). Fig. 3 gives the average step size for rows of W and columns of H in FGD optimization compared with the single step size for W and H in OFGD. It shows that OFGD fails to accelerate MUR on MNIST dataset while FGD does. B. Classification Under Image Occlusion In order to make statistical comparisons between classification performances of different algorithms, Dietterich [32] proposed an empirical method which uses five twofold crossvalidations followed by a t-test. Alpaydin [7] subsequently proposed to modify the Dietterich’s method by removing the unsatisfactory aspect of the result depending on the ordering of the folds, which was called the 5 × 2 cv F-test by the author. In particular, five replications of twofold crossj validation were performed. Assuming pi is the difference between the classification error rates of two algorithms on fold j = 1, 2 of replication i = 1, . . . , 5, the average on replication i was denoted by p¯ i = ( pi1 + pi2 )/2, and the estimated variance was si2 = ( pi1 − p¯ i )2 + ( pi2 − p¯ i )2 . According to [7], the    j statistic F = 5i=1 2j =1 ( pi )2 /2 5i=1 si2 was approximately F-distributed with 10 and 5 degrees of freedom. Throughout this paper, we reject the hypothesis that the algorithms have statistically identical error rate with 95% confidence if the F-statistic is greater than 4.74. In this paper, we use the F-statistic defined above to statistically compare classification performances of NDLA and six other representative algorithms including PCA [1], FLDA [2], DLA [14], NMF [3], LNMF [6], and DNMF [4] on two popular face image datasets, i.e., ORL [29], and UMIST [30], and the popular handwritten dataset MNIST [31]. All face images of both ORL and UMIST were aligned according to the eye position. Each pixel of images was linearly rescaled to the gray level of 256, and each image was rearranged to

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1227

TABLE III AVERAGE E RROR R ATE (%) F OLLOWED BY F -S TATISTIC VALUE OF NDLA V ERSUS R EPRESENTATIVE A GORITHMS ON ORL D ATASET U NDER √ D IFFERENT PARTIAL O CCLUSIONS . T ICK ( ) I NDICATES T HAT NDLA I S S TATISTICALLY S UPERIOR TO THE C OMPARATOR A LGORITHMS DLA [14] 11.7(43) 2.163 18.1(68) 4.243 √ 31.3(120) 5.587( √) 45.6(106) 19.657( )

(a)

NMF [3] √ 22.1(49) 10.827(√) 28.3(73) 39.219(√) 36.1(82) 14.099(√ ) 42.9(75) 6.076( )

(b)

1 0.8 0.6 0.4 0.2 0

LNMF [6] √ 25.3(116) 31.318(√) 37.1(111) 33.557( ) 26.3(120) 3.496 √ 57.3(90) 40.630 ( )

NDLA NMF LNMF DNMF

20

40

(c)

60 80 100 120 dimension

1 0.8 0.6 0.4 0.2 0

error rate

0.6 0.4 20

40

60 80 100 120 dimension

(c)

NDLA 10.8(120) 14.1(119) 23.3(120) 33.3(120)

NDLA NMF LNMF DNMF

20

40

60 80 100 120 dimension

(b) NDLA NMF LNMF DNMF

0.8

0.2

a long vector. For DNMF, we set the parameters in (30) as γ = 10 and δ = 0.01 according to [4]. According to [7], we randomly select an equal number of images from each individual to constitute two folds, signified as training set and test set, and the rest of the images make up the validation set. The training set was used to learn bases for the low-dimensional space and the validation set was used to select the best model parameters, and then the error rate was calculated as the percentage of samples in the test set that were improperly classified using the nearest neighbor rule. To evaluate NDLAs robustness to image occlusion, a randomly positioned square partial occlusion of different size x × x, wherein x is the side length, was added to each image in the test set during the classification phase. Fig. 4 shows the examples of image and the occluded images of three different datasets. By using the model parameters selected on the validation set, the bases for the low-dimensional space was learned on the test set and the error rate was calculated on the training set with a randomly positioned partial occlusion added to each image. Such a trail was independently performed five times, which allowed us to compute a F-statistic from which we decided whether to reject the hypothesis that the classification performances of two algorithms were identical. 1) ORL Dataset: The Cambridge ORL [29] dataset consists of 400 images collected from 40 individuals. There are 10 images for each individual with varying lighting, facial expressions, and facial details (with glasses or no glasses). All images were taken in the same dark background, and each image was normalized to a 112 × 92 pixel array and reshaped to a long vector. We randomly selected eight images from each individual to constitute the twofold training set and test set and the rest makes up the validation set. Fig. 5 shows the average error rate versus the dimension of the subspace on the test set when the side length of partial occlusion x = 20, 25, 30, and 35. Table III gives the average error rates on the two folds and the dimension corresponding to the

DNMF [4] √ 20.0(120) 35.508(√) 28.2(120) 27.492(√) 38.8(120) 28.899(√) 50.2(118) 18.916( )

(a) 1

Fig. 4. Image examples of (a) ORL, (b) UMIST, and (c) MNIST dataset under different occlusions.

error rate

FLDA [2] √ 16.6(39) 48.000(√ ) 24.8(39) 6.599( √) 33.4(39) 24.136( ) 39.9(39) 2.480

1 error rate

PCA [1] 12.8(119) 2.294 √ 20.4(117) 11.234(√) 33.7(111) 14.928(√) 45.8(113) 22.284( )

error rate

Occlusion 20 × 20 25 × 25 30 × 30 35 × 35

NDLA NMF LNMF DNMF

0.8 0.6 0.4 0.2

20

40

60 80 100 120 dimension

(d)

Fig. 5. Average error rate on test set when the partial occlusions size are (a) 20 × 20, (b) 25 × 25, (c) 30 × 30, and (d) 35 × 35 on the ORL dataset.

best performance for all the algorithms under different partial occlusions. Fig. 5 shows that NDLA outperforms all the representative NMF-related algorithms on the test set under different partial occlusions. Table III shows that the average error rates of NDLA are superior to all the comparator algorithms on the training set and test set. It also shows that NDLA is statistically superior to all the comparator algorithms. 2) UMIST Dataset: The UMIST [30] database contains 575 face images collected from 20 people. At least 41 and at most 82 images were taken from each person varying in poses from profile to frontal views. Each photo was transformed into an image in 256 gray levels, and each image was normalized to a 40 × 40 pixel array and reshaped to a vector. Fifteen images from each individual were randomly selected to constitute the dataset, wherein 10 images were randomly selected to constitute the two folds and the rest to make up the validation set. Fig. 5 shows the average error rate versus the dimension of the subspace on the test set when the side length of partial occlusion x = 12, 14, 16, and 18. Table IV gives the average error rates on the two folds and the dimension corresponding to the best performance for all the algorithms under different partial occlusions. Fig. 5 shows that the classification error rate of NDLA on the test set is comparable to that of DNMF when the side length of partial occlusion x = 12 and superior to that of NMF and LNMF. When x = 14, 16, and 18, NDLA outperforms all the representative NMF-related algorithms. Table IV shows that the average error rate of NDLA on the two folds is superior to all competitor algorithms under different partial

1228

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

TABLE IV AVERAGE E RROR R ATE (%) F OLLOWED BY F -S TATISTIC VALUE OF NDLA V ERSUS R EPRESENTATIVE A LGORITHMS ON UMIST D ATASET U NDER √ D IFFERENT PARTIAL O CCLUSIONS . T ICK ( ) I NDICATES T HAT NDLA I S S TATISTICALLY S UPERIOR TO THE C OMPARATOR A LGORITHMS

NDLA NMF LNMF DNMF

20

40 60 dimension

80

1 0.8 0.6 0.4 0.2 0

20

4.333 3.157 √ 18.209( ) √ 27.471( )

40 60 dimension

NMF [3] 18.6(28) 21.2(49) 24.6(77) 31.4(79)

NDLA NMF LNMF DNMF

0.6

40 60 dimension

(c)

80

0.6 0.4 20

40

DNMF [4] 13.6(78) 18.0(79) 24.5(77) 33.4(80)

0.6 0.4 0.2

20

40

60 80 100 120 dimension

40 60 dimension

0.4 0.2

80

20

(d)

Fig. 6. Average error rate on the test set when the partial occlusions size is (a) 12 × 12, (b) 14 × 14, (c) 16 × 16, and (d) 18 × 18 on the UMIST dataset.

40

60 80 100 120 dimension

0.9 0.8 0.7 0.6 0.5 0.4

NDLA NMF LNMF DNMF

20

40

60 80 100 120 dimension

(c)

(d)

Fig. 7. Average error rate on the test set when the partial occlusion size is (a) 6 × 6, (b) 8 × 8, (c) 10 × 10, and (d) 12 × 12 on the MNIST dataset. 1

1

NDLA

0.6 0.4 0.2 0

NDLA

0.8 error rate

0.8 error rate

occlusions. It also shows that NDLA is statistically superior to the competitor algorithms in most cases. 3) MNIST Dataset: The MNIST [31] database contains a training set of 60 000 binary images and a test set of 10 000 binary images of handwritten digits 0 −9. These images are collected from 250 highschool students, each of which is centered in a 28 × 28 image by computing the center of mass of the pixels and translated to position this point at the center of the 28 × 28 image. We select 1500 images from both the training set and the test set, and thus the experiment is based on a sub-dataset that includes 3000 images. Sixty images were randomly selected to constitute the two folds and the rest comprise the validation set. Fig. 7 shows the average error rate versus the dimension of the subspace on the test set when the side length of partial occlusion x = 6, 8, 10, and 12. Table V gives the average error rates on the two folds and the dimension corresponding to the best performance for all the algorithms under different partial occlusions. Fig. 7 shows that the average error rates of NDLA are comparable to those of NMF and DNMF on the test set and superior to those of LNMF under different partial occlusions. Table V shows that the average error rates of NDLA on the two folds are statistically superior to those of FLDA, NMF, and DNMF. From Table V, we can find that the average error rates of NDLA are lower than those of PCA, DLA, and LNMF, but it performs statistically comparable with them. That is because, for such binary images in handwritten dataset, the noise introduced by occlusions does not affect the classification stage very much.

(b) NDLA NMF LNMF DNMF

0.6

09.6(80) 13.1(79) 17.6(79) 22.5(79)

NDLA NMF LNMF DNMF

0.8

(a) 0.8

NDLA

3.324 √ 35.667( ) 3.262 3.747

1

60 80 100 120 dimension

1

0.4 20

√ 11.837( ) √ 5.359( ) √ 57.341( ) √ 23.866( )

NDLA NMF LNMF DNMF

0.8

0.2

80

NDLA NMF LNMF DNMF

0.8

LNMF [6] 16.1(80) 23.1(80) 32.5(80) 51.0(80)

1

0.2 20

√ 15.500( ) 3.380 3.159 √ 7.256( )

(b) 1 error rate

error rate

10.8(74) 18.4(72) 29.9(70) 43.0(73)

NDLA NMF LNMF DNMF

(a) 1 0.8 0.6 0.4 0.2 0

DLA [14]

3.378 2.937 √ 11.371( ) √ 6.672( )

error rate

FLDA [2] 14.3(19) 19.2(19) 29.4(19) 35.3(19)

error rate

4.220 √ 7.934( ) √ 64.941( ) √ 21.922( )

error rate

1 0.8 0.6 0.4 0.2 0

PCA [1] 14.6(59) 22.1(63) 32.0(78) 50.1(71)

error rate

error rate

12 × 12 14 × 14 16 × 16 18 × 18

error rate

Occlusion

0.6 0.4 0.2

50

100 150 200 250 k2 (a)

0

5

10

15 k1 (b)

20

25

Fig. 8. Average error rate on the validation set of the MNIST dataset versus neighborhood size (a) k2 and (b) k1 of the patches.

C. Discussions This section gives some discussions on several problems in this experiment. 1) Subspace Dimension Selection: We experimentally selected these best subspace dimensions for representative algorithms. In Tables III–V, each number in parentheses shows the best subspace dimension for the corresponding algorithm on each dataset. Curves in Figs. 5–7 show that the error rate curve of NDLA is usually below those of other NMF-related algorithms, and thus NDLA usually performs better than the compared baseline algorithms, and NDLA generalizes better than the compared baseline algorithms in terms of subspace dimension. 2) Parameter Selection: The neighborhood size of the patches is a critical parameter in NPAF. In this paper, we selected the parameters, i.e., k1 and k2 , in NDLA by using

GUAN et al.: NON-NEGATIVE PATCH ALIGNMENT FRAMEWORK

1229

TABLE V AVERAGE E RROR R ATE (%) F OLLOWED BY F -S TATISTIC VALUE OF NDLA V ERSUS R EPRESENTATIVE A LGORITHMS ON MNIST D ATASET U NDER √ D IFFERENT PARTIAL O CCLUSIONS . T ICK ( ) I NDICATES T HAT NDLA I S S TATISTICALLY S UPERIOR T O THE C OMPARATOR A LGORITHMS Occlusion 6×6 8×8 10 × 10 12 × 12

PCA [1] 15.4(50) 30.6(50) 38.0(61) 48.3(29)

3.253 2.718 1.998 1.526

FLDA [2] 49.5(9) 52.9(9) 56.4(9) 64.2(9)

√ 92.959( ) √ 93.080( ) √ 27.497( ) √ 51.420( )

DLA [14] 24.8(50) 29.5(36) 37.1(63) 47.7(61)

√ 8.498( ) 4.527 3.546 2.362

cross-validation, which has been adopted in many related papers, e.g., [14]. Fig. 8 shows the average error rate on the validation set of the MNIST dataset versus the neighborhood size (k1 , k2 ) of the patches. By definition of the with-in class patch and the between-class patch, k1 varies from 1 to N/C − 1, where N is the number of samples in the training set and C is the class number, and k2 varies from 1 to N − N/C. In this experiment, N = 300 and C = 10, 1 ≤ k1 ≤ 29, and 1 ≤ k2 ≤ 270. Fig. 8(a) presents that the error rate versus k2 when k1 is set to 5. A foot arises when k2 = 20. Fig. 8(b) presents the error rate versus k1 when k2 is set to 20. There appears a foot when k1 = 5. Fig. 8 shows that NDLA performs robustly with k2 in a wide range of [10, 270], but the performance varies severely when k1 > 9. It means that the classification performance of NDLA is sensitive to the data local geometric structure on the validation set of the MNIST dataset. The classification results on the the training set and the test set of the MNIST dataset show that the selected NDLA model is effective compared to other NMF-related algorithms. 3) NDLA Versus NMF: In Figs. 5 and 6, NMF performs better than NDLA when the dimension r is low, usually r < 40, on both the ORL and UMIST datasets, because the bases learned by NDLA are much sparser than those learned by NMF on both datasets. When the dimension r is low, the classifier constructed by using the first r bases contains insufficient discriminative information, whereas NMF may include most energy in its first r bases. A detailed discussion can be found in [17] and [33]. VI. C ONCLUSION In this paper, we proposed an NPAF that unifies NMFrelated dimension reduction algorithms. It can be applied to better understand the common properties and intrinsic differences in various NMF-related dimension reduction algorithms. We proposed the FGD that uses the Newton method to search the optimal step size for one factor when fixing another in each iteration round and showed NDLA to incorporate both the data local geometric structure and margin maximizationbased discriminative information. In summary, we make the following remarks. 1) We used KLD to measure the difference between original samples and their approximations. KLD can be replaced by other distance metrics, e.g., the Frobenius norm. 2) To ensure the convexity of a specific implementation of NPAF, e.g., DNMF, it is important to choose a

NMF [3] 42.8(55) 47.5(62) 52.7(64) 60.0(48)

√ 61.269( ) √ 85.937( ) √ 26.737( ) √ 15.375( )

LNMF [6] 26.1(112) 29.3(103) 36.6(120) 45.3(103)

1.922 2.053 3.611 3.833

DNMF [4] 30.3(92) 34.0(94) 40.5(104) 49.4(114)

1.521 1.601 √ 4.795( ) √ 39.361( )

NDLA 24.1(109) 28.3(119) 35.0(120) 44.9(120)

suitable tradeoff parameter γ in the objective of NPAF (2). However, the proposed NDLA model [see (35)] is stable over a wide range of γ with small value. For all experiments, we set γ = 0.001. 3) MUR can be applied to NPAF if the alignment matrix in (2) is symmetric. FGD can be applied to NPAF if the alignment matrix is symmetric and positive semidefinite. Both conditions are not difficult to meet in practice, so we can usually apply FGD to reduce the training time. ACKNOWLEDGMENT The authors would like to thank EiC D. Liu, the handling associate editor and anonymous reviewers for their support and constructive comments on this paper. R EFERENCES [1] H. Hotelling, “Analysis of a complex of statistical variables into principal components,” J. Educ. Psychol., vol. 24, no. 6, pp. 417–441, Sep. 1933. [2] R. A. Fisher, “The use of multiple measurements in taxonomic problems,” Ann. Eugen., vol. 7, no. 7, pp. 179–188, 1936. [3] D. D. Lee and H. S. Seung, “Learning the parts of objects by nonnegative matrix factorization,” Nature, vol. 401, no. 21, pp. 788–791, Oct. 1999. [4] S. Zafeiriou, A. Tefas, I. Buciu, and I. Pitas, “Exploiting discriminant information in nonnegative matrix factorization with application to frontal face verification,” IEEE Trans. Neural Netw., vol. 17, no. 3, pp. 683–695, May 2006. [5] D. Cai, X. He, J. Han, and T. Huang, “Graph regularized nonnegative matrix factorization for data representation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. PP, no. 99, p. 1, Dec. 2010. [6] S. Z. Li, X. W. Hou, H. J. Zhang, and Q. S. Cheng, “Learning spatially localized, parts-based representation,” in Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit., vol. 2. Dec. 2001, pp. 207–212. [7] E. Alpaydin, “Combined 5 × 2 cv F test for comparing supervised classification learning algorithms,” Neural Comput., vol. 11, no. 8, pp. 1885–1892, 1999. [8] T. Zhang, D. Tao, X. Li, and J. Yang, “Patch alignment for dimensionality reduction,” IEEE Trans. Knowl. Data Eng., vol. 21, no. 9, pp. 1299–1313, Sep. 2009. [9] W. Bian and D. Tao, “Max-min distance analysis by using sequential SDP relaxation for dimension reduction,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 33, no. 5, pp. 1037–1050, May 2011. [10] D. Tao, X. Li, X. Wu, and S. J. Maybank, “Geometric mean for subspace selection,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 31, no. 2, pp. 260–274, Feb. 2009. [11] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by locally linear embedding,” Science, vol. 290, no. 5500, pp. 2323–2326, Dec. 2000. [12] J. Tenenbaum, V. Silva, and J. Langford, “A global geometric framework for nonlinear dimensionality reduction,” Science, vol. 290, no. 5500, pp. 2319–2323, Dec. 2000. [13] X. He and P. Niyogi, “Locality preserving projections,” in Proc. Adv. Neural Inf. Process. Syst., vol. 15. 2003, pp. 1–8. [14] T. Zhang, D. Tao, and J. Yang, “Discriminative locality alignment,” in Proc. 10th Eur. Conf. Comput. Vis., vol. 5302. 2008, pp. 725–738.

1230

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 22, NO. 8, AUGUST 2011

[15] Z. Zhang and H. Zha, “Principal manifolds and nonlinear dimension reduction via local tangent space alignment,” SIAM J. Sci. Comput., vol. 26, no. 1, pp. 313–338, 2005. [16] D. P. Bertsekas, Nonlinear Programming, 2nd ed. Belmont, MA: Athena Scientific, 1999. [17] N. Guan, D. Tao, Z. Luo, and B. Yuan, “Fast gradient descent for nonnegative patch alignment framework,” Univ. Technol., Sydney (UTS), Sydney, Australia, Tech. Rep. -2010-11, 2010. [18] C. H. Q. Ding, T. Li, and M. Jordan, “Convex and semi-nonnegative matrix factorizations,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 1, pp. 45–55, Jan. 2010. [19] R. Zdunek and A. Cichocki, “Non-negative matrix factorization with quasi-Newton optimization,” in Proc. 8th Int. Conf. Artif. Intell. Soft Comput., vol. 4029. 2006, pp. 870–879. [20] N. Guan, D. Tao, Z. Luo, and B. Yuan, “Manifold regularized discriminative non-negative matrix factorization with fast gradient descent,” IEEE Trans. Image Process., vol. 20, no. 7, pp. 2030–2048, Jul. 2011. [21] R. A. Horn and C. R. Johnson, Matrix Analysis. Cambridge, U.K.: Cambridge Univ. Press, 1985. [22] M. A. Woodbury, “Inverting modified matrices,” Stat. Res. Group, Princeton Univ., Princeton, NJ, Memo. Rep. 42, 1950. [23] C.-J. Lin, “On the convergence of multiplicative update algorithms for nonnegative matrix factorization,” IEEE Trans. Neural Netw., vol. 18, no. 6, pp. 1589–1596, Nov. 2007. [24] J. Yang, S. Yan, Y. Fu, X. Li, and T. Huang, “Non-negative graph embedding,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., vol. 4. Anchorage, AK, Jun. 2008, pp. 1–8. [25] C. Wang, Z. Song, S. Yan, L. Zhang, and H. Zhang, “Multiplicative nonnegative graph embedding,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., vol. 5. Miami, FL, Jun. 2009, pp. 389–396. [26] D. Cai, X. He, and J. Han, “SRDA: An efficient algorithm for largescale discriminant analysis,” IEEE Trans. Knowl. Data Eng., vol. 20, no. 1, pp. 1–12, Jan. 2008. [27] C.-J. Lin, “Projected gradient methods for nonnegative matrix factorization,” Neural Comput., vol. 19, no. 10, pp. 2756–2779, Oct. 2007. [28] M. Belkin, “Problems of learning on manifolds,” Ph.D. thesis, Dept. Math., Univ. Chicago, Chicago, IL, Aug. 2003. [29] F. S. Samaria and A. C. Harter, “Parameterisation of a stochastic model for human face identification,” in Proc. 2nd IEEE Workshop Appl. Comput. Vis., Sarasota, FL, Dec. 1994, pp. 138–142. [30] D. B. Graham and N. M. Allinson, “Characterizing virtual eigensignatures for general purpose face recognition,” in Proc. Face Recognit.: Theory Appl., vol. 163. 1998, pp. 446–456. [31] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278– 2324, Nov. 1998. [32] T. Dietterich, “Approximate statistical tests for comparing supervised classification learning algorithms,” Neural Comput., vol. 10, no. 7, pp. 1895–1923, Oct. 1998. [33] T. Zhou, D. Tao, and X. Wu, “Manifold elastic net: A unified framework for sparse dimension reduction,” Data Min. Knowl. Disc., vol. 22, no. 3, pp. 340–371, May 2011.

Naiyang Guan received the B.S. and M.S. degrees from the National University of Defense Technology, Changsha, China, where he is currently pursuing the Ph.D. degree in the School of Computer Science. He was a Visiting Student with the School of Computer Engineering, Nanyang Technological University, Singapore, from October 2009 to 2010. His current research interests include computer vision, image processing, and convex optimization.

Dacheng Tao (M’07) is a Professor of computer science, Center for Quantum Computation and Information Systems, Faculty of Engineering and Information Technology, University of Technology, Sydney, Australia. He has authored and coauthored more than 100 scientific articles published or presented at top venues including IEEE T RANSACTIONS ON PATTERN A NALYSIS AND M ACHINE I NTELLIGENCE , T RANSACTIONS ON N EURAL N ETWORKS , T ECHNICAL I NTEREST P ROFILE, Network Intrusion Prevention Systems, the International Conference on Machine Learning, Conference on Artificial Intelligence and Statistics, the International Conference on Data Mining the International Joint Conference on Artificial Intelligence, Association for the Advancement of Artificial Intelligence, Conference on Computer Vision and Pattern Recognition, European Conference on Computer Vision, Association for Computing Machinery Transactions on Knowledge Discovery from Data, Multimedia and KDD. His current research interests include application of statistics and mathematics for data analysis problems in data mining, computer vision, machine learning, multimedia, and video surveillance. He received the Best Theory/Algorithm Paper Runner Up Award in IEEE ICDM in 2007.

Zhigang Luo received the B.S., M.S., and Ph.D. degrees from the National University of Defense Technology, Changsha, China, in 1981, 1993, and 2000, respectively. He is currently a Professor with the School of Computer Science, National University of Defense Technology. His current research interests include parallel computing, computer simulation, and bioinformatics.

Bo Yuan received the Bachelors degree from Peking University Medical School, Beijing, China, in 1983, the M.S. degree in biochemistry, and the Ph.D. degree in molecular genetics from the University of Louisville, Louisville, KY, in 1990 and 1995, respectively. He is currently a Professor in the Department of Computer Science and Engineering, Shanghai Jiao Tong University (SJTU), Shanghai, China. Before joining SJTU, he was a Tenure-Track Assistant Professor with Ohio State University (OSU), Columbus, OH, in 2006, and served as a Co-Director for the OSU Program in Pharmacogenomics. At OSU, he was the Founding Director of OSU’s genome initiative during the early 2000s, leading one of the only three independent efforts in the world (besides the Human Genome Project and the Celera Company), having assembled and deciphered the entire human and mouse genomes. His current research interests include biological networks, network evolution, stochastic process, biologically inspired computing, and bioinformatics, particularly on how these frameworks might impact the development of intelligent algorithms and systems.

Non-negative Patch Alignment Framework

In recent years, different NMF-related dimension ... N. Guan and Z. Luo are with the School of Computer Science, National ... sharing the same class label with vi ; and 2) samples taking ...... F-distributed with 10 and 5 degrees of freedom.

941KB Sizes 1 Downloads 185 Views

Recommend Documents

NONNEGATIVE MATRIX FACTORIZATION AND SPATIAL ...
ABSTRACT. We address the problem of blind audio source separation in the under-determined and convolutive case. The contribution of each source to the mixture channels in the time-frequency domain is modeled by a zero-mean Gaussian random vector with

Split alignment
Apr 13, 2012 - I use the standard affine-gap scoring scheme, with one additional parameter: a .... Ai,j: the alignment score for query base j in alignment i.

Multiplicative Nonnegative Graph Embedding
data factorization, graph embedding, and tensor represen- ... tion brings a group of novel nonnegative data factorization ...... Basis matrix visualization of the algorithms PCA (1st .... is a tool and can be used to design new nonnegative data.

Unsupervised Feature Selection Using Nonnegative ...
trix A, ai means the i-th row vector of A, Aij denotes the. (i, j)-th entry of A, ∥A∥F is ..... 4http://www.cs.nyu.edu/∼roweis/data.html. Table 1: Dataset Description.

FAST NONNEGATIVE MATRIX FACTORIZATION
FAST NONNEGATIVE MATRIX FACTORIZATION: AN. ACTIVE-SET-LIKE METHOD AND COMPARISONS∗. JINGU KIM† AND HAESUN PARK†. Abstract. Nonnegative matrix factorization (NMF) is a dimension reduction method that has been widely used for numerous application

1.5 patch mac.pdf
Mac mini only runs 1.5 gigabit on ssd mac minimodela1283 ifixit. Unibox 1.5.2 mac os x imojado free. download macsoftwareand. Blocs 1.5.3 cracked for ...

Joint Weighted Nonnegative Matrix Factorization for Mining ...
Joint Weighted Nonnegative Matrix Factorization for Mining Attributed Graphs.pdf. Joint Weighted Nonnegative Matrix Factorization for Mining Attributed Graphs.

Vehicle alignment system
Jun 13, 1983 - tionally employs rather sophisticated equipment and specialized devices in order to align vehicle wheels. It has been recognized, as shown in the Modern Tire. Dealer, Volume 63, Number 7, June of 1982 at page 31, that alignment of the

Toward Faster Nonnegative Matrix Factorization: A New Algorithm and ...
College of Computing, Georgia Institute of Technology. Atlanta, GA ..... Otherwise, a complementary ba- ...... In Advances in Neural Information Pro- cessing ...

Patch 10 - FAQ.pdf
Problem 2 : Problems like "application terminated "/ "unresolvable" error while clicking. 1. SSA >>> Account Open for ,. 2. RD Default month deposit option after ...

Pesedit 2011 patch
... befour-year-olds who watched theaverageamount oftelevision e. 3-5 hours per ... Friedrich nietzsche pdf.To ppt. converter. Hajimete no oishasan.Nfl 2015 phi.

Downlink Interference Alignment - Stanford University
cellular networks, multi-user MIMO. I. INTRODUCTION. ONE of the key performance metrics in the design of cellular systems is that of cell-edge spectral ...

Downlink Interference Alignment - Stanford University
Paper approved by N. Jindal, the Editor for MIMO Techniques of the. IEEE Communications ... Interference-free degrees-of-freedom ...... a distance . Based on ...

yagi patch template.pdf
1mm nylon. spacers. 5mm nylon bolts. 25mm. 25mm. Page 1 of 1. yagi patch template.pdf. yagi patch template.pdf. Open. Extract. Open with. Sign In. Main menu.

Downlink Interference Alignment
Wireless Foundations. U.C. Berkeley. GLOBECOM 2010. Dec. 8. Joint work .... Downlink: Implementation Benefits. 2. 1. 1. K. Fix K-dim reference plane, indep. of ...

Manifold Alignment Determination
examples from which it is capable to recover a global alignment through a prob- ... only pre-aligned data for the purpose of multiview learning. Rather, we exploit ...

Design of Microstrip Patch Antenna f Design of Microstrip Patch ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 3, ... Since the advantages of the microstrip patch antenna are mainly to .... Wireless Pers Commun 62: pgs.247–256 Springer Science+Business Media,. 2010.

Design of Microstrip Patch Antenna f Design of Microstrip Patch ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 3, March ... Keywords: Microstrip Patch antennas, Wireless capsule Endoscopy, FEKO, Return loss. 1. ... Since the advantages of the microstrip patch antenna are mainly

Nonnegative Polynomials and Sums of Squares Why ...
Jan 24, 2011 - Why do bad things happen to good polynomials? Grigoriy .... There exist constants c1(d) and c2(d), dependent on the degree d only, such that.

Nonnegative Polynomials and Sums of Squares Why ...
Mar 17, 2011 - We can compute instead the best sos lower bound: γ. ∗. = max f −γ is sos ... For small n and 2d find good bounds on degree of q and explicit.

Design of Fractal Patch Antenna for Multiband esign of Fractal Patch ...
3Assistant Professor, Department of Electronics and Communication Engineering, Karunya University ... Micro strip antennas have several advantages over conventional microwave antenna and therefore are ... in various fields such as in medical applicat

switch and patch panel.pdf
Page 3 of 5. switch and patch panel.pdf. switch and patch panel.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying switch and patch panel.pdf.

Metro last light patch
].pdf.Lovestory japanese.Arrowseason 1 all.Thechemicals weforever shallmetro last light patch using areirritants metro last light patch be ... Tungevaag raaban.

Scouting Religious Square Knot Patch Requirements.pdf ...
Scouting Religious Square Knot Patch Requirements.pdf. Scouting Religious Square Knot Patch Requirements.pdf. Open. Extract. Open with. Sign In.