24. On SVD-free Latent Semantic Indexing for Iris Recognition of Large Databases Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

Summary. This chapter presents a method for an automatic identification of persons by iris recognition. A raster image of a human iris is represented as a sequence of color pixels. Information retrieval is conducted by the Latent Semantic Indexing (LSI) method. The pattern recognition algorithm is powered very effectively when the time consuming Singular Value Decomposition (SVD) of LSI is replaced by the partial symmetric eigenproblem. Numerical experiments on a real 488 MB biometric data collection indicates feasibility of the presented approach as a tool for automated image recognition without special preprocessing.

24.1 Introduction Methods of human identification using biometric features like fingerprint, hand geometry, face, voice and iris are widely studied. A human eye iris has its unique structure given by pigmentation spots, furrows and other tiny features that are stable throughout life, see [1, 2]. It is possible to scan an iris without physical contact in spite of wearing eyeglasses or contact lens. The iris can be hardly forged, e.g. replaced or copied. This makes the iris a suitable object for the identification of persons. Iris recognition seems to be more reliable than other biometric techniques like face recognition [3]. Iris biometrics systems for public and personal use have been designed and deployed commercially by British Telecom, US Sandia Labs, UK National Physical Laboratory, NCR, Oki, IriScan, and others. Applications of these systems are expected in personal identification, access control, computer and Internet security, etc. Studies about iris recognition were published in [1, 2, 3, 4, 5, 6, 7]. The method proposed by Daugman [1, 2] is based on the transformation of elementary regions of the iris image into polar coordinates. Then, using twodimensional optimal Gabor functions, a binary iris code is generated. The iris identification consists of comparisons of the generated codes using Hamming distance. In [6] the field of interest is transformed into standardized polar coordinates similarly as in [2]. The characteristic iris vector is computed from the mean brightness levels of elementary ring sectors of the iris image.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

2

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

Liam et al. [5] use a trained Self-Organizing Map Neural Network to recognize iris patterns. The iris of doughnut shape is converted into a rectangular form and fed to the neural network. Roche et al. [7] propose an iris recognition method where the features of the iris are represented by fine-to-coarse approximations at different resolution levels. In this technique the discrete dyadic wavelet transform was used. In this article, we present an alternative approach to the recognition of human iris images. The aim is to show that the Latent Semantic Indexing [8, 9] is as a good way for recognizing images as other above mentioned methods [10]. Moreover, the pattern recognition can be powered very effectively when the time consuming Singular Value Decomposition of LSI is replaced by the partial symmetric eigenproblem, which can be solved by using fast iterative solvers [11].

24.2 Image Retrieval Using Latent Semantic Indexing The numerical linear algebra, especially Singular Value Decomposition is used as a basis for information retrieval in the retrieval strategy called Latent Semantic Indexing, see [8, 9]. Originally, LSI was used as an efficient tool for semantic analysis of large amount of text documents. The main reason is that more conventional retrieval strategies (such as vector space, probabilistic and extended Boolean) are not very efficient for real data, because they retrieve information solely on the basis of keywords and polysemy (words having multiple meanings) and synonymy (multiple words having the same meaning) are not correctly detected. LSI can be viewed as a variant of the vector space model with a low-rank approximation of the original data matrix via the SVD or the other numerical methods [8]. The ”classical” LSI application in information retrieval algorithm has the following basic steps: i) The Singular Value Decomposition of the term matrix using numerical linear algebra. SVD is used to identify and remove redundant noise information from data. ii) The computation of the similarity coefficients between the transformed vectors of data and thus reveal some hidden (latent) structures of data. Numerical experiments pointed out that some kind of dimension reduction, which is applied to the original data, brings to the information retrieval following two main advantages: (i) automatic noise filtering and (ii) natural clustering of data with ”similar” semantic, see Figure 24.1 and Figure 24.2. 24.2.1 Image Coding Recently, the methods of numerical linear algebra, especially SVD, are also successfully used for the face recognition and reconstruction [12], image retrieval [10, 11] and as a tool for information extraction from HTML product

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

3

Fig. 24.1. An example of LSI image retrieval results. Images are automatically sorted by their content using the partial eigenproblem.

Fig. 24.2. An example of a similarity measure is the cosine similarity. Here A, Q, B represents the vectors. Symbols ϕA and ϕB denotes the angle between the vectors A, Q and B, Q, respectively. The vector A is more similar to Q than vector B, because ϕA < ϕB . A small angle is equivalent to a large similarity.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

4

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

Fig. 24.3. Image coding of an 3 × 2 pixels image example and the corresponding six-dimensional vector.

catalogues [13]. In this paper, the Latent Semantic Indexing is used as the tool for solving iris recognition problem. In our approach [10, 11], a raster image is coded as a sequence of pixels, see Figure 24.3. Then the coded image can be understood as a vector of an m-dimensional space, where m denotes the number of pixels (attributes). Let the symbol A denote a m × n term-document matrix related to m keywords (pixels) in n documents (images). Let us remind that the (i, j)-element of the term-document matrix A represents the color of i-th position in the j-th image document, see Figure 24.4. 24.2.2 Document Matrix Scaling Our numerical results pointed out that there is a possibility to increase the ability of the LSI method to extract details from images by scaling of the document matrix. This feature of the method was also exploited to iris recognition. Let the symbol A(:, i) denote the i-th column of the document matrix A. We implemented the following scaling1 : for i = 1:n % Loop over all images A(:,i) = A(:,i)/norm(A(:,i)) end;

1

For the description of matrix-oriented algorithms we used the well-known Matlablike notations [14]. For instance, the symbol A(:, i) denotes the i-th column of the matrix A.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

5

Fig. 24.4. An example of document matrix coding. The document matrix A is represented as a sequence of coded images.

24.3 Implementation Details of Latent Semantic Indexing In this section we will derive the SVD-free Latent Semantic Indexing and we will present the implementation of the LSI algorithms. We will show two modifications of the original LSI algorithm presented in [9]. Here presented the SVD-free modifications of the original LSI algorithm replaced the time-expensive SVD decomposition of a general document matrix by the approximate solution of a partial symmetric eigenproblem of a square matrix. 24.3.1 LSI and Singular Value Decomposition Let the symbol A denotes the m × n document matrix related to m pixels in n images. The aim of SVD is to compute decomposition A = U SV T ,

(24.1)

where S ∈ Rm×n is a diagonal matrix with nonnegative diagonal elements called the singular values, U ∈ Rm×m and V ∈ Rn×n are orthogonal matrices2 . The columns of matrices U and V are called the left singular vectors and the right singular vectors respectively. The decomposition can be computed so that the singular values are sorted in decreasing order. The full SVD decomposition (24.1) is memory and time consuming operation, especially for large problems. Although the document matrix A is 2

A matrix Q ∈ Rn×n is said to be orthogonal if the condition Q−1 = QT holds.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

6

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

often sparse, the matrices U and V have a dense structure. Due these facts, only a few k-largest singular values of A and the corresponding left and right singular vectors are computed and stored in memory. The number of singular values and vectors which are computed and kept in memory can be chosen experimentally as a compromise between the speed/precision ratio of the LSI procedure. We implemented and tested LSI procedure in the Matlab system by Mathworks. The document matrix A was decomposed by the Matlab command svds. Using the svds command brings following advantages: • The document matrix A can be effectively stored in memory by using the Matlab storage format for sparse matrices. • The number of singular values and vectors computed by the partial SVD decomposition can be easily set by the user. Following [9] the Latent Semantic Indexing procedure can be written in Matlab by the following way. Procedure Original LSI [Latent Semantic Indexing] function sim = lsi(A,q,k) % Input: % A ... the m × n matrix % q ... the query vector % k ... Compute k largest singular values and vectors; k ≤ n % Output: % sim ... the vector of similarity coefficients [m,n] = size(A); 1. Compute the co-ordinates of all images in the k-dimensional space by the partial SVD of a document matrix A. [U,S,V] = svds(A,k); % Compute the k largest singular values of A; The rows of V contain the co-ordinates of images. 2. Compute the co-ordinate of a query vector q qc = q’ * U * pinv(S); % The vector qc includes the co-ordinate of the query vector q; The matrix pinv(S) contains reciprocals of nonzeros singular values (an pseudoinverse); The symbol ’ denotes the transpose superscript. 3. Compute the similarity coefficients between the co-ordinates of the query vector and images. for i = 1:n % Loop over all images sim(i)=(qc*V(i,:)’)/(norm(qc)*norm(V(i,:))); end; % Compute the similarity coefficient for i-th image; V (i, :) denotes the

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

7

i-th row of V . The procedure lsi returns to a user the vector of similarity coefficients sim. The i-th element of the vector sim contains a value which indicate a ”measure” of a semantic similarity between the i-th document and the query document. The increasing value of the similarity coefficient indicates the increasing semantic similarity. 24.3.2 LSI1 - The First Enhancement of the LSI Implementation In this chapter, we will remind the well-known fact how to express matrices S and V without the SVD. Of course, there is an efficient SVD iterative algorithm based on Lanczos algorithm [14], but the SVD of a large document matrix remains still time-expensive. Let us assume the well-known relationship between the singular value decomposition of the matrix A and the symmetric eigenproblem of the symmetric square matrices AT A: A = U SV T (24.2) AT = (U SV T )T = V S T U T T

T

T

A A = V S (U U )SV

T

T

= V S SV

(24.3) T

(24.4)

Furthermore, let us assume the SVD decomposition (24.1) again. Because of the fact that the matrix V is orthogonal, the following matrix identity holds: AV = U S.

(24.5)

Finally, we can express the matrix U in the following way: AV S + ≈ U

(24.6)

Here the symbol S + denotes the Moore-Penrose pseudoinverse (pinv). Let us accent that the diagonal matrix S contains only strictly positive singular values for real cases; The singular values less than tol ≈ 0 are cut off by the Matlab eigs(A’*A, k) command, since the number of computed singular values k << n for real problems. Following these observations, the SVD-free Latent Semantic Indexing procedure can be written by the following way. LSI1 function % Input: % A ... % q ... % k ... vectors;

[The SVD-free LSI, Version 1] sim = lsi1(A,q,k) the m × n matrix the query vector Compute k largest singular values and k≤n

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

8

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

% Output: % sim ...

the vector of similarity coefficients

[m,n] = size(A); 1. Compute the co-ordinates of all images in the k-dimensional space by the SVD-free approach. [V,S2] = eigs(A’*A,k); % Compute the k largest eigenvalues and eigenvectors of AT A to obtain V and S 2 without SVD. S = sqrt(S2); U = A * V * pinv(S); % Compute S and U without SVD. 2. Compute the co-ordinate of the query vector q. qc = q’ * U * pinv(S); 3. Compute the similarity coefficients between the co-ordinates of the query vector and images. for i = 1:n % Loop over all images sim(i)=(qc*V(i,:)’)/(norm(qc)*norm(V(i,:))); end; Compared to the previous version of the LSI, we can see modifications in the first step of the algorithm. The second and the third step of the algorithm remains the same. 24.3.3 LSI2 - The Second Enhancement of LSI Analyzing the LSI1 procedure deeply, we can see that the matrix U does not have to be explicitly computed and stored in memory during the second step of the LSI. The exploitation of this observation brings us the additional accelerating of the speed and decreasing the memory requirements of the LSI: LSI2 [The SVD-free LSI, Version 2] function sim = lsi2(A,q,k) % Input: % A ... the m × n matrix % q ... the query vector % k ... Compute k largest singular values and vectors; k ≤ n % Output: % sim ... the vector of similarity coefficients [m,n] = size(A);

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

9

1. Compute the co-ordinates of all images in the k-dimensional space by the SVD-free approach. [V,S2] = eigs(A’*A,k); S = sqrt(S2); % Compute S and V without SVD. 2. Compute the co-ordinate of the query vector q without using the matrix U . qc = (((q’ * A) * V) * pinv(S)) * pinv(S); 3. Compute the similarity coefficients between the co-ordinates of the query vector and images. for i = 1:n % Loop over all images sim(i)=(qc*V(i,:)’)/(norm(qc)*norm(V(i,:))); end; Compared to the LSI1, we can see modifications in the first step and in the second step of the algorithm. The presented choice of parenthesis in the second step of LSI2 is the key implementation detail which extremely influences the efficiency of the LSI speed.

24.4 Iris Recognition Experiments The iris is scanned by TOPCON optical device connected to the CCD Sony camera. The acquired digitized image is RGB of size 576 × 768 pixels [15]. Only the red (R) component of the RGB image participates in our experiments. The reason is that the recognition based on the red component appears to be more reliable than recognition based on green (G) or blue (B) components or grayscale images. It is in accord with the study of Daugman [4], where near-infrared wavelengths are used. 24.4.1 The Small-Size Database Query The testing database contains 24 images – 3 images of the left eye and 3 images of the right eye from 4 persons. The collection of images required 81 MB of RAM. The images were opened using the Matlab (tm) command imread. The queries were represented as images from the collection. We used the LSI2 algorithm for image retrieval in all cases. For example, the name “001L3.tif” implies that the left iris of Person No.1, the series 3 is assumed. The most time consuming part of LSI was multiplying of matrices AT ×A. It takes 3.38 seconds on an Pentium Celeron 2.4 GHz with 256 MB RAM running MS Windows XP. The computation of the partial eigenproblem by the Matlab command eigs() takes only 0.26 seconds, see Table 24.1.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

10

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

In the current computer implementation of the Latent Semantic Indexing method, no pre-processing of images and/or a-priori information is assumed. However, the numerical experiments presented here indicate quite optimistic availability of the proposed algorithm for automated iris recognition.

Fig. 24.5. Image retrieval results related to the query image 002L1 .tif (Query1).

Properties of the document matrix A Number of keywords: 576×768=442 368 Number of documents: 24 Size in memory: 81 MB The SVD-Free LSI processing parameters Dim. of the original space 24 Dim. of the reduced space (k) 10 Time for AT A operation 3.38 secs. Results of the eigensolver 0.26 secs. The total time 3.64 secs. Table 24.1. Iris recognition using the SVD-free Latent Semantic Indexing method; Properties of the document matrix (up) and LSI processing parameters (down) for Query1 and Query2.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

11

Fig. 24.6. Image retrieval results related to the query image 002P1 .tif (Query2).

Query1 All results are sorted out by the similarity coefficient. In order to achieve lucid results, only 16 of the most significant images are presented. In all cases, the remaining 8 images with the negligible similarity coefficient were removed from the presentation. When the left iris of Person No.2, the series 1, is assumed as the query image (002L1 .tif), the most similar image is of course the query image itself with the similarity coefficient 1, see Fig. 24.5. The subsequent image is 002L2 .tif with the similarity coefficient 0.8472. The images 002P3.tif, 002L3 .tif and 002P2.tif follow. Analyzing results we can see that these four most similar images are connected with Person No. 2. Query2 When the right iris of Person No.2, the series 1, is assumed as the query image (002P1.tif), the two most similar images are related to Person No.2, 002P2.tif and 002P3.tif, see Fig. 24.6. The image 001L3 .tif, which is connected to Person No.1, follow. But two subsequent images 002L3 .tif and 002L2 .tif are again relative to Person No.2. 24.4.2 The Large-Scale Database Query Statistically significant numbers of the similarity coefficients for compared “the same” and “different” irides calculated for a database enable to quantify

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

12

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

Table 24.2. The large-scale iris recognition example using the SVD-free Latent Semantic Indexing method; Properties of the document matrix (up) and LSI processing parameters (down). Properties of the document matrix A Number of keywords: 52 488 Number of documents: 384 Size in memory: 157 464 Kbytes The SVD-Free LSI processing parameters Dim. of the original space 384 Dim. of the reduced space (k) 10 Time for AT A operation 13.906 secs. Results of the eigensolver 4.843 secs. The total time 18.749 secs.

the probability of incorrect iris recognition for the method. For this purpose, a database containing 384 images - 3 images of the left eye and 3 images of the right eye taken at different times from 64 persons were used [15]. The large database query on the same irides was performed by 3 × 128 comparisons (the maximum possible number of combination of the same irides). Similarly, the test of different irides was made as 72 576 comparisons of irides from different persons.

Fig. 24.7. The process of the iris recognition for the large-scale iris recognition example.

Figure 24.7 shows the order of steps of the recognition process schematically. The image is acquired and digitized. The positions of the iris boundaries are localized in the image and approximated by concentric circles. The inner circle encircles a pupil and forms the inner boundary and the outer circle en-

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

13

circles the iris and represents the outer boundary. The region of interest (in Figure 24.7 marked as several rectangles) that participates in the recognition is chosen within the iris boundaries to avoid the influence of eyelashes and the reflectance of the camera flash. Then the region of interest of the template image of the iris is stored in the database together with the information describing the iris geometrical properties (the diameter and the center). Of course, the positions of the iris are usually different on the scanning device and the images are mutually rotated and shifted. These differences are not treated in the current simple implementation of the image preprocessing. The properties of the document matrix and LSI processing parameters are summarized in Table 24.2. 24.4.3 Results of the Large-Scale Database Query The quality of the recognition is usually quantified by two types of errors: The False Reject Rate (FRR) and The False Accept Rate (FAR). The FRR represents the percentage of the authentic objects that were rejected. FAR represents the percentage of the impostors that were accepted. FRR and FAR generally depend on the given similarity coefficient threshold of ρ. The lower the threshold ρ the higher is the probability of accepting the impostor. Similarly, the higher the threshold ρ the higher is the probability of rejecting the authentic object. The intersection of the “impostor” and the “authentic” distributions gives the value of the threshold ρ = ρ0 , where the sum FAR(ρ0 ) + FRR(ρ0 ) is minimal. Such distributions obtained from the similarity coefficients through the database are presented in the form of histograms in Figure 24.8 and Figure 24.9. Analyses of the intersection of the distributions gave the value of threshold ρ0 = 0.766 and the errors FAR = 1.12 %, FRR = 2.34 %, see Figure 24.10. Of course, the probability of incorrect iris recognition can be significantly reduced by a preprocessing step including a proper geometrical alignment of any two compared irides before the similarity coefficient is computed.

24.5 Conclusions This article presents the SVD-free Latent Semantic Indexing method for an automatic verification of persons by iris recognition. Although no special preprocessing of images and/or a-priori information is assumed, our numerical experiments indicate ability of the proposed algorithm to solve the large-scale iris recognition problems. Of course, the quality of irides images and proper localization influence the resulting errors. We have found that in our case the errors were caused by an inaccurate localization of images. The proper localization is a subject for future work.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

14

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel Different irides

3500

Absolute frequency

3000 2500 2000 1500 1000 500 0

−1

−0.5

0 sim

0.5

1

Fig. 24.8. The distribution of the similarity coefficient sim for the compared different irides through the database. The big majority of comparing has the similarity coefficient close to zero. This feature indicates the ability of the algorithm to properly recognize irides belonging to a different person. Same irides

250

Absolute frequency

200

150

100

50

0

0.4

0.5

0.6

sim

0.7

0.8

0.9

1

Fig. 24.9. The distribution of the similarity coefficient sim for the same irides through the database. The big majority of comparing has the similarity coefficient close to one. This feature indicates the ability of the algorithm to properly recognize irides belonging to the same person.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

On SVD-free LSI for Iris Recognition of Large databases

15

Detection Error Trade−off curve

10

FRR (%)

5

2

1

1

2

5 FAR (%)

10

20

Fig. 24.10. The Detection Error Trade-off curve (DET curve) was computed by using the DET-Curve Plotting software [16].

Acknowledgments We would like to thank very much to SIAM Linear Algebra 2003 community for valuable comments, especially Zlatko Drmaˇc (Croatia) and William Ferng (U.S.A.) for proposing the scaling of the document matrix. The research has been partially supported by the program ”Information Society” of the Academy of Sciences of the Czech Republic, project No. T401940412 ”Dynamic Reliability Quantification and Modelling”. The work leading to this chapter and future developments have been and will be partially supported by the EU under the IST 6th FP, Network of Excellence K-Space.

References 1. Daugman JG. Biometric Personal Identification System Based on Iris Analysis. US Patent No 5,291,560 US Government 1993;. 2. Daugman JG. High confidence visual recognition of persons by a test of statistical independence. IEEE Trans Pattern Analysis and Machine Intelligence 1993;15(11):11481161. 3. Daugman JG. Statistical richness of visual phase information: Update on recognizing persons by iris patterns. International Journal of Computer Vision 2001;45(1):25–38.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

16

Pavel Praks, Libor Machala, and V´ aclav Sn´ aˇsel

4. Daugman JG. The importance of being random: statistical principles of iris recognition. Pattern Recognition 2003;36(2):279–291. 5. Liam LW, Chekima A, Fan LCh, Dargham JA. Iris recognition using selforganizing neural network. In: Student Conference on Research and Development. Shah Alam, Malaysia; 2001. p. 169–172. 6. Machala L, Posp´ıˇsil J. Proposal and verification of two methods for evaluation of the human iris video-camera images. Optik 2001;112(8):335–340. 7. Roche DM, Sanchez-Avila C, Sanchez-Reillo R. Iris recognition for biometric identification using dyadic wavelet transform zero-crossing. IEEE Press 2001;p. 272–277. 8. Berry WM, Drmaˇc Z, Jessup JR. Matrices, Vector Spaces, and Information Retrieval. SIAM Review 1999;41(2):336–362. 9. Grossman DA, OFrieder. Information retrieval: Algorithms and heuristics. Second edition: Kluwer Academic Publishers; 2000. 10. Praks P, Dvorsk´ y J, Sn´ aˇsel V. Latent Semantic Indexing for Image Retrieval Systems. In: SIAM Conference on Applied Linear Algebra. The College of William and Mary, Williamsburg, USA, http://www.siam.org/meetings/la03/proceedings; 2003. ˇ 11. Praks P, Dvorsk´ y J, Sn´ aˇsel V, Cernohorsk´ y J. On SVD-free Latent Semantic Indexing for Image Retrieval for application in a hard industrial environment. In: IEEE International Conference on Industrial Technology ICIT 2003. IEEE Press, Maribor, Slovenia; 2003. p. 466–471. 12. Muller N, Magaia L, Herbst BM. Singular Value Decomposition, Eigenfaces, and 3D Reconstructions. SIAM REVIEW 2004;46(3):518545. ˇ ab O. Praks P. Information extraction from HTML 13. Sv´ atek V, Labsk´ y M, Sv´ product catalogues: coupling quantitative and knowledge-based approaches. In: Dagstuhl Seminar on Machine Learning for the Semantic Web. Ed. N. Kushmerick, F. Ciravegna, A. Doan, C. Knoblock and S. Staab. Research Center for Computer Science, Wadern, Germany, http://www.smi.ucd.ie/DagstuhlMLSW/proceedings/labsky-svatek-praks-svab.pdf; 2005. 14. Golub GH, Loan CFVan. Matrix Computations. Third edition: The Johns Hopkins University Press; 1996. 15. Dobeˇs M, Machala L. The database of Iris Images. http://phoenix.inf.upol.cz/iris: Palack´ y University, Olomouc, Czech Republic; 2004. 16. DET-Curve Plotting software for use with MATLAB. http://www.nist.gov/speech/tools: National Institute of Standards and Technology; 2000 – 2005.

PDF compression, OCR, web optimization using a watermarked evaluation copy of CVISION PDFCompressor

24. On SVD-free Latent Semantic Indexing for Iris ...

semantic analysis of large amount of text documents. The main ... model with a low-rank approximation of the original data matrix via the SVD or the other ... following two main advantages: (i) automatic noise filtering and (ii) natural clustering ...

304KB Sizes 2 Downloads 181 Views

Recommend Documents

Regularized Latent Semantic Indexing
optimization problems which can be optimized in parallel, for ex- ample via .... edge discovery, relevance ranking in search, and document classifi- cation [23, 35] ..... web search engine, containing about 1.6 million documents and 10 thousand.

Regularized Latent Semantic Indexing: A New ...
particularly propose adopting l1 norm on topics and l2 norm on document representations to create a model with compact and .... to constrain the solutions. In batch ... with limited storage. In that sense, online RLSI has an even better scalability t

Distributed Indexing for Semantic Search - Semantic Web
Apr 26, 2010 - 3. INDEXING RDF DATA. The index structures that need to be built for any par- ticular search ... simplicity, we will call this a horizontal index on the basis that RDF ... a way to implement a secondary sort on values by rewriting.

Indexing Dataspaces - Semantic Scholar
and simple structural requirements, such as “a paper with title 'Birch', authored ... documents, Powerpoint presentations, emails and contacts,. RDB. Docs. XML.

A Short Survey on P2P Data Indexing - Semantic Scholar
Department of Computer Science and Engineering. Fudan University .... mines the bound of hops of a lookup operation, and the degree which determines the ...

Shape Indexing and Semantic Image Retrieval Based on Ontological ...
Retrieval Engine by NEC USA Inc.) provides image retrieval in Web by ...... The design and implementation of the Redland RDF application framework, Proc.

Shape Indexing and Semantic Image Retrieval Based on Ontological ...
Center retrieves images, graphics and video data from online collections using color, .... ular class of image collection, and w(i,j) is semantic weight associated with a class of images to which .... Mn is defined by means of two coordinates (x;y).

A Short Survey on P2P Data Indexing - Semantic Scholar
Department of Computer Science and Engineering. Fudan University ... existing schemes fall into two categories: the over-DHT index- ing paradigm, which as a ...

JUST-IN-TIME LATENT SEMANTIC ADAPTATION ON ...
SPEECH RECOGNITION USING WEB DATA. Qin Gao, Xiaojun ... Development of World Wide Web makes it a huge data source. The Web .... as the decoding history changes, every access to the trigram probability requires (7) to be computed. In this work, Web da

Iris Data Indexing Method Using Gabor ieee.pdf
Whoops! There was a problem previewing this document. Retrying... Download. Connect more apps... Iris Data Ind ... abor ieee.pdf. Iris Data Ind ... abor ieee.pdf.

A Latent Semantic Pattern Recognition Strategy for an ...
Abstract—Target definition is a process aimed at partitioning the potential ...... blog texts and its application to event discovery,” Data Mining and Knowledge ...

LATENT SEMANTIC RATIONAL KERNELS FOR TOPIC ...
Chao Weng, Biing-Hwang (Fred) Juang. Center for Signal and Image Processing, Georgia Institute of Technology, Atlanta, USA. 1chao.weng,[email protected]. ABSTRACT. In this work, we propose latent semantic rational kernels. (LSRK) for topic spotti

LATENT SEMANTIC RETRIEVAL OF SPOKEN ...
dia/spoken documents out of the huge quantities of Internet content is becoming more and more important. Very good results for spoken document retrieval have ...

Polynomial Semantic Indexing - Research at Google
In particular, we will consider the following degree k = 2 model: ..... 5We removed links to calendar years as they provide little information while ... technology.

Review of Iris Recognition System Iris Recognition System Iris ... - IJRIT
Abstract. Iris recognition is an important biometric method for human identification with high accuracy. It is the most reliable and accurate biometric identification system available today. This paper gives an overview of the research on iris recogn

Indexing certain physiological parameters on drought management
S. VINCENT, R. RETHINARAJA AND S. RAJARATHINAM. Coconut Research Station, Veppankulam 614 906, Tamil Nadu. Abstract: The genotype of coconut East Coast Tall was studied for its tolerance of drought through the physiological parameters by using differ

A Survey on Efficiently Indexing Graphs for Similarity ...
Keywords: Similarity Search, Indexing Graphs, Graph Edit Distance. 1. Introduction. Recently .... graph Q, we also generate its k-ATs, and for each graph G in the data set we calculate the number of common k-ATs of Q and G. Then we use inequality (1)

moment restrictions on latent
In policy and program evaluation (Manski (1990)) and more general contexts ..... Let P = N (0,1), U = Y = R, Vθ = {ν : Eν(U)=0}, and Γθ (y) = {1} for all y ∈ Y, and ...