Image and Vision Computing 28 (2010) 744–753

Contents lists available at ScienceDirect

Image and Vision Computing journal homepage: www.elsevier.com/locate/imavis

Probabilistic learning for fully automatic face recognition across pose M. Saquib Sarfraz *, Olaf Hellwich Computer Vision and Remote Sensing, Berlin University of Technology, Sekr. FR 3-1, Franklinstr. 28/29, 10587 Berlin, Germany

a r t i c l e

i n f o

Article history: Received 8 March 2009 Received in revised form 19 July 2009 Accepted 26 July 2009

Keywords: Face recognition Recognition across pose Bayesian face modeling Face-GLOH-Signature

a b s t r a c t Recent pose invariant methods try to model the subject specific appearance change across pose. For this, however, almost all of the existing methods require a perfect alignment between a gallery and a probe image. In this paper we present a pose invariant face recognition method that does not require the facial landmarks to be detected as such and is able to work with only single training image of the subject. We propose novel extensions by introducing to use a more robust feature description as opposed to pixelbased appearances. Using such features we put forward to synthesize the non-frontal views to frontal. Furthermore, using local kernel density estimation, instead of commonly used normal density assumption, is suggested to derive the prior models. Our method does not require any strict alignment between gallery and probe images which makes it particularly attractive as compared to the existing state of the art methods. Improved recognition across a wide range of poses has been achieved using these extensions. Ó 2009 Elsevier B.V. All rights reserved.

1. Introduction Recent approaches to face recognition are able to achieve very low error rates in the context of frontal faces. A more challenging task is to recognize a face at a non-frontal view when only one (e.g. frontal) training image is available. Several previous studies have presented algorithms which can take a single probe image at one pose and attempt to match it to a single gallery image at a different pose. Pose variation in terms of pixel appearance, is highly non-linear in 2D, but linear in 3D. Recently [1] shows good results, in the presence of pose mismatch, by generating a full 3D head model of the subject based on just one image. A drawback of this, however, is a precise registration of the probe in order to guide the fitting process and moreover the computation involved is too restrictive for a practical face recognition system. From a practical stand point 3D technology cannot be easily retrofitted to existing applications, e.g. surveillance systems, that contains inherently 2D data. 2D methods, therefore, have to be further investigated for view independent recognition. An alternative approach in the 2D context is to treat this as a learning problem in which we aim to predict frontal images from non-frontal ones. The other approach in this context is to only model what is discriminative between images of the same subject in different poses and images of the different subjects in these poses. Lucey and Chen [2] categorized these into view-point gener-

* Corresponding author. E-mail addresses: [email protected], [email protected] (M. Saquib Sarfraz), [email protected] (O. Hellwich). URL: http://www.cv.tu-berlin.de (M. Saquib Sarfraz). 0262-8856/$ - see front matter Ó 2009 Elsevier B.V. All rights reserved. doi:10.1016/j.imavis.2009.07.008

ative and view-point discriminative approaches, respectively. The emphasis in view-point generative methods is to find a mapping function that can be used to generate a given non-frontal image to its frontal counterpart [7–9,1]. A simple distance metric is then utilized to compute similarity between a frontal gallery image and the transformed probe image. The assumption in generative methods, of finding a transformation function that can be used to generate a near perfect frontal image of its non-frontal counterpart in any pose, is weak and in practice the transformed images exhibit strong variations that accounts for degradation in recognition performance. The view-point discriminative approaches, on the other hand, has some inherent advantages over the viewpoint-generative approach as more emphasis is given to discrimination, rather than the generation of a gallery view image from the probe view appearance [5,3,4,2]. This, however, is a very naive assumption because it assumes that the appearance variations among different subjects across different pose mismatches is always larger than variations among same subject in different poses. This practically may not be true since the discrimination of appearance across large pose differences of the same subject becomes significant enough that it can not preserve the identity [6]. We can overcome these associated problems by first finding a generative function for each pose and then following a view-point discriminative approach to model the remaining associated appearance variations specific to each pose explicitly. The goal of such an approach is creating a model that can predict how a given face will appear when viewed at different poses. This seems a natural formulation for a recognition task especially in unconstrained scenarios. In this paper we develop this idea in a full Bayesian probabilistic setting.

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

In the context of the recent discriminative or generative appearance based methods the emphasis has been to directly model the local appearance change, due to pose, across same subjects and among different subjects. Differences exist among different methods, in how these models are built, but the goal of all is same, i.e. trying to approximate the joint probability of a gallery and probe face across different pose. Such an approach is particularly attractive in that it automatically solves the one training image problem in a principled way as these appearance models can be learned effectively from an offline database of representative faces. Another benefit of such a line of work lies in the fact that adding a new person’s image in the database does not require training the models again. We note, however, almost all of these methods proposed in literature until now intrinsically assume a perfect alignment between a gallery and probe face in each pose. This alignment is needed, because, otherwise, in current appearance-based methods it is not possible to discern between the change of appearance due to pose and change of appearance due to the local movement of facial parts across pose. In this contribution, we introduce novel methods and propose to build models on features that are invariant with respect to misalignments and thus do not require the facial landmarks to be detected as such. Our approach, briefly, is to first learn the generative functions to transform non-frontal views to frontal and then follow a discriminative approach to model remaining appearance variations between the frontal and transformed non-frontal poses. This is done by learning probabilistic models describing the approximated joint probability distribution of a gallery and probe image at different poses, when identity of a person is same and when it is different across pose. This is achieved by computing similarities between extracted features of faces at frontal and all other views. The distribution of these similarities is then used to obtain the likelihood functions of the form PðIg ; Ip jCÞ, where C refers to classes when the gallery Ig and probe Ip images are similar ‘S’ and dissimilar ‘D’ in terms of subject identity. For this purpose an independent generic set of faces, at views we want to model, is used for offline training. A contribution is made in this paper towards improved recognition performance across pose without the need of properly aligning gallery and probe images. To achieve this, we propose to use an extension of SIFT features [10], that are specifically adapted for the purpose of face recognition in this work. This feature description captures the whole appearance of a face in a rotation and scale invariant manner, and is shown robust with regards to variation of facial appearance due to localization problems. Furthermore, we propose to synthesize these features at non-frontal views to frontal by using multivariate regression techniques. The benefit of this in recognition performance is demonstrated empirically. Local kernel density estimation as opposed to commonly used Gaussian model is suggested for deriving prior models. 1.1. Related work Our contribution lies in the body of work that concerns estimating the joint likelihood for the purpose of recognition in the presence of pose mismatch. Here, therefore, we introduce the related existing work in this direction in order to put our work into the right context. There are three main methodologies in this domain. The first tries to model the joint likelihood function PðIg ; Ip jSÞ. The likelihood PðIg ; Ip jDÞ is typically omitted due to the complexity associated with its estimation. Due to the large dimensionality of the whole face, subspace methods based on PCA are employed to approximate the likelihood from a generic face dataset. The Tensorface [11] and Eigen Light Field [9] are very recent techniques that fall into this category. The second methodology attempts to model the differential appearance between gallery and probe images. An offline generic set of examples is used in order to make the approximation

745

PðIg ; Ip jCÞ ffi PðjIg  Ip jjCÞ. These likelihoods are attempting to model the whole face, for both similar and dissimilar classes by using absolute difference in pixels. The most well known method in using differential appearance has been the intra-personal and extra-personal approach of Moghaddam and Pentland [12]. The similar and dissimilar classes of the differential appearance likelihood are modeled through a normal distribution. These distributions are estimated within a subspace, found using PCA. Techniques centered on LDA [4] also employ a similar paradigm, in terms of differential appearance, although they are not framed within a strict probabilistic framework. The third methodology is to decompose the face into an ensemble of salient patches/regions. [13,14] reported superior recognition performance, in the presence of occlusion and expression variations, with respect to approaches that treat the face as a whole. Recently Kanade and Yamada [5] proposed an effective technique for pose invariant recognition within this framework. Their extension was centered on the hypothesis that individual patches/regions can be treated as independent and modeling the change of appearance of these small regions is more effective as opposed to the whole face appearance.Their approach, therefore, can be thought of a direct extension of the differential appearance paradigm in that they combine the differential appearance likelihoods of several local patches to approximate the joint likelihood. Some extensions to this approach have been reported in the literature [3,2] with improved performance. We argue, however, that the strong assumption of patch independence is not statistically right, since face is a highly symmetrical object and different regions of a face are not independent. Such an assumption, nonetheless, was needed in order to overcome the problems that arise, in modeling differential change of appearance across pose, in holistic appearance-based methods. Some recent proposals employs the similar paradigm by first transforming the problem into a pose invariant feature space. Two most recent successful attempts are image-based rendering of Chai et al. [25] and local feature based generative models of Prince et al. [26]. Chai et al. render a posed image to frontal by using image based regression. Their results although encouraging, suggests that their method may not be useful when the images are not properly aligned and there exist appearance variations as typically expected from a face detector output. The reason for that is, their method works directly on the image pixels and therefore sensitive to such variations. Prince et al. on the other hand use a pose contingent linear transform of the extracted features by learning a generative density model across each pose. They use 21 manually marked points on the face to extract local features and then model their variation through a Gaussian density. The general framework of our proposal is similar to their method, however the main differences are: (1) We use a whole-face representation which is robust against misalignments and hence does not need proper alignment or detecting some key points on the face. (2) We use multivariate linear regression to learn the generative function for each pose. (3) The variation of frontal and transformed posed features is modeled by estimating a kernel density directly. Prince et al. reports the best performance across pose so far, their method however, relies heavily on detecting several fiducial points on the face. For a fully automatic face recognition system such a requirement is hard to met in practice, because detecting several fiducial points automatically on a facial image in different poses is not always possible. This is largely due to the fact that when a face rotates from frontal to left or right profile view, the appearance of individual facial parts also change considerably and some parts simply disappear, e.g. when a face moves from frontal to right profile the left eye will not be visible any more and hence, any methods that try to detect center of both eyes will eventually fail. In this paper, we suggest that one should, instead, derive such whole-face appearance representa-

746

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

tions that are easily tractable across pose and can take into account, to certain extent, the change of appearance of different parts of face due to pose. 2. Modeling whole-face appearance change across pose Our approach is to extract whole appearance of the face in a manner which is robust against misalignment. For this we use feature description [15] that is specifically adapted for the purpose of face recognition in this work. It models the local parts of the face and combines them into a global description. 2.1. Feature extraction: the Face-GLOH-Signature Commonly used facial representations are related directly to pixel intensities and, as such, are not invariant to changes in scale, position, orientation, brightness and contrast of a face [16]. Since these types of transformations are to be expected after a face detector stage, alignment by using several facial landmarks is needed. We propose to use a representation based on gradient location-orientation histogram (GLOH) [15], which is more sophisticated and is specifically designed to reduce in-class variance by providing some degree of invariance to the aforementioned transformations. GLOH features are an extension to the descriptors used in the scale invariant feature transform (SIFT) [10], and have been reported to outperform other types of descriptors in object recognition tasks [15]. The extraction procedure has been specifically adapted to the task of face recognition and will be described in the remainder of this section. The extraction process begins with the computation of scale adaptive spatial gradients for a given image Iðx; yÞ. These gradients are given by

clude a central region and seven radial sectors. The radius of the central region is chosen to make the areas of all partitions equal. Each partition is then processed to yield a histogram of gradient magnitude over gradient orientations. The histogram for each partition has 16 bins corresponding to orientations between 0 and 2p, and all histograms are concatenated to give the final 128-dimensional feature vector which we term as Face-GLOH-Signature, see Fig. 1d. The dimensionality of the feature vector depends on the number of partitions used. A higher number of partitions results in a longer vector and vice versa. The choice has to be made with respect to some experimental evidence and the effect on the recognition performance. We have assessed the recognition performance on a validation face dataset. By varying the partition sizes from 3 (one central region and two sectors), 5, 8, 12 and 17, we found that increasing number of partitions results in degrading performance especially with respect to misalignments while using coarse partitions affects recognition performance with more pose variations. Based on the results, eight partitions seem to be the optimal choice and a good trade off between achieving better recognition performance and minimizing the effect of misalignment [6]. It should be noted that, in practice, the quality of the descriptor improves when care is taken to minimize aliasing artifacts. The recommended measures include the use of smooth partition boundaries as well as a soft assignment of gradient vectors to orientation histogram bins. 2.2. Generative pose models for synthesizing features

ð2Þ

As noted earlier, following a view-point discriminative approach directly by modeling the appearance across each pose may not preserve the identity in general, across large pose differences, especially when we do not assume a strict alignment between images. We therefore, synthesize the obtained feature vectors at non-frontal views to frontal by finding a generative function for each pose and then follow a view-point discriminative approach to model the remaining associated appearance variations specific to each pose explicitly. Computing similarities using these synthesized features between frontal and other poses provides us with prior distribution for each pose.

The gradient magnitudes obtained for two example images (Fig. 1e) are shown in Fig. 1b. The gradient image is then partitioned on a grid in polar coordinates, as illustrated in Fig. 1c. The partitions in-

2.2.1. Synthesizing features at non-frontal views to frontal It is well known that when a large number of subjects are considered, the recognition performance of appearance-based

rxy 

X

pffiffi wðx; y; tÞ t rtxy Lðx; y; tÞ

ð1Þ

t

where Lðx; y; tÞ denotes the linear Gaussian scale space of Iðx; yÞ [17] and wðx; y; t) is a weighting obtained as

pffiffi j trtxy Lðx; y; tÞj4 wðx; y; tÞ ¼ P pffiffi t 4 t j t rxy Lðx; y; tÞj

Fig. 1. Face-GLOH-Signature: (a, b) Gradient magnitudes; (c) polar-grid partitions; (d) 128-dimentional feature vector; (e) example image.

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

methods deteriorates significantly. It is due to the fact that distribution of face patterns is no longer convex as assumed by linear models. By transforming the image in the previous section into a scale and rotation invariant manner, we assume that there exists a certain relation between these features of frontal and posed image that we can linearly transform. We justify this assumption by comparing the similarity distributions estimated from non-synthesized features and synthesized features. One simple and powerful way of relating these features is to use the regression techniques. Let us suppose that we have the following multivariate linear regression model, for finding relation between the feature vectors of frontal IF and any other angle IP .

IF ¼ IP B 2 3 2 ~ ~ IT IF1T 6 7 6 p1 6 .. 7 6 . 6 . 7 ¼ 6 .. 4 5 4 ~ ~ IF Tn ITpn

32 bð1;1Þ 76 76 .. 74 . 5 bðDþ1;1Þ 1 1

3 bð1;DÞ 7 .. 7 . 5 . . . bðDþ1;1Þ ... .. .

ð3Þ

where n > D þ 1, with D being the dimensionality of each ~ IF and ~ IP . B is a pose transformation matrix of unknown regression parameters, under the sum-of-least-squares regression criterion, B can be found using Moor–Penrose inverse.

B ¼ ðITP IP Þ1 ITP IF

Ig  Ip kIg k2  kIp k2

ð7Þ

where k  k denotes the Euclidean norm of the vectors. Fig. 3 depicts the histograms for the prior same and different distributions of the similarity c, for gallery and probe images across a number of pose mismatches. These distributions depicted here are obtained by using images of half of the subjects in nine main poses from PIE database [22]. To make the estimation of pose transformation matrix feasible, 200 images in each pose of these subjects (illumination and expression variants) are used. The images used in our evaluation are cropped from the database without using any commonly employed normalization procedure. The face images therefore contain typical variations that may arise from miss localization such as back ground, part clippings and scale. Example images at frontal and four pose differences of a subject from PIE database are shown in Fig. 2. Note that, in Fig. 3, the more separated the two distributions are the more discriminative power it has to tell if the two faces are of the same person or not for that particular pose. It is clear that the discriminative power decreases as the pose moves away from frontal. As shown in Fig. 3, synthesizing features to frontal significantly improves this discrimination ability over a wide range of poses.

ð4Þ

This transformation matrix B is found for each of the poses IP (e.g. ±22.5°, ±45°, ±65°, ±90°) with frontal 0° IF . Given a set of a priori feature vectors representing faces at frontal IF and other poses IP , we can thus find the relation between them. Any incoming probe feature vector can now be transformed to its frontal counterpart using:

IbP ¼ BP  ½ITP

cðIg ; Ip Þ ¼

747

1T

ð5Þ

3. Obtaining prior appearance models for recognition The likelihood of joint occurrence of a probe and gallery face at different poses is obtained by using an offline generic set of faces at views we want to model. These models explicitly describe the appearance variations between frontal and other poses for when the identity is same and when it is different. These prior models can be used to compute a match score between an online probe and gallery image in a Bayesian setting. We approximate the joint likelihood of a probe and gallery face as

3.1. Local kernel density estimation In order to compute Pðcpg jS; /p Þ and Pðcpg jD; /p Þ (note that angle /g is typically omitted since the gallery pose is fixed to frontal), i.e. conditional probabilities describing similarity distributions when subject identity is same ‘S’ and when it is different ‘D’, these distributions must be described by some form. The most common assumption is the Gaussian. Functional density estimators like the Gaussian assume a functional form of the distribution and therefore depend heavily on the accuracy of that assumption. It is especially deceitful that a functional estimation always ‘looks’ correct, no matter how poor the assumption is for the underlying distribution. In [18] authors noted that describing such prior appearance models by a normal density is not optimal and results in biased recognition results. We also note that employing a normal density results in a poor fit, see Fig. 4. We therefore propose to use local kernel density estimate.

PðcÞ ¼

N c  c  1 X i k n Nr i¼1 r

ð8Þ

2

PðIg ; Ip jC; /g ; /p Þ  Pðcpg jC; /g ; /p Þ

ð6Þ

where ‘C 2 fS; Dg’ refers to classes when the gallery Ig and probe IP images are similar ðSÞ and dissimilar ðDÞ in terms of subject identity. / is the pose angle for the corresponding gallery and probe image and cpg is the similarity between gallery and probe image. These likelihoods for the similar and dissimilar class are then found by modeling the distribution of similarities of extracted features between frontal and every pose from offline training set. Cosine measure is used as a similarity metric.

where kðmÞ ¼ ð2p1 Þn em =2 , there exist various methods to automatically estimate appropriate values for the width r of the kernel function. In this work, we simply set r to be the average nearest neighbor distance:

r2 ¼

N 1X min jc  cj j2 N i¼1 i–j i

ð9Þ

As depicted in Fig. 4, the kernel density estimate is a better fit, this is because the assumption of Gaussian distribution in such scenar-

Fig. 2. Example images of a subject at five poses.

748

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

Fig. 3. x-Axis denotes the similarity measure c and y-axis denotes the density approximation. First row depicts histograms for the same and different classes on nonsynthesized features across four pose mismatches (see Fig. 5 for the approximate pose angles). Second row depicts the kind of separation and improvement we get by using feature synthesis.

ios is generally not fulfilled. Kernel density estimator, on the other hand, is known to approximate arbitrary distributions [19]. 3.2. Recognition across pose Obtained likelihood estimates Pðcpg jS; /p Þ and Pðcpg jD; /p Þ in the previous section, can now be directly used to compute the posterior probability. For a probe image IP at some pose /p , of unknown identity, we can now decide if it is coming from the same subject as gallery Ig , with each of the gallery image, by using this posterior as a match score. Employing these likelihoods, using Bayes rule, we write

PðSjcpg ; /p Þ ¼

Pðcpg jS; /p ÞPðSÞ Pðcpg jS; /p ÞPðSÞ þ Pðcpg jD; /p ÞPðDÞ

ð10Þ

Since the pose /p of the probe image is in general not known, we can marginalize over it. In this case the conditional densities for similarity value cpg can be written as

Pðcpg jSÞ ¼

X

Pð/p ÞPðcpg jS; /p Þ

ð11Þ

p

PðSjcpg Þ ¼

Pðcpg jSÞPðSÞ Pðcpg jSÞPðSÞ þ Pðcpg jDÞPðDÞ

ð13Þ

If no other knowledge about the probe pose is given, one can assume the pose prior Pð/p Þ to be uniformly distributed. We, however, use the pose estimates for a given probe face by our developed front-end pose estimation procedure [20,21]. The pose estimation system provides us with probability scores for each pose that can be used directly as priors Pð/p Þ in Eqs. (11) and (12). Due to a reasonably high accuracy of these pose estimates, these probabilities can act as very strong priors and thus increase the chances of a probe to be recognized correctly. Note that this front-end pose estimation system also does not require any kind of land mark detection for geometric normalization of the detected face windows in order to estimate the pose. We compute this posterior for an unknown probe image with all of the gallery images and choose the identity of the gallery image with the highest score as recognition result. 4. Experiments and recognition results

and

Pðcpg jDÞ ¼

X

Pð/p ÞPðcpg jD; /p Þ

ð12Þ

p

Similar to the posterior defined in Eq. (10), we can compute the probability of the unknown probe image coming from the same subject (given similarity cpg ) as

In this section we test the method and models developed in the previous sections on two of the large databases, i.e. CMU-PIE database [22] and FERET [23] database. Where Images of half of the subjects from PIE and FERET are used as generic offline set for training of the models, while remaining are used for gallery and probe sets.

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

749

Fig. 4. First row shows fitting a normal density, second row shows the kernel density fits on the distribution of similarities obtained previously.

4.1. Experimental setup The pose subsets of both the databases are used in our experiments. From CMU-PIE all images of 68 subjects imaged under 13

poses in neutral expression are used, see Fig. 5. The approximate pose difference between images of the same subject is 22.5° varying from frontal 0° to ±90° profile. FERET set includes images of 200 subjects in nine pose variations with an approximate pose differ-

Fig. 5. Thirteen poses covering left profile (9), frontal (1) to right profile (5), and slightly up/down tilt in pose 10, 11, 13 and 12 of corresponding poses in 8, 1 and 4, respectively.

750

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

ence of 15°, varying from 0° frontal to ±60° profile view. We test our system for the automatic case without any manual localization. We, therefore, use Viola–Jones face detector [24] to automatically localize faces. Detected face windows are then cropped from the database without employing any commonly used normalization procedure. Therefore images contain typical variations that may arise due to miss-localization like scale, part clippings and background. All images are then resized to 128  128 pixels. Typical variations present in the database are depicted in few of the example images in Fig. 6. Note that, since we do not employ any kind of normalization such as detecting and fixing eye location or eye distance, face images across pose suffer from typical misalignments. As mentioned, half of the subjects are used for training the models, this amount to images of 32 subjects for the experiments on PIE database and 100 subjects for the experiments on FERET. As the gallery, the frontal images of all the subjects are used. For the tests on PIE database, since we do not assume any alignment between gallery and probe images, therefore models are trained for the main nine poses, i.e. poses 1–9 in Fig. 5. While pose 10, 11, 12 and 13 corresponding to up/down tilt of the face are treated as the variations due to misalignment for corresponding poses in the test set. All 13 poses for a subject in the test set are therefore considered. Class priors PðSÞ and PðDÞ are set to PðSÞ 1 and PðDÞ ¼ 1  PðSÞ in all of our experiments. 4.2. Test results We provide results of several experiments demonstrating the effectiveness of our method. The first set of results are obtained by using PIE database followed by experiment on FERET database. 4.2.1. Experiment 1: Known probe pose For our first experiment, we assume probe pose to be known and therefore use Eq. (10) to compute the posterior. In order to show the effectiveness of our method, we include the results of Kanade and Yamada’s [5] Bayesian Face Sub-region ‘BFS’ algorithm

Fig. 6. Examples of detected face windows depicting typical variations due to misalignments, e.g. scale, part clipping, background, etc.

and Eigenface algorithm [12] for comparison, where the former is considered since our method is similar to it in the principled approach while Eigenface is included as it is the common benchmark in facial image processing. Results are reported on PIE database. In order to use BFS on our dataset, the face image is divided into 32  32 pixel overlapping patches, with an overlap of 16 pixels. This is done since we do not assume aligned probe and gallery with respect to eyes and mouth positions. Prior models are obtained as described in the original paper [5] by using sum of square difference measure for each patch. Our results, as depicted in Fig. 7, show the robustness of our method against misalignments between probe and gallery while the results of BFS are much worse on misaligned images as compared to what they originally reported on the same database, see Table 1.

4.2.2. Experiment 2: Unknown probe pose For our second experiment, we report results for the fully automatic case, where pose of the probe images is not known a priori. For an incoming probe image, we extract Face-GLOH-Signature as described in Section 2.1. In order to synthesize these features to frontal we need to know the probe pose, as we have to use the corresponding pose transformation matrix B. Since we use the frontend pose estimation step that provides us with the probabilities for different possible poses, we therefore use marginalization (Eqs. (11) and (12)) by transforming the extracted feature vector of given probe to frontal for all poses. Note that, still using marginalization after a pose estimation step may seem counterproductive at first but since our system is based on models learned from synthesized features, and as shown in Fig. 3, the distributions depicting the similar class are almost same for the nearest pose mismatches, therefore this in fact improves the recognition performance in most cases. This is due to the fact that these pose prior probabilities, obtained from the pose estimation system, act as weights and they are only high for the nearest poses. As shown in Fig. 8 performance of our method with marginalization, by using strong pose priors, improves recognition accuracy. Note that, if pose estimation step has not been performed and pose priors are assumed equal, the performance of our method will be compromised. This is due to the fact that marginalization, in this case, will be introducing more noise due to the equal weighing of erroneous pose transformations.

Fig. 7. Recognition performance for each of the 13 poses for the test set. Results of our method and comparison with BFS and Eigenface for known probe pose (see Fig. 5 for the corresponding poses).

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

4.2.3. Experiment 3: Comparison with and without feature synthesis We compare the performance with and without feature synthesis. Fig. 9 shows the performance gain achieved by using feature synthesis. As much as 20% of performance gain is observed for probe poses moving away from frontal. These results provide an insight into the effectiveness of the learned generative pose models. 4.2.4. Experiment 4: Evaluation on FERET Following a similar procedure, the results on FERET are summarized here. We use 100 subjects as probe where frontal images of all 200 subjects are used as gallery. Note that there is a significant pose difference between FERET and PIE images, the average pose difference across FERET is 15°. The results reported here are obtained in a fully automatic setting where probe poses are obtained by using marginalization. The recognition accuracy up till ±45° of pose difference is above 90%. The overall average recognition accuracy on FERET is 92.1%. Fig. 10(a) plots the average recognition accuracy across each probe pose. 4.2.5. Experiment 5: Recognition across databases For all experiments shown so far the training and gallery/probe subjects were taken from the same database. Here we demonstrate the generalization ability of the learned models across different databases. For this purpose prior models are learned by using FER-

751

ET database and tested on PIE database. Note, however, there is a significant pose difference among the two databases. In order to cope with that we use 7 PIE poses (pose 1, 2, 3, 4, 6, 7 and 8) that loosely correspond to the corresponding FERET poses (0°, ±25°; ±45°, ±60°). In Fig. 10b we show recognition accuracies for tests with seven poses of the PIE database. Training for both prior appearance models and pose generative model is performed by using all the 200 subjects of the FERET database. Correspondence between FERET and PIE poses is determined manually. The results indicate the good generalization ability of our method. The overall accuracy is 87.7%, that compares favorably with the results obtained previously using the same respective database for training and testing. 4.3. Comparison with contemporary methods We summarize and compare the average recognition accuracy of our method with that of some of the most representative algorithms that achieved state-of-the art results on face identification studies on the same databases so far. When comparing identification results, one should keep in mind the over restrictive assumptions behind these methods that hinders a direct generalization of these methods to fully automatic case. In particular the degree of manual intervention should be noted. Almost all of the previous studies rely on manual localization of some points on the face in order to align images or to establish a direct correspondence to model local patches around these points. Our method does not re-

Fig. 8. Comparison of recognition performance with and without marginalization (see Fig. 5 for the corresponding poses).

Fig. 9. Comparison of our method for with and without feature synthesis (results reported here are obtained using marginalization).

Fig. 10. (a) Average recognition accuracy across each pose on FERET. (b) Recognition accuracy for test on seven PIE poses. Prior models are obtained using all the FERET subjects and tested using only the PIE subjects.

752

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753

Table 1 Comparison with state-of-the-art face identification studies across pose. Method

Alignment

Database

Pose diff.

% Correct

Yamada [5] Gross et al. [9] Gross et al. Blanz et al. [1] Chai et al. [25] Prince et al. [26] Prince et al. Our Method [27]

3-point 3-points 39-points 11-points 3-point 21-point 21-point No alignment

CMU-PIE FERET CMU-PIE FRVT CMU-PIE CMU-PIE FERET CMU-PIE

Our method

No alignment

FERET

Average all 13 Average all 9 Average all 13 45° 23°/45° 23°/67.5° 15°/45° Average all 13 23°/45°/67.5° Average all 9 15°/45°/60°

81 75 78.8 86 98.5/89.7 100/91 100/99 80.7 91.5/87.9/81 92.1 100/90.3/82.5

quire any manual registration and is fully automatic in this sense. With these considerations in mind, we present a summary of identification performance from other studies in Table 1. 4.4. Discussion Our results compares favorably with the previous approaches. Gross et al. (Eigen light filed method) [9] reports an overall 75 percent first-match results over 100 subjects from the FERET database, by using three manually marked feature points. Our system achieves 92.1% performance, with out manual registration. In the same study, they also report 39% and 93% performance for the PIE database conditions 67.5° and 23°, respectively, with a large number (>39) of manually labeled key points. For the same conditions, we report 81% and 91.5%, respectively, with no annotation. Kanade and Yamada’s BFS method achieves an average of approximately 81% performance on PIE database. Their method however, is sensitive to the manually annotated points on the face. As has been shown in our experiments (Fig. 7) the performance of their method is much worse when we do not assume any alignment. Blanz et al. [1] report results for a test database of 87 subjects with a horizontal pose variation of 45° from the Face Recognition Vendor Test FRVT 2002 [28] database, using, on the average, 11 manually established feature points. They investigate estimating the 3D model and creating a frontal image to compare to the test database (86.25% correct). Our system produces better performance at larger pose differences for comparable databases. Probably the best results reported up till now are those of Chai et al. [25] and Prince et al. [26]. Both of the methods use a similar pose contingent linear transformation of non-frontal views to frontal, where Chai et al. synthesize raw pixels and thus it cannot generalize directly to the automatic case where typical variations due to miss localization are expected, while Prince et al. transforms the local features extracted from manually located 21 points on the face image (at a different pose) to frontal and then model the variation of the corresponding local features across pose, their method therefore puts a hard constraint on the precise correspondence of these points across each pose. The comparison in Table 1 shows that our method is able to achieve comparable or better results in a fully automatic sense even without the need of properly aligning the gallery and probe images. That is especially attractive in the context of fully automatic pose invariant face recognition. 5. Conclusion We have presented a pose invariant face recognition method that requires only a single image of the person to be recognized in the gallery. The proposed approach is centered on modeling joint appearance of gallery and probe images across pose in a Bayesian framework. We have proposed novel methods in this direction by introducing to use a more robust feature description

as opposed to pixel-based appearances. The variation of these features across pose is modeled by a multivariate regression approach. Furthermore using kernel density estimate, instead of commonly used normal density assumption, is proposed to derive the prior models. Our method does not require any strict alignment between gallery and probe images and that makes it particularly attractive as compared to the existing state of the art methods. Several experiments and comparisons with previous state-ofthe-art approaches, demonstrates the effectiveness and weakness of the proposed approach. Our method is able to achieve above 80% of performance within a pose difference of approximately 65°. The performance of our system, as depicted by experiments on PIE database, on full profile views, i.e. 90° of pose difference is around 50%. A relatively low performance on these conditions depends on a number of factors and among others suggests that using a linear model for pose transformation is not able to cope well with these extreme pose differences. Note that when a probe is at frontal, the scores are 100% since exactly the same images are used in the gallery for frontal pose. The results reported here are for the fully automatic case, where faces are localized by using a face detector and we do not assume probe pose to be known. Our results show that one can achieve comparable performance without requiring the facial landmarks to be detected. Current methods rely on this information for the purpose of registration. For the purpose of automatic pose invariant face recognition, this is a major bottleneck for the current methods. Our approach tries to lift off this barrier and works directly on the output of a face detector. Although, we have presented results by using gallery as fixed at frontal pose, we note that it is straight forward to use our method for any pose as gallery.

References [1] V. Blanz, P. Grother, P.J. Phillips, T. Vetter, Face recognition based on frontal views generated from non-frontal images, CVPR (2005) 454–461. [2] S. Lucey, T. Chen, A viewpoint invariant, sparsely registered, patch based, face verifier, IJCV (January) (2008). [3] X. Liu, T. Chen, Pose-robust face recognition using geometry assisted probabilistic modeling, in: International Conference on Computer Vision and Pattern Recognition (CVPR) vol. 1, 2005, pp. 502–509. [4] T. Kim, J. Kittler, Locally linear discriminant analysis for multimodally distributed classes for face recognition with a single model image, IEEE Transactions on Pattern Analysis and Machine Intelligence 27 (3) (2005) 318–327. [5] T. Kanade, A. Yamada, Multi-subregion based probabilistic approach towards pose-invariant face recognition, in: IEEE International Symposium on Computational Intelligence in Robotics Automation, vol. 2, 2003, pp. 954–959. [6] M. Saquib Sarfraz, Towards automatic face recognition in unconstrained scenarios, Ph.D. Dissertation, Technische Universitat Berlin, 2008. urn:nbn:de: kobv:83-opus-20689. [7] D. Beymer, T. Poggio, Face recognition from one model view, in: International Conference on Computer Vision, 1995. [8] W. Zhao, R. Chellappa, SFS based view synthesis for robust face recognition, in: International Conference on Automatic Face and Gesture Recognition, 2000, pp. 285–292. [9] R. Gross, I. Matthews, S. Baker, Appearance-based face recognition and lightfields, IEEE Transactions on Pattern Analysis and Machine Intelligence 26 (4) (2004) 449–465. [10] D. Lowe, Distinctive image features from scale-invariant keypoints, International Journal of Computer Vision 2 (60) (2004) 91–110. [11] M.A.O. Vasilescu, D. Terzopoulos, Multilinear analysis of image ensembles: TensorFaces, ECCV 2350 (2002) 447–460. [12] B. Moghaddam, A. Pentland, Probabilistic visual learning for object recognition, IEEE Transactions on PAMI 19 (7) (1997) 696–710. [13] R. Brunelli, T. Poggio, Face recognition: features versus templates, IEEE Transactions on PAMI 15 (10) (1993) 1042–1052. [14] A.M. Martinez, Recognizing imprecisely localized, partially occluded, and expression variant faces from a single sample per class, IEEE Transactions on PAMI 24 (6) (2002) 748–763. [15] Mikolajczyk, C. Schmid, Performance evaluation of local descriptors, PAMI 27 (10) (2005) 31–47. [16] M.S. Sarfraz, M. Jger, O. Hellwich, Performance analysis of classifiers on face recognition, in: Proc. of the 5th IEEE Advances in Cybernetics Systems AICS Conference, 2006, pp. 255–264. [17] T. Lindeberg, Feature detection with automatic scale selection, International Journal of Computer Vision 30 (2) (1998) 79–116.

M. Saquib Sarfraz, O. Hellwich / Image and Vision Computing 28 (2010) 744–753 [18] S. Lucey, T. Chen, Learning patch dependencies for improved pose mismatched face verification, in: Proc. of the IEEE Int’l Conf. on Computer Vision and Pattern Recognition, vol. 1, 2006, pp. 17–22. [19] BW. Silverman, Density Estimation for Statistics and Data Analysis, Chapman & Hall, London, 1992. [20] M.S. Sarfraz, O. Hellwich, Head pose estimation in face recognition across pose scenarios, in: Int. Conference on Computer Vision Theory and Applications VISAPP, vol. 1, January 2008, pp. 235–242. [21] M.S. Sarfraz, O. Hellwich, An efficient front-end facial pose estimation system for face recognition, International Journal of Pattern Recognition and Image Analysis, dist. by Springer 18 (3) (2008) 434–441. [22] T. Sim, S. Baker, M. Bsat, The CMU pose, illumination and expression (PIE) database, in: Proc. of the Fifth IEEE FG, May 2002, pp. 46–51. [23] P.J. Phillips, H. Moon, S.A. Rizvi, P.J. Rauss, The FERET evaluation methodology for face-recognition algorithms, IEEE Transactions on Pattern Analysis and Machine Intelligence 22 (10) (2000) 1090–1104.

753

[24] M.J. Jones, P. Viola, Fast multiview face detection, Technical report TR2003-96 MERL, 2003. [25] X. Chai, S. Shan, X. Chen, W. Gao, Locally linear regression for pose-invariant face recognition, IEEE Transactions in Image Processing 16 (7) (2007) 1716–1725. [26] S.J.D. Prince, J.H. Elder, J. Warrell, F.M. Felisberti, Tied factor analysis for face recognition across large pose differences, IEEE Transactions on PAMI 30 (6) (2008) 970–984. [27] M.S. Sarfraz, O. Hellwich, Statistical appearance models for automatic pose invariant face recognition, in: Proc. of the 8th IEEE Int. Conference on Face and Gesture Recognition, FG September 2008. [28] FRVT 2002: Overview and summary, P. Phillips, P. Grother, R. Micheals, D. Blackburn, E. Tabassi, J. Bone, 2003. Available at .

Probabilistic learning for fully automatic face ...

We propose novel extensions by introducing to use a more robust feature description as opposed to pixel- based appearances. Using such features we put forward ..... Thirteen poses covering left profile (9), frontal (1) to right profile (5), and slightly up/down tilt in pose 10, 11, 13 and 12 of corresponding poses in 8, 1 and 4,.

2MB Sizes 1 Downloads 199 Views

Recommend Documents

Fully Automatic hp-Adaptivity for Electromagnetics. Application to the ...
variable Order of approximation Supporting anisotropy and hanging ... adaptive) hp-strategy applied to the analysis of H-plane ..... Computer Methods in Applied.

Fully Automatic hp-Adaptivity for Electromagnetics. Application to the ...
adaptive) hp-strategy applied to the analysis of H-plane and E-plane ..... 3 -, "A two-dimensional hp-adaptive finite element package for electromagnetics ...

FULLY AUTOMATIC GOAL-ORIENTED hp ...
also the same lines of code. Comparative 2D numerical .... We define Php : V −→ Vhp ..... 2Dhp90: A Fully automatic hp-adaptive Finite Element code. Figure 4: ...

A fully automatic method for the reconstruction of ...
based on a mixture density network (MDN), in the search for a ... (pairs of C and S vectors) using a neural network which can be any ..... Recovery of fundamental ...

Hybrid Generative/Discriminative Learning for Automatic Image ...
1 Introduction. As the exponential growth of internet photographs (e.g. ..... Figure 2: Image annotation performance and tag-scalability comparison. (Left) Top-k ...

FULLY AUTOMATIC hp-ADAPTIVITY IN THREE ...
way, will deliver exponential convergence rates in terms of error vs. the number of degrees-of-freedom .... 3 with faces f and edges e. The idea of the projection-based interpolation is based on three ..... we call, the direction of re£nement: h or

Active Learning for Probabilistic Hypotheses Using the ...
Department of Computer Science. National University of Singapore .... these settings, we prove that maxGEC is near-optimal compared to the best policy that ...

Automatic Face Annotation in News Images by ...
Google Images, Bing Images and Yahoo! Image Search. Our .... Third, for each name n and each face f a procedure is executed in order to compute a ...

Trusted Machine Learning for Probabilistic Models
Computer Science Laboratory, SRI International. Xiaojin Zhu. [email protected]. Department of Computer Sciences, University of Wisconsin-Madison.

Learning Probabilistic Relational Dynamics for Multiple ...
Algorithms have been developed for learning relational probabilistic planning ...... For a more detailed explanation, see the master's thesis by Deshpande [2007].

Osborne - Probabilistic numerics for deep learning - DLSS 2017.pdf ...
Sign in. Page. 1. /. 28. Loading… Page 1 of 28. http://rspk.paksociety.com. http://rspk.paksociety.com. http://rspk.paksociety.com. Page 1 of 28. Page 2 of 28. Page 2 of 28. Page 3 of 28. Page 3 of 28. Osborne - Probabilistic numerics for deep lear

Unsupervised Learning of Probabilistic Grammar ...
1Department of Statistics, 3Psychology and 4Computer Science .... Thirdly, the approach is able to learn from noisy data (where half of the training data is only ...

Learning Auxiliary Dictionaries for Undersampled Face ...
Jul 17, 2013 - dictionary updates. • Sparse Coding for Updating X. ➢ Given fixed A, we rewrite our formulation as: where. ➢ The above is the standard L1-minimization problem. ➢ Can be solved via techniques like Homotopy, Iterative Shrinkage-.

Unsupervised Learning of Probabilistic Grammar ...
Index Terms. Computer Vision, Structural Models, Grammars, Markov Random Fields, Object Recognition .... Ten of the object categories from Caltech 101 which we learn in this paper. of PGMMs. Section (V) ...... Ketch Laptop. LeopardsM.

Learning Prototype Hyperplanes for Face Verification in the Wild
(referred to as the prototype hyperplane) of one Support Vector. Machine (SVM) model, in which a sparse set of support vectors is selected from the unlabeled ...

LNCS 4233 - Fast Learning for Statistical Face Detection - Springer Link
Department of Computer Science and Engineering, Shanghai Jiao Tong University,. 1954 Hua Shan Road, Shanghai ... SNoW (sparse network of winnows) face detection system by Yang et al. [20] is a sparse network of linear ..... International Journal of C