scispace - formally typeset
Search or ask a question

Showing papers on "Three-dimensional face recognition published in 2005"


Proceedings ArticleDOI
15 Oct 2005
TL;DR: It is shown that the direct 3D counterparts to commonly used 2D interest point detectors are inadequate, and an alternative is proposed, and a recognition algorithm based on spatio-temporally windowed data is devised.
Abstract: A common trend in object recognition is to detect and leverage the use of sparse, informative feature points. The use of such features makes the problem more manageable while providing increased robustness to noise and pose variation. In this work we develop an extension of these ideas to the spatio-temporal case. For this purpose, we show that the direct 3D counterparts to commonly used 2D interest point detectors are inadequate, and we propose an alternative. Anchoring off of these interest points, we devise a recognition algorithm based on spatio-temporally windowed data. We present recognition results on a variety of datasets including both human and rodent behavior.

2,699 citations


Journal ArticleDOI
Seong G. Kong1, Jingu Heo1, Besma Abidi1, Joonki Paik1, Mongi A. Abidi1 
TL;DR: This paper provides an up-to-date review of research efforts in face recognition techniques based on two-dimensional images in the visual and infrared (IR) spectra.

650 citations


Journal ArticleDOI
TL;DR: The result is an efficient and accurate face recognition algorithm, robust to facial expressions, that can distinguish between identical twins and compare its performance to classical face recognition methods.
Abstract: An expression-invariant 3D face recognition approach is presented. Our basic assumption is that facial expressions can be modelled as isometries of the facial surface. This allows to construct expression-invariant representations of faces using the bending-invariant canonical forms approach. The result is an efficient and accurate face recognition algorithm, robust to facial expressions, that can distinguish between identical twins (the first two authors). We demonstrate a prototype system based on the proposed algorithm and compare its performance to classical face recognition methods. The numerical methods employed by our approach do not require the facial surface explicitly. The surface gradients field, or the surface metric, are sufficient for constructing the expression-invariant representation of any given face. It allows us to perform the 3D face recognition task while avoiding the surface reconstruction stage.

569 citations


Journal ArticleDOI
TL;DR: A new privacy-enabling algorithm is presented, named k-Same, that guarantees face recognition software cannot reliably recognize deidentified faces, even though many facial details are preserved.
Abstract: In the context of sharing video surveillance data, a significant threat to privacy is face recognition software, which can automatically identify known people, such as from a database of drivers' license photos, and thereby track people regardless of suspicion. This paper introduces an algorithm to protect the privacy of individuals in video surveillance data by deidentifying faces such that many facial characteristics remain but the face cannot be reliably recognized. A trivial solution to deidentifying faces involves blacking out each face. This thwarts any possible face recognition, but because all facial details are obscured, the result is of limited use. Many ad hoc attempts, such as covering eyes, fail to thwart face recognition because of the robustness of face recognition methods. This work presents a new privacy-enabling algorithm, named k-Same, that guarantees face recognition software cannot reliably recognize deidentified faces, even though many facial details are preserved. The algorithm determines similarity between faces based on a distance metric and creates new faces by averaging image components, which may be the original image pixels (k-Same-Pixel) or eigenvectors (k-Same-Eigen). Results are presented on a standard collection of real face images with varying k.

504 citations


Journal ArticleDOI
01 Aug 2005
TL;DR: Experiments show that the hallucinated face images are not only very helpful for recognition by humans, but also make the automatic recognition procedure easier, since they emphasize the face difference by adding more high-frequency details.
Abstract: In video surveillance, the faces of interest are often of small size. Image resolution is an important factor affecting face recognition by human and computer. In this paper, we propose a new face hallucination method using eigentransformation. Different from most of the proposed methods based on probabilistic models, this method views hallucination as a transformation between different image styles. We use Principal Component Analysis (PCA) to fit the input face image as a linear combination of the low-resolution face images in the training set. The high-resolution image is rendered by replacing the low-resolution training images with high-resolution ones, while retaining the same combination coefficients. Experiments show that the hallucinated face images are not only very helpful for recognition by humans, but also make the automatic recognition procedure easier, since they emphasize the face difference by adding more high-frequency details.

503 citations


Proceedings ArticleDOI
10 Oct 2005
TL;DR: This paper presents a method for fully automatic detection of 20 facial feature points in images of expressionless faces using Gabor feature based boosted classifiers using GentleBoost templates built from both gray level intensities and Gabor wavelet features.
Abstract: Locating facial feature points in images of faces is an important stage for numerous facial image interpretation tasks. In this paper we present a method for fully automatic detection of 20 facial feature points in images of expressionless faces using Gabor feature based boosted classifiers. The method adopts fast and robust face detection algorithm, which represents an adapted version of the original Viola-Jones face detector. The detected face region is then divided into 20 relevant regions of interest, each of which is examined further to predict the location of the facial feature points. The proposed facial feature point detection method uses individual feature patch templates to detect points in the relevant region of interest. These feature models are GentleBoost templates built from both gray level intensities and Gabor wavelet features. When tested on the Cohn-Kanade database, the method has achieved average recognition rates of 93%.

304 citations


Journal ArticleDOI
TL;DR: An efficient two-dimensional-to-three-dimensional integrated face reconstruction approach is introduced to reconstruct a personalized 3D face model from a single frontal face image with neutral expression and normal illumination and the synthesized virtual faces significantly improve the accuracy of face recognition with changing PIE.

251 citations


Journal ArticleDOI
TL;DR: This work applies a diagnostic test for crowding to a word and a face, and finds that the critical spacing of the parts required for recognition is proportional to distance from fixation and independent of size and kind.
Abstract: Do we identify an object as a whole or by its parts? This simple question has been surprisingly hard to answer. It has been suggested that faces are recognized as wholes and words are recognized by parts. Here we answer the question by applying a test for crowding. In crowding, a target is harder to identify in the presence of nearby flankers. Previous work has described crowding between objects. We show that crowding also occurs between the parts of an object. Such internal crowding severely impairs perception, identification, and fMRI face-area activation. We apply a diagnostic test for crowding to a word and a face, and we find that the critical spacing of the parts required for recognition is proportional to distance from fixation and independent of size and kind. The critical spacing defines an isolation field around the target. Some objects can be recognized only when each part is isolated from the rest of the object by the critical spacing. In that case, recognition is by parts. Recognition is holistic if the observer can recognize the object even when the whole object fits within a critical spacing. Such an object has only one part. Multiple parts within an isolation field will crowd each other and spoil recognition. To assess the robustness of the crowding test, we manipulated familiarity through inversion and the face- and word-superiority effects. We find that threshold contrast for word and face identification is the product of two factors: familiarity and crowding. Familiarity increases sensitivity by a factor of x1.5, independent of eccentricity, while crowding attenuates sensitivity more and more as eccentricity increases. Our findings show that observers process words and faces in much the same way: The effects of familiarity and crowding do not distinguish between them. Words and faces are both recognized by parts, and their parts -- letters and facial features -- are recognized holistically. We propose that internal crowding be taken as the signature of recognition by parts.

241 citations


Proceedings ArticleDOI
20 Jun 2005
TL;DR: The impact of eye locations on face recognition accuracy is studied, and an automatic technique for eye detection is introduced, and the face recognition performance is shown to be comparable to that of using manually given eye positions.
Abstract: The accuracy of face alignment affects the performance of a face recognition system. Since face alignment is usually conducted using eye positions, an accurate eye localization algorithm is therefore essential for accurate face recognition. In this paper, we first study the impact of eye locations on face recognition accuracy, and then introduce an automatic technique for eye detection. The performance of our automatic eye detection technique is subsequently validated using FRGC 1.0 database. The validation shows that our eye detector has an overall 94.5% eye detection rate, with the detected eyes very close to the manually provided eye positions. In addition, the face recognition performance based on the automatic eye detection is shown to be comparable to that of using manually given eye positions.

237 citations


Book ChapterDOI
Kazuhiro Fukui1, Osamu Yamaguchi1
01 Jan 2005
TL;DR: A face recognition method based on the constrained mutual subspace method (CMSM) using multi-viewpoint face patterns attributable to the movement of a robot or a subject using multiple face patterns obtained in various views for robot vision is introduced.
Abstract: This paper introduces a novel approach for face recognition using multiple face patterns obtained in various views for robot vision. A face pattern may change dramatically due to changes in the relation between the positions of a robot, a subject and light sources. As a robot is not generally able to ascertain such changes by itself, face recognition in robot vision must be robust against variations caused by the changes. Conventional methods using a single face pattern are not capable of dealing with the variations of face pattern. In order to overcome the problem, we have developed a face recognition method based on the constrained mutual subspace method (CMSM) using multi-viewpoint face patterns attributable to the movement of a robot or a subject. The effectiveness of our method for robot vision is demonstrated by means of a preliminary experiment.

226 citations


Proceedings ArticleDOI
20 Jun 2005
TL;DR: The algorithm used to preprocess the images in FRVT 2002 is explained, additional FRVT2002 results are presented, and these results to recognition from the model coefficients are compared.
Abstract: This paper presents a method for face recognition across large changes in viewpoint. Our method is based on a morphable model of 3D faces that represents face-specific information extracted from a dataset of 3D scans. For non-frontal face recognition in 2D still images, the morphable model can be incorporated in two different approaches: in the first, it serves as a preprocessing step by estimating the 3D shape of novel faces from the non-frontal input images, and generating frontal views of the reconstructed faces at a standard illumination using 3D computer graphics. The transformed images are then fed into state-of-the-art face recognition systems that are optimized for frontal views. This method was shown to be extremely effective in the Face Recognition Vendor Test FRVT 2002. In the process of estimating the 3D shape of a face from an image, a set of model coefficients are estimated. In the second method, face recognition is performed directly from these coefficients. In this paper we explain the algorithm used to preprocess the images in FRVT 2002, present additional FRVT 2002 results, and compare these results to recognition from the model coefficients.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: It is demonstrated that high recall rates can be achieved whilst maintaining good precision (over 93%) and a recognition method based on a cascade of processing steps that normalize for the effects of the changing imaging environment is developed.
Abstract: The objective of this work is to recognize all the frontal faces of a character in the closed world of a movie or situation comedy, given a small number of query faces. This is challenging because faces in a feature-length film are relatively uncontrolled with a wide variability of scale, pose, illumination, and expressions, and also may be partially occluded. We develop a recognition method based on a cascade of processing steps that normalize for the effects of the changing imaging environment. In particular there are three areas of novelty: (i) we suppress the background surrounding the face, enabling the maximum area of the face to be retained for recognition rather than a subset; (ii) we include a pose refinement step to optimize the registration between the test image and face exemplar; and (iii) we use robust distance to a sub-space to allow for partial occlusion and expression change. The method is applied and evaluated on several feature length films. It is demonstrated that high recall rates (over 92%) can be achieved whilst maintaining good precision (over 93%).

Proceedings ArticleDOI
10 Oct 2005
TL;DR: An approach to automatic visual emotion recognition from two modalities: face and body is presented, in which individual classifiers are trained from individual modalities and facial expression and affective body gesture information are fuse first at a feature-level, and later at a decision-level.
Abstract: This paper presents an approach to automatic visual emotion recognition from two modalities: face and body. Firstly, individual classifiers are trained from individual modalities. Secondly, we fuse facial expression and affective body gesture information first at a feature-level, in which the data from both modalities are combined before classification, and later at a decision-level, in which we integrate the outputs of the monomodal systems by the use of suitable criteria. We then evaluate these two fusion approaches, in terms of performance over monomodal emotion recognition based on facial expression modality only. In the experiments performed the emotion classification using the two modalities achieved a better recognition accuracy outperforming the classification using the individual facial modality. Moreover, fusion at the feature-level proved better recognition than fusion at the decision-level.

Book ChapterDOI
TL;DR: The main purpose of this overview is to describe the recent 3D face recognition algorithms, which hold more information of the face, like surface information, that can be used for face recognition or subject discrimination.
Abstract: Many researches in face recognition have been dealing with the challenge of the great variability in head pose, lighting intensity and direction,facial expression, and aging. The main purpose of this overview is to describe the recent 3D face recognition algorithms. The last few years more and more 2D face recognition algorithms are improved and tested on less than perfect images. However, 3D models hold more information of the face, like surface information, that can be used for face recognition or subject discrimination. Another major advantage is that 3D face recognition is pose invariant. A disadvantage of most presented 3D face recognition methods is that they still treat the human face as a rigid object. This means that the methods aren't capable of handling facial expressions. Although 2D face recognition still seems to outperform the 3D face recognition methods, it is expected that this will change in the near future.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: This paper uses the face recognition grand challenge dataset to evaluate hierarchical graph matching (HGM), an universal approach to 2D and 3D face recognition, and shows that HGM yields the best results presented at the recent FRGC workshop, and that 2d face recognition is significantly more accurate than 3DFace recognition and that the fusion of both modalities leads to a further improvement of the 2D results.
Abstract: The extension of 2D image-based face recognition methods with respect to 3D shape information and the fusion of both modalities is one of the main topics in the recent development of facial recognition. In this paper we discuss different strategies and their expected benefit for the fusion of 2D and 3D face recognition. The face recognition grand challenge (FRGC) provides for the first time ever a public benchmark dataset of a suitable size to evaluate the accuracy of both 2D and 3D face recognition. We use this benchmark to evaluate hierarchical graph matching (HGM), an universal approach to 2D and 3D face recognition, and demonstrate the benefit of different fusion strategies. The results show that HGM yields the best results presented at the recent FRGC workshop, that 2D face recognition is significantly more accurate than 3D face recognition and that the fusion of both modalities leads to a further improvement of the 2D results.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: The Geometrix FaceVision 3D face recognition technology is presented, then how shape and texture are fused is described, and the effect of fusion on error rates is discussed and the remaining errors are analyzed.
Abstract: We present and analyze the performance of the Geometrix ActiveIDTM Biometric Identity System on the FRGC data from Experiment 3, consisting of 4007 3D faces. The performance is presented and analyzed for different facial expression categories. We first present the Geometrix FaceVision 3D face recognition technology, then describe how shape and texture are fused. The effect of fusion on error rates is discussed, and the remaining errors are analyzed. We conclude with lessons learned, general findings, and what we expect in the near future.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: Two algorithms for detecting face anchor points in the context of face verification are presented; One for frontal images and one for arbitrary pose, demonstrating the challenges in 3D face recognition under arbitrary pose and expression.
Abstract: This paper outlines methods to detect key anchor points in 3D face scanner data. These anchor points can be used to estimate the pose and then match the test image to a 3D face model. We present two algorithms for detecting face anchor points in the context of face verification; One for frontal images and one for arbitrary pose. We achieve 99% success in finding anchor points in frontal images and 86% success in scans with large variations in pose and changes in expression. These results demonstrate the challenges in 3D face recognition under arbitrary pose and expression. We are currently working on robust ?tting algorithms to localize more precisely the anchor points for arbitrary pose images.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: This work uses wavelet analysis to extract a compact biometric signature of selected localized face areas using an annotated face model, allowing for rapid comparisons on either a global or a per area basis.
Abstract: From a user’s perspective, face recognition is one of the most desirable biometrics, due to its non-intrusive nature; however, variables such as face expression tend to severely affect recognition rates. We have applied to this problem our previous work on elastically adaptive deformable models to obtain parametric representations of the geometry of selected localized face areas using an annotated face model. We then use wavelet analysis to extract a compact biometric signature, thus allowing us to perform rapid comparisons on either a global or a per area basis. To evaluate the performance of our algorithm, we have conducted experiments using data from the Face Recognition Grand Challenge data corpus, the largest and most established data corpus for face recognition currently available. Our results indicate that our algorithm exhibits high levels of accuracy and robustness, and is not gender biased. In addition, it is minimally affected by facial expressions.

Proceedings ArticleDOI
17 Oct 2005
TL;DR: This paper presents a Bayesian framework to perform multimodal (such as variations in viewpoint and illumination) face image super-resolution for recognition in tensor space, and integrates the tasks of super- resolution and recognition by directly computing a maximum likelihood identity parameter vector in high-resolution Tensor space for recognition.
Abstract: Face images of non-frontal views under poor illumination resolution reduce dramatically face recognition accuracy. This is evident most compellingly by the very low recognition rate of all existing face recognition systems when applied to live CCTV camera input. In this paper, we present a Bayesian framework to perform multimodal (such as variations in viewpoint and illumination) face image super-resolution for recognition in tensor space. Given a single modal low-resolution face image, we benefit from the multiple factor interactions of training sensor and super-resolve its high-resolution reconstructions across different modalities for face recognition. Instead of performing pixel-domain super-resolution and recognition independently as two separate sequential processes, we integrate the tasks of super-resolution and recognition by directly computing a maximum likelihood identity parameter vector in high-resolution tensor space for recognition. We show results from multi-modal super-resolution and face recognition experiments across different imaging modalities, using low-resolution images as testing inputs and demonstrate improved recognition rates over standard tensorface and eigenface representations

Proceedings ArticleDOI
10 Oct 2005
TL;DR: The method presented here attempts to handle a large range of human facial behavior by recognizing facial action units and their temporal segments that produce expressions and introduces AU-dynamics recognition using temporal rules.
Abstract: The recognition of facial expressions in image sequences is a difficult problem with many applications in human-machine interaction. Facial expression analyzers achieve good recognition rates, but virtually all of them deal only with prototypic facial expressions of emotions and cannot handle temporal dynamics of facial displays. The method presented here attempts to handle a large range of human facial behavior by recognizing facial action units (AUs) and their temporal segments (i.e., onset, apex, offset) that produce expressions. We exploit particle filtering to track 20 facial points in an input face video and we introduce AU-dynamics recognition using temporal rules. When tested on Cohn-Kanade and MMI facial expression databases, the proposed method achieved a recognition rate of 90% when detecting 27 AUs occurring alone or in a combination in an input face image sequence.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: A geometry assisted probabilistic approach to improve face recognition under pose variation by approximate a human head with a 3D ellipsoid model, which enables the recognition to be conducted by comparing the texture maps instead of the original images, as done in traditional face recognition.
Abstract: Researchers have been working on human face recognition for decades. Face recognition is hard due to different types of variations in face images, such as pose, illumination and expression, among which pose variation is the hardest one to deal with. To improve face recognition under pose variation, this paper presents a geometry assisted probabilistic approach. We approximate a human head with a 3D ellipsoid model, so that any face image is a 2D projection of such a 3D ellipsoid at a certain pose. In this approach, both training and test images are back projected to the surface of the 3D ellipsoid, according to their estimated poses, to form the texture maps. Thus the recognition can be conducted by comparing the texture maps instead of the original images, as done in traditional face recognition. In addition, we represent the texture map as an array of local patches, which enables us to train a probabilistic model for comparing corresponding patches. By conducting experiments on the CMU PIE database, we show that the proposed algorithm provides better performance than the existing algorithms.

Journal ArticleDOI
TL;DR: A novel approach to handle the illumination problem that can restore a face image captured under arbitrary lighting conditions to one with frontal illumination by using a ratio-image between the face image and a reference face image, both of which are blurred by a Gaussian filter.

Proceedings ArticleDOI
09 May 2005
TL;DR: The paper introduces the IIT-NRC video-based database of faces which consists of pairs of low-resolution video clips of unconstrained facial motions, and offers to use the neuro-associative principle employed by human brain, according to which both memorization and recognition of data are done based on a flow of frames rather than on one frame.
Abstract: This paper presents a number of new views and techniques claimed to be very important for the problem of face recognition in video (FRiV). First, a clear differentiation is made between photographic facial data and video-acquired facial data as being two different modalities: one providing hard biometrics, the other providing softer biometrics. Second, faces which have the resolution of at least 12 pixels between the eyes are shown to be recognizable by computers just as they are by humans. As a way to deal with low resolution and quality of each individual video frame, the paper offers to use the neuro-associative principle employed by human brain, according to which both memorization and recognition of data are done based on a flow of frames rather than on one frame: synaptic plasticity provides a way to memorize from a sequence, while the collective decision making over time is very suitable for recognition of a sequence. As a benchmark for FRiV approaches, the paper introduces the IIT-NRC video-based database of faces which consists of pairs of low-resolution video clips of unconstrained facial motions. The recognition rate of over 95%, which we achieve on this database, as well as the results obtained on real-time annotation of people on TV allow us to believe that the proposed framework brings us closer to the ultimate benchmark for the FRiV approaches, which is "if you are able to recognize a person, so should the computer".

Proceedings ArticleDOI
20 Jun 2005
TL;DR: 3D face recognition has lately been attracting ever increasing attention and this paper complements other reviews in the face biometrics area by focusing on the sensor technology, and by detailing the efforts in 3D face modelling and 3D assisted 2D face matching.
Abstract: 3D face recognition has lately been attracting ever increasing attention. In this paper we review the full spectrum of 3D face processing technology, from sensing to recognition. The review covers 3D face modelling, 3D to 3D and 3D to 2D registration, 3D based recognition and 3D assisted 2D based recognition. The fusion of 2D and 3D modalities is also addressed. The paper complements other reviews in the face biometrics area by focusing on the sensor technology, and by detailing the efforts in 3D face modelling and 3D assisted 2D face matching.

Proceedings ArticleDOI
05 Jan 2005
TL;DR: A face recognition system that utilizes three-dimensional shape information to make the system more robust to arbitrary view, lighting, and facial appearance is developed and the results show the feasibility of the proposed matching scheme.
Abstract: The performance of face recognition systems that use two-dimensional images depends on consistent conditions w.r.t. lighting, pose, and facial appearance. We are developing a face recognition system that utilizes three-dimensional shape information to make the system more robust to arbitrary view, lighting, and facial appearance. For each subject, a 3D face model is constructed by integrating several 2.5D face scans from different viewpoints. A 2.5D scan is composed of one range image along with a registered 2D color image. The recognition engine consists of two components, surface matching and appearance-based matching. The surface matching component is based on a modified Iterative Closest Point (ICP) algorithm.The candidate list used for appearance matching is dynamically generated based on the output of the surface matching component, which reduces the complexity of the appearance-based matching stage. The 3D model in the gallery is used to synthesize new appearance samples with pose and illumination variations that are used for discriminant subspace analysis. The weighted sum rule is applied to combine the two matching components. A hierarchical matching structure is designed to further improve the system performance in both accuracy and efficiency. Experimental results are given for matching a database of 100 3D face models with 598 2.5D independent test scans acquired in different pose and lighting conditions, and with some smiling expression. The results show the feasibility of the proposed matching scheme.

Proceedings ArticleDOI
20 Jun 2005
TL;DR: This paper presents an effective method to automatically extract ROI of facial surface, which mainly depends on automatic detection of facial bilateral symmetry plane and localization of nose tip, and builds a reference plane through the nose tip for calculating the relative depth values.
Abstract: This paper addresses 3D face recognition from facial shape. Firstly, we present an effective method to automatically extract ROI of facial surface, which mainly depends on automatic detection of facial bilateral symmetry plane and localization of nose tip. Then we build a reference plane through the nose tip for calculating the relative depth values. Considering the non-rigid property of facial surface, the ROI is triangulated and parameterized into an isomorphic 2D planar circle, attempting to preserve the intrinsic geometric properties. At the same time the relative depth values are also mapped. Finally we perform eigenface on the mapped relative depth image. The entire scheme is insensitive to pose variance. The experiment using FRGC database v1.0 obtains the rank-1 identification score of 95%, which outperforms the result of the PCA base-line method by 4%, which demonstrates the effectiveness of our algorithm.

Proceedings ArticleDOI
N. Uchida1, Takuma Shibahara1, Takafumi Aoki1, H. Nakajima, K. Kobayashi 
14 Nov 2005
TL;DR: This paper proposes a face recognition system that uses passive stereo vision to capture three-dimensional (3D) facial information and 3D matching using a simple ICP (iterative closest point) algorithm and develops a high-accuracy 3D measurement system based on Passive stereo vision, where phase-based image matching is employed for sub-pixel disparity estimation.
Abstract: This paper proposes a face recognition system that uses (i) passive stereo vision to capture three-dimensional (3D) facial information and (ii) 3D matching using a simple ICP (iterative closest point) algorithm. So far, the reported 3D face recognition techniques assume the use of active 3D measurement for 3D facial capture. However, active methods employ structured illumination (structure projection, phase shift, gray-code demodulation, etc.) or laser scanning, which is not desirable in many human recognition applications. A major problem of using passive stereo vision for 3D measurement is its low accuracy, and thus no passive methods for 3D face recognition have been reported previously. Addressing this problem, we have newly developed a high-accuracy 3D measurement system based on passive stereo vision, where phase-based image matching is employed for sub-pixel disparity estimation. This paper presents the first attempt to create a practical face recognition system based on fully passive 3D reconstruction.

Proceedings ArticleDOI
24 Oct 2005
TL;DR: Thirty local geometrical features extracted from 3D hitman face surfaces have been used to model the face for face recognition, with the most discriminating ones selected from a set of 86.
Abstract: Thirty local geometrical features extracted from 3D hitman face surfaces have been used to model the face for face recognition. They are the most discriminating ones selected from a set of 86. We have experimented with 420 3D-facial meshes (without texture) of 60 individuals. There are 7 images per subject including views presenting fight rotations and facial expressions. The HK algorithm, based in the signs of the mean and Gaussian curvatures, has been used for region segmentation. Experiments under controlled and non-controlled acquisition conditions, considering pose variations and facial expressions, have been achieved to analyze the robustness of the selected characteristics. Success recognition results of 82.0% and 90.16% were obtained when the images are frontal views with neutral expression using PCA and SVM, respectively. The recognition rates only decrease to 76.2% and 77.9% using PCA and SVM matching schemes respectively, under gesture and light face rotation.

Book ChapterDOI
16 Oct 2005
TL;DR: This paper proposes to use Local Binary Pattern features to represent 3D faces and presents a statistical learning procedure for feature selection and classifier learning, which leads to a matching engine for 3D face recognition.
Abstract: 2D intensity images and 3D shape models are both useful for face recognition, but in different ways. While algorithms have long been developed using 2D or 3D data, recently has seen work on combining both into multi-modal face biometrics to achieve higher performance. However, the fusion of the two modalities has mostly been at the decision level, based on scores obtained from independent 2D and 3D matchers. In this paper, we propose a systematic framework for fusing 2D and 3D face recognition at both feature and decision levels, by exploring synergies of the two modalities at these levels. The novelties are the following. First, we propose to use Local Binary Pattern (LBP) features to represent 3D faces and present a statistical learning procedure for feature selection and classifier learning. This leads to a matching engine for 3D face recognition. Second, we propose a statistical learning approach for fusing 2D and 3D based face recognition at both feature and decision levels. Experiments show that the fusion at both levels yields significantly better performance than fusion at the decision level.

Book ChapterDOI
16 Oct 2005
TL;DR: The proposed AdaBoost Gabor Fisher Classifier (AGFC) can effectively reduce the dimensionality of Gabor features and greatly increase the recognition accuracy, and its experimental results show its robustness to variations in facial expression and accessories.
Abstract: This paper proposes the AdaBoost Gabor Fisher Classifier (AGFC) for robust face recognition, in which a chain AdaBoost learning method based on Bootstrap re-sampling is proposed and applied to face recognition with impressive recognition performance. Gabor features have been recognized as one of the most successful face representations, but it is too high dimensional for fast extraction and accurate classification. In AGFC, AdaBoost is exploited to select optimally the most informative Gabor features (hereinafter as AdaGabor features). The selected low-dimensional AdaGabor features are then classified by Fisher discriminant analysis for final face identification. Our experiments on two large-scale face databases, FERET and CAS-PEAL (with 5789 images of 1040 subjects), have shown that the proposed method can effectively reduce the dimensionality of Gabor features and greatly increase the recognition accuracy. In addition, our experimental results show its robustness to variations in facial expression and accessories.