scispace - formally typeset
Search or ask a question

Showing papers by "Tieniu Tan published in 2009"


Journal ArticleDOI
TL;DR: Experimental results on three challenging iris image databases demonstrate that the proposed algorithm outperforms state-of-the-art methods in both accuracy and speed.
Abstract: Iris segmentation is an essential module in iris recognition because it defines the effective image region used for subsequent processing such as feature extraction. Traditional iris segmentation methods often involve an exhaustive search of a large parameter space, which is time consuming and sensitive to noise. To address these problems, this paper presents a novel algorithm for accurate and fast iris segmentation. After efficient reflection removal, an Adaboost-cascade iris detector is first built to extract a rough position of the iris center. Edge points of iris boundaries are then detected, and an elastic model named pulling and pushing is established. Under this model, the center and radius of the circular iris boundaries are iteratively refined in a way driven by the restoring forces of Hooke's law. Furthermore, a smoothing spline-based edge fitting scheme is presented to deal with noncircular iris boundaries. After that, eyelids are localized via edge detection followed by curve fitting. The novelty here is the adoption of a rank filter for noise elimination and a histogram filter for tackling the shape irregularity of eyelids. Finally, eyelashes and shadows are detected via a learned prediction model. This model provides an adaptive threshold for eyelash and shadow detection by analyzing the intensity distributions of different iris regions. Experimental results on three challenging iris image databases demonstrate that the proposed algorithm outperforms state-of-the-art methods in both accuracy and speed.

402 citations


Journal ArticleDOI
TL;DR: This paper develops multilobe differential filters to compute ordinal measures with flexible intralobe and interlobe parameters such as location, scale, orientation, and distance and demonstrates the effectiveness of the proposed ordinal feature models.
Abstract: Images of a human iris contain rich texture information useful for identity authentication. A key and still open issue in iris recognition is how best to represent such textural information using a compact set of features (iris features). In this paper, we propose using ordinal measures for iris feature representation with the objective of characterizing qualitative relationships between iris regions rather than precise measurements of iris image structures. Such a representation may lose some image-specific information, but it achieves a good trade-off between distinctiveness and robustness. We show that ordinal measures are intrinsic features of iris patterns and largely invariant to illumination changes. Moreover, compactness and low computational complexity of ordinal measures enable highly efficient iris recognition. Ordinal measures are a general concept useful for image analysis and many variants can be derived for ordinal feature extraction. In this paper, we develop multilobe differential filters to compute ordinal measures with flexible intralobe and interlobe parameters such as location, scale, orientation, and distance. Experimental results on three public iris image databases demonstrate the effectiveness of the proposed ordinal feature models.

295 citations


Journal ArticleDOI
TL;DR: The prior knowledge extracted from the psychological experiments can be combined with an automatic method to further improve classification accuracy, and the proposed method achieves higher performance than some other methods, and is even more accurate than human observers.
Abstract: Gender is an important cue in social activities. In this correspondence, we present a study and analysis of gender classification based on human gait. Psychological experiments were carried out. These experiments showed that humans can recognize gender based on gait information, and that contributions of different body components vary. The prior knowledge extracted from the psychological experiments can be combined with an automatic method to further improve classification accuracy. The proposed method which combines human knowledge achieves higher performance than some other methods, and is even more accurate than human observers. We also present a numerical analysis of the contributions of different human components, which shows that head and hair, back, chest and thigh are more discriminative than other components. We also did challenging cross-race experiments that used Asian gait data to classify the gender of Europeans, and vice versa. Encouraging results were obtained. All the above prove that gait-based gender classification is feasible in controlled environments. In real applications, it still suffers from many difficulties, such as view variation, clothing and shoes changes, or carrying objects. We analyze the difficulties and suggest some possible solutions.

277 citations


Proceedings ArticleDOI
07 Nov 2009
TL;DR: A color image splicing detection method based on gray level co-occurrence matrix (GLCM) of thresholded edge image of image chroma is proposed and the effectiveness of the proposed method has been demonstrated by the experimental results.
Abstract: A color image splicing detection method based on gray level co-occurrence matrix (GLCM) of thresholded edge image of image chroma is proposed in this paper. Edge images are generated by subtracting horizontal, vertical, main and minor diagonal pixel values from current pixel values respectively and then thresholded with a predefined threshold T. The GLCMs of edge images along the four directions serve as features for image splicing detection. Boosting feature selection is applied to select optimal features and Support Vector Machine (SVM) is utilized as classifier in our approach. The effectiveness of the proposed method has been demonstrated by our experimental results.

139 citations


Journal ArticleDOI
TL;DR: A novel hierarchical selecting scheme embedded in linear discriminant analysis (LDA) and AdaBoost learning is proposed to select the most effective and most robust features and to construct a strong classifier for face recognition systems.

108 citations


Book ChapterDOI
04 Jun 2009
TL;DR: A kernel density estimation scheme is proposed to complement the insufficiency of counterfeit iris images during Adaboost training and outperforms state-of-the-art methods in both accuracy and speed.
Abstract: Recently, spoof detection has become an important and challenging topic in iris recognition. Based on the textural differences between the counterfeit iris images and the live iris images, we propose an efficient method to tackle this problem. Firstly, the normalized iris image is divided into sub-regions according to the properties of iris textures. Local binary patterns (LBP) are then adopted for texture representation of each sub-region. Finally, Adaboost learning is performed to select the most discriminative LBP features for spoof detection. In particular, a kernel density estimation scheme is proposed to complement the insufficiency of counterfeit iris images during Adaboost training. The comparison experiments indicate that the proposed method outperforms state-of-the-art methods in both accuracy and speed.

95 citations


Book ChapterDOI
01 Oct 2009
TL;DR: A simple but efficient approach for blind image splicing detection by analyzing the discontinuity of image pixel correlation and coherency caused by splicing in terms of image run-length representation and sharp image characteristics.
Abstract: In this paper, a simple but efficient approach for blind image splicing detection is proposed. Image splicing is a common and fundamental operation used for image forgery. The detection of image splicing is a preliminary but desirable study for image forensics. Passive detection approaches of image splicing are usually regarded as pattern recognition problems based on features which are sensitive to splicing. In the proposed approach, we analyze the discontinuity of image pixel correlation and coherency caused by splicing in terms of image run-length representation and sharp image characteristics. The statistical features extracted from image run-length representation and image edge statistics are used for splicing detection. The support vector machine (SVM) is used as the classifier. Our experimental results demonstrate that the two proposed features outperform existing ones both in detection accuracy and computational complexity.

92 citations


Proceedings ArticleDOI
07 Nov 2009
TL;DR: A novel method for rapid and robust human detection and tracking based on the omega-shape features of people's head-shoulder parts, which shows great robustness in scenarios of crowding, background distractors and partial occlusions is proposed.
Abstract: This paper proposes a novel method for rapid and robust human detection and tracking based on the omega-shape features of people's head-shoulder parts. There are two modules in this method. In the first module, a Viola-Jones type classifier and a local HOG (Histograms of Oriented Gradients) feature based AdaBoost classifier are combined to detect head-shoulders rapidly and effectively. Then, in the second module, each detected head-shoulder is tracked by a particle filter tracker using local HOG features to model target's appearance, which shows great robustness in scenarios of crowding, background distractors and partial occlusions. Experimental results demonstrate the effectiveness and efficiency of the proposed approach.

79 citations


Book ChapterDOI
18 Aug 2009
TL;DR: An overview of passive digital image tampering detection methods in three levels, that is low level, middle level, and high level in semantic sense, are presented.
Abstract: Digital images can be easily tampered with image editing tools. The detection of tampering operations is of great importance. Passive digital image tampering detection aims at verifying the authenticity of digital images without any a prior knowledge on the original images. There are various methods proposed in this filed in recent years. In this paper, we present an overview of these methods in three levels, that is low level, middle level, and high level in semantic sense. The main ideas of the proposed approaches at each level are described in detail, and some comments are given.

58 citations


Proceedings ArticleDOI
04 Dec 2009
TL;DR: An iris recognition system at a distance about 3 meters is designed, which uses screens and audio signals to direct users to stand on the right position and give them mul- timedia feedback, and a self-adaptive machine to automatically adapt to different people is designed.
Abstract: Iris recognition is a powerful biometrics for personal identification, but it is difficult to acquire good-quality iris images in real time. For making iris recognition more convenient to use, we design an iris recognition system at a distance about 3 meters. There are many key issues to design such a system, including iris image acquisition, human-machine-interface and image processing. In this paper, we respectively introduce how we deal with these problems and accomplish the engineering design. Experiments show that our system is convenient to use at the distance of 3 meters and the recognition rate is not worse than the state-of-the-art close-range systems. 1)Iris image acquisition: The human iris is very small and the required resolution for iris recognition is large, so it is diffi- cult to design the optical path for iris imaging at a distance. We carefully calculate the parameters of cameras, lens and illumi- nation intensity and elaborately select their types to set up the optical system. 2)Human-machine-interface: Because users are of differ- ent height, it is impossible for a single camera to cover so large range or for users to cooperate with the camera at a distance. We design a self-adaptive machine to automatically adapt to different people. Moreover, we use screens and audio signals to direct users to stand on the right position and give them mul- timedia feedback. All these devices are installed into a cabinet

51 citations


Book ChapterDOI
04 Jun 2009
TL;DR: A robust key extraction approach which consists of concatenated coding scheme and bit masking scheme based on iris database is proposed to minimize and randomize the errors occur in the iris codes, making the error pattern more suitable for the coding scheme.
Abstract: Biometric cryptosystem can not only provide an efficient mechanism for template protection, but also facilitate cryptographic key management, thus becomes a promising direction in information security field. In this paper, we propose a robust key extraction approach which consists of concatenated coding scheme and bit masking scheme based on iris database. The concatenated coding scheme that combines Reed-Solomon code and convolutional code is proposed so that much longer keys can be extracted from the iris data, while the bit masking scheme is proposed to minimize and randomize the errors occur in the iris codes, making the error pattern more suitable for the coding scheme. The experiment results show that the system can achieve a FRR of 0.52% with the key length of 938 bits.

Proceedings ArticleDOI
26 Jul 2009
TL;DR: A novel method to text-independent writer identification from online handwriting with high accuracy, fast speed and low requirement of the amount of characters in handwriting samples is proposed.
Abstract: This paper proposes a novel method to text-independent writer identification from online handwriting. The main contributions of our method include two parts: shape primitive representation and hierarchical structure. Both shape primitive's features are developed to represent the robust and distinctive characteristics of handwriting in two hierarchies. In first hierarchy, the shape primitives probability distribution function (SPPDF)is defined as the static features, to characterize orientation information of writing style. For each shape primitive, the statistics of pressure is defined as the dynamic shape primitives probability distribution function (DSPPDF) and the second hierarchy we build Gaussian model in dynamic attributes (DA) according to curvature of shape primitives. Experiments were conducted on the NLPR handwriting database collected from 242 persons. The results show that the new method achieves high accuracy, fast speed and low requirement of the amount of characters in handwriting samples. We achieve a writer identification rate of 91.5 \% with datasets in Chinese text and 93.6 \% in English text.

Proceedings ArticleDOI
26 Jul 2009
TL;DR: Experimental results show that the proposed method for online text-independent writer identification can improve the identification accuracy with a small number of characters and is even effective for cross-language (English & Chinese) writer identification.
Abstract: In this paper we present a novel method for online text-independent writer identification. Most of the existing writer identification techniques require the data to be from a specific text which is not applicable to cases where such text is not available, such as in criminal justice systems when text documents with different content need to be compared. Text-independent approaches often require a large amount of data to be confident of good results. We propose temporal sequence and shape codes to encode online handwriting. Temporal sequence codes (TSC) are to characterize trajectory in speed and pressure change in writing, and shape codes (SC) are to characterize direction of trajectory in writing handwriting. For TSC, we use two different codes to encode speed and pressure to codebook: stroke temporal sequence codes (STSC) and neighbor temporal sequence codes (NTSC). At identification stage, we implement decision and fusion strategy to identify writer. Experimental results show that our proposed method can improve the identification accuracy with a small number of characters. Moreover, we find that the proposed method is even effective for cross-language (English & Chinese) writer identification.

Journal ArticleDOI
01 Aug 2009
TL;DR: This paper introduces a view-independent behavior-analysis framework based on decision fusion in which distance and view angle factors are analyzed and a first corresponding video database is built.
Abstract: The motion analysis of the human body is an important topic of research in computer vision devoted to detecting, tracking, and understanding people's physical behavior. This strong interest is driven by a wide spectrum of applications in various areas such as smart video surveillance. Most research in behavior (or gesture) representation focusses on view-dependent representation, and some research on view invariance considers only information from 3-D models, which is effective under considerable changes of viewpoint. This paper introduces a view-independent behavior-analysis framework based on decision fusion in which distance and view angle factors are analyzed. This is a first effort to tackle the problem of behaviors under significant changes in view angle, and a first corresponding video database is built.

Journal ArticleDOI
TL;DR: A novel motion descriptor, which is an improved feature based on optical flow, is proposed for motion representation that is improved with a motion filter, and feature fusion with the shape and trajectory information.
Abstract: Human behavior analysis is an important area of research in computer vision and is also driven by a wide spectrum of applications, such as smart video surveillance and human-computer interface. In this paper, we present a novel approach for human behavior analysis. Two research challenges, motion representation and behavior recognition, are addressed. A novel motion descriptor, which is an improved feature based on optical flow, is proposed for motion representation. Optical flow is improved with a motion filter, and feature fusion with the shape and trajectory information. To recognize the behavior, the support vector machine is employed to train the classifier where the concatenation of histograms is formed as the input features. Experimental results on the Weizmann behavior database and the Institute of Automation, Chinese Academy of Science real-world multiview behavior database demonstrate the robustness and effectiveness of our method.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: A compact motion representation for human activity recognition is presented with the employment of efficient features extracted from optical flow as the main part, together with global information, which can achieve satisfying recognition performance with low computational cost as well as robustness against both horizontal and vertical viewpoint changes.
Abstract: We address the problem of action recognition. Our aim is to recognize single person activities in surveillance scenes. To meet the requirements of real scene action recognition, we present a compact motion representation for human activity recognition. With the employment of efficient features extracted from optical flow as the main part, together with global information, our motion representation is compact and discriminative. We also build a novel human action dataset(CASIA) in surveillance scene with three vertically different viewpoints and distant people. Experiments on CASIA dataset and WEIZMANN dataset show that our method can achieve satisfying recognition performance with low computational cost as well as robustness against both horizontal(panning) and vertical(tilting) viewpoint changes.

Patent
24 Jun 2009
TL;DR: In this paper, a method for analyzing crowd density based on statistical characteristics is proposed, which consists of the following steps: video input and frame extraction are carried out; the mosaic image difference MID characteristics are extracted from the video-frequency frame sequence, and the subtle movements in the crowd are detected.
Abstract: The invention discloses a method for analyzing crowd density base on statistical characteristics. The method comprises the following steps: video input and frame extraction are carried out; the mosaic image difference MID characteristics are extracted from the video-frequency frame sequence, and the subtle movements in the crowd are detected; the uniform distribution of the sequence time of the mosaic image difference MID characteristics is checked; the geometric correction is performed to the crowd and the scene with obvious perspective phenomenon, and a contribution factor of each picture element to the crowd density is obtained on the image plane; and the weighting process is performed to the crowd space area, so as to obtain the crowd density. Compared with the prior method, the method has no need of the reference background, also has no need of the background modeling and can self-adapt change of either morning or evening light, the algorithm is quite robust, and the application is convenient; the mathematical model is simple and effective, the spatial distribution and the size of the crowd can be accurately located, and the vivacity is strong; the calculation amount is small, and the method is suitable for real-time visual monitoring. The invention can be widely applied to the monitoring and the management of the public places with detained crowd density such as the public transportation, the subway, the square and the like.

Book ChapterDOI
23 Sep 2009
TL;DR: Compared to the state of arts feature detectors: SIFT and SURF, HLSIFD shows high performance, and Harris-like Scale Invariant Feature Detector uses Hessian Matrix which is proved to be more stable in scale-space than Harris matrix.
Abstract: Image feature detection is a fundamental issue in computer vision. SIFT[1] and SURF[2] are very effective in scale-space feature detection, but their stabilities are not good enough because unstable features such as edges are often detected even if they use edge suppression as a post-treatment. Inspired by Harris function[3], we extend Harris to scale-space and propose a novel method - Harris-like Scale Invariant Feature Detector (HLSIFD). Different to Harris-Laplace which is a hybrid method of Harris and Laplace, HLSIFD uses Hessian Matrix which is proved to be more stable in scale-space than Harris matrix. Unlike other methods suppressing edges in a sudden way(SIFT) or ignoring it(SURF), HLSIFD suppresses edges smoothly and uniformly, so fewer fake points are detected by HLSIFD. The approach is evaluated on public databases and in real scenes. Compared to the state of arts feature detectors: SIFT and SURF, HLSIFD shows high performance of HLSIFD.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: The core idea of the method is to dynamically adjust the decision threshold of iris matching module based on the quality measure of input iris image so that the poor quality iris images also have a chance to match template database under the controlled false accept rate.
Abstract: Current iris recognition systems usually regard poor quality iris images useless since defocused or partially occluded iris images may cause false acceptance. However, such a strategy may lose an opportunity to correctly report a genuine match with poor-quality samples. This paper proposes an adaptive iris matching method to improve the throughput of iris recognition systems. The core idea of the method is to dynamically adjust the decision threshold of iris matching module based on the quality measure of input iris image. So that the poor quality iris images also have a chance to match template database under the controlled false accept rate. Experiment results on the real system demonstrate the effectiveness of the proposed method and the recognition time is expected to be greatly reduced.

Book ChapterDOI
18 Aug 2009
TL;DR: Experimental results demonstrate that the previously developed run-length based features for multi-class blind image steganalysis are able to classify different stego images according to their embedding techniques based on appropriate supervised learning.
Abstract: In this paper, we investigate our previously developed run-length based features for multi-class blind image steganalysis. We construct a Support Vector Machine classifier for multi-class recognition for both spatial and frequency domain based steganographic algorithms. We also study hierarchical and non-hierarchical multi-class schemes and compare their performance for steganalysis. Experimental results demonstrate that our approach is able to classify different stego images according to their embedding techniques based on appropriate supervised learning. It is also shown that the hierarchical scheme performs better in our experiments.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: A set of simplified biologically inspired features is proposed for object representation and the Bhattacharyya coefficient is used to measure the similarity between the target model and candidate targets and the proposed appearance model is combined into a Bayesian state inference tracking framework utilizing the SIR particle filter to propagate sample distributions over time.
Abstract: We address the problem of robust appearance-based visual tracking. First, a set of simplified biologically inspired features (SBIF) is proposed for object representation and the Bhattacharyya coefficient is used to measure the similarity between the target model and candidate targets. Then, the proposed appearance model is combined into a Bayesian state inference tracking framework utilizing the SIR (sampling importance resampling) particle filter to propagate sample distributions over time. Numerous experiments are conducted and experimental results demonstrate that our algorithm is robust to partial occlusions and variations of illumination and pose, resistent to nearby distractors, as well as possesses the state-of-the-art tracking accuracy.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: A model based on Salient Contrast Change (SCC) feature is proposed, which applies learning process to enhance adaptability and analyzes trajectories to improve the effectiveness of detection.
Abstract: Night surveillance is a challenging task because of low brightness, low contrast, low Signal to Noise Ratio (SNR) and low appearance information. Most existing models for night surveillance share the following problems: a lack of adaptability for different scenes and separation between detection and tracking. To solve these problems we propose a model based on Salient Contrast Change (SCC) feature, which applies learning process to enhance adaptability and analyzes trajectories to improve the effectiveness of detection. Empirical studies on several real night videos show that the proposed model is more effective than the original CC model and other traditional models.

Book ChapterDOI
04 Jun 2009
TL;DR: The main contribution of this paper is that it view ethnicity categorization as a fuzzy problem and give an effective solution to assign the 3D face a reasonable membership degree.
Abstract: In this paper, we propose a novel fuzzy 3D face ethnicity categorization algorithm, which contains two stages, learning and mapping. In learning stage, the visual codes are first learned for both the eastern and western individuals using the learned visual codebook (LVC) method, then from these codes we can learn two distance measures, merging distance and mapping distance. Using the merging distance, we can learn the eastern, western and human codes based on the visual codes. In mapping stage, we compute the probabilities for each 3D face mapped to eastern and western individuals using the mapping distance. And the membership degree is determined by our defined membership function. The main contribution of this paper is that we view ethnicity categorization as a fuzzy problem and give an effective solution to assign the 3D face a reasonable membership degree. All experiments are based on the challenging FRGC2.0 3D Face Database. Experimental results illustrate the efficiency and accuracy of our fuzzy 3D face ethnicity categorization method.

Book ChapterDOI
23 Sep 2009
TL;DR: A novel visual organization is proposed, termed computational topological perceptual organization (CTPO), which pioneers the early holistic registration in computational vision and is invariant to global transformation such as translation, scaling, rotation and insensitive to topological deformation.
Abstract: What are the primitives of visual perception? The early feature-analysis theory insists on it being a local-to-global process which has acted as the foundation of most computer vision applications for the past 30 years. The early holistic registration theory, however, considers it as a global-to-local process, of which Chen’s theory of topological perceptual organization (TPO) has been strongly supported by psychological and physiological proofs. In this paper, inspired by Chen’s theory, we propose a novel visual organization, termed computational topological perceptual organization (CTPO), which pioneers the early holistic registration in computational vision. Empirical studies on synthetic datasets prove that CTPO is invariant to global transformation such as translation, scaling, rotation and insensitive to topological deformation. We also extend it to other applications by integrating it with local features. Experiments show that our algorithm achieves competitive performance compared with some popular algorithms.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: A new method of computing invariants in videos captured from different views to achieve view-invariant action recognition that outperforms the state-of-the-art methods in effectiveness and stability is presented.
Abstract: We present a new method of computing invariants in videos captured from different views to achieve view-invariant action recognition. To avoid the constraints of collinearity or coplanarity of image points for constructing invariants, we consider several neighboring frames to compute cross ratios, namely cross ratios across frames (CRAF), as our invariant representation of action. For every five points sampled with different intervals from the trajectories of action, we construct a pair of cross ratios (CRs). Afterwards, we transform the CRs to histograms as the feature vectors for classification. Experimental results demonstrate that the proposed method outperforms the state-of-the-art methods in effectiveness and stability.

Proceedings ArticleDOI
01 Sep 2009
TL;DR: It is shown experimentally that the PSOF descriptor is much more robust to image blur and noise than the HOG (Histograms of Oriented Gradients) descriptor, as well as possesses excellent detection performance in normal imaging condition as HOG does.
Abstract: We study the problem of robust pedestrian detection. A new descriptor, Pyramidal Statistics of Oriented Filtering (PSOF), is proposed for shape representation. Unlike one-scale gradient-based methods, the PSOF descriptor constructs an image pyramid and uses a Gabor filter bank to obtain multi-scale pixel-level orientation information. Then, locally normalized pyramidal statistics of these Gabor responses are used to represent object shape. After feature extraction, the AdaBoost training algorithm is adopted to train a classifier for the final pedestrian detector. We show experimentally that the PSOF descriptor is much more robust to image blur and noise than the HOG (Histograms of Oriented Gradients) descriptor, as well as possesses excellent detection performance in normal imaging condition as HOG does. We also study the influence of various parameter settings, concluding that multi-scale information and statistic combination are two important factors for the robustness of the PSOF descriptor.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: This paper proposes a coarse-to-fine cascade scheme, which makes use of information redundancy of statistical texture descriptions between different spatial scales, and reduces most of computational burden and achieves accurate classification simultaneously.
Abstract: Recent literatures have revealed that statistics of local texture measures can provide accurate descriptions of palmprint appearances. In this framework, one palmprint image is divided into local blocks with multiple spatial resolutions. The statistical texture descriptions of each block are then concatenated to form a multi-scale image representation. However, resultant high-dimensional statistical features lead to increasing of computational cost. In this paper, we tackle this problem by performing a coarse-to-fine cascade scheme, which makes use of information redundancy of statistical texture descriptions between different spatial scales. In contrast with non-cascade strategies, the proposed method reduces most of computational burden and achieves accurate classification simultaneously.

Book ChapterDOI
04 Jun 2009
TL;DR: A novel palmprint representation named Spatial Bags of Local Layered Descriptors (SBLLD) is presented, which works by partitioning the whole palmprint image into sub-regions and describing distributions of layered palmprint descriptors inside each sub-region.
Abstract: State-of-the-art palmprint recognition algorithms achieve high accuracy based on component based texture analysis. However, they are still sensitive to local variations of appearances introduced by deformation of skin surfaces or local contrast variations. To tackle this problem, this paper presents a novel palmprint representation named Spatial Bags of Local Layered Descriptors (SBLLD). This technique works by partitioning the whole palmprint image into sub-regions and describing distributions of layered palmprint descriptors inside each sub-region. Through the procedure of partitioning and disordering, local statistical palmprint descriptions and spatial information of palmprint patterns are integrated to achieve accurate image description. Furthermore, to remove irrelevant and attributes from the proposed feature representation, we apply a simple but efficient ranking based feature selection procedure to construct compact and descriptive statistical palmprint representation, which improves classification ability of the proposed method in a further step. Our idea is verified through verification test on large-scale PolyU Palmprint Database Version 2.0. Extensive experimental results testify efficiency of our proposed palmprint representation.

Book ChapterDOI
04 Jun 2009
TL;DR: A novel palmprint representation is proposed in this paper, which describes palmprint images by constructing rank correlation statistics of appearance patterns within local image areas within localimage areas.
Abstract: Automatic personal identification based on palmprints has been considered as a promising technology in biometrics family during recent years. In pursuit of accurate palmprint recognition approaches, it is a key issue to design proper image representation to describe skin textures in palm regions. According to previous achievements, directional texture measurement provides a powerful tool for depicting palmprint appearances. Most of successful approaches can be ranged into this framework. Following this idea, we propose a novel palmprint representation in this paper, which describes palmprint images by constructing rank correlation statistics of appearance patterns within local image areas. Promising experimental results on two large scale palmprint databases demonstrate that the proposed method achieves even better performances than the state-of-the-art approaches.

Proceedings ArticleDOI
07 Nov 2009
TL;DR: This paper proposes a computational model to imitate the primitives of visual perception based on the pyschological theory of topological perceptual organization, and adopts geodesic distance based descriptor to describe an independent topological structure.
Abstract: Great stride has been made in psychological research about primitives of visual perception, which is important to computer vision and image processing. In this paper, we propose a computational model to imitate the primitives of visual perception based on the pyschological theory of topological perceptual organization. First, we adopt geodesic distance based descriptor to describe an independent topological structure. Then, we consider the spatial relationship of two independent structures. Experiments on structures classification demonstrates that the propose model is consistent with the psychological theory. Further experiments on patches clustering prove that our approach can be used to enhance other algorithms.