scispace - formally typeset
Search or ask a question

Showing papers on "Feature extraction published in 2002"


Journal ArticleDOI
TL;DR: This paper presents work on computing shape models that are computationally fast and invariant basic transformations like translation, scaling and rotation, and proposes shape detection using a feature called shape context, which is descriptive of the shape of the object.
Abstract: We present a novel approach to measuring similarity between shapes and exploit it for object recognition. In our framework, the measurement of similarity is preceded by: (1) solving for correspondences between points on the two shapes; (2) using the correspondences to estimate an aligning transform. In order to solve the correspondence problem, we attach a descriptor, the shape context, to each point. The shape context at a reference point captures the distribution of the remaining points relative to it, thus offering a globally discriminative characterization. Corresponding points on two similar shapes will have similar shape contexts, enabling us to solve for correspondences as an optimal assignment problem. Given the point correspondences, we estimate the transformation that best aligns the two shapes; regularized thin-plate splines provide a flexible class of transformation maps for this purpose. The dissimilarity between the two shapes is computed as a sum of matching errors between corresponding points, together with a term measuring the magnitude of the aligning transform. We treat recognition in a nearest-neighbor classification framework as the problem of finding the stored prototype shape that is maximally similar to that in the image. Results are presented for silhouettes, trademarks, handwritten digits, and the COIL data set.

6,693 citations


Journal ArticleDOI
TL;DR: In this article, the authors categorize and evaluate face detection algorithms and discuss relevant issues such as data collection, evaluation metrics and benchmarking, and conclude with several promising directions for future research.
Abstract: Images containing faces are essential to intelligent vision-based human-computer interaction, and research efforts in face processing include face recognition, face tracking, pose estimation and expression recognition. However, many reported methods assume that the faces in an image or an image sequence have been identified and localized. To build fully automated systems that analyze the information contained in face images, robust and efficient face detection algorithms are required. Given a single image, the goal of face detection is to identify all image regions which contain a face, regardless of its 3D position, orientation and lighting conditions. Such a problem is challenging because faces are non-rigid and have a high degree of variability in size, shape, color and texture. Numerous techniques have been developed to detect faces in a single image, and the purpose of this paper is to categorize and evaluate these algorithms. We also discuss relevant issues such as data collection, evaluation metrics and benchmarking. After analyzing these algorithms and identifying their limitations, we conclude with several promising directions for future research.

3,894 citations


Journal ArticleDOI
TL;DR: A face detection algorithm for color images in the presence of varying lighting conditions as well as complex backgrounds is proposedBased on a novel lighting compensation technique and a nonlinear color transformation, this method detects skin regions over the entire image and generates face candidates based on the spatial arrangement of these skin patches.
Abstract: Human face detection plays an important role in applications such as video surveillance, human computer interface, face recognition, and face image database management. We propose a face detection algorithm for color images in the presence of varying lighting conditions as well as complex backgrounds. Based on a novel lighting compensation technique and a nonlinear color transformation, our method detects skin regions over the entire image and then generates face candidates based on the spatial arrangement of these skin patches. The algorithm constructs eye, mouth, and boundary maps for verifying each face candidate. Experimental results demonstrate successful face detection over a wide range of facial variations in color, position, scale, orientation, 3D pose, and expression in images from several photo collections (both indoors and outdoors).

2,075 citations


Journal ArticleDOI
TL;DR: Zhang et al. as discussed by the authors introduced a novel Gabor-Fisher (1936) classifier (GFC) for face recognition, which is robust to changes in illumination and facial expression, applies the enhanced Fisher linear discriminant model (EFM) to an augmented Gabor feature vector derived from the Gabor wavelet representation of face images.
Abstract: This paper introduces a novel Gabor-Fisher (1936) classifier (GFC) for face recognition. The GFC method, which is robust to changes in illumination and facial expression, applies the enhanced Fisher linear discriminant model (EFM) to an augmented Gabor feature vector derived from the Gabor wavelet representation of face images. The novelty of this paper comes from (1) the derivation of an augmented Gabor feature vector, whose dimensionality is further reduced using the EFM by considering both data compression and recognition (generalization) performance; (2) the development of a Gabor-Fisher classifier for multi-class problems; and (3) extensive performance evaluation studies. In particular, we performed comparative studies of different similarity measures applied to various classifiers. We also performed comparative experimental studies of various face recognition schemes, including our novel GFC method, the Gabor wavelet method, the eigenfaces method, the Fisherfaces method, the EFM method, the combination of Gabor and the eigenfaces method, and the combination of Gabor and the Fisherfaces method. The feasibility of the new GFC method has been successfully tested on face recognition using 600 FERET frontal face images corresponding to 200 subjects, which were acquired under variable illumination and facial expressions. The novel GFC method achieves 100% accuracy on face recognition using only 62 features.

1,759 citations


07 Apr 2002
TL;DR: An updated review covering the years 1996 2001 will summarize the outcome of an updated review of the structural health monitoring literature, finding that although there are many more SHM studies being reported, the investigators, in general, have not yet fully embraced the well-developed tools from statistical pattern recognition.
Abstract: Staff members at Los Alamos National Laboratory (LANL) produced a summary of the structural health monitoring literature in 1995. This presentation will summarize the outcome of an updated review covering the years 1996 2001. The updated review follows the LANL statistical pattern recognition paradigm for SHM, which addresses four topics: 1. Operational Evaluation; 2. Data Acquisition and Cleansing; 3. Feature Extraction; and 4. Statistical Modeling for Feature Discrimination. The literature has been reviewed based on how a particular study addresses these four topics. A significant observation from this review is that although there are many more SHM studies being reported, the investigators, in general, have not yet fully embraced the well-developed tools from statistical pattern recognition. As such, the discrimination procedures employed are often lacking the appropriate rigor necessary for this technology to evolve beyond demonstration problems carried out in laboratory setting.

1,467 citations


Journal ArticleDOI
TL;DR: An unsupervised feature selection algorithm suitable for data sets, large in both dimension and size, based on measuring similarity between features whereby redundancy therein is removed, which does not need any search and is fast.
Abstract: In this article, we describe an unsupervised feature selection algorithm suitable for data sets, large in both dimension and size. The method is based on measuring similarity between features whereby redundancy therein is removed. This does not need any search and, therefore, is fast. A new feature similarity measure, called maximum information compression index, is introduced. The algorithm is generic in nature and has the capability of multiscale representation of data sets. The superiority of the algorithm, in terms of speed and performance, is established extensively over various real-life data sets of different sizes and dimensions. It is also demonstrated how redundancy and information loss in feature selection can be quantified with an entropy measure.

1,432 citations


Journal ArticleDOI
TL;DR: This work derives a sequence of analytical results which show that the reconstruction constraints provide less and less useful information as the magnification factor increases, and proposes a super-resolution algorithm which attempts to recognize local features in the low-resolution images and then enhances their resolution in an appropriate manner.
Abstract: Nearly all super-resolution algorithms are based on the fundamental constraints that the super-resolution image should generate low resolution input images when appropriately warped and down-sampled to model the image formation process. (These reconstruction constraints are normally combined with some form of smoothness prior to regularize their solution.) We derive a sequence of analytical results which show that the reconstruction constraints provide less and less useful information as the magnification factor increases. We also validate these results empirically and show that, for large enough magnification factors, any smoothness prior leads to overly smooth results with very little high-frequency content. Next, we propose a super-resolution algorithm that uses a different kind of constraint in addition to the reconstruction constraints. The algorithm attempts to recognize local features in the low-resolution images and then enhances their resolution in an appropriate manner. We call such a super-resolution algorithm a hallucination or reconstruction algorithm. We tried our hallucination algorithm on two different data sets, frontal images of faces and printed Roman text. We obtained significantly better results than existing reconstruction-based algorithms, both qualitatively and in terms of RMS pixel error.

1,418 citations


Journal ArticleDOI
TL;DR: The various applications of neural networks in image processing are categorised into a novel two-dimensional taxonomy for image processing algorithms and their specific conditions are discussed in detail.

1,100 citations


Book
18 Feb 2002
TL;DR: The new edition of Feature Extraction and Image Processing provides an essential guide to the implementation of image processing and computer vision techniques, explaining techniques and fundamentals in a clear and concise manner, and features a companion website that includes worksheets, links to free software, Matlab files, solutions and new demonstrations.
Abstract: Image processing and computer vision are currently hot topics with undergraduates and professionals alike. "Feature Extraction and Image Processing" provides an essential guide to the implementation of image processing and computer vision techniques, explaining techniques and fundamentals in a clear and concise manner. Readers can develop working techniques, with usable code provided throughout and working Matlab and Mathcad files on the web. Focusing on feature extraction while also covering issues and techniques such as image acquisition, sampling theory, point operations and low-level feature extraction, the authors have a clear and coherent approach that will appeal to a wide range of students and professionals.The new edition includes: a new coverage of curvature in low-level feature extraction (SIFT and saliency) and features (phase congruency); geometric active contours; morphology; and camera models and an updated coverage of image smoothing (anistropic diffusion); skeletonization; edge detection; curvature; and shape descriptions (moments). It is an essential reading for engineers and students working in this cutting edge field. It is an ideal module text and background reference for courses in image processing and computer vision. It features a companion website that includes worksheets, links to free software, Matlab files, solutions and new demonstrations.

929 citations


Journal ArticleDOI
TL;DR: It is shown that dynamic-scale ridge traversal is insensitive to its initial parameter settings, operates with little additional computational overhead, tracks centerlines with subvoxel accuracy, passes branch points, and handles significant image noise.
Abstract: The extraction of the centerlines of tubular objects in two and three-dimensional images is a part of many clinical image analysis tasks. One common approach to tubular object centerline extraction is based on intensity ridge traversal. In this paper, we evaluate the effects of initialization, noise, and singularities on intensity ridge traversal and present multiscale heuristics and optimal-scale measures that minimize these effects. Monte Carlo experiments using simulated and clinical data are used to quantify how these "dynamic-scale" enhancements address clinical needs regarding speed, accuracy, and automation. In particular, we show that dynamic-scale ridge traversal is insensitive to its initial parameter settings, operates with little additional computational overhead, tracks centerlines with subvoxel accuracy, passes branch points, and handles significant image noise. We also illustrate the capabilities of the method for medical applications involving a variety of tubular structures in clinical data from different organs, patients, and imaging modalities.

660 citations


Journal ArticleDOI
TL;DR: A new method of calculating mutual information between input and class variables based on the Parzen window is proposed, and this is applied to a feature selection algorithm for classification problems.
Abstract: Mutual information is a good indicator of relevance between variables, and have been used as a measure in several feature selection algorithms. However, calculating the mutual information is difficult, and the performance of a feature selection algorithm depends on the accuracy of the mutual information. In this paper, we propose a new method of calculating mutual information between input and class variables based on the Parzen window, and we apply this to a feature selection algorithm for classification problems.

Journal ArticleDOI
TL;DR: Experimental results suggest that using invariant features decreases the probability of being trapped in a local minimum and may be an effective solution for difficult range image registration problems where the scene is very small compared to the model.
Abstract: Investigates the use of Euclidean invariant features in a generalization of iterative closest point (ICP) registration of range images. Pointwise correspondences are chosen as the closest point with respect to a weighted linear combination of positional and feature distances. It is shown that, under ideal noise-free conditions, correspondences formed using this distance function are correct more often than correspondences formed using the positional distance alone. In addition, monotonic convergence to at least a local minimum is shown to hold for this method. When noise is present, a method that automatically sets the optimal relative contribution of features and positions is described. This method trades off the error in feature values due to noise against the error in positions due to misalignment. Experimental results suggest that using invariant features decreases the probability of being trapped in a local minimum and may be an effective solution for difficult range image registration problems where the scene is very small compared to the model.

Journal ArticleDOI
Lie Lu1, Hong-Jiang Zhang1, Hao Jiang1
TL;DR: A robust approach that is capable of classifying and segmenting an audio stream into speech, music, environment sound, and silence is proposed, and an unsupervised speaker segmentation algorithm using a novel scheme based on quasi-GMM and LSP correlation analysis is developed.
Abstract: We present our study of audio content analysis for classification and segmentation, in which an audio stream is segmented according to audio type or speaker identity. We propose a robust approach that is capable of classifying and segmenting an audio stream into speech, music, environment sound, and silence. Audio classification is processed in two steps, which makes it suitable for different applications. The first step of the classification is speech and nonspeech discrimination. In this step, a novel algorithm based on K-nearest-neighbor (KNN) and linear spectral pairs-vector quantization (LSP-VQ) is developed. The second step further divides nonspeech class into music, environment sounds, and silence with a rule-based classification scheme. A set of new features such as the noise frame ratio and band periodicity are introduced and discussed in detail. We also develop an unsupervised speaker segmentation algorithm using a novel scheme based on quasi-GMM and LSP correlation analysis. Without a priori knowledge, this algorithm can support the open-set speaker, online speaker modeling and real time segmentation. Experimental results indicate that the proposed algorithms can produce very satisfactory results.

Journal ArticleDOI
TL;DR: A new automated method that performs unsupervised pixel purity determination and endmember extraction from multidimensional datasets; this is achieved by using both spatial and spectral information in a combined manner.
Abstract: Spectral mixture analysis provides an efficient mechanism for the interpretation and classification of remotely sensed multidimensional imagery. It aims to identify a set of reference signatures (also known as endmembers) that can be used to model the reflectance spectrum at each pixel of the original image. Thus, the modeling is carried out as a linear combination of a finite number of ground components. Although spectral mixture models have proved to be appropriate for the purpose of large hyperspectral dataset subpixel analysis, few methods are available in the literature for the extraction of appropriate endmembers in spectral unmixing. Most approaches have been designed from a spectroscopic viewpoint and, thus, tend to neglect the existing spatial correlation between pixels. This paper presents a new automated method that performs unsupervised pixel purity determination and endmember extraction from multidimensional datasets; this is achieved by using both spatial and spectral information in a combined manner. The method is based on mathematical morphology, a classic image processing technique that can be applied to the spectral domain while being able to keep its spatial characteristics. The proposed methodology is evaluated through a specifically designed framework that uses both simulated and real hyperspectral data.

Journal ArticleDOI
TL;DR: This review paper is to provide a summary and guidelines for using the most widely used pattern analysis techniques, as well as to identify research directions that are at the frontier of sensor-based machine olfaction.
Abstract: Pattern analysis constitutes a critical building block in the development of gas sensor array instruments capable of detecting, identifying, and measuring volatile compounds, a technology that has been proposed as an artificial substitute for the human olfactory system. The successful design of a pattern analysis system for machine olfaction requires a careful consideration of the various issues involved in processing multivariate data: signal-preprocessing, feature extraction, feature selection, classification, regression, clustering, and validation. A considerable number of methods from statistical pattern recognition, neural networks, chemometrics, machine learning, and biological cybernetics have been used to process electronic nose data. The objective of this review paper is to provide a summary and guidelines for using the most widely used pattern analysis techniques, as well as to identify research directions that are at the frontier of sensor-based machine olfaction.

Proceedings ArticleDOI
10 Dec 2002
TL;DR: The feature extraction method has been applied for both image segmentation as well as histogram generation applications - two distinct approaches to content based image retrieval (CBIR), showing better identification of objects in an image.
Abstract: We have analyzed the properties of the HSV (hue, saturation and value) color space with emphasis on the visual perception of the variation in hue, saturation and intensity values of an image pixel. We extract pixel features by either choosing the hue or the intensity as the dominant property based on the saturation value of a pixel. The feature extraction method has been applied for both image segmentation as well as histogram generation applications - two distinct approaches to content based image retrieval (CBIR). Segmentation using this method shows better identification of objects in an image. The histogram retains a uniform color transition that enables us to do a window-based smoothing during retrieval. The results have been compared with those generated using the RGB color space.

Journal ArticleDOI
TL;DR: Three methods for parameterizing lip image sequences for recognition using hidden Markov models are compared and two are top-down approaches that fit a model of the inner and outer lip contours and derive lipreading features from a principal component analysis of shape or shape and appearance, respectively.
Abstract: The multimodal nature of speech is often ignored in human-computer interaction, but lip deformations and other body motion, such as those of the head, convey additional information. We integrate speech cues from many sources and this improves intelligibility, especially when the acoustic signal is degraded. The paper shows how this additional, often complementary, visual speech information can be used for speech recognition. Three methods for parameterizing lip image sequences for recognition using hidden Markov models are compared. Two of these are top-down approaches that fit a model of the inner and outer lip contours and derive lipreading features from a principal component analysis of shape or shape and appearance, respectively. The third, bottom-up, method uses a nonlinear scale-space analysis to form features directly from the pixel intensity. All methods are compared on a multitalker visual speech recognition task of isolated letters.

Journal ArticleDOI
TL;DR: A new approach for watermarking of digital images providing robustness to geometrical distortions by proposing an embedding and detection scheme where the mark is bound with a content descriptor defined by salient points.
Abstract: This paper presents a new approach for watermarking of digital images providing robustness to geometrical distortions. The weaknesses of classical watermarking methods to geometrical distortions are outlined first. Geometrical distortions can be decomposed into two classes: global transformations such as rotations and translations and local transformations such as the StirMark attack. An overview of existing self-synchronizing schemes is then presented. Theses schemes can use periodical properties of the mark, invariant properties of transforms, template insertion, or information provided by the original image to counter geometrical distortions. Thereafter, a new class of watermarking schemes using the image content is presented. We propose an embedding and detection scheme where the mark is bound with a content descriptor defined by salient points. Three different types of feature points are studied and their robustness to geometrical transformations is evaluated to develop an enhanced detector. The embedding of the signature is done by extracting feature points of the image and performing a Delaunay tessellation on the set of points. The mark is embedded using a classical additive scheme inside each triangle of the tessellation. The detection is done using correlation properties on the different triangles. The performance of the presented scheme is evaluated after JPEG compression, geometrical attack and transformations. Results show that the fact that the scheme is robust to these different manipulations. Finally, in our concluding remarks, we analyze the different perspectives of such content-based watermarking scheme.

Journal ArticleDOI
TL;DR: This work applies the multiresolution wavelet transform to extract the waveletface and performs the linear discriminant analysis on waveletfaces to reinforce discriminant power.
Abstract: Feature extraction, discriminant analysis, and classification rules are three crucial issues for face recognition. We present hybrid approaches to handle three issues together. For feature extraction, we apply the multiresolution wavelet transform to extract the waveletface. We also perform the linear discriminant analysis on waveletfaces to reinforce discriminant power. During classification, the nearest feature plane (NFP) and nearest feature space (NFS) classifiers are explored for robust decisions in presence of wide facial variations. Their relationships to conventional nearest neighbor and nearest feature line classifiers are demonstrated. In the experiments, the discriminant waveletface incorporated with the NFS classifier achieves the best face recognition performance.

Journal ArticleDOI
TL;DR: The dyadic discrete wavelet transform approach is shown to significantly increase the overall classification accuracy and is tested using hyperspectral data for various agricultural applications.
Abstract: In this paper, the dyadic discrete wavelet transform is proposed for feature extraction from a high-dimensional data space. The wavelet's inherent multiresolutional properties are discussed in terms related to multispectral and hyperspectral remote sensing. Furthermore, various wavelet-based features are applied to the problem of automatic classification of specific ground vegetations from hyperspectral signatures. The wavelet transform features are evaluated using an automated statistical classifier. The system is tested using hyperspectral data for various agricultural applications. The experimental results demonstrate the promising discriminant capability of the wavelet-based features. The automated classification system consistently provides over 95% and 80% classification accuracy for endmember and mixed-signature applications, respectively. When compared to conventional feature extraction methods, the wavelet transform approach is shown to significantly increase the overall classification accuracy.

Journal ArticleDOI
TL;DR: A fuzzy logic approach, UFM (unified feature matching), for region-based image retrieval, which greatly reduces the influence of inaccurate segmentation and provides a very intuitive quantification.
Abstract: This paper proposes a fuzzy logic approach, UFM (unified feature matching), for region-based image retrieval. In our retrieval system, an image is represented by a set of segmented regions, each of which is characterized by a fuzzy feature (fuzzy set) reflecting color, texture, and shape properties. As a result, an image is associated with a family of fuzzy features corresponding to regions. Fuzzy features naturally characterize the gradual transition between regions (blurry boundaries) within an image and incorporate the segmentation-related uncertainties into the retrieval algorithm. The resemblance of two images is then defined as the overall similarity between two families of fuzzy features and quantified by a similarity measure, UFM measure, which integrates properties of all the regions in the images. Compared with similarity measures based on individual regions and on all regions with crisp-valued feature representations, the UFM measure greatly reduces the influence of inaccurate segmentation and provides a very intuitive quantification. The UFM has been implemented as a part of our experimental SIMPLIcity image retrieval system. The performance of the system is illustrated using examples from an image database of about 60,000 general-purpose images.

01 Jan 2002
TL;DR: Although there are many more SHM studies being reported, the investigators, in general, have not yet fully embraced the well-developed tools from statistical pattern recognition, and the discrimination procedures employed are often lacking the appropriate rigor necessary for this technology to evolve beyond demonstration problems carried out in laboratory setting.
Abstract: Staff members at Los Alamos National Laboratory (LANL) produced a summary of the structural health monitoring literature in 1995. This presentation will summarize the outcome of an updated review covering the years 1996 - 2001. The updated review follows the LANL statistical pattern recognition paradigm for SHM, which addresses four topics: (1) Operational Evaluation; (2) Data Acquisition and Cleansing; (3) Feature Extraction; and (4) Statistical Modeling for Feature Discrimination. The literature has been reviewed based on how a particular study addresses these four topics. A significant observation from this review is that although there are many more SHM studies being reported, the investigators, in general, have not yet fully embraced the well-developed tools from statistical pattern recognition. As such, the discrimination procedures employed are often lacking the appropriate rigor necessary for this technology to evolve beyond demonstration problems carried out in laboratory setting.

Journal ArticleDOI
TL;DR: I Enhancement Fundametal Enhancement Techniques Adaptive Image Filtering Enhancement by Multiscale Nonlinear Operators Medical Image Enhancement with Hybrid Filters and Segmentation Overview and Fundamentals of Medical Image Se segmentation by Fuzzy Clustering.
Abstract: I Enhancement Fundametal Enhancement Techniques Adaptive Image Filtering Enhancement by Multiscale Nonlinear Operators Medical Image Enhancement with Hybrid Filters. II Segmentation Overview and Fundamentals of Medical Image Segmentation Image Segmentation by Fuzzy Clustering: Methods and Issues Segmentation with Neural Networks Deformable Models Shape Constraints in Deformable Models Gradient Vector Flow Deformable Models Fully Automated Hybrid Segmentation of the Brain Volumetric Segmentation Partial Volume Segmentation with Voxel Histograms. III Quantification Two-dimensional Shape and Texture Quantification Texture Analysis in Three Dimensions as a Cue to Medical Diagnosis Computational Neuroanatomy Using Shape Transformation Morphometry of Arterial Trees and Microvascular Networks Image-Based Computational Biomechanics of the Musculoskeletal System Three-Dimensional Bone Angle Quantification Database Selection and Feature Extraction for Neural Networks Quantitative Image Analysis for Estimation of Breast Cancer Risk Classification of Breast Lesions in Mammograms Quantitative Analysis of Cardiac Function Image Processing and Analysis in Tagged Cardiac MRI Image Interpolation and Resampling. IV Registration The Physical Basis of Spatial Distortions in Magnetic Resonance Images Physical and Biological Bases of Spatial Distortiobns in Positron Emission Tomography Images Biological Underpinnings of Anatomic Consistency and Variability in the Human Brain Spatial Transformation Models Validation of Registration Accuracy Landmark-based Registration using Features Identified Through Differential Geometry Registration of Contours using Chamfer Matching Within-Modality Registration using Intensity-Based Cost Functions Across-Modality Registration using Intensity-Based Cost Functions Talairach Space as a Tool for Intersubject Standardization in the Brain Warping Strategies for Intersubject Registration Optimizing the Resampling of Registered Images Clinical Applications of Image Registration Registration for Image-Guided Surgery Image Registration and the Construction of Multidimensional Brain Atlases. V Visualization Visualization Pathways in Biomedicine Three-Dimensional Visualization in Medicine and Biology Volume Visualization in Medicine Fast Isosurface Extraction Methods for Large Image Data Sets Morphometric Methods for Virtual Endoscopy. VI Compression Storage and Communication Fundamentals and Standards of Compression and Communication Medical Image Archive and Retrieval Image Standardization in PACS Quality Evaluation for Compressed Medical Images: Fundamentals Quality Evaluation for Compressed Medical Images: Diagnostic Accuracy Quality Evaluation for Compressed Medical Images: Statistical Issues Three-Dimensional Image Compression with Wavelet Transforms Medical Image Processing and Analysis Software. Index

Proceedings ArticleDOI
10 Dec 2002
TL;DR: The method consists of three major components: image preprocessing, feature extraction and classifier design, which uses an efficient approach called nearest feature line (NFL) for iris matching.
Abstract: Proposes a method for personal identification based on iris recognition. The method consists of three major components: image preprocessing, feature extraction and classifier design. A bank of circular symmetric filters is used to capture local iris characteristics to form a fixed length feature vector. In iris matching, an efficient approach called nearest feature line (NFL) is used. Constraints are imposed on the original NFL method to improve performance. Experimental results show that the proposed method has an encouraging performance.

Proceedings ArticleDOI
20 May 2002
TL;DR: Algorithms and a prototype system for hand tracking and hand posture recognition in terms of hierarchies of multi-scale colour image features at different scales, with qualitative inter-relations in Terms of scale, position and orientation are presented.
Abstract: This paper presents algorithms and a prototype system for hand tracking and hand posture recognition. Hand postures are represented in terms of hierarchies of multi-scale colour image features at different scales, with qualitative inter-relations in terms of scale, position and orientation. In each image, detection of multi-scale colour features is performed. Hand states are then simultaneously detected and tracked using particle filtering, with an extension of layered sampling referred to as hierarchical layered sampling. Experiments are presented showing that the performance of the system is substantially improved by performing feature detection in colour space and including a prior with respect to skin colour. These components have been integrated into a real-time prototype system, applied to a test problem of controlling consumer electronics using hand gestures. In a simplified demo scenario, this system has been successfully tested by participants at two fairs during 2001.

Journal ArticleDOI
TL;DR: Experimental results demonstrate the effectiveness of SVMs in texture classification, and it is shown that SVMs can incorporate conventional texture feature extraction methods within their own architecture, while also providing solutions to problems inherent in these methods.
Abstract: This paper investigates the application of support vector machines (SVMs) in texture classification. Instead of relying on an external feature extractor, the SVM receives the gray-level values of the raw pixels, as SVMs can generalize well even in high-dimensional spaces. Furthermore, it is shown that SVMs can incorporate conventional texture feature extraction methods within their own architecture, while also providing solutions to problems inherent in these methods. One-against-others decomposition is adopted to apply binary SVMs to multitexture classification, plus a neural network is used as an arbitrator to make final classifications from several one-against-others SVM outputs. Experimental results demonstrate the effectiveness of SVMs in texture classification.

Journal ArticleDOI
TL;DR: A new feature extraction method by converting a palmprint image from a spatial domain to a frequency domain using Fourier Transform is proposed, which shows that palmprint identification based on feature extraction in the frequency domain is effective in terms of accuracy and efficiency.
Abstract: Palmprint identification refers to searching in a database for the palmprint template, which is from the same palm as a given palmprint input. The identification process involves preprocessing, feature extraction, feature matching and decision-making. As a key step in the process, in this paper, we propose a new feature extraction method by converting a palmprint image from a spatial domain to a frequency domain using Fourier Transform. The features extracted in the frequency domain are used as indexes to the palmprint templates in the database and the searching process for the best match is conducted by a layered fashion. The experimental results show that palmprint identification based on feature extraction in the frequency domain is effective in terms of accuracy and efficiency.

Journal ArticleDOI
TL;DR: A novel algorithm for segmentation of moving objects in video sequences and extraction of video object planes (VOPs) based on connected components analysis and smoothness of VO displacement in successive frames is proposed.
Abstract: The new video-coding standard MPEG-4 enables content-based functionality, as well as high coding efficiency, by taking into account shape information of moving objects. A novel algorithm for segmentation of moving objects in video sequences and extraction of video object planes (VOPs) is proposed . For the case of multiple video objects in a scene, the extraction of a specific single video object (VO) based on connected components analysis and smoothness of VO displacement in successive frames is also discussed. Our algorithm begins with a robust double-edge map derived from the difference between two successive frames. After removing edge points which belong to the previous frame, the remaining edge map, moving edge (ME), is used to extract the VOP. The proposed algorithm is evaluated on an indoor sequence captured by a low-end camera as well as MPEG-4 test sequences and produces promising results.

Journal ArticleDOI
TL;DR: A robust hierarchical algorithm for fully-automatic registration of a pair of images of the curved human retina photographed by a fundus microscope, making the algorithm robust to unmatchable image features and mismatches between features caused by large interframe motions.
Abstract: This paper describes a robust hierarchical algorithm for fully-automatic registration of a pair of images of the curved human retina photographed by a fundus microscope. Accurate registration is essential for mosaic synthesis, change detection, and design of computer-aided instrumentation. Central to the algorithm is a 12-parameter interimage transformation derived by modeling the retina as a rigid quadratic surface with unknown parameters. The parameters are estimated by matching vascular landmarks by recursively tracing the blood vessel structure. The parameter estimation technique, which could be generalized to other applications, is a hierarchy of models and methods, making the algorithm robust to unmatchable image features and mismatches between features caused by large interframe motions. Experiments involving 3,000 image pairs from 16 different healthy eyes were performed. Final registration errors less than a pixel are routinely achieved. The speed, accuracy, and ability to handle small overlaps compare favorably with retinal image registration techniques published in the literature.

01 Jan 2002
TL;DR: A new approach for personal identification based on iris recognition is presented, which uses a bank of Gabor filters to capture both local and global iris characteristics to form a fixed length feature vector.
Abstract: A new approach for personal identification based on iris recognition is presented in this paper The body of this paper details the steps of iris recognition, including image preprocessing, feature extraction and classifier design The proposed algorithm uses a bank of Gabor filters to capture both local and global iris characteristics to form a fixed length feature vector Iris matching is based on the weighted Euclidean distance between the two corresponding iris vectors and is therefore very fast Experimental results are reported to demonstrate the performance of the algorithm