scispace - formally typeset
Search or ask a question

Showing papers on "Contextual image classification published in 2003"


Proceedings ArticleDOI
18 Jun 2003
TL;DR: The flexible nature of the model is demonstrated by excellent results over a range of datasets including geometrically constrained classes (e.g. faces, cars) and flexible objects (such as animals).
Abstract: We present a method to learn and recognize object class models from unlabeled and unsegmented cluttered scenes in a scale invariant manner. Objects are modeled as flexible constellations of parts. A probabilistic representation is used for all aspects of the object: shape, appearance, occlusion and relative scale. An entropy-based feature detector is used to select regions and their scale within the image. In learning the parameters of the scale-invariant object model are estimated. This is done using expectation-maximization in a maximum-likelihood setting. In recognition, this model is used in a Bayesian manner to classify images. The flexible nature of the model is demonstrated by excellent results over a range of datasets including geometrically constrained classes (e.g. faces, cars) and flexible objects (such as animals).

2,411 citations


Proceedings ArticleDOI
13 Oct 2003
TL;DR: A two-class classification model for grouping is proposed that defines a variety of features derived from the classical Gestalt cues, including contour, texture, brightness and good continuation, and trains a linear classifier to combine these features.
Abstract: We propose a two-class classification model for grouping. Human segmented natural images are used as positive examples. Negative examples of grouping are constructed by randomly matching human segmentations and images. In a preprocessing stage an image is over-segmented into super-pixels. We define a variety of features derived from the classical Gestalt cues, including contour, texture, brightness and good continuation. Information-theoretic analysis is applied to evaluate the power of these grouping cues. We train a linear classifier to combine these features. To demonstrate the power of the classification model, a simple algorithm is used to randomly search for good segmentations. Results are shown on a wide range of images.

1,708 citations


Journal ArticleDOI
TL;DR: This paper implemented and tested the ALIP (Automatic Linguistic Indexing of Pictures) system on a photographic image database of 600 different concepts, each with about 40 training images and demonstrated the good accuracy of the system and its high potential in linguistic indexing of photographic images.
Abstract: Automatic linguistic indexing of pictures is an important but highly challenging problem for researchers in computer vision and content-based image retrieval. In this paper, we introduce a statistical modeling approach to this problem. Categorized images are used to train a dictionary of hundreds of statistical models each representing a concept. Images of any given concept are regarded as instances of a stochastic process that characterizes the concept. To measure the extent of association between an image and the textual description of a concept, the likelihood of the occurrence of the image based on the characterizing stochastic process is computed. A high likelihood indicates a strong association. In our experimental implementation, we focus on a particular group of stochastic processes, that is, the two-dimensional multiresolution hidden Markov models (2D MHMMs). We implemented and tested our ALIP (Automatic Linguistic Indexing of Pictures) system on a photographic image database of 600 different concepts, each with about 40 training images. The system is evaluated quantitatively using more than 4,600 images outside the training database and compared with a random annotation scheme. Experiments have demonstrated the good accuracy of the system and its high potential in linguistic indexing of photographic images.

1,163 citations


Journal ArticleDOI
TL;DR: It is seen that relatively few features are needed to achieve the same classification accuracies as in the original feature space when classification of panchromatic high-resolution data from urban areas using morphological and neural approaches.
Abstract: Classification of panchromatic high-resolution data from urban areas using morphological and neural approaches is investigated. The proposed approach is based on three steps. First, the composition of geodesic opening and closing operations of different sizes is used in order to build a differential morphological profile that records image structural information. Although, the original panchromatic image only has one data channel, the use of the composition operations will give many additional channels, which may contain redundancies. Therefore, feature extraction or feature selection is applied in the second step. Both discriminant analysis feature extraction and decision boundary feature extraction are investigated in the second step along with a simple feature selection based on picking the largest indexes of the differential morphological profiles. Third, a neural network is used to classify the features from the second step. The proposed approach is applied in experiments on high-resolution Indian Remote Sensing 1C (IRS-1C) and IKONOS remote sensing data from urban areas. In experiments, the proposed method performs well in terms of classification accuracies. It is seen that relatively few features are needed to achieve the same classification accuracies as in the original feature space.

756 citations


Journal ArticleDOI
TL;DR: Some recent results in statistical modeling of natural images that attempt to explain patterns of non-Gaussian behavior of image statistics, i.e. high kurtosis, heavy tails, and sharp central cusps are reviewed.
Abstract: Statistical analysis of images reveals two interesting properties: (i) invariance of image statistics to scaling of images, and (ii) non-Gaussian behavior of image statistics, i.e. high kurtosis, heavy tails, and sharp central cusps. In this paper we review some recent results in statistical modeling of natural images that attempt to explain these patterns. Two categories of results are considered: (i) studies of probability models of images or image decompositions (such as Fourier or wavelet decompositions), and (ii) discoveries of underlying image manifolds while restricting to natural images. Applications of these models in areas such as texture analysis, image classification, compression, and denoising are also considered.

561 citations


Proceedings ArticleDOI
13 Oct 2003
TL;DR: This work presents discriminative random fields (DRFs), a discrim inative framework for the classification of image regions by incorporating neighborhood interactions in the labels as well as the observed data that offers several advantages over the conventional Markov random field framework.
Abstract: In this work we present discriminative random fields (DRFs), a discriminative framework for the classification of image regions by incorporating neighborhood interactions in the labels as well as the observed data. The discriminative random fields offer several advantages over the conventional Markov random field (MRF) framework. First, the DRFs allow to relax the strong assumption of conditional independence of the observed data generally used in the MRF framework for tractability. This assumption is too restrictive for a large number of applications in vision. Second, the DRFs derive their classification power by exploiting the probabilistic discriminative models instead of the generative models used in the MRF framework. Finally, all the parameters in the DRF model are estimated simultaneously from the training data unlike the MRF framework where likelihood parameters are usually learned separately from the field parameters. We illustrate the advantages of the DRFs over the MRF framework in an application of man-made structure detection in natural images taken from the Corel database.

512 citations


Proceedings ArticleDOI
18 Jun 2003
TL;DR: A novel texton based representation is developed, which is suited to modeling this joint neighborhood distribution for MRFs, and it is demonstrated that textures can be classified using the joint distribution of intensity values over extremely compact neighborhoods.
Abstract: We question the role that large scale filter banks have traditionally played in texture classification. It is demonstrated that textures can be classified using the joint distribution of intensity values over extremely compact neighborhoods (starting from as small as 3 /spl times/ 3 pixels square), and that this outperforms classification using filter banks with large support. We develop a novel texton based representation, which is suited to modeling this joint neighborhood distribution for MRFs. The representation is learnt from training images, and then used to classify novel images (with unknown viewpoint and lighting) into texture classes. The power of the method is demonstrated by classifying over 2800 images of all 61 textures present in the Columbia-Utrecht database. The classification performance surpasses that of recent state-of-the-art filter bank based classifiers such as Leung & Malik, Cula & Dana, and Varma & Zisserman.

504 citations


Journal ArticleDOI
TL;DR: An independent Gabor features (IGFs) method and its application to face recognition is presented, which achieves 98.5% correct face recognition accuracy when using 180 features for the FERET dataset, and 100% accuracy for the ORL dataset using 88 features.
Abstract: We present an independent Gabor features (IGFs) method and its application to face recognition. The novelty of the IGF method comes from 1) the derivation of independent Gabor features in the feature extraction stage and 2) the development of an IGF features-based probabilistic reasoning model (PRM) classification method in the pattern recognition stage. In particular, the IGF method first derives a Gabor feature vector from a set of downsampled Gabor wavelet representations of face images, then reduces the dimensionality of the vector by means of principal component analysis, and finally defines the independent Gabor features based on the independent component analysis (ICA). The independence property of these Gabor features facilitates the application of the PRM method for classification. The rationale behind integrating the Gabor wavelets and the ICA is twofold. On the one hand, the Gabor transformed face images exhibit strong characteristics of spatial locality, scale, and orientation selectivity. These images can, thus, produce salient local features that are most suitable for face recognition. On the other hand, ICA would further reduce redundancy and represent independent features explicitly. These independent features are most useful for subsequent pattern discrimination and associative recall. Experiments on face recognition using the FacE REcognition Technology (FERET) and the ORL datasets, where the images vary in illumination, expression, pose, and scale, show the feasibility of the IGF method. In particular, the IGF method achieves 98.5% correct face recognition accuracy when using 180 features for the FERET dataset, and 100% accuracy for the ORL dataset using 88 features.

488 citations


Journal ArticleDOI
TL;DR: This work proposes a content-based soft annotation procedure for providing images with semantical labels, and experiments with two learning methods, support vector machines (SVMs) and Bayes point machines (BPMs), to select a base binary-classifier for CBSA.
Abstract: We propose a content-based soft annotation (CBSA) procedure for providing images with semantical labels. The annotation procedure starts with labeling a small set of training images, each with one single semantical label (e.g., forest, animal, or sky). An ensemble of binary classifiers is then trained for predicting label membership for images. The trained ensemble is applied to each individual image to give the image multiple soft labels, and each label is associated with a label membership factor. To select a base binary-classifier for CBSA, we experiment with two learning methods, support vector machines (SVMs) and Bayes point machines (BPMs), and compare their class-prediction accuracy. Our empirical study on a 116-category 25K-image set shows that the BPM-based ensemble provides better annotation quality than the SVM-based ensemble for supporting multimodal image retrievals.

479 citations


Proceedings ArticleDOI
13 Oct 2003
TL;DR: This work exploits a recently proposed approximation technique, locality-sensitive hashing (LSH), to reduce the computational complexity of adaptive mean shift and implements the implementation of LSH, where the optimal parameters of the data structure are determined by a pilot learning procedure, and the partitions are data driven.
Abstract: Feature space analysis is the main module in many computer vision tasks. The most popular technique, k-means clustering, however, has two inherent limitations: the clusters are constrained to be spherically symmetric and their number has to be known a priori. In nonparametric clustering methods, like the one based on mean shift, these limitations are eliminated but the amount of computation becomes prohibitively large as the dimension of the space increases. We exploit a recently proposed approximation technique, locality-sensitive hashing (LSH), to reduce the computational complexity of adaptive mean shift. In our implementation of LSH the optimal parameters of the data structure are determined by a pilot learning procedure, and the partitions are data driven. As an application, the performance of mode and k-means based textons are compared in a texture classification study.

465 citations


Proceedings ArticleDOI
13 Oct 2003
TL;DR: Large-scale recognition results are presented, which demonstrate that SVMs with the proposed kernel perform better than standard matching techniques on local features and that local feature representations significantly outperform global approaches.
Abstract: Recent developments in computer vision have shown that local features can provide efficient representations suitable for robust object recognition. Support vector machines have been established as powerful learning algorithms with good generalization capabilities. We combine these two approaches and propose a general kernel method for recognition with local features. We show that the proposed kernel satisfies the Mercer condition and that it is, suitable for many established local feature frameworks. Large-scale recognition results are presented on three different databases, which demonstrate that SVMs with the proposed kernel perform better than standard matching techniques on local features. In addition, experiments on noisy and occluded images show that local feature representations significantly outperform global approaches.

Journal ArticleDOI
TL;DR: An object-based approach for urban land cover classification from high-resolution multispectral image data that builds upon a pixel-based fuzzy classification approach is presented and is able to identify buildings, impervious surface, and roads in dense urban areas with 76, 81, and 99% classification accuracies.
Abstract: In this paper, we present an object-based approach for urban land cover classification from high-resolution multispectral image data that builds upon a pixel-based fuzzy classification approach. This combined pixel/object approach is demonstrated using pan-sharpened multispectral IKONOS imagery from dense urban areas. The fuzzy pixel-based classifier utilizes both spectral and spatial information to discriminate between spectrally similar road and building urban land cover classes. After the pixel-based classification, a technique that utilizes both spectral and spatial heterogeneity is used to segment the image to facilitate further object-based classification. An object-based fuzzy logic classifier is then implemented to improve upon the pixel-based classification by identifying one additional class in dense urban areas: nonroad, nonbuilding impervious surface. With the fuzzy pixel-based classification as input, the object-based classifier then uses shape, spectral, and neighborhood features to determine the final classification of the segmented image. Using these techniques, the object-based classifier is able to identify buildings, impervious surface, and roads in dense urban areas with 76%, 81%, and 99% classification accuracies, respectively.

Journal ArticleDOI
TL;DR: It is concluded that general robust PV segmentation of MR brain images requires statistical models that describe the spatial distribution of brain tissues more accurately than currently available models.
Abstract: Accurate brain tissue segmentation by intensity-based voxel classification of magnetic resonance (MR) images is complicated by partial volume (PV) voxels that contain a mixture of two or more tissue types. In this paper, we present a statistical framework for PV segmentation that encompasses and extends existing techniques. We start from a commonly used parametric statistical image model in which each voxel belongs to one single tissue type, and introduce an additional downsampling step that causes partial voluming along the borders between tissues. An expectation-maximization approach is used to simultaneously estimate the parameters of the resulting model and perform a PV classification. We present results on well-chosen simulated images and on real MR images of the brain, and demonstrate that the use of appropriate spatial prior knowledge not only improves the classifications, but is often indispensable for robust parameter estimation as well. We conclude that general robust PV segmentation of MR brain images requires statistical models that describe the spatial distribution of brain tissues more accurately than currently available models.

Journal ArticleDOI
TL;DR: A novel technique to automatically find lesion margins in ultrasound images, by combining intensity and texture with empirical domain specific knowledge along with directional gradient and a deformable shape-based model is presented.
Abstract: Breast cancer is the most frequently diagnosed malignancy and the second leading cause of mortality in women . In the last decade, ultrasound along with digital mammography has come to be regarded as the gold standard for breast cancer diagnosis. Automatically detecting tumors and extracting lesion boundaries in ultrasound images is difficult due to their specular nature and the variance in shape and appearance of sonographic lesions. Past work on automated ultrasonic breast lesion segmentation has not addressed important issues such as shadowing artifacts or dealing with similar tumor like structures in the sonogram. Algorithms that claim to automatically classify ultrasonic breast lesions, rely on manual delineation of the tumor boundaries. In this paper, we present a novel technique to automatically find lesion margins in ultrasound images, by combining intensity and texture with empirical domain specific knowledge along with directional gradient and a deformable shape-based model. The images are first filtered to remove speckle noise and then contrast enhanced to emphasize the tumor regions. For the first time, a mathematical formulation of the empirical rules used by radiologists in detecting ultrasonic breast lesions, popularly known as the "Stavros Criteria" is presented in this paper. We have applied this formulation to automatically determine a seed point within the image. Probabilistic classification of image pixels based on intensity and texture is followed by region growing using the automatically determined seed point to obtain an initial segmentation of the lesion. Boundary points are found on the directional gradient of the image. Outliers are removed by a process of recursive refinement. These boundary points are then supplied as an initial estimate to a deformable model. Incorporating empirical domain specific knowledge along with low and high-level knowledge makes it possible to avoid shadowing artifacts and lowers the chance of confusing similar tumor like structures for the lesion. The system was validated on a database of breast sonograms for 42 patients. The average mean boundary error between manual and automated segmentation was 6.6 pixels and the normalized true positive area overlap was 75.1%. The algorithm was found to be robust to 1) variations in system parameters, 2) number of training samples used, and 3) the position of the seed point within the tumor. Running time for segmenting a single sonogram was 18 s on a 1.8-GHz Pentium machine.

Proceedings ArticleDOI
TL;DR: An innovative image annotation tool for classifying image regions in one of seven classes - sky, skin, vegetation, snow, water, ground, and buildings - or as unknown is described.
Abstract: The paper describes an innovative image annotation tool for classifying image regions in one of seven classes - sky, skin, vegetation, snow, water, ground, and buildings - or as unknown. This tool could be productively applied in the management of large image and video databases where a considerable volume of images/frames there must be automatically indexed. The annotation is performed by a classification system based on a multi-class Support Vector Machine. Experimental results on a test set of 200 images are reported and discussed.

Journal ArticleDOI
TL;DR: In this paper, the authors assess the accuracy of three different methods for extracting land-cover/land-use information from high-resolution imagery of urban environments: (1) combined supervised/ unsupervised spectral classification, (2) raster-based spatial modeling, and (3) image segmentation classification using classification tree analysis.
Abstract: Recent advances in digital airborne sensors and satellite platforms make spatially accurate, high-resolution multispectral imagery readily available. These advances provide the opportunity for a host of new applications to address and solve old problems. High-resolution imagery is particularly well suited to urban applications. Previous data sources (such as Landsat TM) did not show the spatial detail necessary to provide many urban planning solutions. This paper provides an overview of a project in which one-meter digital imagery was used to produce a map of pervious and impervious surfaces to be used by the city of Scottsdale, Arizona for storm-water runoff estimation. The increased spatial information in onemeter or less resolution imagery strains the usefulness of image classification using traditional supervised and unsupervised spectral classification algorithms. This study assesses the accuracy of three different methods for extracting land-cover/land-use information from high-resolution imagery of urban environments: (1) combined supervised/ unsupervised spectral classification, (2) raster-based spatial modeling, and (3) image segmentation classification using classification tree analysis. A discussion of the results and relative merits of each method is included.

Journal ArticleDOI
01 Sep 2003
TL;DR: A computer-aided diagnostic (CAD) system for the classification of hepatic lesions from computed tomography (CT) images is presented and shows that genetic algorithms result in lower dimension feature vectors and improved classification performance.
Abstract: In this paper, a computer-aided diagnostic (CAD) system for the classification of hepatic lesions from computed tomography (CT) images is presented. Regions of interest (ROIs) taken from nonenhanced CT images of normal liver, hepatic cysts, hemangiomas, and hepatocellular carcinomas have been used as input to the system. The proposed system consists of two modules: the feature extraction and the classification modules. The feature extraction module calculates the average gray level and 48 texture characteristics, which are derived from the spatial gray-level co-occurrence matrices, obtained from the ROIs. The classifier module consists of three sequentially placed feed-forward neural networks (NNs). The first NN classifies into normal or pathological liver regions. The pathological liver regions are characterized by the second NN as cyst or "other disease". The third NN classifies "other disease" into hemangioma or hepatocellular carcinoma. Three feature selection techniques have been applied to each individual NN: the sequential forward selection, the sequential floating forward selection, and a genetic algorithm for feature selection. The comparative study of the above dimensionality reduction methods shows that genetic algorithms result in lower dimension feature vectors and improved classification performance.

Journal ArticleDOI
TL;DR: The proposed classification scheme using log-polar wavelet signatures outperforms two other texture classification methods, its overall accuracy rate for joint rotation and scale invariance being 90.8 percent, and its robustness to noise also performs better than the other methods.
Abstract: Classification of texture images is important in image analysis and classification This paper proposes an effective scheme for rotation and scale invariant texture classification using log-polar wavelet signatures The rotation and scale invariant feature extraction for a given image involves applying a log-polar transform to eliminate the rotation and scale effects, but at same time produce a row shifted log-polar image, which is then passed to an adaptive row shift invariant wavelet packet transform to eliminate the row shift effects So, the output wavelet coefficients are rotation and scale invariant The adaptive row shift invariant wavelet packet transform is quite efficient with only O(n /spl middot/ log n) complexity A feature vector of the most dominant log-polar wavelet energy signatures extracted from each subband of wavelet coefficients is constructed for rotation and scale invariant texture classification In the experiments, we employed a Mahalanobis classifier to classify a set of 25 distinct natural textures selected from the Brodatz album The experimental results, based on different testing data sets for images with different orientations and scales, show that the proposed classification scheme using log-polar wavelet signatures outperforms two other texture classification methods, its overall accuracy rate for joint rotation and scale invariance being 908 percent, demonstrating that the extracted energy signatures are effective rotation and scale invariant features Concerning its robustness to noise, the classification scheme also performs better than the other methods

Journal ArticleDOI
TL;DR: This paper investigates the usefulness of high-resolution multispectral satellite imagery for classification of urban and suburban areas and presents a fuzzy logic methodology to improve classification accuracy and shows a hierarchical fuzzy classification approach that makes use of both spectral and spatial information.
Abstract: In this paper, we investigate the usefulness of high-resolution multispectral satellite imagery for classification of urban and suburban areas and present a fuzzy logic methodology to improve classification accuracy. Panchromatic and multispectral IKONOS image datasets are analyzed for two urban locations in this study. Both multispectral and pan-sharpened multispectral images are first classified using a traditional maximum-likelihood approach. Maximum-likelihood classification accuracies between 79% to 87% were achieved with significant misclassification error between the spectrally similar Road and Building urban land cover types. A number of different texture measures were investigated, and a length-width contextual measure is developed. These spatial measures were used to increase the discrimination between spectrally similar classes, thereby yielding higher accuracy urban land cover maps. Finally, a hierarchical fuzzy classification approach that makes use of both spectral and spatial information is presented. This technique is shown to increase the discrimination between spectrally similar urban land cover classes and results in classification accuracies that are 8% to 11% larger than those from the traditional maximum-likelihood approach.

Proceedings ArticleDOI
21 Jul 2003
TL;DR: It appears that modified kernels are presented to take into consideration the spectral similarity between support vectors to outperform SVM-based classification of hyperspectral data cube and reduce false alarms that were induced by illumination effects with classical kernels.
Abstract: Support vector machines (SVM) has been recently used with success for the classification of hyperspectral images. This method appears to be a robust alternative for pattern recognition with hyperspectral data: since the method is based on a geometric point of view, no statistical estimation has to be achieved. Then, SVM outperforms classical supervised classification algorithms such as the maximum likelihood when the number of spectral bands increases or when the number of training samples remains limited. Nevertheless, those kernel-based methods do not take into consideration the spectral similarity between support vectors. Then, some modified kernels are presented to take into consideration the spectral similarity between support vectors to outperform SVM-based classification of hyperspectral data cube. Those kernels (that still suit Mercer's conditions) are based on the use of spectral angle to evaluate the distance between support vectors. Classifiers to compare have been applied to an image from the CASI sensor including 17 bands from 450 to 950nm representing an intensive agricultural region (Brittany, France). It appears that those kernels reduce false alarms that were induced by illumination effects with classical kernels.

Journal ArticleDOI
TL;DR: This work addresses two problems that are often encountered in object recognition: object segmentation, for which a distance sets shape filter is formulated, and shape matching, which is illustrated on printed and handwritten character recognition and detection of traffic signs in complex scenes.
Abstract: We introduce a novel rich local descriptor of an image point, we call the (labeled) distance set, which is determined by the spatial arrangement of image features around that point. We describe a two-dimensional (2D) visual object by the set of (labeled) distance sets associated with the feature points of that object. Based on a dissimilarity measure between (labeled) distance sets and a dissimilarity measure between sets of (labeled) distance sets, we address two problems that are often encountered in object recognition: object segmentation, for which we formulate a distance sets shape filter, and shape matching. The use of the shape filter is illustrated on printed and handwritten character recognition and detection of traffic signs in complex scenes. The shape comparison procedure is illustrated on handwritten character classification, COIL-20 database object recognition and MPEG-7 silhouette database retrieval.

Journal ArticleDOI
TL;DR: The problem of classifying an image into different homogeneous regions is viewed as the task of clustering the pixels in the intensity space and real-coded variable string length genetic fuzzy clustering with automatic evolution of clusters is used.
Abstract: The problem of classifying an image into different homogeneous regions is viewed as the task of clustering the pixels in the intensity space. Real-coded variable string length genetic fuzzy clustering with automatic evolution of clusters is used for this purpose. The cluster centers are encoded in the chromosomes, and the Xie-Beni index is used as a measure of the validity of the corresponding partition. The effectiveness of the proposed technique is demonstrated for classifying different landcover regions in remote sensing imagery. Results are compared with those obtained using the well-known fuzzy C-means algorithm.

Proceedings ArticleDOI
27 Oct 2003
TL;DR: Experiments with artificial data containing varying levels of noise and occlusion of the objects show that Kullback-Leibler and likelihood matching yield robust recognition rates.
Abstract: A statistical representation of three-dimensional shapes is introduced, based on a novel four-dimensional feature. The feature parameterizes the intrinsic geometrical relation of an oriented surface-point pair. The set of all such features represents both local and global characteristics of the surface. We compress this set into a histogram. A database of histograms, one per object, is sampled in a training phase. During recognition, sensed surface data, as may be acquired by stereo vision, a laser range-scanner, etc., are processed and compared to the stored histograms. We evaluate the match quality by six different criteria that are commonly used in statistical settings. Experiments with artificial data containing varying levels of noise and occlusion of the objects show that Kullback-Leibler and likelihood matching yield robust recognition rates. We propose histograms of the geometric relation between two oriented surface points (surflets) as a compact yet distinctive representation of arbitrary three-dimensional shapes.

Journal ArticleDOI
TL;DR: In a study to assess the map-updating capabilities of such sensors in urban areas, some modern texture measures were investigated and the conclusion is that texture improves the classification accuracy.
Abstract: In single-band and single-polarized synthetic aperture radar (SAR) image classification, texture holds useful information. In a study to assess the map-updating capabilities of such sensors in urban areas, some modern texture measures were investigated. Among them were histogram measures, wavelet energy, fractal dimension, lacunarity, and semivariograms. The latter were chosen as an alternative for the well-known gray-level cooccurrence family of features. The area that was studied using a European Remote Sensing Satellite 1 (ERS-1) SAR image was the conurbation around Rotterdam and The Hague in The Netherlands. The area can be characterized as a well-planned dispersed urban area with residential areas, industry, greenhouses, pasture, arable land, and some forest. The digital map to be updated was a 1:250000 Vector Map (VMap1). The study was done on the basis of nonparametric separability measures and classification techniques because most texture distributions were not normal. The conclusion is that texture improves the classification accuracy. The measures that performed best were mean intensity (actually no texture), variance, weighted-rank fill ratio, and semivariogram, but the accuracies vary for different classes. Despite the improvement, the overall classification accuracy indicates that the land-cover information content of ERS-1 leaves something to be desired.

Journal ArticleDOI
TL;DR: A new feedback approach with progressive learning capability combined with a novel method for the feature subspace extraction based on a Bayesian classifier that treats positive and negative feedback examples with different strategies to improve the retrieval accuracy.
Abstract: Research has been devoted in the past few years to relevance feedback as an effective solution to improve performance of content-based image retrieval (CBIR). In this paper, we propose a new feedback approach with progressive learning capability combined with a novel method for the feature subspace extraction. The proposed approach is based on a Bayesian classifier and treats positive and negative feedback examples with different strategies. Positive examples are used to estimate a Gaussian distribution that represents the desired images for a given query; while the negative examples are used to modify the ranking of the retrieved candidates. In addition, feature subspace is extracted and updated during the feedback process using a principal component analysis (PCA) technique and based on user's feedback. That is, in addition to reducing the dimensionality of feature spaces, a proper subspace for each type of features is obtained in the feedback process to further improve the retrieval accuracy. Experiments demonstrate that the proposed method increases the retrieval speed, reduces the required memory and improves the retrieval accuracy significantly.

Journal ArticleDOI
TL;DR: A filter selection algorithm is proposed to maximize classification performance of a given dataset and the spectral histogram representation provides a robust feature statistic for textures and generalizes well.
Abstract: Based on a local spatial/frequency representation,we employ a spectral histogram as a feature statistic for texture classification. The spectral histogram consists of marginal distributions of responses of a bank of filters and encodes implicitly the local structure of images through the filtering stage and the global appearance through the histogram stage. The distance between two spectral histograms is measured using /spl chi//sup 2/-statistic. The spectral histogram with the associated distance measure exhibits several properties that are necessary for texture classification. A filter selection algorithm is proposed to maximize classification performance of a given dataset. Our classification experiments using natural texture images reveal that the spectral histogram representation provides a robust feature statistic for textures and generalizes well. Comparisons show that our method produces a marked improvement in classification performance. Finally we point out the relationships between existing texture features and the spectral histogram, suggesting that the latter may provide a unified texture feature.

Journal ArticleDOI
TL;DR: It is shown that automatic wavelet reduction yields better or comparable classification accuracy for hyperspectral data, while achieving substantial computational savings.
Abstract: Hyperspectral imagery provides richer information about materials than multispectral imagery. The new larger data volumes from hyperspectral sensors present a challenge for traditional processing techniques. For example, the identification of each ground surface pixel by its corresponding spectral signature is still difficult because of the immense volume of data. Conventional classification methods may not be used without dimension reduction preprocessing. This is due to the curse of dimensionality, which refers to the fact that the sample size needed to estimate a function of several variables to a given degree of accuracy grows exponentially with the number of variables. Principal component analysis (PCA) has been the technique of choice for dimension reduction. However, PCA is computationally expensive and does not eliminate anomalies that can be seen at one arbitrary band. Spectral data reduction using automatic wavelet decomposition could be useful. This is because it preserves the distinctions among spectral signatures. It is also computed in automatic fashion and can filter data anomalies. This is due to the intrinsic properties of wavelet transforms that preserves high- and low-frequency features, therefore preserving peaks and valleys found in typical spectra. Compared to PCA, for the same level of data reduction, we show that automatic wavelet reduction yields better or comparable classification accuracy for hyperspectral data, while achieving substantial computational savings.

Proceedings ArticleDOI
Ce Liu1, Hueng-Yeung Shum1
18 Jun 2003
TL;DR: To apply KLBoosting to high-dimensional image space, this paper proposes a data-driven Kullback-Leibler Analysis (KLA) approach to find KL features for image objects (e.g., face patches) and promising experimental results on face detection demonstrate the effectiveness of KLBoosted.
Abstract: In this paper, we develop a general classification framework called Kullback-Leibler Boosting, or KLBoosting. KLBoosting has following properties. First, classification is based on the sum of histogram divergences along corresponding global and discriminating linear features. Second, these linear features, called KL features, are iteratively learnt by maximizing the projected Kullback-Leibler divergence in a boosting manner. Third, the coefficients to combine the histogram divergences are learnt by minimizing the recognition error once a new feature is added to the classifier. This contrasts conventional AdaBoost where the coefficients are empirically set. Because of these properties, KLBoosting classifier generalizes very well. Moreover, to apply KLBoosting to high-dimensional image space, we propose a data-driven Kullback-Leibler Analysis (KLA) approach to find KL features for image objects (e.g., face patches). Promising experimental results on face detection demonstrate the effectiveness of KLBoosting.

Journal ArticleDOI
TL;DR: A comparison between NMF, WNMF and the well-known principal component analysis (PCA) in the context of image patch classification has been carried out and it is claimed that all three techniques can be combined in a common and unique classifier.

Proceedings ArticleDOI
13 Oct 2003
TL;DR: A human recognition algorithm by combining static and dynamic body biometrics, fused on the decision level using different combinations of rules to improve the performance of both identification and verification is described.
Abstract: Human identification at a distance has recently gained growing interest from computer vision researchers. This paper aims to propose a visual recognition algorithm based upon fusion of static and dynamic body biometrics. For each sequence involving a walking figure, pose changes of the segmented moving silhouettes are represented as an associated sequence of complex vector configurations, and are then analyzed using the Procrustes shape analysis method to obtain a compact appearance representation, called static information of body. Also, a model-based approach is presented under a condensation framework to track the walker and to recover joint-angle trajectories of lower limbs, called dynamic information of gait. Both static and dynamic cues are respectively used for recognition using the nearest exemplar classifier. They are also effectively fused on decision level using different combination rules to improve the performance of both identification and verification. Experimental results on a dataset including 20 subjects demonstrate the validity of the proposed algorithm.