scispace - formally typeset
Search or ask a question

Showing papers on "Segmentation-based object categorization published in 2003"


Journal ArticleDOI
TL;DR: A parametric model for an implicit representation of the segmenting curve is derived by applying principal component analysis to a collection of signed distance representations of the training data to minimize an objective function for segmentation.
Abstract: We propose a shape-based approach to curve evolution for the segmentation of medical images containing known object types. In particular, motivated by the work of Leventon, Grimson, and Faugeras (2000), we derive a parametric model for an implicit representation of the segmenting curve by applying principal component analysis to a collection of signed distance representations of the training data. The parameters of this representation are then manipulated to minimize an objective function for segmentation. The resulting algorithm is able to handle multidimensional data, can deal with topological changes of the curve, is robust to noise and initial contour placements, and is computationally efficient. At the same time, it avoids the need for point correspondences during the training phase of the algorithm. We demonstrate this technique by applying it to two medical applications; two-dimensional segmentation of cardiac magnetic resonance imaging (MRI) and three-dimensional segmentation of prostate MRI.

974 citations


Journal ArticleDOI
TL;DR: Results show how visual categorization based directly on low-level features, without grouping or segmentation stages, can benefit object localization and identification.
Abstract: In this paper we study the statistical properties of natural images belonging to different categories and their relevance for scene and object categorization tasks. We discuss how second-order statistics are correlated with image categories, scene scale and objects. We propose how scene categorization could be computed in a feedforward manner in order to provide top-down and contextual information very early in the visual processing chain. Results show how visual categorization based directly on low-level features, without grouping or segmentation stages, can benefit object localization and identification. We show how simple image statistics can be used to predict the presence and absence of objects in the scene before exploring the image.

858 citations


Proceedings ArticleDOI
17 Sep 2003
TL;DR: A new algorithm for fuzzy segmentation of MR brain images is presented, starting from the standard FCM and its bias-corrected version BCFCM algorithm, and by introducing a new factor, the amount of required calculations is considerably reduced.
Abstract: This paper presents a new algorithm for fuzzy segmentation of MR brain images. Starting from the standard FCM and its bias-corrected version BCFCM algorithm, by splitting up the two major steps of the latter, and by introducing a new factor, the amount of required calculations is considerably reduced. The algorithm provides good-quality segmented brain images a very quick way, which makes it an excellent tool to support virtual brain endoscopy.

430 citations


Proceedings Article
Jean Serra1
01 Jan 2003
TL;DR: An axiomatic definition for the notion of "segmentation" in image processing is proposed, which is based on the idea of a maximal partition and a key theorem links segmentation with connection, on the one hand, and with connective criteria on the other one.
Abstract: Firstly, the paper proposes an axiomatic definition for the notion of "segmentation" in image processing, which is based on the idea of a maximal partition. Then a key theorem links segmentation with connection, on the one hand, and with connective criteria on the other one. A series of lattice properties are then developed. In a last part, two examples of segmentations are proposed.

386 citations


Journal ArticleDOI
TL;DR: The segmentation of the kidney from CT and the hippocampus from MRI serve as the major examples in this paper and the accuracy of segmentation as compared to manual, slice-by-slice segmentation is reported.
Abstract: M-reps (formerly called DSLs) are a multiscale medial means for modeling and rendering 3D solid geometry. They are particularly well suited to model anatomic objects and in particular to capture prior geometric information effectively in deformable models segmentation approaches. The representation is based on figural models, which define objects at coarse scale by a hierarchy of figures—each figure generally a slab representing a solid region and its boundary simultaneously. This paper focuses on the use of single figure models to segment objects of relatively simple structure. A single figure is a sheet of medial atoms, which is interpolated from the model formed by a net, i.e., a mesh or chain, of medial atoms (hence the name m-reps), each atom modeling a solid region via not only a position and a width but also a local figural frame giving figural directions and an object angle between opposing, corresponding positions on the boundary implied by the m-rep. The special capability of an m-rep is to provide spatial and orientational correspondence between an object in two different states of deformation. This ability is central to effective measurement of both geometric typicality and geometry to image match, the two terms of the objective function optimized in segmentation by deformable models. The other ability of m-reps central to effective segmentation is their ability to support segmentation at multiple levels of scale, with successively finer precision. Objects modeled by single figures are segmented first by a similarity transform augmented by object elongation, then by adjustment of each medial atom, and finally by displacing a dense sampling of the m-rep implied boundary. While these models and approaches also exist in 2D, we focus on 3D objects. The segmentation of the kidney from CT and the hippocampus from MRI serve as the major examples in this paper. The accuracy of segmentation as compared to manual, slice-by-slice segmentation is reported.

384 citations


Journal ArticleDOI
TL;DR: This work has shown that tight clustering of nuclei in 3D confocal microscope images is a common source of segmentation error, and a compelling need to minimize these errors for constructing highly automated scoring systems.
Abstract: Background Automated segmentation of fluorescently-labeled cell nuclei in 3D confocal microscope images is essential to many studies involving morphological and functional analysis. A common source of segmentation error is tight clustering of nuclei. There is a compelling need to minimize these errors for constructing highly automated scoring systems. Methods A combination of two approaches is presented. First, an improved distance transform combining intensity gradients and geometric distance is used for the watershed step. Second, an explicit mathematical model for the anatomic characteristics of cell nuclei such as size and shape measures is incorporated. This model is constructed automatically from the data. Deliberate initial over-segmentation of the image data is performed, followed by statistical model-based merging. A confidence score is computed for each detected nucleus, measuring how well the nucleus fits the model. This is used in combination with the intensity gradient to control the merge decisions. Results Experimental validation on a set of rodent brain cell images showed 97% concordance with the human observer and significant improvement over prior methods. Conclusions Combining a gradient-weighted distance transform with a richer morphometric model significantly improves the accuracy of automated segmentation and FISH analysis. Cytometry Part A 56A:23–36, 2003. © 2003 Wiley-Liss, Inc.

338 citations


Journal ArticleDOI
TL;DR: A new cost function, cut ratio, for segmenting images using graph-based methods that allows the image perimeter to be segmented, guarantees that the segments produced by bipartitioning are connected, and does not introduce a size, shape, smoothness, or boundary-length bias.
Abstract: This paper proposes a new cost function, cut ratio, for segmenting images using graph-based methods. The cut ratio is defined as the ratio of the corresponding sums of two different weights of edges along the cut boundary and models the mean affinity between the segments separated by the boundary per unit boundary length. This new cost function allows the image perimeter to be segmented, guarantees that the segments produced by bipartitioning are connected, and does not introduce a size, shape, smoothness, or boundary-length bias. The latter allows it to produce segmentations where boundaries are aligned with image edges. Furthermore, the cut-ratio cost function allows efficient iterated region-based segmentation as well as pixel-based segmentation. These properties may be useful for some image-segmentation applications. While the problem of finding a minimum ratio cut in an arbitrary graph is NP-hard, one can find a minimum ratio cut in the connected planar graphs that arise during image segmentation in polynomial time. While the cut ratio, alone, is not sufficient as a baseline method for image segmentation, it forms a good basis for an extended method of image segmentation when combined with a small number of standard techniques. We present an implemented algorithm for finding a minimum ratio cut, prove its correctness, discuss its application to image segmentation, and present the results of segmenting a number of medical and natural images using our techniques.

303 citations


Proceedings ArticleDOI
18 Jun 2003
TL;DR: A variational framework is proposed that incorporates a small set of good features for texture segmentation based on the structure tensor and nonlinear diffusion in a level set based unsupervised segmentation process that adaptively takes into account their estimated statistical information inside and outside the region to segment.
Abstract: We propose a novel and efficient approach for active unsupervised texture segmentation. First, we show how we can extract a small set of good features for texture segmentation based on the structure tensor and nonlinear diffusion. Then, we propose a variational framework that incorporates these features in a level set based unsupervised segmentation process that adaptively takes into account their estimated statistical information inside and outside the region to segment. The approach has been tested on various textured images, and its performance is favorably compared to recent studies.

296 citations


Book ChapterDOI
01 Jan 2003
TL;DR: A novel method for the categorization of unfamiliar objects in difficult real-world scenes is presented, which uses a probabilistic formulation to incorporate knowledge about the recognized category as well as the supporting information in the image to segment the object from the background.
Abstract: Historically, figure-ground segmentation has been seen as an important and even necessary precursor for object recognition In that context, segmentation is mostly defined as a data driven, that is bottom-up, process As for humans object recognition and segmentation are heavily intertwined processes, it has been argued that top-down knowledge from object recognition can and should be used for guiding the segmentation process In this paper, we present a method for the categorization of unfamiliar objects in difficult real-world scenes The method generates object hypotheses without prior segmentation that can be used to obtain a category-specific figure-ground segmentation In particular, the proposed approach uses a probabilistic formulation to incorporate knowledge about the recognized category as well as the supporting information in the image to segment the object from the background This segmentation can then be used for hypothesis verification, to further improve recognition performance Experimental results show the capacity of the approach to categorize and segment object categories as diverse as cars and cows

232 citations


Journal ArticleDOI
TL;DR: The performance of this segmentation algorithm is superior to traditional single resolution techniques such as texture spectrum, co-occurrences, local linear transforms, etc.

219 citations


Journal ArticleDOI
TL;DR: This paper shows how to represent watershed segmentation as an energy minimization problem using the distance-based definition of the watershed line, and proposes a new segmentation method called watersnakes, integrating the strengths of watershed segmentsation and energy based segmentation.
Abstract: The watershed algorithm from mathematical morphology is powerful for segmentation. However, it does not allow incorporation of a priori information as segmentation methods that are based on energy minimization. In particular, there is no control of the smoothness of the segmentation result. In this paper, we show how to represent watershed segmentation as an energy minimization problem using the distance-based definition of the watershed line. A priori considerations about smoothness can then be imposed by adding the contour length to the energy function. This leads to a new segmentation method called watersnakes, integrating the strengths of watershed segmentation and energy based segmentation. Experimental results show that, when the original watershed segmentation has noisy boundaries or wrong limbs attached to the object of interest, the proposed method overcomes those drawbacks and yields a better segmentation.

Proceedings ArticleDOI
18 Jun 2003
TL;DR: A framework for the image segmentation problem based on a new graph-theoretic formulation of clustering, motivated by the analogies between the intuitive concept of a cluster and that of a dominant set of vertices, which establishes a correspondence between dominant sets and the extrema of a quadratic form over the standard simplex.
Abstract: We develop a framework for the image segmentation problem based on a new graph-theoretic formulation of clustering. The approach is motivated by the analogies between the intuitive concept of a cluster and that of a dominant set of vertices, a notion that generalizes that of a maximal complete subgraph to edge-weighted graphs. We also establish a correspondence between dominant sets and the extrema of a quadratic form over the standard simplex, thereby allowing us the use of continuous optimization techniques such as replicator dynamics from evolutionary game theory. Such systems are attractive as they can be coded in a few lines of any high-level programming language, can easily be implemented in a parallel network of locally interacting units, and offer the advantage of biological plausibility. We present experimental results on real-world images which show the effectiveness of the proposed approach.

Proceedings ArticleDOI
24 Nov 2003
TL;DR: The proposed approach bridges the gap between keyword-based approaches, which assume the existence of rich image captions or require manual evaluation and annotation of every image of the collection, and query-by-example approaches,Which assume that the user queries for images similar to one that already is at his disposal.
Abstract: In this paper, an image retrieval methodology suited for search in large collections of heterogeneous images is presented. The proposed approach employs a fully unsupervised segmentation algorithm to divide images into regions. Low-level features describing the color, position, size and shape of the resulting regions are extracted and are automatically mapped to appropriate intermediate-level descriptors forming a simple vocabulary termed object ontology. The object ontology is used to allow the qualitative definition of the high-level concepts the user queries for (semantic objects, each represented by a keyword) in a human-centered fashion. When querying, clearly irrelevant image regions are rejected using the intermediate-level descriptors; following that, a relevance feedback mechanism employing the low-level features is invoked to produce the final query results. The proposed approach bridges the gap between keyword-based approaches, which assume the existence of rich image captions or require manual evaluation and annotation of every image of the collection, and query-by-example approaches, which assume that the user queries for images similar to one that already is at his disposal.

Proceedings ArticleDOI
TL;DR: A novel objective segmentation evaluation method based on information theory that uses entropy as the basis for measuring the uniformity of pixel characteristics (luminance is used in this paper) within a segmentation region.
Abstract: Accurate image segmentation is important for many image, video and computer vision applications. Over the last few decades, many image segmentation methods have been proposed. However, the results of these segmentation methods are usually evaluated only visually, qualitatively, or indirectly by the effectiveness of the segmentation on the subsequent processing steps. Such methods are either subjective or tied to particular applications. They do not judge the performance of a segmentation method objectively, and cannot be used as a means to compare the performance of different segmentation techniques. A few quantitative evaluation methods have been proposed, but these early methods have been based entirely on empirical analysis and have no theoretical grounding. In this paper, we propose a novel objective segmentation evaluation method based on information theory. The new method uses entropy as the basis for measuring the uniformity of pixel characteristics (luminance is used in this paper) within a segmentation region. The evaluation method provides a relative quality score that can be used to compare different segmentations of the same image. This method can be used to compare both various parameterizations of one particular segmentation method as well as fundamentally different segmentation techniques. The results from this preliminary study indicate that the proposed evaluation method is superior to the prior quantitative segmentation evaluation techniques, and identify areas for future research in objective segmentation evaluation.

Journal ArticleDOI
TL;DR: A new Bayesian formulation forParametric image segmentation is presented, based on the key idea of using a doubly stochastic prior model for the label field, which allows one to find exact optimal estimators for both this field and the model parameters by the minimization of a differentiable function.
Abstract: Parametric image segmentation consists of finding a label field that defines a partition of an image into a set of nonoverlapping regions and the parameters of the models that describe the variation of some property within each region. A new Bayesian formulation for the solution of this problem is presented, based on the key idea of using a doubly stochastic prior model for the label field, which allows one to find exact optimal estimators for both this field and the model parameters by the minimization of a differentiable function. An efficient minimization algorithm and comparisons with existing methods on synthetic images are presented, as well as examples of realistic applications to the segmentation of Magnetic Resonance volumes and to motion segmentation.

Journal ArticleDOI
TL;DR: Experimental results show that the determinant of the covariance matrix appears to be a very relevant tool for segmentation of homogeneous color regions for image and video segmentation using active contours.
Abstract: This paper deals with image and video segmentation using active contours. We propose a general form for the energy functional related to region-based active contours. We compute the associated evolution equation using shape derivation tools and accounting for the evolving region-based terms. Then we apply this general framework to compute the evolution equation from functionals that include various statistical measures of homogeneity for the region to be segmented. Experimental results show that the determinant of the covariance matrix appears to be a very relevant tool for segmentation of homogeneous color regions. As an example, it has been successfully applied to face segmentation in real video sequences.

Journal ArticleDOI
TL;DR: The most frequently used approach-based on a modified Hidden Markov Model (HMM) phonetic recognizer is analyzed, and a general framework for the local refinement of boundaries is proposed, and the performance of several pattern classification approaches is compared within this framework.
Abstract: This paper presents the results and conclusions of a thorough study on automatic phonetic segmentation. It starts with a review of the state of the art in this field. Then, it analyzes the most frequently used approach-based on a modified Hidden Markov Model (HMM) phonetic recognizer. For this approach, a statistical correction procedure is proposed to compensate for the systematic errors produced by context-dependent HMMs, and the use of speaker adaptation techniques is considered to increase segmentation precision. Finally, this paper explores the possibility of locally refining the boundaries obtained with the former techniques. A general framework is proposed for the local refinement of boundaries, and the performance of several pattern classification approaches (fuzzy logic, neural networks and Gaussian mixture models) is compared within this framework. The resulting phonetic segmentation scheme was able to increase the performance of a baseline HMM segmentation tool from 27.12%, 79.27%, and 97.75% of automatic boundary marks with errors smaller than 5, 20, and 50 ms, respectively, to 65.86%, 96.01%, and 99.31% in speaker-dependent mode, which is a reasonably good approximation to manual segmentation.

Journal ArticleDOI
TL;DR: This paper reviews various segmentation proposals that integrate edge and region information and highlights different strategies and methods for fusing such information.

Journal ArticleDOI
TL;DR: Numerical experiments on multispectral images show that the proposed algorithm is much faster than a similar reference algorithm based on "flat" MRF models, and its performance, in terms of segmentation accuracy and map smoothness, is comparable or even superior.
Abstract: We present a new image segmentation algorithm based on a tree-structured binary MRF model. The image is recursively segmented in smaller and smaller regions until a stopping condition, local to each region, is met. Each elementary binary segmentation is obtained as the solution of a MAP estimation problem, with the region prior modeled as an MRF. Since only binary fields are used, and thanks to the tree structure, the algorithm is quite fast, and allows one to address the cluster validation problem in a seamless way. In addition, all field parameters are estimated locally, allowing for some spatial adaptivity. To improve segmentation accuracy, a split-and-merge procedure is also developed and a spatially adaptive MRF model is used. Numerical experiments on multispectral images show that the proposed algorithm is much faster than a similar reference algorithm based on "flat" MRF models, and its performance, in terms of segmentation accuracy and map smoothness, is comparable or even superior.

Journal ArticleDOI
TL;DR: An improvement on the adaptivity is proposed by introducing an enhancement to control the adaptive properties of the segmentation process, which takes the form of a weighting function accounting for both local and global statistics, and is introduced in the minimisation.

Proceedings ArticleDOI
27 Oct 2003
TL;DR: Experimental results confirm the effectiveness of the proposed focus measures and the segmentation algorithm and are especially suitable for high resolution microscopic computer vision tasks in high precision micromanipulation and microassembly applications.
Abstract: This paper reports on the construction of two new focus measure operators M/sub WT//sup 1/ an M/sub WT//sup 2/ defined in the wavelet transform domain. M/sub WT//sup 2/ provides significantly better focus performance in depth resolution than previously reported spatial domain operators. M/sub WT//sup 1/ provides performance equivalent to that of the best spatial domain operator but has lower computational cost than M/sub WT//sup 2/. Both operators can be used with a wide variety of wavelet bases optimized for different applications. Selection of wavelet bases is studied based on their number of vanishing moments, size of support and symmetry. The depth resolution of these operators makes them an important cue in the segmentation of low depth-of-field microscopic images. An unsupervised segmentation technique based on graph partition is then introduced. It uses M/sub WT//sup 2/ together with proximity and image intensity as segmentation features. This segmentation method does not depend on the connection of local image features and remains robust under defocusing. Experimental results confirm the effectiveness of the proposed focus measures and the segmentation algorithm. These techniques are especially suitable for high resolution microscopic computer vision tasks in high precision micromanipulation and microassembly applications.

Journal ArticleDOI
TL;DR: A multi-scale level set framework for segmentation of endocardial boundaries at each frame in a multiframe echocardiographic image sequence is presented and it is pointed out that the intensity distribution of an ultrasound image at a very coarse scale can be approximately modeled by Gaussian.

Journal ArticleDOI
TL;DR: The proposed spatial fuzzy clustering algorithm is able to take into account both the distributions of data in feature space and the spatial interactions between neighboring pixels during clustering.
Abstract: In this paper, we describe the application of a novel spatial fuzzy clustering algorithm to the lip segmentation problem. The proposed spatial fuzzy clustering algorithm is able to take into account both the distributions of data in feature space and the spatial interactions between neighboring pixels during clustering. By appropriate pre- and postprocessing utilizing the color and shape properties of the lip region, successful segmentation of most lip images is possible. Comparative study with some existing lip segmentation algorithms such as the hue filtering algorithm and the fuzzy entropy histogram thresholding algorithm has demonstrated the superior performance of our method.

Proceedings ArticleDOI
21 Jul 2003
TL;DR: A relatively new segmentation approach, multiresolution segmentation, is being examined using two data sets (Landsat and IRS) to find optimum segmentation parameters for extracting different land cover classes.
Abstract: The main aim of this research is to find optimum segmentation parameters for extracting different land cover classes. A relatively new segmentation approach, multiresolution segmentation, is being examined using two data sets (Landsat and IRS).

Journal ArticleDOI
TL;DR: Experimental results demonstrate robustness not only in the variation of luminance conditions and changes in environmental conditions, but also for occlusions among multiple moving objects.

Book ChapterDOI
25 Aug 2003
TL;DR: This paper integrates colour, texture, and motion into a segmentation process using a variational framework for vector-valued data using a level set approach and a statistical model to describe the interior and the complement of a region.
Abstract: In this paper we integrate colour, texture, and motion into a segmentation process. The segmentation consists of two steps, which both combine the given information: a pre-segmentation step based on nonlinear diffusion for improving the quality of the features, and a variational framework for vector-valued data using a level set approach and a statistical model to describe the interior and the complement of a region. For the nonlinear diffusion we apply a novel diffusivity closely related to the total variation diffusivity, but being strictly edge enhancing. A multi-scale implementation is used in order to obtain more robust results. In several experiments we demonstrate the usefulness of integrating many kinds of information. Good results are obtained for both object segmentation and tracking of multiple objects.

Journal ArticleDOI
TL;DR: Both stand alone and relative evaluation metrics are developed to cover the cases for which a reference segmentation is missing or available for comparison, as well as for complete segmentation partitions.
Abstract: Video segmentation assumes a major role in the context of object-based coding and description applications. Evaluating the adequacy of a segmentation result for a given application is a requisite both to allow the appropriate selection of segmentation algorithms as well as to adjust their parameters for optimal performance. Subjective testing, the current practice for the evaluation of video segmentation quality, is an expensive and time-consuming process. Objective segmentation quality evaluation techniques can alternatively be used; however, it is recognized that, so far, much less research effort has been devoted to this subject than to the development of segmentation algorithms. This paper discusses the problem of video segmentation quality evaluation, proposing evaluation methodologies and objective segmentation quality metrics for individual objects as well as for complete segmentation partitions. Both standalone and relative evaluation metrics are developed to cover the cases for which a reference segmentation is missing or available for comparison.

Journal ArticleDOI
TL;DR: A novel marker location algorithm is subsequently used to locate significant homogeneous textured or non textured regions and a marker driven watershed transform is then used to segment the identified regions properly.
Abstract: The segmentation of images into meaningful and homogenous regions is a key method for image analysis within applications such as content based retrieval. The watershed transform is a well established tool for the segmentation of images. However, watershed segmentation is often not effective for textured image regions that are perceptually homogeneous. In order to segment such regions properly, the concept of the "texture gradient" is introduced. Texture information and its gradient are extracted using a novel nondecimated form of a complex wavelet transform. A novel marker location algorithm is subsequently used to locate significant homogeneous textured or non textured regions. A marker driven watershed transform is then used to segment the identified regions properly. The combined algorithm produces effective texture and intensity based segmentation for application to content based image retrieval.

01 Jan 2003
TL;DR: A new algorithm that represents each pixel in the frame by a group of clusters that is adapted to deal with background and lighting variations and demonstrated equal or better segmentation than the other techniques.
Abstract: Automatic analysis of digital video scenes often requires the segmentation of moving objects from the background. Historically, algorithms developed for this purpose have been restricted to small frame sizes, low frame rates or offline processing. The simplest approach involves subtracting the current frame from the known background. However, as the background is unknown, the key is how to learn and model it. This paper proposes a new algorithm that represents each pixel in the frame by a group of clusters. The clusters are ordered according the likelihood they model the background and are adapted to deal with background and lighting variations. Incoming pixels are matched against the corresponding cluster group and are classified according to whether the matching cluster is considered part of the background. The algorithm has been subjectively evaluated against three other techniques. It demonstrated equal or better segmentation than the other techniques and proved capable of processing 320 /spl times/ 240 video at 28 fps, excluding post-processing.

Journal ArticleDOI
TL;DR: The presented method can be applied to the segmentation of noise or degraded images as well as reduce over-segmentation and is applied to human face detection with accurate and closed boundaries.