scispace - formally typeset
Search or ask a question

Showing papers on "Image processing published in 2003"


Book
01 Dec 2003
TL;DR: 1. Fundamentals of Image Processing, 2. Intensity Transformations and Spatial Filtering, and 3. Frequency Domain Processing.
Abstract: 1. Introduction. 2. Fundamentals. 3. Intensity Transformations and Spatial Filtering. 4. Frequency Domain Processing. 5. Image Restoration. 6. Color Image Processing. 7. Wavelets. 8. Image Compression. 9. Morphological Image Processing. 10. Image Segmentation. 11. Representation and Description. 12. Object Recognition.

6,306 citations


Proceedings ArticleDOI
09 Nov 2003
TL;DR: This paper proposes a multiscale structural similarity method, which supplies more flexibility than previous single-scale methods in incorporating the variations of viewing conditions, and develops an image synthesis method to calibrate the parameters that define the relative importance of different scales.
Abstract: The structural similarity image quality paradigm is based on the assumption that the human visual system is highly adapted for extracting structural information from the scene, and therefore a measure of structural similarity can provide a good approximation to perceived image quality. This paper proposes a multiscale structural similarity method, which supplies more flexibility than previous single-scale methods in incorporating the variations of viewing conditions. We develop an image synthesis method to calibrate the parameters that define the relative importance of different scales. Experimental comparisons demonstrate the effectiveness of the proposed method.

4,333 citations


Journal ArticleDOI
TL;DR: An overview is presented of the medical image processing literature on mutual-information-based registration, an introduction for those new to the field, an overview for those working in the field and a reference for those searching for literature on a specific application.
Abstract: An overview is presented of the medical image processing literature on mutual-information-based registration. The aim of the survey is threefold: an introduction for those new to the field, an overview for those working in the field, and a reference for those searching for literature on a specific application. Methods are classified according to the different aspects of mutual-information-based registration. The main division is in aspects of the methodology and of the application. The part on methodology describes choices made on facets such as preprocessing of images, gray value interpolation, optimization, adaptations to the mutual information measure, and different types of geometrical transformations. The part on applications is a reference of the literature available on different modalities, on interpatient registration and on different anatomical objects. Comparison studies including mutual information are also considered. The paper starts with a description of entropy and mutual information and it closes with a discussion on past achievements and some future challenges.

3,121 citations


Journal ArticleDOI
TL;DR: A fully probabilistic framework is presented for estimating local probability density functions on parameters of interest in a model of diffusion, allowing for the quantification of belief in tractography results and the estimation of the cortical connectivity of the human thalamus.
Abstract: A fully probabilistic framework is presented for estimating local probability density functions on parameters of interest in a model of diffusion. This technique is applied to the estimation of parameters in the diffusion tensor model, and also to a simple partial volume model of diffusion. In both cases the parameters of interest include parameters defining local fiber direction. A technique is then presented for using these density functions to estimate global connectivity (i.e., the probability of the existence of a connection through the data field, between any two distant points), allowing for the quantification of belief in tractography results. This technique is then applied to the estimation of the cortical connectivity of the human thalamus. The resulting connectivity distributions correspond well with predictions from invasive tracer methods in nonhuman primate.

2,970 citations


Journal ArticleDOI
TL;DR: This work suggests acquiring two images for each diffusion gradient; one with bottom-up and one with top-down traversal of k-space in the phase-encode direction, which achieves the simultaneous goals of providing information on the underlying displacement field and intensity maps with adequate spatial sampling density even in distorted areas.

2,409 citations


Journal ArticleDOI
TL;DR: Two computer programs are presented, CTFFIND3 and CTFTILT, which determine defocus parameters from images of untilted specimens, as well as defocus and tilt parameters from image of tilted specimens, respectively, using a simple algorithm.

1,480 citations


Proceedings ArticleDOI
TL;DR: A new dataset, UCID (pronounced "use it") - an Uncompressed Colour Image Dataset which tries to bridge the gap between standardised image databases and objective evaluation of image retrieval algorithms that operate in the compressed domain.
Abstract: Standardised image databases or rather the lack of them are one of the main weaknesses in the field of content based image retrieval (CBIR). Authors often use their own images or do not specify the source of their datasets. Naturally this makes comparison of results somewhat difficult. While a first approach towards a common colour image set has been taken by the MPEG 7 committee 1 their database does not cater for all strands of research in the CBIR community. In particular as the MPEG-7 images only exist in compressed form it does not allow for an objective evaluation of image retrieval algorithms that operate in the compressed domain or to judge the influence image compression has on the performance of CBIR algorithms. In this paper we introduce a new dataset, UCID (pronounced ”use it”) - an Uncompressed Colour Image Dataset which tries to bridge this gap. The UCID dataset currently consists of 1338 uncompressed images together with a ground truth of a series of query images with corresponding models that an ideal CBIR algorithm would retrieve. While its initial intention was to provide a dataset for the evaluation of compressed domain algorithms, the UCID database also represents a good benchmark set for the evaluation of any kind of CBIR method as well as an image set that can be used to evaluate image compression and colour quantisation algorithms.

1,117 citations


Journal ArticleDOI
TL;DR: This work has expanded and refined existing algorithms for the subdivision of MRI datasets into anatomical structures, and used the normalized superimposed atlases to create a maximum probability map in stereotaxic space, which retains quantitative information regarding inter‐subject variability.
Abstract: Probabilistic atlases of neuroanatomy are more representative of population anatomy than single brain atlases. They allow anatomical labeling of the results of group studies in stereotaxic space, automated anatomical labeling of individual brain imaging datasets, and the statistical assessment of normal ranges for structure volumes and extents. No such manually constructed atlas is currently available for the frequently studied group of young adults. We studied 20 normal subjects (10 women, median age 31 years) with high-resolution magnetic resonance imaging (MRI) scanning. Images were nonuniformity corrected and reoriented along both the anterior-posterior commissure (AC-PC) line horizontally and the midsagittal plane sagittally. Building on our previous work, we have expanded and refined existing algorithms for the subdivision of MRI datasets into anatomical structures. The resulting algorithm is presented in the Appendix. Forty-nine structures were interactively defined as three-dimensional volumes-of-interest (VOIs). The resulting 20 individual atlases were spatially transformed (normalized) into standard stereotaxic space, using SPM99 software and the MNI/ICBM 152 template. We evaluated volume data for all structures both in native space and after spatial normalization, and used the normalized superimposed atlases to create a maximum probability map in stereotaxic space, which retains quantitative information regarding inter-subject variability. Its potential applications range from the automatic labeling of new scans to the detection of anatomical abnormalities in patients. Further data can be extracted from the atlas for the detailed analysis of individual structures.

1,084 citations


Journal ArticleDOI
TL;DR: The novel contribution of this paper is the combination of these three previously developed components, image decomposition with inpainting and texture synthesis, which permits the simultaneous use of filling-in algorithms that are suited for different image characteristics.
Abstract: An algorithm for the simultaneous filling-in of texture and structure in regions of missing image information is presented in this paper. The basic idea is to first decompose the image into the sum of two functions with different basic characteristics, and then reconstruct each one of these functions separately with structure and texture filling-in algorithms. The first function used in the decomposition is of bounded variation, representing the underlying image structure, while the second function captures the texture and possible noise. The region of missing information in the bounded variation image is reconstructed using image inpainting algorithms, while the same region in the texture image is filled-in with texture synthesis techniques. The original image is then reconstructed adding back these two sub-images. The novel contribution of this paper is then in the combination of these three previously developed components, image decomposition with inpainting and texture synthesis, which permits the simultaneous use of filling-in algorithms that are suited for different image characteristics. Examples on real images show the advantages of this proposed approach.

1,024 citations


Journal ArticleDOI
TL;DR: The proposed framework includes some novel low-level processing algorithms, such as dominant color region detection, robust shot boundary detection, and shot classification, as well as some higher-level algorithms for goal detection, referee detection,and penalty-box detection.
Abstract: We propose a fully automatic and computationally efficient framework for analysis and summarization of soccer videos using cinematic and object-based features. The proposed framework includes some novel low-level processing algorithms, such as dominant color region detection, robust shot boundary detection, and shot classification, as well as some higher-level algorithms for goal detection, referee detection, and penalty-box detection. The system can output three types of summaries: i) all slow-motion segments in a game; ii) all goals in a game; iii) slow-motion segments classified according to object-based features. The first two types of summaries are based on cinematic features only for speedy processing, while the summaries of the last type contain higher-level semantics. The proposed framework is efficient, effective, and robust. It is efficient in the sense that there is no need to compute object-based features when cinematic features are sufficient for the detection of certain events, e.g., goals in soccer. It is effective in the sense that the framework can also employ object-based features when needed to increase accuracy (at the expense of more computation). The efficiency, effectiveness, and robustness of the proposed framework are demonstrated over a large data set, consisting of more than 13 hours of soccer video, captured in different countries and under different conditions.

943 citations


Journal ArticleDOI
TL;DR: Based on this approach, two methods were developed to significantly improve the performance of dynamic imaging, named k‐t BLAST (Broad‐use Linear Acquisition Speed‐up Technique) and k-t SENSE (SENSitivity Encoding) for use with a single or multiple receiver coils, respectively.
Abstract: Dynamic images of natural objects exhibit significant correlations in k-space and time. Thus, it is feasible to acquire only a reduced amount of data and recover the missing portion afterwards. This leads to an improved temporal resolution, or an improved spatial resolution for a given amount of acquisition. Based on this approach, two methods were developed to significantly improve the performance of dynamic imaging, named k-t BLAST (Broad-use Linear Acquisition Speed-up Technique) and k-t SENSE (SENSitivity Encoding) for use with a single or multiple receiver coils, respectively. Signal correlations were learned from a small set of training data and the missing data were recovered using all available information in a consistent and integral manner. The general theory of k-t BLAST and k-t SENSE is applicable to arbitrary k-space trajectories, timevarying coil sensitivities, and under- and overdetermined reconstruction problems. Examples from ungated cardiac imaging demonstrate a 4-fold acceleration (voxel size 2.42 2.52 mm 2 , 38.4 fps) with either one or six receiver coils. k-t BLAST and k-t SENSE are applicable to many areas, especially those exhibiting quasiperiodic motion, such as imaging of the heart, the lungs, the abdomen, and the brain under periodic stimulation. Magn Reson Med 50:1031–1042, 2003. © 2003 Wiley-Liss, Inc.

Journal ArticleDOI
TL;DR: A new method for image smoothing based on a fourth-order PDE model that demonstrates good noise suppression without destruction of important anatomical or functional detail, even at poor signal-to-noise ratio is introduced.
Abstract: We introduce a new method for image smoothing based on a fourth-order PDE model. The method is tested on a broad range of real medical magnetic resonance images, both in space and time, as well as on nonmedical synthesized test images. Our algorithm demonstrates good noise suppression without destruction of important anatomical or functional detail, even at poor signal-to-noise ratio. We have also compared our method with related PDE models.

Journal ArticleDOI
TL;DR: A progressive morphological filter was developed to detect nonground LIDAR measurements and shows that the filter can remove most of the nong round points effectively.
Abstract: Recent advances in airborne light detection and ranging (LIDAR) technology allow rapid and inexpensive measurements of topography over large areas. This technology is becoming a primary method for generating high-resolution digital terrain models (DTMs) that are essential to numerous applications such as flood modeling and landslide prediction. Airborne LIDAR systems usually return a three-dimensional cloud of point measurements from reflective objects scanned by the laser beneath the flight path. In order to generate a DTM, measurements from nonground features such as buildings, vehicles, and vegetation have to be classified and removed. In this paper, a progressive morphological filter was developed to detect nonground LIDAR measurements. By gradually increasing the window size of the filter and using elevation difference thresholds, the measurements of vehicles, vegetation, and buildings are removed, while ground data are preserved. Datasets from mountainous and flat urbanized areas were selected to test the progressive morphological filter. The results show that the filter can remove most of the nonground points effectively.

Journal ArticleDOI
TL;DR: Results show how visual categorization based directly on low-level features, without grouping or segmentation stages, can benefit object localization and identification.
Abstract: In this paper we study the statistical properties of natural images belonging to different categories and their relevance for scene and object categorization tasks. We discuss how second-order statistics are correlated with image categories, scene scale and objects. We propose how scene categorization could be computed in a feedforward manner in order to provide top-down and contextual information very early in the visual processing chain. Results show how visual categorization based directly on low-level features, without grouping or segmentation stages, can benefit object localization and identification. We show how simple image statistics can be used to predict the presence and absence of objects in the scene before exploring the image.

Proceedings ArticleDOI
24 Nov 2003
TL;DR: Three variants of a new quality metric for image fusion based on an image quality index recently introduced by Wang and Bovik are presented, which are compliant with subjective evaluations and can therefore be used to compare different image fusion methods or to find the best parameters for a given fusion algorithm.
Abstract: We present three variants of a new quality metric for image fusion. The interest of our metrics, which are based on an image quality index recently introduced by Wang and Bovik in [Z. Wang et al., March 2002], lies in the fact that they do not require a ground-truth or reference image. We perform several simulations which show that our metrics are compliant with subjective evaluations and can therefore be used to compare different image fusion methods or to find the best parameters for a given fusion algorithm.

Patent
30 Apr 2003
TL;DR: In this article, an imaging system for a vehicle includes an imaging sensor and a control, which is operable to capture an image of a scene occurring exteriorly of the vehicle.
Abstract: An imaging system for a vehicle includes an imaging sensor and a control. The imaging sensor is operable to capture an image of a scene occurring exteriorly of the vehicle. The control receives the captured image, which comprises an image data set representative of the exterior scene. The control may apply an edge detection algorithm to a reduced image data set of the image data set. The reduced image data set is representative of a target zone of the captured image. The control may be operable to process the reduced image data set more than other image data, which are representative of areas of the captured image outside of the target zone, to detect objects present within the target zone. The imaging system may be associated with a side object detection system, a lane change assist system, a lane departure warning system and/or the like.

Journal ArticleDOI
TL;DR: This work proposes an orthonormal version of the ridgelet transform for discrete and finite-size images and uses the finite Radon transform (FRAT) as a building block to overcome the periodization effect of a finite transform.
Abstract: The ridgelet transform was introduced as a sparse expansion for functions on continuous spaces that are smooth away from discontinuities along lines. We propose an orthonormal version of the ridgelet transform for discrete and finite-size images. Our construction uses the finite Radon transform (FRAT) as a building block. To overcome the periodization effect of a finite transform, we introduce a novel ordering of the FRAT coefficients. We also analyze the FRAT as a frame operator and derive the exact frame bounds. The resulting finite ridgelet transform (FRIT) is invertible, nonredundant and computed via fast algorithms. Furthermore, this construction leads to a family of directional and orthonormal bases for images. Numerical results show that the FRIT is more effective than the wavelet transform in approximating and denoising images with straight edges.

Journal ArticleDOI
TL;DR: This paper decomposes a given (possible textured) image f into a sum of two functions u+v, where u∈BV is a function of bounded variation (a cartoon or sketchy approximation of f), while v is afunction representing the texture or noise.
Abstract: This paper is devoted to the modeling of real textured images by functional minimization and partial differential equations. Following the ideas of Yves Meyer in a total variation minimization framework of L. Rudin, S. Osher, and E. Fatemi, we decompose a given (possible textured) image f into a sum of two functions u+v, where u∈BV is a function of bounded variation (a cartoon or sketchy approximation of f), while v is a function representing the texture or noise. To model v we use the space of oscillating functions introduced by Yves Meyer, which is in some sense the dual of the BV space. The new algorithm is very simple, making use of differential equations and is easily solved in practice. Finally, we implement the method by finite differences, and we present various numerical results on real textured images, showing the obtained decomposition u+v, but we also show how the method can be used for texture discrimination and texture segmentation.

Journal ArticleDOI
TL;DR: The state of the art in machine vision inspection and a critical overview of real-world applications are presented and two independent ways to classify applications are proposed.

Proceedings ArticleDOI
01 Jul 2003
TL;DR: A novel hierarchical mesh decomposition algorithm that computes a decomposition into the meaningful components of a given mesh, which generally refers to segmentation at regions of deep concavities and the utility of the algorithm in control-skeleton extraction is demonstrated.
Abstract: Cutting up a complex object into simpler sub-objects is a fundamental problem in various disciplines. In image processing, images are segmented while in computational geometry, solid polyhedra are decomposed. In recent years, in computer graphics, polygonal meshes are decomposed into sub-meshes. In this paper we propose a novel hierarchical mesh decomposition algorithm. Our algorithm computes a decomposition into the meaningful components of a given mesh, which generally refers to segmentation at regions of deep concavities. The algorithm also avoids over-segmentation and jaggy boundaries between the components. Finally, we demonstrate the utility of the algorithm in control-skeleton extraction.

Journal ArticleDOI
TL;DR: A public domain program under the GNU license for PC computers is developed that can be run on a variety of hardware and software platforms and was tested on human lymphocytes exposed to gamma-rays and found to yield reproducible results.
Abstract: The single-cell gel electrophoresis, also known as the comet assay, has gained wide-spread popularity as a simple and reliable method to measure genotoxic and cytotoxic effects of physical and chemical agents as well as kinetics of DNA repair. Cells are generally stained with fluorescent dyes. The analysis of comets--damaged cells which form a typical comet-shaped pattern--is greatly facilitated by the use of a computer image-analysis program. Although several image-analysis programs are available commercially, they are expensive and their source codes are not provided. For Macintosh computers a cost-free public domain macro is available on the Internet. No ready for use, cost-free program exists for the PC platform. We have, therefore, developed such a public domain program under the GNU license for PC computers. The program is called CASP and can be run on a variety of hardware and software platforms. Its practical merit was tested on human lymphocytes exposed to gamma-rays and found to yield reproducible results. The binaries for Windows 95 and Linux, together with the source code can be obtained from: http://www.casp.of.pl.

Journal ArticleDOI
TL;DR: Simulation results with the chosen feature set and well-known watermarking and steganographic techniques indicate that the proposed approach is able with reasonable accuracy to distinguish between cover and stego images.
Abstract: We present techniques for steganalysis of images that have been potentially subjected to steganographic algorithms, both within the passive warden and active warden frameworks. Our hypothesis is that steganographic schemes leave statistical evidence that can be exploited for detection with the aid of image quality features and multivariate regression analysis. To this effect image quality metrics have been identified based on the analysis of variance (ANOVA) technique as feature sets to distinguish between cover-images and stego-images. The classifier between cover and stego-images is built using multivariate regression on the selected quality metrics and is trained based on an estimate of the original image. Simulation results with the chosen feature set and well-known watermarking and steganographic techniques indicate that our approach is able with reasonable accuracy to distinguish between cover and stego images.

Journal ArticleDOI
TL;DR: This paper presents an overview of image processing and analysis tools used in traffic applications and relates these tools with complete systems developed for specific traffic applications, and categorizes processing methods based on the intrinsic organization of their input data and the domain of processing.

Journal ArticleDOI
TL;DR: A new model for image restoration and image decomposition into cartoon and texture is proposed, based on the total variation minimization of Rudin, Osher, and Fatemi, and on oscillatory functions, which follows results of Meyer.
Abstract: In this paper, we propose a new model for image restoration and image decomposition into cartoon and texture, based on the total variation minimization of Rudin, Osher, and Fatemi [Phys. D, 60 (1992), pp. 259--268], and on oscillatory functions, which follows results of Meyer [Oscillating Patterns in Image Processing and Nonlinear Evolution Equations, Univ. Lecture Ser. 22, AMS, Providence, RI, 2002]. This paper also continues the ideas introduced by the authors in a previous work on image decomposition models into cartoon and texture [L. Vese and S. Osher, J. Sci. Comput., to appear]. Indeed, by an alternative formulation, an initial image f is decomposed here into a cartoon part u and a texture or noise part v. The u component is modeled by a function of bounded variation, while the v component is modeled by an oscillatory function, bounded in the norm dual to $|\cdot|_{H^1_0}$. After some transformation, the resulting PDE is of fourth order, envolving the Laplacian of the curvature of level lines. Fina...

Book
24 Jul 2003
TL;DR: Morphological image processing, a standard part of the imaging scientist's toolbox, can be applied to a wide range of industrial applications and shows how to analyse the problems and then develop successful algorithms to solve them.
Abstract: Morphological image processing, a standard part of the imaging scientist's toolbox, can be applied to a wide range of industrial applications. Concentrating on applications, this text shows how to analyse the problems and then develop successful algorithms to solve them.

Proceedings ArticleDOI
18 Jun 2003
TL;DR: The novel contribution of the paper is the combination of these three previously developed components: image decomposition with inpainting and texture synthesis, which permits the simultaneous use of filling-in algorithms that are suited for different image characteristics.
Abstract: An algorithm for the simultaneous filling-in of texture and structure in regions of missing image information is presented. The basic idea is to first decompose the image into the sum of two functions with different basic characteristics, and then reconstruct each one of these functions separately with structure and texture filling-in algorithms. The first function used in the decomposition is of bounded variation, representing the underlying image structure, while the second function captures the texture and possible noise. The region of missing information in the bounded variation image is reconstructed using image inpainting algorithms, while the same region in the texture image is filled-in with texture synthesis techniques. The original image is then reconstructed adding back these two sub-images. The novel contribution of the paper is then in the combination of these three previously developed components: image decomposition with inpainting and texture synthesis, which permits the simultaneous use of filling-in algorithms that are suited for different image characteristics. Examples on real images show the advantages of this proposed approach.

Journal ArticleDOI
TL;DR: A new method for contrast enhancement based on the curvelet transform is presented, which out-performs other enhancement methods on noisy images, but on noiseless or nearNoiseless images curvelet based enhancement is not remarkably better than wave let based enhancement.
Abstract: We present a new method for contrast enhancement based on the curvelet transform. The curvelet transform represents edges better than wavelets, and is therefore well-suited for multiscale edge enhancement. We compare this approach with enhancement based on the wavelet transform, and the multiscale retinex. In a range of examples, we use edge detection and segmentation, among other processing applications, to provide for quantitative comparative evaluation. Our findings are that curvelet based enhancement out-performs other enhancement methods on noisy images, but on noiseless or near noiseless images curvelet based enhancement is not remarkably better than wavelet based enhancement.

Journal ArticleDOI
TL;DR: An algorithm for fast elastic multidimensional intensity-based image registration with a parametric model of the deformation that is computationally more efficient than other alternatives and capable of accepting expert hints in the form of soft landmark constraints.
Abstract: We present an algorithm for fast elastic multidimensional intensity-based image registration with a parametric model of the deformation. It is fully automatic in its default mode of operation. In the case of hard real-world problems, it is capable of accepting expert hints in the form of soft landmark constraints. Much fewer landmarks are needed and the results are far superior compared to pure landmark registration. Particular attention has been paid to the factors influencing the speed of this algorithm. The B-spline deformation model is shown to be computationally more efficient than other alternatives. The algorithm has been successfully used for several two-dimensional (2-D) and three-dimensional (3-D) registration tasks in the medical domain, involving MRI, SPECT, CT, and ultrasound image modalities. We also present experiments in a controlled environment, permitting an exact evaluation of the registration accuracy. Test deformations are generated automatically using a random hierarchical fractional wavelet-based generator.

Book ChapterDOI
TL;DR: Statistical considerations are frequently to the fore in the analysis of microarray data, as researchers sift through massive amounts of data and adjust for various sources of variability in order to identify the important genes amongst the many which are measured.
Abstract: Statistical considerations are frequently to the fore in the analysis of microarray data, as researchers sift through massive amounts of data and adjust for various sources of variability in order to identify the important genes amongst the many which are measured. This article summarizes some of the issues involved and provides a brief review of the analysis tools which are available to researchers to deal with them. Any microarray experiment involves a number of distinct stages. Firstly there is the design of the experiment. The researchers must decide which genes are to be printed on the arrays, which sources of RNA are to be hybridized to the arrays and on how many arrays the hybridizations will be replicated. Secondly, after hybridization, there follows a number of data-cleaning steps or `low-level analysis’ of the microarray data. The microarray images must be processed to acquire red and green foreground and background intensities for each spot. The acquired red/green ratios must be normalized to adjust for dye-bias and for any systematic variation other than that due to the differences between the RNA samples being studied. Thirdly, the normalized ratios are analyzed by various graphical and numerical means to select differentially expressed (DE) genes or to find groups of genes whose expression profiles can reliably classify the different RNA sources into meaningful groups. The sections of this article correspond roughly to the various analysis steps. The following notation will be used throughout the article. The foreground red and green

Journal ArticleDOI
TL;DR: This work proposes a content-based soft annotation procedure for providing images with semantical labels, and experiments with two learning methods, support vector machines (SVMs) and Bayes point machines (BPMs), to select a base binary-classifier for CBSA.
Abstract: We propose a content-based soft annotation (CBSA) procedure for providing images with semantical labels. The annotation procedure starts with labeling a small set of training images, each with one single semantical label (e.g., forest, animal, or sky). An ensemble of binary classifiers is then trained for predicting label membership for images. The trained ensemble is applied to each individual image to give the image multiple soft labels, and each label is associated with a label membership factor. To select a base binary-classifier for CBSA, we experiment with two learning methods, support vector machines (SVMs) and Bayes point machines (BPMs), and compare their class-prediction accuracy. Our empirical study on a 116-category 25K-image set shows that the BPM-based ensemble provides better annotation quality than the SVM-based ensemble for supporting multimodal image retrievals.