scispace - formally typeset
Search or ask a question

Showing papers on "Content-based image retrieval published in 2013"


Journal ArticleDOI
TL;DR: An extensive evaluation of local invariant features for image retrieval of land-use/land-cover classes in high-resolution aerial imagery using a bag-of-visual-words (BOVW) representation and describes interesting findings such as the performance-efficiency tradeoffs that are possible through the appropriate pairings of different-sized codebooks and dissimilarity measures.
Abstract: This paper investigates local invariant features for geographic (overhead) image retrieval. Local features are particularly well suited for the newer generations of aerial and satellite imagery whose increased spatial resolution, often just tens of centimeters per pixel, allows a greater range of objects and spatial patterns to be recognized than ever before. Local invariant features have been successfully applied to a broad range of computer vision problems and, as such, are receiving increased attention from the remote sensing community particularly for challenging tasks such as detection and classification. We perform an extensive evaluation of local invariant features for image retrieval of land-use/land-cover (LULC) classes in high-resolution aerial imagery. We report on the effects of a number of design parameters on a bag-of-visual-words (BOVW) representation including saliency- versus grid-based local feature extraction, the size of the visual codebook, the clustering algorithm used to create the codebook, and the dissimilarity measure used to compare the BOVW representations. We also perform comparisons with standard features such as color and texture. The performance is quantitatively evaluated using a first-of-its-kind LULC ground truth data set which will be made publicly available to other researchers. In addition to reporting on the effects of the core design parameters, we also describe interesting findings such as the performance-efficiency tradeoffs that are possible through the appropriate pairings of different-sized codebooks and dissimilarity measures. While the focus is on image retrieval, we expect our insights to be informative for other applications such as detection and classification.

338 citations


Proceedings Article
01 Jul 2013
TL;DR: This paper analyzes the effectiveness of the fusion of global and local features in automatic image annotation and content based image retrieval community, including some classic models and their illustrations in the literature.
Abstract: Feature extraction and representation is a crucial step for multimedia processing. How to extract ideal features that can reflect the intrinsic content of the images as complete as possible is still a challenging problem in computer vision. However, very little research has paid attention to this problem in the last decades. So in this paper, we focus our review on the latest development in image feature extraction and provide a comprehensive survey on image feature representation techniques. In particular, we analyze the effectiveness of the fusion of global and local features in automatic image annotation and content based image retrieval community, including some classic models and their illustrations in the literature. Finally, we summarize this paper with some important conclusions and point out the future potential research directions.

248 citations


Journal ArticleDOI
TL;DR: This paper presents a review of state-of-the-art medical CBIR approaches in five main categories: two-dimensional image retrieval, retrieval of images with three or more dimensions, the use of nonimage data to enhance the retrieval, multimodality image retrieved, and retrieval from diverse datasets.
Abstract: Medical imaging is fundamental to modern healthcare, and its widespread use has resulted in the creation of image databases, as well as picture archiving and communication systems. These repositories now contain images from a diverse range of modalities, multidimensional (three-dimensional or time-varying) images, as well as co-aligned multimodality images. These image collections offer the opportunity for evidence-based diagnosis, teaching, and research; for these applications, there is a requirement for appropriate methods to search the collections for images that have characteristics similar to the case(s) of interest. Content-based image retrieval (CBIR) is an image search technique that complements the conventional text-based retrieval of images by using visual features, such as color, texture, and shape, as search criteria. Medical CBIR is an established field of study that is beginning to realize promise when applied to multidimensional and multimodality medical data. In this paper, we present a review of state-of-the-art medical CBIR approaches in five main categories: two-dimensional image retrieval, retrieval of images with three or more dimensions, the use of nonimage data to enhance the retrieval, multimodality image retrieval, and retrieval from diverse datasets. We use these categories as a framework for discussing the state of the art, focusing on the characteristics and modalities of the information used during medical image retrieval.

194 citations


Journal ArticleDOI
TL;DR: This paper investigates various combinations of mid-level features to build an effective image retrieval system based on the bag-of-features (BoF) model and shows that the integrations of these features yield complementary and substantial improvement on image retrieval even with noisy background and ambiguous objects.

151 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed structure elements' descriptor (SED), a novel texture descriptor, which can effectively describe images and represent image local features and extract and describe color and texture features.

137 citations


Journal ArticleDOI
TL;DR: A hybrid approach to reduce the semantic gap between low level visual features and high level semantics, through simultaneous feature adaptation and feature selection is presented.
Abstract: In content-based image retrieval (CBIR) applications, each database needs its corresponding parameter setting for feature extraction. However, most of the CBIR systems perform indexing by a set of fixed and pre-specific parameters. On the other hand, feature selection methods have currently gained considerable popularity to reduce semantic gap. In this regard, this paper is devoted to present a hybrid approach to reduce the semantic gap between low level visual features and high level semantics, through simultaneous feature adaptation and feature selection. In the proposed approach, a hybrid meta-heuristic swarm intelligence-based search technique, called mixed gravitational search algorithm (MGSA), is employed. Some feature extraction parameters (i.e. the parameters of a 6-tap parameterized orthogonal mother wavelet in texture features and quantization levels in color histogram) are optimized to reach a maximum precision of the CBIR systems. Meanwhile, feature subset selection is done for the same purpose. A comparative experimental study with the conventional CBIR system is reported on a database of 1000 images. The obtained results confirm the effectiveness of the proposed adaptive indexing method in the field of CBIR.

126 citations


Journal ArticleDOI
TL;DR: The quantized histogram statistical texture features are extracted from the DCT blocks of the image using the significant energy of the DC and the first three AC coefficients of the blocks for the effective matching of images in the compressed domain.
Abstract: The effective content-based image retrieval (CBIR) needs efficient extraction of low level features like color, texture and shapes for indexing and fast query image matching with indexed images for the retrieval of similar images. Features are extracted from images in pixel and compressed domains. However, now most of the existing images are in compressed formats like JPEG using DCT (discrete cosine transformation). In this paper we study the issues of efficient extraction of features and the effective matching of images in the compressed domain. In our method the quantized histogram statistical texture features are extracted from the DCT blocks of the image using the significant energy of the DC and the first three AC coefficients of the blocks. For the effective matching of the image with images, various distance metrics are used to measure similarities using texture features. The analysis of the effective CBIR is performed on the basis of various distance metrics in different number of quantization bins. The proposed method is tested by using Corel image database and the experimental results show that our method has robust image retrieval for various distance metrics with different histogram quantization in a compressed domain.

110 citations


Proceedings ArticleDOI
15 Apr 2013
TL;DR: A survey on content based image retrieval presented and performance literature review by using principles of Content Based Image Retrieval based unlabelled images to determine the efficient imaging for CBIR.
Abstract: Literature survey is most important for understanding and gaining much more knowledge about specific area of a subject. In this paper a survey on content based image retrieval presented. Content Based Image Retrieval (CBIR) is a technique which uses visual features of image such as color, shape, texture, etc... to search user required image from large image database according to user's requests in the form of a query image. We consider Content Based Image Retrieval viz. labelled and unlabelled images for analyzing efficient image for different image retrieval process viz. D-EM, SVM, RF, etc. To determining the efficient imaging for Content Based Image Retrieval, We performance literature review by using principles of Content Based Image Retrieval based unlabelled images. And also give some recommendations for improve the CBIR system using unlabelled images.

101 citations


Journal ArticleDOI
TL;DR: This paper presents a novel context-based approach for redefining distances and later re-ranking images aiming to improve the effectiveness of CBIR systems, where distances among images are redefined based on the similarity of their ranked lists.

88 citations


Journal ArticleDOI
TL;DR: A new semantic feature extracted from dominant colors (weight for each DC) is proposed, which helps reduce the effect of image background on image matching decision where an object's colors receive much more focus.

76 citations


Journal ArticleDOI
TL;DR: A new and effective image indexing technique that extracts features from JPEG compressed images using vector quantization techniques and a codebook generated using a K -means clustering algorithm that can accelerate the work of indexing images.

Proceedings ArticleDOI
09 Mar 2013
TL;DR: An efficient algorithm for Content Based Image Retrieval (CBIR) based on Discrete Wavelet Transform (DWT) and Edge Histogram Descriptor (EHD) feature of MPEG-7 and compared with various other proposed schemes to show the superiority of this scheme.
Abstract: This paper describes an efficient algorithm for Content Based Image Retrieval (CBIR) based on Discrete Wavelet Transform (DWT) and Edge Histogram Descriptor (EHD) feature of MPEG-7. The proposed algorithm is explained for image retrieval based on shape and texture features only not on the basis of color information. Here input image is first decomposed into wavelet coefficients. These wavelet coefficients give mainly horizontal, vertical and diagonal features in the image. After wavelet transform, Edge Histogram Descriptor is then used on selected wavelet coefficients to gather the information of dominant edge orientations. The combination of DWT and EHD techniques increases the performance of image retrieval system for shape and texture based search. The performance of various wavelets is also compared to find the suitability of particular wavelet function for image retrieval. The proposed algorithm is trained and tested for Wang image database. The results of retrieval are expressed in terms of Precision and Recall and compared with various other proposed schemes to show the superiority of our scheme.

Journal ArticleDOI
TL;DR: The state of the art of content-based retrieval in Earth observation image archives focusing on complete systems showing promise for operational implementation is analyzed, focusing in particular on the phases after primitive feature extraction.
Abstract: We analyze the state of the art of content-based retrieval in Earth observation image archives focusing on complete systems showing promise for operational implementation. The different paradigms at the basis of the main system families are introduced. The approaches taken are considered, focusing in particular on the phases after primitive feature extraction. The solutions envisaged for the issues related to feature simplification and synthesis, indexing, semantic labeling are reviewed. The methodologies for query specification and execution are evaluated. Conclusions are drawn on the state of published research in Earth observation (EO) mining.

Book
01 Jan 2013
TL;DR: This tutorial presents an overview of visual information retrieval concepts, techniques, algorithms, and applications and discusses the challenges associated with building real-world, large scale VIR solutions, including a brief overview of publicly available datasets used in worldwide challenges, contests, and benchmarks.
Abstract: Visual information retrieval (VIR) is an active and vibrant research area, which attempts at providing means for organizing, indexing, annotating, and retrieving visual information (images and videos) form large, unstructured repositories. The goal of VIR is to retrieve the highest number of relevant matches to a given query (often expressed as an example image and/or a series of keywords). In its early years (1995-2000) the research efforts were dominated by content-based approaches contributed primarily by the image and video processing community. During the past decade, it was widely recognized that the challenges imposed by the semantic gap (the lack of coincidence between an image's visual contents and its semantic interpretation) required a clever use of textual metadata (in addition to information extracted from the image's pixel contents) to make image and video retrieval solutions efficient and effective. The need to bridge (or at least narrow) the semantic gap has been one of the driving forces behind current VIR research. Additionally, other related research problems and market opportunities have started to emerge, offering a broad range of exciting problems for computer scientists and engineers to work on. In this tutorial, we present an overview of visual information retrieval (VIR) concepts, techniques, algorithms, and applications. Several topics are supported by examples written in Java, using Lucene (an open-source Java-based indexing and search implementation) and LIRE (Lucene Image REtrieval), an open-source Java-based library for content-based image retrieval (CBIR) written by Mathias Lux.After motivating the topic, we briefly review the fundamentals of information retrieval, present the most relevant and effective visual descriptors currently used in VIR, the most common indexing approaches for visual descriptors, the most prominent machine learning techniques used in connection with contemporary VIR solutions, as well as the challenges associated with building real-world, large scale VIR solutions, including a brief overview of publicly available datasets used in worldwide challenges, contests, and benchmarks. Throughout the tutorial, we integrate examples using LIRE, whose main features and design principles are also discussed. Finally, we conclude the tutorial with suggestions for deepening the knowledge in the topic, including a brief discussion of the most relevant advances, open challenges, and promising opportunities in VIR and related areas.The tutorial is primarily targeted at experienced Information Retrieval researchers and practitioners interested in extending their knowledge of document-based IR to equivalent concepts, techniques, and challenges in VIR. The acquired knowledge should allow participants to derive insightful conclusions and promising avenues for further investigation.

Journal ArticleDOI
TL;DR: This paper proposes a rotation invariant partial duplicate image retrieval (PDIR) approach, which effectively and efficiently retrieves the partial duplicate images by accurately matching the representative SIFT features.
Abstract: Partial duplicate images often have large non-duplicate regions and small duplicate regions with random rotation, which lead to the following problems: 1) large number of noisy features from the non-duplicate regions; 2) small number of representative features from the duplicate regions; 3) randomly rotated or deformed duplicate regions. These problems challenge many content based image retrieval (CBIR) approaches, since most of them cannot distinguish the representative features from a large proportion of noisy features in a rotation invariant way. In this paper, we propose a rotation invariant partial duplicate image retrieval (PDIR) approach, which effectively and efficiently retrieves the partial duplicate images by accurately matching the representative SIFT features. Our method is based on the Combined-Orientation-Position (COP) consistency graph model, which consists of the following two parts: 1) The COP consistency, which is a rotation invariant measurement of the relative spatial consistency among the candidate matches of SIFT features; it uses a coarse-to-fine family of evenly sectored polar coordinate systems to softly quantize and combine the orientations and positions of the SIFT features. 2) The consistency graph model, which robustly rejects the spatially inconsistent noisy features by effectively detecting the group of candidate feature matches with the largest average COP consistency. Extensive experiments on five large scale image data sets show promising retrieval performances.

Journal ArticleDOI
TL;DR: The ODBTC method offers an effective way to index an image in a content-based image retrieval system, and simultaneously it is able to compress an image efficiently, and can be a very competitive candidate in image retrieval applications.

Journal ArticleDOI
TL;DR: This paper proposes a new content-based image retrieval technique using color and texture information, which achieves higher retrieval efficiency and provides a robust feature set for color image retrieval.

01 Jan 2013
TL;DR: This survey covers approaches used for extracting low level features; various distance measures for measuring the similarity of images, the mechanisms for reducing the semantic gap and about invariant image retrieval.
Abstract: Content Based Image Retrieval (CBIR) is a very important research area in the field of image processing, and comprises of low level feature extraction such as color, texture and shape and similarity measures for the comparison of images. Recently, the research focus in CBIR has been in reducing the semantic gap, between the low level visual features and the high level image semantics. This paper provides a comprehensive survey of all these aspects. This survey covers approaches used for extracting low level features; various distance measures for measuring the similarity of images, the mechanisms for reducing the semantic gap and about invariant image retrieval. In addition to these, various data sets used in CBIR and the performance measures, are also addressed. Finally, future research directions are also suggested. (I. Felci Rajam, S. Valli. A Survey on Content Based Image Retrieval. Life Sci J 2013; 10(2): 2475-2487). (ISSN: 1097-8135). http://www.lifesciencesite.com 343

Journal ArticleDOI
TL;DR: In a compact descriptor for visual search only a limited number of local features may be included, and the estimated probability for correct match between keypoints provides a good criterion for selection of a subset.
Abstract: In a compact descriptor for visual search only a limited number of local features may be included. The estimated probability for correct match between keypoints provides a good criterion for selection of a subset.

Journal ArticleDOI
TL;DR: An active SVM-based RF using multiple classifiers ensemble and features reweighting is proposed, and extensive simulations on large databases show that the proposed algorithm is significantly more effective than the state-of-the-art approaches.

Journal ArticleDOI
TL;DR: An automatic system to annotate and retrieve images that assumes that regions in an image can be described using a vocabulary of blobs, and that a structural approach requires a smaller vocabulary size to reach its best performance.

Proceedings ArticleDOI
16 Apr 2013
TL;DR: This paper proposes a novel distance function, the signature matching distance, which matches coincident visual properties of images based on their signatures and shows that this approach is able to outperform other signature-based approaches to content-based image retrieval.
Abstract: We propose a simple yet effective approach to content-based image retrieval: the signature matching distance. While recent approaches to content-based image retrieval utilize the bag-of-visual-words model, where image descriptors are matched through a common visual vocabulary, signature-based approaches use a distance between signatures, i.e. between image-specific bags of locally aggregated descriptors, in order to quantify image dissimilarity. In this paper, we focus on the signature-based approach to content-based image retrieval and propose a novel distance function, the signature matching distance. This distance matches coincident visual properties of images based on their signatures. In particular, by investigating different descriptor matching strategies and their suitability to match signatures, we show that our approach is able to outperform other signature-based approaches to content-based image retrieval. Moreover, in combination with a simple color and texture-based image descriptor, our approach is able to compete with the majority of bag-of-visual-words approaches.

Journal ArticleDOI
TL;DR: An approach that simultaneously clusters images and learns dictionaries from the clusters and provides both in-plane rotation and scale invariant clustering, which is useful in numerous applications, including content-based image retrieval (CBIR).
Abstract: In this paper, we present an approach that simultaneously clusters images and learns dictionaries from the clusters. The method learns dictionaries and clusters images in the radon transform domain. The main feature of the proposed approach is that it provides both in-plane rotation and scale invariant clustering, which is useful in numerous applications, including content-based image retrieval (CBIR). We demonstrate the effectiveness of our rotation and scale invariant clustering method on a series of CBIR experiments. Experiments are performed on the Smithsonian isolated leaf, Kimia shape, and Brodatz texture datasets. Our method provides both good retrieval performance and greater robustness compared to standard Gabor-based and three state-of-the-art shape-based methods that have similar objectives.

Proceedings ArticleDOI
21 Oct 2013
TL;DR: An overview on LIRE is given, its use, capabilities and reports on retrieval and runtime performance, that provides a simple way to index and retrieve millions of images based on the images' contents.
Abstract: Content based image retrieval has been around for some time. There are lots of different test data sets, lots of published methods and techniques, and manifold retrieval challenges, where content based image retrieval is of interest. LIRE is a Java library, that provides a simple way to index and retrieve millions of images based on the images' contents. LIRE is robust and well tested and is not only recommended by the websites of ImageCLEF and MediaEval, but is also employed in industry. This paper gives an overview on LIRE, its use, capabilities and reports on retrieval and runtime performance.

Journal ArticleDOI
TL;DR: This paper presents a system used in the medical domain for three distinct tasks: image annotation, semantic based image retrieval and content based image retrieved.

Proceedings ArticleDOI
01 Jul 2013
TL;DR: A novel framework for Content Based Image Retrieval (CBIR), which combines color, texture and spatial structure of image, which integrates three features to enhance the retrieval performance.
Abstract: This paper presents a novel framework for Content Based Image Retrieval(CBIR), which combines color, texture and spatial structure of image. The proposed method uses color, texture and spatial structure descriptors to form a feature vector. Images are segmented into regions to extract local color, texture and CENTRIST(CENsus Transform hISTogram) features respectively. Multiple-instance learning (MIL) and Diverse Density(DD) are incorporated with regions as instances to find the objective instance. In addition, to denote the whole structure of image better, we perform PCA to CENTRIST features of all images, i.e. spatial Principal component Analysis of Census Transform(spatial PACT). This framework integrates three features to enhance the retrieval performance. Experiments on COREL standard database invalidate the proposed method by comparing with some state-of-the-art methods. (4 pages)

Book ChapterDOI
01 Jan 2013
TL;DR: This chapter addresses the omission of a comprehensive survey of both short-term and long-term learning RF techniques in the published literature, and offers suggestions for future work.
Abstract: In content-based image retrieval, relevance feedback is an interactive process, which builds a bridge to connect users with a search engine. It leads to much improved retrieval performance by updating a query and similarity measures according to a user’s preference; and recently techniques have matured to some extent. Most previous relevance feedback approaches exploit short-term learning (intraquery learning) that deals with the current feedback session but ignoring historical data from other users, which potentially results in a great loss of useful information. In the last few years, long-term learning (inter-query learning), by recording and collecting feedback knowledge from different users over a variety of query sessions has played an increasingly important role in multimedia information searching. It can further improve the retrieval performance in terms of effectiveness and efficiency. In the published literature, no comprehensive survey of both short-term learning and long-term learning RF techniques has been conducted. To this end, the goal of this chapter is to address this omission and offer suggestions for future work.

Journal ArticleDOI
06 May 2013-PLOS ONE
TL;DR: A methodology for parametric CBIR based on similarity profiles is proposed that provides the basis for a CBIR expansion mechanism and the solution developed integrates with DICOM based PACS networks where it provides CBIR functionality in a seamless manner.
Abstract: Content-based image retrieval (CBIR) has been heralded as a mechanism to cope with the increasingly larger volumes of information present in medical imaging repositories. However, generic, extensible CBIR frameworks that work natively with Picture Archive and Communication Systems (PACS) are scarce. In this article we propose a methodology for parametric CBIR based on similarity profiles. The architecture and implementation of a profiled CBIR system, based on query by example, atop Dicoogle, an open-source, full-fletched PACS is also presented and discussed. In this solution, CBIR profiles allow the specification of both a distance function to be applied and the feature set that must be present for that function to operate. The presented framework provides the basis for a CBIR expansion mechanism and the solution developed integrates with DICOM based PACS networks where it provides CBIR functionality in a seamless manner.

Proceedings ArticleDOI
01 Oct 2013
TL;DR: The experimental results show that the proposed CBIR using SIFT algorithm producing excellent retrieval result for images with many corners as compared to retrieving image with less corners.
Abstract: This paper presents an alternative approach for Content Based Image Retrieval (CBIR) using Scale Invariant Feature Transform (SIFT) algorithm for binary and gray scale images. The motivation to use SIFT algorithm for CBIR is due to the fact that SIFT is invariant to scale, rotation and translation as well as partially invariant to affine distortion and illumination changes. Inspired by these facts, this paper investigates the fundamental properties of SIFT for robust CBIR by using MPEG-7, COIL-20 and ZuBuD image databases. Our approach uses detected keypoints and its descriptors to match between the query images and images from the database. Our experimental results show that the proposed CBIR using SIFT algorithm producing excellent retrieval result for images with many corners as compared to retrieving image with less corners.

Journal ArticleDOI
TL;DR: A novel QBME method for fast image retrieval based on transductive learning framework that accepts one input example at a time and all the input examples are processed at the same time in this strategy.