scispace - formally typeset
Search or ask a question
Topic

Content-based image retrieval

About: Content-based image retrieval is a research topic. Over the lifetime, 6916 publications have been published within this topic receiving 150696 citations. The topic is also known as: CBIR.


Papers
More filters
Journal ArticleDOI
Wei Yang1, Zhentai Lu1, Mei Yu1, Meiyan Huang1, Qianjin Feng1, Wufan Chen1 
TL;DR: Preliminary results demonstrate that the BoW representation is effective and feasible for retrieval of liver lesions in contrast-enhanced CT images.
Abstract: This paper is aimed at developing and evaluating a content-based retrieval method for contrast-enhanced liver computed tomographic (CT) images using bag-of-visual-words (BoW) representations of single and multiple phases. The BoW histograms are extracted using the raw intensity as local patch descriptor for each enhance phase by densely sampling the image patches within the liver lesion regions. The distance metric learning algorithms are employed to obtain the semantic similarity on the Hellinger kernel feature map of the BoW histograms. The different visual vocabularies for BoW and learned distance metrics are evaluated in a contrast-enhanced CT image dataset comprised of 189 patients with three types of focal liver lesions, including 87 hepatomas, 62 cysts, and 60 hemangiomas. For each single enhance phase, the mean of average precision (mAP) of BoW representations for retrieval can reach above 90 % which is significantly higher than that of intensity histogram and Gabor filters. Furthermore, the combined BoW representations of the three enhance phases can improve mAP to 94.5 %. These preliminary results demonstrate that the BoW representation is effective and feasible for retrieval of liver lesions in contrast-enhanced CT images.

77 citations

Journal ArticleDOI
TL;DR: A new semantic feature extracted from dominant colors (weight for each DC) is proposed, which helps reduce the effect of image background on image matching decision where an object's colors receive much more focus.

76 citations

Journal ArticleDOI
TL;DR: A novel scheme of online multi-modal distance metric learning (OMDML) is investigated, which explores a unified two-level online learning scheme that learns to optimize a distance metric on each individual feature space; and then it learns to find the optimal combination of diverse types of features.
Abstract: Distance metric learning (DML) is an important technique to improve similarity search in content-based image retrieval. Despite being studied extensively, most existing DML approaches typically adopt a single-modal learning framework that learns the distance metric on either a single feature type or a combined feature space where multiple types of features are simply concatenated. Such single-modal DML methods suffer from some critical limitations: (i) some type of features may significantly dominate the others in the DML task due to diverse feature representations; and (ii) learning a distance metric on the combined high-dimensional feature space can be extremely time-consuming using the naive feature concatenation approach. To address these limitations, in this paper, we investigate a novel scheme of online multi-modal distance metric learning (OMDML), which explores a unified two-level online learning scheme: (i) it learns to optimize a distance metric on each individual feature space; and (ii) then it learns to find the optimal combination of diverse types of features. To further reduce the expensive cost of DML on high-dimensional feature space, we propose a low-rank OMDML algorithm which not only significantly reduces the computational cost but also retains highly competing or even better learning accuracy. We conduct extensive experiments to evaluate the performance of the proposed algorithms for multi-modal image retrieval, in which encouraging results validate the effectiveness of the proposed technique.

76 citations

Journal ArticleDOI
TL;DR: A novel method based on non-negative matrix factorization to generate multimodal image representations that integrate visual features and text information that highly outperforms the response of the system in both tasks, when compared to multi-modal latent semantic spaces generated by a singular value decomposition.

76 citations

Journal ArticleDOI
TL;DR: A generalisation of the multiple-instance learning model in which a bag's label is not based on a single instance's proximity to a single target point, but on a collection of instances, each near one of a set of target points.
Abstract: We describe a generalisation of the multiple-instance learning model in which a bag's label is not based on a single instance's proximity to a single target point. Rather, a bag is positive if and only if it contains a collection of instances, each near one of a set of target points. We then adapt a learning-theoretic algorithm for learning in this model and present empirical results on data from robot vision, content-based image retrieval, and protein sequence identification.

76 citations


Network Information
Related Topics (5)
Feature extraction
111.8K papers, 2.1M citations
90% related
Feature (computer vision)
128.2K papers, 1.7M citations
88% related
Image segmentation
79.6K papers, 1.8M citations
87% related
Convolutional neural network
74.7K papers, 2M citations
87% related
Deep learning
79.8K papers, 2.1M citations
86% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202358
2022141
2021180
2020163
2019224
2018270