scispace - formally typeset
Open AccessJournal ArticleDOI

A Weighted Topic Model Learned From Local Semantic Space for Automatic Image Annotation

TLDR
A novel annotation method based on topic model, namely local learning-based probabilistic latent semantic analysis (LL-PLSA) that significantly outperforms the state-of-the-art especially in terms of overall metrics.
Abstract
Automatic image annotation plays a significant role in image understanding, retrieval, classification, and indexing. Today, it is becoming increasingly important in order to annotate large-scale social media images from content-sharing websites and social networks. These social images are usually annotated by user-provided low-quality tags. The topic model is considered as a promising method to describe these weak-labeling images by learning latent representations of training samples. The recent annotation methods based on topic models have two shortcomings. First, they are difficult to scale to a large-scale image dataset. Second, they can not be used to online image repository because of continuous addition of new images and new tags. In this paper, we propose a novel annotation method based on topic model, namely local learning-based probabilistic latent semantic analysis (LL-PLSA), to solve the above problems. The key idea is to train a weighted topic model for a given test image on its semantic neighborhood consisting of a fixed number of semantically and visually similar images. This method can scale to a large-scale image database, as training samples involved in modeling are a few nearest neighbors rather than the entire database. Moreover, this proposed topic model, online customized for the test image, naturally addresses the issue of continuous addition of new images and new tags in a database. Extensive experiments on three benchmark datasets demonstrate that the proposed method significantly outperforms the state-of-the-art especially in terms of overall metrics.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Top-k Partial Label Machine

TL;DR: In this article , the authors proposed a top-and convex top-Hierarchical Partial Label Machine (TPLM) to deal with ambiguities in partial label learning.
Journal ArticleDOI

Generalized Large Margin $k$NN for Partial Label Learning

TL;DR: In this paper , a generalized large margin nearest neighbor (GLMNN-PLL) method is proposed to deal with noisy labels in partial label learning, which adapts the framework of LMNN to PLL by modifying the constraint from "the same class" to "similarly-labeled".
Journal ArticleDOI

Automatic Image Annotation by Sequentially Learning From Multi-Level Semantic Neighborhoods

TL;DR: Zhang et al. as mentioned in this paper proposed a three-pass KNN (k-Nearest Neighbor) model to solve the problems of semantic gap, label imbalance, wider range labels and weak-labeling.
Proceedings Article

Partial Label Learning via Label Influence Function

TL;DR: In this article , a Partial Label Learning via Label Influence Function (PLL-IF) framework is proposed to deal with ambiguities in partial label learning.
References
More filters
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Proceedings Article

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition

TL;DR: DeCAF as discussed by the authors is an open-source implementation of these deep convolutional activation features, along with all associated network parameters, to enable vision researchers to conduct experimentation with deep representations across a range of visual concept learning paradigms.
Posted Content

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition

TL;DR: DeCAF, an open-source implementation of deep convolutional activation features, along with all associated network parameters, are released to enable vision researchers to be able to conduct experimentation with deep representations across a range of visual concept learning paradigms.
Book ChapterDOI

Object Recognition as Machine Translation: Learning a Lexicon for a Fixed Image Vocabulary

TL;DR: This work shows how to cluster words that individually are difficult to predict into clusters that can be predicted well, and cannot predict the distinction between train and locomotive using the current set of features, but can predict the underlying concept.
Related Papers (5)