Book ChapterDOI
Utilizing Locality-Sensitive Hash Learning for Cross-Media Retrieval
Jia Yuhua,Bai Liang,Wang Peng,Guo Jinlin,Xie Yuxiang,Yu Tianyuan +5 more
- pp 550-561
TLDR
Experimental results show that, in the set of the queries’ near neighbors obtained by the proposed FCMR, the proportions of relevant documents can be much boosted, and it indicates that the retrieval based on near neighbors can be effectively conducted.Abstract:
Cross-media retrieval is an imperative approach to handle the explosive growth of multimodal data on the web. However, existed approaches to cross-media retrieval are computationally expensive due to the curse of dimensionality. To efficiently retrieve in multimodal data, it is essential to reduce the proportion of irrelevant documents. In this paper, we propose a cross-media retrieval approach (FCMR) based on locality-sensitive hashing (LSH) and neural networks. Multimodal information is projected by LSH algorithm to cluster similar objects into the same hash bucket and dissimilar objects into different ones, using hash functions learned through neural networks. Once given a textual or visual query, it can be efficiently mapped to a hash bucket in which objects stored can be near neighbors of this query. Experimental results show that, in the set of the queries’ near neighbors obtained by the proposed method, the proportions of relevant documents can be much boosted, and it indicates that the retrieval based on near neighbors can be effectively conducted. Further evaluations on two public datasets demonstrate the effectiveness of the proposed retrieval method compared to the baselines.read more
Citations
More filters
Journal ArticleDOI
Locating similar names through locality sensitive hashing and graph theory
Fernando Turrado García,Luis Javier García Villalba,Ana Lucila Sandoval Orozco,Francisco Damián Aranda Ruiz,Andrés Aguirre Juárez,Tai-hoon Kim +5 more
TL;DR: This paper shows how can Locality Sensitive Hashing be applied to identify misspelled people names (name, middle name and last name) or near duplicates.
References
More filters
Journal ArticleDOI
Latent dirichlet allocation
TL;DR: This work proposes a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hofmann's aspect model.
Proceedings Article
Latent Dirichlet Allocation
TL;DR: This paper proposed a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams, and Hof-mann's aspect model, also known as probabilistic latent semantic indexing (pLSI).
Book
Learning Deep Architectures for AI
TL;DR: The motivations and principles regarding learning algorithms for deep architectures, in particular those exploiting as building blocks unsupervised learning of single-layer modelssuch as Restricted Boltzmann Machines, used to construct deeper models such as Deep Belief Networks are discussed.
Journal ArticleDOI
Canonical Correlation Analysis: An Overview with Application to Learning Methods
TL;DR: A general method using kernel canonical correlation analysis to learn a semantic representation to web images and their associated text and compares orthogonalization approaches against a standard cross-representation retrieval technique known as the generalized vector space model is presented.
Proceedings ArticleDOI
Modeling annotated data
David M. Blei,Michael I. Jordan +1 more
TL;DR: Three hierarchical probabilistic mixture models which aim to describe annotated data with multiple types, culminating in correspondence latent Dirichlet allocation, a latent variable model that is effective at modeling the joint distribution of both types and the conditional distribution of the annotation given the primary type.