scispace - formally typeset
Search or ask a question

Showing papers on "Word embedding published in 2012"


Patent
02 Feb 2012
TL;DR: In this article, a set of word embedding transforms are applied to transform text words of an input document into K-dimensional word vectors in order to generate a set or sequence of word vectors representing the input document.
Abstract: A set of word embedding transforms are applied to transform text words of a set of documents into K-dimensional word vectors in order to generate sets or sequences of word vectors representing the documents of the set of documents. A probabilistic topic model is learned using the sets or sequences of word vectors representing the documents of the set of documents. The set of word embedding transforms are applied to transform text words of an input document into K-dimensional word vectors in order to generate a set or sequence of word vectors representing the input document. The learned probabilistic topic model is applied to assign probabilities for topics of the probabilistic topic model to the set or sequence of word vectors representing the input document. A document processing operation such as annotation, classification, or similar document retrieval may be performed using the assigned topic probabilities.

46 citations


Proceedings ArticleDOI
07 Nov 2012
TL;DR: A novel word embedding model is proposed in which both images and words can be represented in the same embedding space in a discriminative nearest neighbor manner such that the annotation information could be propagated among neighbors.
Abstract: Automatically annotating words for images is a key to semantic-level image retrieval. Recently, several embedding learning based methods achieve good performance in this task which inspires this paper. Here we propose a novel word embedding model in which both images and words can be represented in the same embedding space. The embedding space is learnt in a discriminative nearest neighbor manner such that the annotation information could be propagated among neighbors. In order to accelerate model learning and testing, approximate-nearest-neighbor search is performed, and word embedding space is learnt in a stochastic manner. The experimental results demonstrate the effectiveness of the proposed method.