scispace - formally typeset
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

Reads0
Chats0
TLDR
A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Abstract
Recent methods for learning vector space representations of words have succeeded in capturing fine-grained semantic and syntactic regularities using vector arithmetic, but the origin of these regularities has remained opaque. We analyze and make explicit the model properties needed for such regularities to emerge in word vectors. The result is a new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods. Our model efficiently leverages statistical information by training only on the nonzero elements in a word-word cooccurrence matrix, rather than on the entire sparse matrix or on individual context windows in a large corpus. The model produces a vector space with meaningful substructure, as evidenced by its performance of 75% on a recent word analogy task. It also outperforms related models on similarity tasks and named entity recognition.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Comparison of neutrosophic approach to various deep learning models for sentiment analysis

TL;DR: In this article, the authors proposed a novel framework to implement neutrosophy in deep learning models, where instead of just predicting a single class as output, they quantified the sentiments using three membership functions to understand them better.
Proceedings ArticleDOI

Mittens: an Extension of GloVe for Learning Domain-Specialized Representations

TL;DR: In this article, a simple extension of the GloVe representation learning model is presented, which can lead to faster learning and better results on a variety of tasks on a specialized domain.
Proceedings ArticleDOI

Learning with Weak Supervision for Email Intent Detection

TL;DR: In this article, the authors propose to leverage user actions as a source of weak supervision, in addition to a limited set of annotated examples, to detect intents in emails.
Proceedings ArticleDOI

Learn to Select via Hierarchical Gate Mechanism for Aspect-Based Sentiment Analysis.

TL;DR: A novel architecture named Hierarchical Gate Memory Network (HGMN) is proposed for ABSA, which employs the proposed hierarchical gate mechanism to learn to select the related part about the given aspect, which can keep the original sequence structure of sentence at the same time.
References
More filters
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Journal ArticleDOI

Indexing by Latent Semantic Analysis

TL;DR: A new method for automatic indexing and retrieval to take advantage of implicit higher-order structure in the association of terms with documents (“semantic structure”) in order to improve the detection of relevant documents on the basis of terms found in queries.
Proceedings Article

Efficient Estimation of Word Representations in Vector Space

TL;DR: Two novel model architectures for computing continuous vector representations of words from very large data sets are proposed and it is shown that these vectors provide state-of-the-art performance on the authors' test set for measuring syntactic and semantic word similarities.
Book

Learning Deep Architectures for AI

TL;DR: The motivations and principles regarding learning algorithms for deep architectures, in particular those exploiting as building blocks unsupervised learning of single-layer modelssuch as Restricted Boltzmann Machines, used to construct deeper models such as Deep Belief Networks are discussed.
Related Papers (5)