scispace - formally typeset
Open AccessJournal ArticleDOI

Lexicon-Enhanced LSTM With Attention for General Sentiment Analysis

TLDR
A lexicon-enhanced LSTM model that first uses sentiment lexicon as an extra information pre-training a word sentiment classifier and then gets the sentiment embeddings of words including the words not in the lexicon to make word representation more accurate.
Abstract
Long short-term memory networks (LSTMs) have gained good performance in sentiment analysis tasks. The general method is to use LSTMs to combine word embeddings for text representation. However, word embeddings carry more semantic information rather than sentiment information. Only using word embeddings to represent words is inaccurate in sentiment analysis tasks. To solve the problem, we propose a lexicon-enhanced LSTM model. The model first uses sentiment lexicon as an extra information pre-training a word sentiment classifier and then get the sentiment embeddings of words including the words not in the lexicon. Combining the sentiment embedding and its word embedding can make word representation more accurate. Furthermore, we define a new method to find the attention vector in general sentiment analysis without a target that can improve the LSTM ability in capturing global sentiment information. The results of experiments on English and Chinese datasets show that our models have comparative or better results than the existing models.

read more

Citations
More filters
Journal ArticleDOI

Sarcasm Detection Using Soft Attention-Based Bidirectional Long Short-Term Memory Model With Convolution Network

TL;DR: A deep learning model called sAtt-BLSTM convNet that is based on the hybrid of soft attention-based bidirectional long short-term memory (sAtt- BLSTM) and convolution neural network (convNet) applying global vectors for word representation (GLoVe) for building semantic word embeddings is proposed.
Journal ArticleDOI

Sentiment Classification Using a Single-Layered BiLSTM Model

TL;DR: This study presents a computationally efficient deep learning model for binary sentiment classification, which aims to decide the sentiment polarity of people’s opinions, attitudes, and emotions expressed in written text and utilizes merely one bidirectional long short-term memory (BiLSTM) layer.
Journal ArticleDOI

Sentiment analysis of student feedback using multi-head attention fusion model of word and context embedding for LSTM

TL;DR: It is concluded that the fusion of multiple layers accompanied with LSTM improves the result over a common Natural Language Processing method.
Journal ArticleDOI

Improving time series forecasting using LSTM and attention models

TL;DR: The results reveal that model based on multi-head attention is the second-best method with regard to AR, which demonstrates the predictive power of attention mechanism in time series forecasting.
Journal ArticleDOI

Sentiment classification and aspect-based sentiment analysis on yelp reviews using deep learning and word embeddings

TL;DR: This research analysed the content of online reviews including the text of reviews and their rankings to support opinion mining and found that opinion mining has significantly supported knowledge and decision-making.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Related Papers (5)