Lexicon-Enhanced LSTM With Attention for General Sentiment Analysis
TLDR
A lexicon-enhanced LSTM model that first uses sentiment lexicon as an extra information pre-training a word sentiment classifier and then gets the sentiment embeddings of words including the words not in the lexicon to make word representation more accurate.Abstract:
Long short-term memory networks (LSTMs) have gained good performance in sentiment analysis tasks. The general method is to use LSTMs to combine word embeddings for text representation. However, word embeddings carry more semantic information rather than sentiment information. Only using word embeddings to represent words is inaccurate in sentiment analysis tasks. To solve the problem, we propose a lexicon-enhanced LSTM model. The model first uses sentiment lexicon as an extra information pre-training a word sentiment classifier and then get the sentiment embeddings of words including the words not in the lexicon. Combining the sentiment embedding and its word embedding can make word representation more accurate. Furthermore, we define a new method to find the attention vector in general sentiment analysis without a target that can improve the LSTM ability in capturing global sentiment information. The results of experiments on English and Chinese datasets show that our models have comparative or better results than the existing models.read more
Citations
More filters
Journal ArticleDOI
Sarcasm Detection Using Soft Attention-Based Bidirectional Long Short-Term Memory Model With Convolution Network
Le Hoang Son,Akshi Kumar,Saurabh Raj Sangwan,Anshika Arora,Anand Nayyar,Mohamed Abdel-Basset +5 more
TL;DR: A deep learning model called sAtt-BLSTM convNet that is based on the hybrid of soft attention-based bidirectional long short-term memory (sAtt- BLSTM) and convolution neural network (convNet) applying global vectors for word representation (GLoVe) for building semantic word embeddings is proposed.
Journal ArticleDOI
Sentiment Classification Using a Single-Layered BiLSTM Model
TL;DR: This study presents a computationally efficient deep learning model for binary sentiment classification, which aims to decide the sentiment polarity of people’s opinions, attitudes, and emotions expressed in written text and utilizes merely one bidirectional long short-term memory (BiLSTM) layer.
Journal ArticleDOI
Sentiment analysis of student feedback using multi-head attention fusion model of word and context embedding for LSTM
K. Sangeetha,D. Prabha +1 more
TL;DR: It is concluded that the fusion of multiple layers accompanied with LSTM improves the result over a common Natural Language Processing method.
Journal ArticleDOI
Improving time series forecasting using LSTM and attention models
Hossein Abbasimehr,Reza Paki +1 more
TL;DR: The results reveal that model based on multi-head attention is the second-best method with regard to AR, which demonstrates the predictive power of attention mechanism in time series forecasting.
Journal ArticleDOI
Sentiment classification and aspect-based sentiment analysis on yelp reviews using deep learning and word embeddings
TL;DR: This research analysed the content of online reviews including the text of reviews and their rankings to support opinion mining and found that opinion mining has significantly supported knowledge and decision-making.
References
More filters
Proceedings Article
Adam: A Method for Stochastic Optimization
Diederik P. Kingma,Jimmy Ba +1 more
TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI
Long short-term memory
TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings ArticleDOI
Glove: Global Vectors for Word Representation
TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings Article
Distributed Representations of Words and Phrases and their Compositionality
TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content
Efficient Estimation of Word Representations in Vector Space
TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.