scispace - formally typeset
Open AccessPosted Content

A C-LSTM Neural Network for Text Classification

Reads0
Chats0
TLDR
C-LSTM is a novel and unified model for sentence representation and text classification that outperforms both CNN and LSTM and can achieve excellent performance on these tasks.
Abstract
Neural network models have been demonstrated to be capable of achieving remarkable performance in sentence and document modeling. Convolutional neural network (CNN) and recurrent neural network (RNN) are two mainstream architectures for such modeling tasks, which adopt totally different ways of understanding natural languages. In this work, we combine the strengths of both architectures and propose a novel and unified model called C-LSTM for sentence representation and text classification. C-LSTM utilizes CNN to extract a sequence of higher-level phrase representations, and are fed into a long short-term memory recurrent neural network (LSTM) to obtain the sentence representation. C-LSTM is able to capture both local features of phrases as well as global and temporal sentence semantics. We evaluate the proposed architecture on sentiment classification and question classification tasks. The experimental results show that the C-LSTM outperforms both CNN and LSTM and can achieve excellent performance on these tasks.

read more

Citations
More filters
Journal ArticleDOI

Deep learning for detecting inappropriate content in text

TL;DR: A novel deep learning architecture called “Convolutional Bi-Directional LSTM (C-BiLSTM)" is proposed which combines the strengths of both Convolution Neural Networks (CNN) and Bi-directional L STMs (BLSTM) and it is revealed that they significantly outperform both pattern-based and other hand-crafted feature-based baselines.
Proceedings ArticleDOI

Detecting Linguistic Characteristics of Alzheimer’s Dementia by Interpreting Neural Models

TL;DR: This paper used NLP techniques to classify and analyze the linguistic characteristics of Alzheimer's disease (AD) patients using the DementiaBank dataset and applied three neural models based on CNNs, LSTM-RNNs and their combination to distinguish between language samples from AD and control patients.
Proceedings ArticleDOI

Prayas at EmoInt 2017: An Ensemble of Deep Neural Architectures for Emotion Intensity Prediction in Tweets

TL;DR: The paper describes the best performing system for EmoInt - a shared task to predict the intensity of emotions in tweets by applying three different deep neural network based models, which approach the problem from essentially different directions.
Journal ArticleDOI

Detecting Malicious URLs via a Keyword-Based Convolutional Gated-Recurrent-Unit Neural Network

TL;DR: A convolutional gated-recurrent-unit (GRU) neural network is designed for the detection of malicious URLs detection based on characters as text classification features, and a GRU is used in place of the original pooling layer to perform feature acquisition on the time dimension, resulting in high-accuracy multicategory results.
Proceedings ArticleDOI

DeepHate: Hate Speech Detection via Multi-Faceted Text Representations

TL;DR: DeepHate is a novel deep learning model that combines multi-faceted text representations such as word embeddings, sentiments, and topical information, to detect hate speech in online social platforms and outperforms the state-of-the-art baselines on the hate speech detection task.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Proceedings ArticleDOI

Learning Phrase Representations using RNN Encoder--Decoder for Statistical Machine Translation

TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.
Proceedings Article

Rectified Linear Units Improve Restricted Boltzmann Machines

TL;DR: Restricted Boltzmann machines were developed using binary stochastic hidden units that learn features that are better for object recognition on the NORB dataset and face verification on the Labeled Faces in the Wild dataset.
Posted Content

Sequence to Sequence Learning with Neural Networks

TL;DR: This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Related Papers (5)