scispace - formally typeset
Open AccessProceedings Article

Coupled Multi-Layer Attentions for Co-Extraction of Aspect and Opinion Terms.

TLDR
A novel deep learning model, named coupled multi-layer attentions, where each layer consists of a couple of attentions with tensor operators that are learned interactively to dually propagate information between aspect terms and opinion terms.
Abstract
The task of aspect and opinion terms co-extraction aims to explicitly extract aspect terms describing features of an entity and opinion terms expressing emotions from user-generated texts. To achieve this task, one effective approach is to exploit relations between aspect terms and opinion terms by parsing syntactic structure for each sentence. However, this approach requires expensive effort for parsing and highly depends on the quality of the parsing results. In this paper, we offer a novel deep learning model, named coupled multi-layer attentions. The proposed model provides an end-to-end solution and does not require any parsers or other linguistic resources for preprocessing. Specifically, the proposed model is a multilayer attention network, where each layer consists of a couple of attentions with tensor operators. One attention is for extracting aspect terms, while the other is for extracting opinion terms. They are learned interactively to dually propagate information between aspect terms and opinion terms. Through multiple layers, the model can further exploit indirect relations between terms for more precise information extraction. Experimental results on three benchmark datasets in SemEval Challenge 2014 and 2015 show that our model achieves stateof-the-art performances compared with several baselines.

read more

Citations
More filters
Journal ArticleDOI

Multilingual aspect clustering for sentiment analysis

TL;DR: This article addresses the novel task of multilingual aspect clustering, which aims at grouping semantically related aspects extracted from reviews written in several languages using an unsupervised method that relies on the contextual information of the aspects, which is represented by word embeddings.
Journal ArticleDOI

A dependency syntactic knowledge augmented interactive architecture for end-to-end aspect-based sentiment analysis

TL;DR: The authors proposed a dependency syntactic knowledge augmented interactive architecture with multi-task learning for end-to-end aspect-based sentiment analysis (ABSA) task, which aims to extract the aspect term and then identify its sentiment orientation.
Proceedings ArticleDOI

Learning Span-Level Interactions for Aspect Sentiment Triplet Extraction

TL;DR: This paper proposed a dual-channel span pruning strategy by incorporating supervision from the Aspect Term Extraction (ATE) and Opinion Term Extraction (OTE) tasks, which not only improves computational efficiency but also distinguishes the opinion and target spans more properly.
Proceedings ArticleDOI

Enhanced Multi-Channel Graph Convolutional Network for Aspect Sentiment Triplet Extraction

TL;DR: This paper proposes an Enhanced Multi-Channel Graph Convolutional Network model (EMC-GCN), which outperforms state-of-the-art methods significantly and considers the implicit results of aspect and opinion extraction when determining whether word pairs match or not.
Proceedings ArticleDOI

Understanding Pre-trained BERT for Aspect-based Sentiment Analysis

TL;DR: It is found that BERT uses very few self-attention heads to encode context words and opinion words for an aspect, and most features in the representation of an aspect are dedicated to the fine-grained semantics of the domain and the aspect itself, instead of carrying summarized opinions from its context.
References
More filters
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Proceedings Article

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: It is conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and it is proposed to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Proceedings ArticleDOI

Learning Phrase Representations using RNN Encoder--Decoder for Statistical Machine Translation

TL;DR: In this paper, the encoder and decoder of the RNN Encoder-Decoder model are jointly trained to maximize the conditional probability of a target sequence given a source sequence.
Posted Content

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: In this paper, the authors propose to use a soft-searching model to find the parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Book

Opinion Mining and Sentiment Analysis

TL;DR: This survey covers techniques and approaches that promise to directly enable opinion-oriented information-seeking systems and focuses on methods that seek to address the new challenges raised by sentiment-aware applications, as compared to those that are already present in more traditional fact-based analysis.
Related Papers (5)