Recurrent Attention Network on Memory for Aspect Sentiment Analysis
Citations
917 citations
491 citations
Cites background from "Recurrent Attention Network on Memo..."
...Rather than using a single level of attention, deep memory networks (Tang, Qin, and Liu 2016) and recurrent attention models (Chen et al. 2017) have achieved superior performance by learning a deep attention over the singlelevel attention, as multiple passes (or hops) over the input sequence could…...
[...]
...Rather than using a single level of attention, deep memory networks (Tang, Qin, and Liu 2016) and recurrent attention models (Chen et al. 2017) have achieved superior performance by learning a deep attention over the singlelevel attention, as multiple passes (or hops) over the input sequence could refine the attended words again and again to find the most important words....
[...]
417 citations
388 citations
Cites background or methods from "Recurrent Attention Network on Memo..."
...Thus, it can be inferred that CRFs can take advantage of the entire sentence sequence to estimate probability for the sentence labelling making CRF a frequent final classification layer of bidirectional RNNs (T. Chen et al., 2017; Irsoy & Cardie, 2014; Lample et al., 2016; P. Liu et al., 2015)....
[...]
...Chen et al. (2017) Restaurant SemEval '16 English BiLSTM + Google WE + CRF F1: 72.44% Restaurant SemEval '16 Spanish F1: 71.70% Restaurant SemEval '16 French F1: 73.50% Restaurant SemEval '16 Russian F1: 67.08% Restaurant SemEval '16 Dutch F1: 64.29 % Restaurant SemEval '16 Turkish F1: 63.76% 3 Liu et al. (2015) Laptop SemEval '14 English LSTM-RNN+ POS + chunk + Amazon WE F1: 75....
[...]
...Chen et al. (2016) also combined LSTM and CNN together for sentiment classification but used LST for generating context embedding and CNN for detecting features....
[...]
...Chen et al. (2017) Twitter data Dong et al. (2014) English Recurrent Attention on Memory (RAM) + attention layers Acc: 69....
[...]
...Chen et al. (2017) and Tay, Tuan, et al. (2017) also focused on attention mechanisms for the LSTM to incorporate aspect information into the model. While P. Chen et al. (2017) adopted a multiple-attention mechanism, Tay, Tuan, et al. (2017) introduced a novel association layer with holographic reduced representation....
[...]
315 citations
References
30,558 citations
"Recurrent Attention Network on Memo..." refers background or methods in this paper
...We use 300-dimension word vectors pre-trained by GloVe (Pennington et al., 2014) (whose vocabulary size is 1.9M2) for our experiments on the English datasets, as previous works did (Tang et al., 2016)....
[...]
...the general embeddings from (Pennington et al., 2014) for all datasets, so that the experimental results can better reveal the model’s capability and...
[...]
...We use 300-dimension word vectors pre-trained by GloVe (Pennington et al., 2014) (whose vocabulary size is 1....
[...]
...In contrast, we prefer to use 2http://nlp.stanford.edu/projects/glove/ the general embeddings from (Pennington et al., 2014) for all datasets, so that the experimental results can better reveal the model’s capability and the figures are directly comparable across different papers....
[...]
...Let L ∈ Rd×|V | be an embedding lookup table generated by an unsupervised method such as GloVe (Pennington et al., 2014) or CBOW...
[...]
20,077 citations
20,027 citations
"Recurrent Attention Network on Memo..." refers background or methods in this paper
...Attention mechanism, which has been used successfully in many areas (Bahdanau et al., 2014; Rush et al., 2015), can be treated as a simplified version of NTM because the size of memory is unlimited and we only need to read from it....
[...]
...…feeling is that the phone, after using it for three months and considering its price, is really cost-effective”.1 Attention mechanism, which has been successfully used in machine translation (Bahdanau et al., 2014), can enforce a model to pay more attention to the important part of a sentence....
[...]
...the states of time steps generated by LSTM) from the input, as bidirectional recurrent neural networks (RNNs) were found effective for a similar purpose in machine translation (Bahdanau et al., 2014)....
[...]
...Specifically, our framework first adopts a bidirectional LSTM (BLSTM) to produce the memory (i.e. the states of time steps generated by LSTM) from the input, as bidirectional recurrent neural networks (RNNs) were found effective for a similar purpose in machine translation (Bahdanau et al., 2014)....
[...]
...1 Attention mechanism, which has been successfully used in machine translation (Bahdanau et al., 2014), can enforce a model to pay more attention to the important part of a sentence....
[...]
14,077 citations
9,270 citations
"Recurrent Attention Network on Memo..." refers background or methods in this paper
...(Mikolov et al., 2013), where d is the dimension of word vectors and |V | is the vocabulary size....
[...]
...The embeddings for Chinese experiments are trained with a corpus of 1.4 billion tokens with CBOW3....
[...]
...Let L ∈ Rd×|V | be an embedding lookup table generated by an unsupervised method such as GloVe (Pennington et al., 2014) or CBOW (Mikolov et al., 2013), where d is the dimension of word vectors and |V | is the vocabulary size....
[...]