Universal Sentence Encoder for English
Citations
512 citations
487 citations
345 citations
344 citations
Additional excerpts
..., 2016), QuickThought (Logeswaran and Lee, 2018), USETrans (Cer et al., 2018), and m-USETrans (Yang et al....
[...]
316 citations
Cites background or methods from "Universal Sentence Encoder for Engl..."
...Multitask learning has been shown to be helpful to learn English sentence embeddings (Subrama- 10We consider the average of en→xx and xx→en nian et al., 2018; Cer et al., 2018)....
[...]
...This was recently extended to multitask learning, combining different training objectives like that of skip-thought, NLI and machine translation (Cer et al., 2018; Subramanian et al., 2018)....
[...]
References
72,897 citations
"Universal Sentence Encoder for Engl..." refers methods in this paper
...This section presents the data used for the transfer learning experiments and word embedding association tests (WEAT): (MR) Movie review sentiment on a five star scale (Pang and Lee, 2005); (CR) Sentiment of customer reviews (Hu and Liu, 2004); (SUBJ) Subjectivity of movie reviews and plot summaries (Pang and Lee, 2004); 5The Skip-Thought like task replaces the LSTM (Hochreiter and Schmidhuber, 1997) in the original formulation with a transformer model....
[...]
...The Skip-Thought like task replaces the LSTM (Hochreiter and Schmidhuber, 1997) in the original formulation with a transformer model....
[...]
52,856 citations
24,012 citations
"Universal Sentence Encoder for Engl..." refers methods in this paper
...For word-level transfer, we incorporate word embeddings from a word2vec skip-gram model trained on a corpus of news data (Mikolov et al., 2013)....
[...]
10,913 citations
9,776 citations
"Universal Sentence Encoder for Engl..." refers methods in this paper
...However, we observe that our best performing model still makes use of transformer sentencelevel transfer but combined with a CNN with no word-level transfer, UT+CNNrnd. Table 5 contrasts Caliskan et al. (2017)’s findings on bias within GloVe embeddings with results from the transformer and DAN encoders....
[...]
...On the STS Benchmark, we compare with InferSent and the state-of-the-art neural STS systems CNN (HCTI) (Shao, 2017) and gConv (Yang et al., 2018)....
[...]
...The pretrained word embeddings are included as input to two model types: a convolutional neural network model (CNN) (Kim, 2014); a DAN....
[...]
...Additional baseline CNN and DAN models are trained without using any pretrained word or sentence embeddings....
[...]
...Training with 1k labeled examples and the transformer sentence embeddings surpasses wordlevel transfer using the full training set, CNNw2v, and approaches the performance of the best model without transfer learning trained on the complete dataset, CNNrnd@67.3k....
[...]