Instance Weighting for Neural Machine Translation Domain Adaptation
Citations
239 citations
182 citations
Cites background or methods from "Instance Weighting for Neural Machi..."
...On the other hand, the model centric category focuses on NMT models that are specialized for domain adaptation, which can be either the training objective (Luong and Manning, 2015; Sennrich et al., 2016b; Servan et al., 2016; Freitag and Al-Onaizan, 2016; Wang et al., 2017b; Chen et al., 2017a; Varga, 2017; Dakwale and Monz, 2017; Chu et al., 2017; Miceli Barone et al., 2017), the NMT architecture (Kobus et al....
[...]
...To address this problem, Wang et al. (2017a) exploit the internal embedding of the source sentence in NMT, and use the sentence embedding similarity to select the sentences that are close to in-domain data from out-of-domain data (Figure 4)....
[...]
...Figure 5: Instance weighting for NMT (Wang et al., 2017b)....
[...]
...…and Zong, 2016b; Cheng et al., 2016; Currey et al., 2017; Domhan and Hieber, 2017), synthetic corpora (Sennrich et al., 2016b; Zhang and Zong, 2016b; Park et al., 2017), or parallel copora (Chu et al., 2017; Sajjad et al., 2017; Britz et al., 2017; Wang et al., 2017a; van der Wees et al., 2017)....
[...]
...Data Selection As mentioned in the SMT section (Section 3.1), the data selection methods in SMT can improve NMT performance modestly, because their criteria of data selection are not very related to NMT (Wang et al., 2017a)....
[...]
146 citations
127 citations
125 citations
Cites background from "Instance Weighting for Neural Machi..."
...Intuitive Weighting Adaptive tuning [98], [99], [100], [101]...
[...]
...Instance re-weighting based domain adaptation was first proposed for natural language processing (NLP) [98], [99]....
[...]
References
21,126 citations
8,055 citations
"Instance Weighting for Neural Machi..." refers background in this paper
...…directly models the conditional probability p(y|x) of translating a source sentence, x = {x1, ..., xn}, to a target sentence, y = {y1, ..., ym} (Luong et al., 2015): p(y|x) = m∏ j=1 softmax(g(yj |yj−1, sj , cj)), (1) with g being the transformation function that outputs a vocabulary-sized…...
[...]
6,189 citations
"Instance Weighting for Neural Machi..." refers methods in this paper
...Each NMT model was trained for 500K batches by using ADADELTA optimizer (Zeiler, 2012)....
[...]
6,008 citations
"Instance Weighting for Neural Machi..." refers background or methods in this paper
...Further training (Luong and Manning, 2015) can be viewed as a special case of the proposed batch weighting method....
[...]
...Recently, Chu et al. (2017) make an empirical comparison of NMT further training (Luong and Manning, 2015) and domain control (Kobus et al., 2016), which applied word-level domain features to word embedding layer....
[...]
...This adaptation corpora settings were the same as those used in (Luong and Manning, 2015)....
[...]
...In Statistical Machine Translation (SMT), unrelated additional corpora, known as out-ofdomain corpora, have been shown not to benefit some domains and tasks, such as TED-talks and IWSLT tasks (Axelrod et al., 2011; Luong and Manning, 2015)....
[...]
...There are two methods for model combination of NMT: i) the in-domain model and out-of-domain model can be ensembled (Jean et al., 2015). ii) an NMT further training (fine-tuning) method (Luong and Manning, 2015)....
[...]
1,690 citations
"Instance Weighting for Neural Machi..." refers methods in this paper
..., 2002), with the paired bootstrap re-sampling test (Koehn, 2004)5....
[...]