Cross-Language Text Classification Using Structural Correspondence Learning
Citations
2,900 citations
454 citations
435 citations
Cites methods from "Cross-Language Text Classification ..."
...Sentiment Classification: We use the Cross-Lingual Sentiment (CLS) dataset5 [36], which is an extended version of the Multi-Domain Sentiment Dataset [2] widely used for domain adaptation....
[...]
315 citations
Cites background or methods from "Cross-Language Text Classification ..."
...Based on structural correspondence learning (Blitzer et al., 2006), two methods (Prettenhofer & Stein, 2010; Wei & Pal, 2010) were recently proposed to extract the so-called pivot features from the source and target domains, which is specifically designed for the cross-language text classification…...
[...]
...The pioneer works (Dai et al., 2009; Prettenhofer & Stein, 2010; Wei & Pal, 2010; Yang et al., 2009; Zhu et al., 2011) are limited to some specific HDA tasks, because they required additional information to transfer the source knowledge to the target domain....
[...]
281 citations
Cites methods from "Cross-Language Text Classification ..."
...With the advances of machine translation techniques, one can easily get different translations for one document (Prettenhofer & Stein, 2010), and the translation in each language can be considered as a view....
[...]
References
16,538 citations
"Cross-Language Text Classification ..." refers background or methods in this paper
...An alternative view of cross-language structural correspondence learning is provided by the framework of structural learning (Ando and Zhang, 2005a)....
[...]
...SCL is related to the structural learning paradigm introduced by Ando and Zhang (2005a)....
[...]
...Following Ando and Zhang (2005a) and Quattoni et al. (2007) we choose w for the target task to be w∗ = θT v∗, where v∗ is defined as follows: v∗ = argmin v∈Rk ∑ (x,y)∈DS L(y, (θTv)Tx) + λ 2 ‖v‖2 (3) Since (θT v)T = vT θ it follows that this view of CL-SCL corresponds to the induction of a new…...
[...]
...Here we propose a different approach to crosslanguage text classification which adopts ideas from the field of multi-task learning (Ando and Zhang, 2005a)....
[...]
...Ando and Zhang (2005b) present a semi-supervised learning method based on this paradigm, which generates related tasks from unlabeled data....
[...]
6,980 citations
6,626 citations
"Cross-Language Text Classification ..." refers result in this paper
...The average accuracy is about 82%, which is consistent with prior work on monolingual sentiment analysis (Pang et al., 2002; Blitzer et al., 2007)....
[...]
2,239 citations
"Cross-Language Text Classification ..." refers background or result in this paper
...The average accuracy is about 82%, which is consistent with prior work on monolingual sentiment analysis (Pang et al., 2002; Blitzer et al., 2007)....
[...]
...Following Blitzer et al. (2007) a review with >3 (<3) stars is labeled as positive (negative); other reviews are discarded....
[...]
...The corpus is extended with English product reviews provided by Blitzer et al. (2007)....
[...]
2,037 citations
"Cross-Language Text Classification ..." refers methods in this paper
...In particular, the learning rate schedule from PEGASOS is adopted (Shalev-Shwartz et al., 2007), and the modified Huber loss, introduced by Zhang (2004), is chosen as loss function L....
[...]
...In particular, the learning rate schedule from PEGASOS is adopted (Shalev-Shwartz et al., 2007), and the modified Huber loss, introduced by Zhang (2004), is chosen as loss function L.3 SGD receives two hyperparameters as input: the number of iterations T , and the regularization parameter λ....
[...]