scispace - formally typeset
Open AccessProceedings ArticleDOI

Cross-Lingual Learning-to-Rank with Shared Representations

TLDR
A large-scale dataset derived from Wikipedia is introduced to support CLIR research in 25 languages and a simple yet effective neural learning-to-rank model is presented that shares representations across languages and reduces the data requirement.
Abstract
Cross-lingual information retrieval (CLIR) is a document retrieval task where the documents are written in a language different from that of the user’s query. This is a challenging problem for data-driven approaches due to the general lack of labeled training data. We introduce a large-scale dataset derived from Wikipedia to support CLIR research in 25 languages. Further, we present a simple yet effective neural learning-to-rank model that shares representations across languages and reduces the data requirement. This model can exploit training data in, for example, Japanese-English CLIR to improve the results of Swahili-English CLIR.

read more

Content maybe subject to copyright    Report

Citations
More filters

Cross-Language Information Retrieval.

TL;DR: This chapter reviews research and practice in CLIR that allows users to state queries in their native language and retrieve documents in any other language supported by the system.
Proceedings ArticleDOI

On Difficulties of Cross-Lingual Transfer with Order Differences: A Case Study on Dependency Parsing

TL;DR: The authors compare encoders and decoders based on Recurrent Neural Networks (RNNs) and modified self-attentive architectures for cross-lingual transfer and show that RNN-based architectures transfer well to languages that are close to English and perform especially well on distant languages.
Proceedings ArticleDOI

XOR QA: Cross-lingual Open-Retrieval Question Answering

TL;DR: This work constructs a large-scale dataset built on 40K information-seeking questions across 7 diverse non-English languages that TyDi QA could not find same-language answers for and introduces a task framework, called Cross-lingual Open-Retrieval Question Answering (XOR QA), that consists of three new tasks involving cross-lingually document retrieval from multilingual and English resources.
Posted Content

On Difficulties of Cross-Lingual Transfer with Order Differences: A Case Study on Dependency Parsing

TL;DR: Investigating crosslingual transfer and posit that an orderagnostic model will perform better when transferring to distant foreign languages shows that RNN-based architectures transfer well to languages that are close to English, while self-attentive models have better overall cross-lingualtransferability and perform especially well on distant languages.
Posted Content

FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding

TL;DR: FILTER is proposed, an enhanced fusion method that takes cross-lingual data as input for XLM finetuning and proposes an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Proceedings ArticleDOI

Learning deep structured semantic models for web search using clickthrough data

TL;DR: A series of new latent semantic models with a deep structure that project queries and documents into a common low-dimensional space where the relevance of a document given a query is readily computed as the distance between them are developed.
Proceedings ArticleDOI

A Deep Relevance Matching Model for Ad-hoc Retrieval

TL;DR: A novel deep relevance matching model (DRMM) for ad-hoc retrieval that employs a joint deep architecture at the query term level for relevance matching and can significantly outperform some well-known retrieval models as well as state-of-the-art deep matching models.
Related Papers (5)