scispace - formally typeset
A

Aditya Siddhant

Researcher at Google

Publications -  34
Citations -  2418

Aditya Siddhant is an academic researcher from Google. The author has contributed to research in topics: Machine translation & Computer science. The author has an hindex of 12, co-authored 30 publications receiving 934 citations. Previous affiliations of Aditya Siddhant include Carnegie Mellon University & Indian Institute of Technology Guwahati.

Papers
More filters
Proceedings ArticleDOI

mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer

TL;DR: This paper proposed a multilingual variant of T5, mT5, which was pre-trained on a new Common Crawl-based dataset covering 101 languages and achieved state-of-the-art performance on many multilingual benchmarks.
Posted Content

XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization

TL;DR: The Cross-lingual TRansfer Evaluation of Multilingual Encoders XTREME benchmark is introduced, a multi-task benchmark for evaluating the cross-lingually generalization capabilities of multilingual representations across 40 languages and 9 tasks.
Proceedings Article

XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalisation

TL;DR: The Cross-lingual TRansfer Evaluation of Multilingual Encoders XTREME benchmark as discussed by the authors is a multi-task benchmark for evaluating the crosslingual generalization capabilities of multilingual representations across 40 languages and 9 tasks.
Proceedings ArticleDOI

Deep Bayesian Active Learning for Natural Language Processing: Results of a Large-Scale Empirical Study

TL;DR: A large-scale empirical study of deep active learning, addressing multiple tasks and, for each, multiple datasets, multiple models, and a full suite of acquisition functions, finds that across all settings, Bayesian active learning by disagreement significantly improves over i.i.d. baselines and usually outperforms classic uncertainty sampling.
Posted Content

mT5: A massively multilingual pre-trained text-to-text transformer

TL;DR: This article proposed a multilingual variant of T5, mT5, which was pre-trained on a new Common Crawl-based dataset covering 101 languages and achieved state-of-the-art performance on many multilingual benchmarks.