A
Aditya Siddhant
Researcher at Google
Publications - 34
Citations - 2418
Aditya Siddhant is an academic researcher from Google. The author has contributed to research in topics: Machine translation & Computer science. The author has an hindex of 12, co-authored 30 publications receiving 934 citations. Previous affiliations of Aditya Siddhant include Carnegie Mellon University & Indian Institute of Technology Guwahati.
Papers
More filters
Proceedings ArticleDOI
mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
Linting Xue,Noah Constant,Adam Roberts,Mihir Kale,Rami Al-Rfou,Aditya Siddhant,Aditya Barua,Colin Raffel +7 more
TL;DR: This paper proposed a multilingual variant of T5, mT5, which was pre-trained on a new Common Crawl-based dataset covering 101 languages and achieved state-of-the-art performance on many multilingual benchmarks.
Posted Content
XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization
TL;DR: The Cross-lingual TRansfer Evaluation of Multilingual Encoders XTREME benchmark is introduced, a multi-task benchmark for evaluating the cross-lingually generalization capabilities of multilingual representations across 40 languages and 9 tasks.
Proceedings Article
XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalisation
TL;DR: The Cross-lingual TRansfer Evaluation of Multilingual Encoders XTREME benchmark as discussed by the authors is a multi-task benchmark for evaluating the crosslingual generalization capabilities of multilingual representations across 40 languages and 9 tasks.
Proceedings ArticleDOI
Deep Bayesian Active Learning for Natural Language Processing: Results of a Large-Scale Empirical Study
TL;DR: A large-scale empirical study of deep active learning, addressing multiple tasks and, for each, multiple datasets, multiple models, and a full suite of acquisition functions, finds that across all settings, Bayesian active learning by disagreement significantly improves over i.i.d. baselines and usually outperforms classic uncertainty sampling.
Posted Content
mT5: A massively multilingual pre-trained text-to-text transformer
Linting Xue,Noah Constant,Adam Roberts,Mihir Kale,Rami Al-Rfou,Aditya Siddhant,Aditya Barua,Colin Raffel +7 more
TL;DR: This article proposed a multilingual variant of T5, mT5, which was pre-trained on a new Common Crawl-based dataset covering 101 languages and achieved state-of-the-art performance on many multilingual benchmarks.