scispace - formally typeset
Open AccessProceedings ArticleDOI

mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer

Reads0
Chats0
TLDR
This paper proposed a multilingual variant of T5, mT5, which was pre-trained on a new Common Crawl-based dataset covering 101 languages and achieved state-of-the-art performance on many multilingual benchmarks.
Abstract
The recent “Text-to-Text Transfer Transformer” (T5) leveraged a unified text-to-text format and scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this paper, we introduce mT5, a multilingual variant of T5 that was pre-trained on a new Common Crawl-based dataset covering 101 languages. We detail the design and modified training of mT5 and demonstrate its state-of-the-art performance on many multilingual benchmarks. We also describe a simple technique to prevent “accidental translation” in the zero-shot setting, where a generative model chooses to (partially) translate its prediction into the wrong language. All of the code and model checkpoints used in this work are publicly available.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Multilingual Detection of Check-Worthy Claims using World Languages and Adapter Fusion

TL;DR: This article proposed cross-training adapters on a subset of world languages, combined by adapter fusion, to detect claims emerging globally in multiple languages, which outperformed the top multilingual approaches in their benchmark tasks.
Proceedings ArticleDOI

CLSE: Corpus of Linguistically Significant Entities

TL;DR: This article released a corpus of linguistically significant entities (CLSE) annotated by linguist experts, which includes 34 languages and covers 74 different semantic types to support various applications from airline ticketing to video games.
Proceedings Article

MUCS@MixMT: IndicTrans-based Machine Translation for Hinglish Text

TL;DR: In the Code-mixed Machine Translation (MixMT) shared task in the WMT shared task as discussed by the authors , the authors have proposed models that translate the codemixed English text to Hinglish text and vice-versa.
Proceedings ArticleDOI

COMPILING: A Benchmark Dataset for Chinese Complexity Controllable Definition Generation

TL;DR: This paper proposes a novel task of generating definitions for a word with controllable complexity levels, and introduces COMPILING, a dataset given detailed information about Chinese de-flnitions, and each de-nition is labeled with its complexity levels.
Proceedings ArticleDOI

GlobalWoZ: Globalizing MultiWoZ to Develop Multilingual Task-Oriented Dialogue Systems

TL;DR: Ding et al. as discussed by the authors presented the 60th Annual Meeting of the Association for Computational Linguistics (ACLL) on the topic of "Linguistics: Long Papers".
References
More filters
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Posted Content

RoBERTa: A Robustly Optimized BERT Pretraining Approach

TL;DR: It is found that BERT was significantly undertrained, and can match or exceed the performance of every model published after it, and the best model achieves state-of-the-art results on GLUE, RACE and SQuAD.
Proceedings ArticleDOI

SQuAD: 100,000+ Questions for Machine Comprehension of Text

TL;DR: The Stanford Question Answering Dataset (SQuAD) as mentioned in this paper is a reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corresponding reading passage.
Proceedings ArticleDOI

Unsupervised Cross-lingual Representation Learning at Scale

TL;DR: It is shown that pretraining multilingual language models at scale leads to significant performance gains for a wide range of cross-lingual transfer tasks, and the possibility of multilingual modeling without sacrificing per-language performance is shown for the first time.
Proceedings ArticleDOI

Universal Language Model Fine-tuning for Text Classification

TL;DR: Universal Language Model Fine-tuning (ULMFiT) as mentioned in this paper is an effective transfer learning method that can be applied to any task in NLP, and introduces techniques that are key for finetuning a language model.
Related Papers (5)
Trending Questions (3)
ISINDEBELE text generation under NLP using MT5 tool

The paper does not specifically mention ISINDEBELE text generation using the MT5 tool. The paper introduces mT5, a multilingual variant of T5, and demonstrates its performance on multilingual benchmarks.

Isindebele text generation under NLP using MT5 tool

The paper does not mention specifically about Isindebele text generation using the MT5 tool.

A Massively Multilingual Pre-trained Text-to-Text Transformer?

The paper introduces mT5, a multilingual variant of T5, which is a massively multilingual pre-trained text-to-text transformer.