scispace - formally typeset
Open AccessPosted Content

Release Strategies and the Social Impacts of Language Models.

TLDR
This report discusses OpenAI's work related to the release of its GPT-2 language model and discusses staged release, which allows time between model releases to conduct risk and benefit analyses as model sizes increased.
Abstract
Large language models have a range of beneficial uses: they can assist in prose, poetry, and programming; analyze dataset biases; and more. However, their flexibility and generative capabilities also raise misuse concerns. This report discusses OpenAI's work related to the release of its GPT-2 language model. It discusses staged release, which allows time between model releases to conduct risk and benefit analyses as model sizes increased. It also discusses ongoing partnership-based research and provides recommendations for better coordination and responsible publication in AI.

read more

Citations
More filters
Proceedings ArticleDOI

Training language models to follow instructions with human feedback

TL;DR: The results show that fine-tuning with human feedback is a promising direction for aligning language models with human intent and showing improvements in truthfulness and reductions in toxic output generation while having minimal performance regressions on public NLP datasets.
Posted Content

Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks

TL;DR: A general-purpose fine-tuning recipe for retrieval-augmented generation (RAG) -- models which combine pre-trained parametric and non-parametric memory for language generation, and finds that RAG models generate more specific, diverse and factual language than a state-of-the-art parametric-only seq2seq baseline.
Proceedings Article

Defending Against Neural Fake News

TL;DR: A model for controllable text generation called Grover, found that best current discriminators can classify neural fake news from real, human-written, news with 73% accuracy, assuming access to a moderate level of training data, and the best defense against Grover turns out to be Grover itself, with 92% accuracy.
References
More filters
Journal ArticleDOI

Semantics derived automatically from language corpora contain human-like biases

TL;DR: This article showed that applying machine learning to ordinary human language results in human-like semantic biases and replicated a spectrum of known biases, as measured by the Implicit Association Test, using a widely used, purely statistical machine-learning model trained on a standard corpus of text from the World Wide Web.
Posted Content

The Curious Case of Neural Text Degeneration

TL;DR: This paper showed that decoding strategies alone alone can dramatically affect the quality of machine text, even when generated from exactly the same neural language model, and they proposed Nucleus Sampling, a simple but effective method to draw the best out of neural generation.
Posted Content

Semi-supervised Sequence Learning

Andrew M. Dai, +1 more
- 04 Nov 2015 - 
TL;DR: This paper used unlabeled data to improve sequence learning with recurrent networks, which can be used as a "pre-training" step for a later supervised sequence learning algorithm, so that the parameters obtained from the unsupervised step can be a starting point for other supervised training models.
Journal ArticleDOI

Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science

TL;DR: It is argued that data statements will help alleviate issues related to exclusion and bias in language technology, lead to better precision in claims about how natural language processing research can generalize and thus better engineering results, protect companies from public embarrassment, and ultimately lead to language technology that meets its users in their own preferred linguistic style.
Proceedings Article

Defending Against Neural Fake News

TL;DR: A model for controllable text generation called Grover, found that best current discriminators can classify neural fake news from real, human-written, news with 73% accuracy, assuming access to a moderate level of training data, and the best defense against Grover turns out to be Grover itself, with 92% accuracy.
Related Papers (5)
Trending Questions (3)
What are the use cases of large language models in ERP?

The paper does not specifically mention the use cases of large language models in ERP (Enterprise Resource Planning). The paper discusses the use cases of language models in prose, poetry, programming, and dataset analysis.

What are the potential uses of large language models for cities?

The paper does not specifically mention the potential uses of large language models for cities. The paper discusses the beneficial uses of language models in general, such as assisting in prose, poetry, programming, and analyzing dataset biases.

What is large language model?

A large language model refers to a powerful AI system that can understand and generate human-like text.