G
Girish Sastry
Researcher at OpenAI
Publications - 17
Citations - 13037
Girish Sastry is an academic researcher from OpenAI. The author has contributed to research in topics: Language model & Usability. The author has an hindex of 8, co-authored 9 publications receiving 3765 citations. Previous affiliations of Girish Sastry include Indian Institute of Technology Kharagpur & University of Oxford.
Papers
More filters
Proceedings Article
Language Models are Few-Shot Learners
Tom B. Brown,Benjamin Mann,Nick Ryder,Melanie Subbiah,Jared Kaplan,Prafulla Dhariwal,Arvind Neelakantan,Pranav Shyam,Girish Sastry,Amanda Askell,Sandhini Agarwal,Ariel Herbert-Voss,Gretchen Krueger,Thomas Henighan,Rewon Child,Aditya Ramesh,Daniel M. Ziegler,Jeffrey Wu,Clemens Winter,Christopher Hesse,Mark Chen,Eric Sigler,Mateusz Litwin,Scott Gray,Benjamin Chess,Jack Clark,Christopher Berner,Samuel McCandlish,Alec Radford,Ilya Sutskever,Dario Amodei +30 more
TL;DR: GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic.
Posted Content
Language Models are Few-Shot Learners
Tom B. Brown,Benjamin Mann,Nick Ryder,Melanie Subbiah,Jared Kaplan,Prafulla Dhariwal,Arvind Neelakantan,Pranav Shyam,Girish Sastry,Amanda Askell,Sandhini Agarwal,Ariel Herbert-Voss,Gretchen Krueger,Thomas Henighan,Rewon Child,Aditya Ramesh,Daniel M. Ziegler,Jeffrey Wu,Clemens Winter,Christopher Hesse,Mark Chen,Eric Sigler,Mateusz Litwin,Scott Gray,Benjamin Chess,Jack Clark,Christopher Berner,Samuel McCandlish,Alec Radford,Ilya Sutskever,Dario Amodei +30 more
TL;DR: This article showed that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches.
Posted Content
Learning Transferable Visual Models From Natural Language Supervision
Alec Radford,Jong Wook Kim,Chris Hallacy,Aditya Ramesh,Gabriel Goh,Sandhini Agarwal,Girish Sastry,Amanda Askell,Pamela Mishkin,Jack Clark,Gretchen Krueger,Ilya Sutskever +11 more
TL;DR: In this article, a pre-training task of predicting which caption goes with which image is used to learn SOTA image representations from scratch on a dataset of 400 million (image, text) pairs collected from the internet.
Posted Content
Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims
Miles Brundage,Shahar Avin,Jasmine Wang,Haydn Belfield,Gretchen Krueger,Gillian K. Hadfield,Gillian K. Hadfield,Heidy Khlaaf,Jingying Yang,Helen Toner,Ruth Fong,Tegan Maharaj,Pang Wei Koh,Sara Hooker,Jade Leung,Andrew Trask,Emma Bluemke,Jonathan Lebensbold,Cullen O'Keefe,Mark Koren,Théo Ryffel,J. B. Rubinovitz,Tamay Besiroglu,Federica Carugati,Jack Clark,Peter Eckersley,Sarah de Haas,Maritza Johnson,Ben Laurie,Alex Ingerman,Igor Krawczuk,Amanda Askell,Rosario Cammarota,Andrew J. Lohn,David Krueger,Charlotte Stix,Peter Henderson,Logan Graham,Carina E. A. Prunkl,Bianca Martin,Elizabeth Seger,Noa Zilberman,Seán Ó hÉigeartaigh,Frens Kroeger,Girish Sastry,Rebecca Kagan,Adrian Weller,Adrian Weller,Brian Tse,Elizabeth A. Barnes,Allan Dafoe,Paul Scharre,Ariel Herbert-Voss,Martijn Rasser,Shagun Sodhani,Carrick Flynn,Thomas Krendl Gilbert,Lisa Dyer,Saif Khan,Yoshua Bengio,Markus Anderljung +60 more
TL;DR: This report suggests various steps that different stakeholders can take to improve the verifiability of claims made about AI systems and their associated development processes, with a focus on providing evidence about the safety, security, fairness, and privacy protection of AI systems.
Posted Content
Trial without Error: Towards Safe Reinforcement Learning via Human Intervention
TL;DR: This work formalizes human intervention for RL and shows how to reduce the human labor required by training a supervised learner to imitate the human's intervention decisions, and outlines extensions of the scheme that are necessary if the authors are to train model-free agents without a single catastrophe.