M
Melanie Subbiah
Publications - 8
Citations - 12032
Melanie Subbiah is an academic researcher. The author has contributed to research in topics: Computer science & Language model. The author has an hindex of 2, co-authored 2 publications receiving 3036 citations.
Papers
More filters
Proceedings Article
Language Models are Few-Shot Learners
Tom B. Brown,Benjamin Mann,Nick Ryder,Melanie Subbiah,Jared Kaplan,Prafulla Dhariwal,Arvind Neelakantan,Pranav Shyam,Girish Sastry,Amanda Askell,Sandhini Agarwal,Ariel Herbert-Voss,Gretchen Krueger,Thomas Henighan,Rewon Child,Aditya Ramesh,Daniel M. Ziegler,Jeffrey Wu,Clemens Winter,Christopher Hesse,Mark Chen,Eric Sigler,Mateusz Litwin,Scott Gray,Benjamin Chess,Jack Clark,Christopher Berner,Samuel McCandlish,Alec Radford,Ilya Sutskever,Dario Amodei +30 more
TL;DR: GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic.
Posted Content
Language Models are Few-Shot Learners
Tom B. Brown,Benjamin Mann,Nick Ryder,Melanie Subbiah,Jared Kaplan,Prafulla Dhariwal,Arvind Neelakantan,Pranav Shyam,Girish Sastry,Amanda Askell,Sandhini Agarwal,Ariel Herbert-Voss,Gretchen Krueger,Thomas Henighan,Rewon Child,Aditya Ramesh,Daniel M. Ziegler,Jeffrey Wu,Clemens Winter,Christopher Hesse,Mark Chen,Eric Sigler,Mateusz Litwin,Scott Gray,Benjamin Chess,Jack Clark,Christopher Berner,Samuel McCandlish,Alec Radford,Ilya Sutskever,Dario Amodei +30 more
TL;DR: This article showed that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches.
Proceedings ArticleDOI
SafeText: A Benchmark for Exploring Physical Safety in Language Models
Sharon Levy,Emily Allaway,Melanie Subbiah,Lydia B. Chilton,Desmond Upton Patton,Kathleen R. McKeown,William Yang Wang +6 more
TL;DR: It is argued that state-of-the-art large language models are susceptible to the generation of unsafe text and haveulty rejecting unsafe advice, and it is argued for further studies of safety and the assessment of commonsense physical safety in models before release.
Proceedings ArticleDOI
Mitigating Covertly Unsafe Text within Natural Language Systems
Alex Mei,Anisha Kabir,Sharon Levy,Melanie Subbiah,Emily Allaway,John Judge,Desmond Upton Patton,Bruce Bimber,Kathleen R. McKeown,William Yang Wang +9 more
TL;DR: This work distinguishes types of text that can lead to physical harm and establishes one particularly underexplored category: covertly unsafe text, which is further broken down with respect to the system’s information and discusses solutions to mitigate the generation of text in each of these subcategories.
Looking Under the Hood of DetectGPT
Benjamin Mann,Melanie Subbiah,Prafulla Dhariwal,Arvind Neelakantan,Pranav Shyam,Girish Sastry,Sandhini Agarwal,Maarten Bosma,Gaurav Mishra,Parker Schuh +9 more
TL;DR: This paper analyzed DetectGPT and showed that selectively masking a combination of nouns, verbs, and adjectives improves the AUROC metric by up to 9.5%, demonstrating the importance of targeted masking strategies.