K
Kevin Robinson
Publications - 3
Citations - 1503
Kevin Robinson is an academic researcher. The author has contributed to research in topics: Computer science. The author has an hindex of 2, co-authored 3 publications receiving 1503 citations.
Papers
More filters
Journal Article
PaLM: Scaling Language Modeling with Pathways
Aakanksha Chowdhery,Sharan Narang,Jacob Devlin,Maarten Bosma,Gaurav Mishra,Adam Roberts,Paul Barham,Hyung Won Chung,Charles Sutton,Sebastian Gehrmann,Parker Schuh,Kensen Shi,Sasha Tsvyashchenko,Joshua Maynez,Abhishek Rao,Parker Barnes,Yi Tay,Noam Shazeer,Velu Prabhakaran,Emily Reif,Nan Du,B. C. Hutchinson,Reiner Pope,James Bradbury,Jacob Austin,Michael Isard,Guy Gur-Ari,Peng Yin,Toju Duke,Anselm Levskaya,Sanjay Ghemawat,Sunipa Dev,Henryk Michalewski,Xavier Garcia,Vedant Misra,Kevin Robinson,L Fedus,Denny Zhou,Daphne Ippolito,David Luan,Hyeontaek Lim,Barret Zoph,Alexander Spiridonov,Ryan Sepassi,David Dohan,Shivani Agrawal,Mark Omernick,Andrew M. Dai,Thanumalayan Sankaranarayana Pillai,Marie Pellat,Aitor Lewkowycz,Erica Oliveira Moreira,Rewon Child,Oleksandr Polozov,Katherine Lee,Zong Tuan Zhou,Xuezhi Wang,Brennan Saeta,Mark Díaz,Orhan Firat,M. Catasta,Jason Loh Seong Wei,Kathleen S. Meier-Hellstern,Douglas Eck,Jeffrey Dean,Slav Petrov,Noah Fiedel +66 more
TL;DR: A 540-billion parameter, densely activated, Transformer language model, which is called PaLM achieves breakthrough performance, outperforming the state-of-the-art on a suite of multi-step reasoning tasks, and outperforming average human performance on the recently released BIG-bench benchmark.
Journal ArticleDOI
PaLM 2 Technical Report
Rohan Anil,Andrew M. Dai,Orhan Firat,Melvin George Johnson,Dmitry Lepikhin,Alexandre Passos,Siamak Shakeri,Emanuel Taropa,Paige Bailey,Zhi Chen,Eric Chu,Jonathan H. Clark,Laurent El Shafey,Yanping Huang,Kathleen S. Meier-Hellstern,Gaurav Mishra,Erica Oliveira Moreira,Mark Omernick,Kevin Robinson,Sebastian Ruder,Yi Pei. Tay,Kefan Xiao,Yuanzhong Xu,Yujing Zhang,Gustavo Hernandez-Abrego,Junwhan Ahn,Jacob Austin,Paul Barham,Jan A. Botha,James Bradbury,Siddhartha Brahma,Kevin Michael Brooks,M. Catasta,Yongzhou Cheng,Colin Cherry,Christopher A. Choquette-Choo,Aakanksha Chowdhery,C Crepy,Shachi Dave,Mostafa Dehghani,Sunipa Dev,Jacob Devlin,M. D'iaz,Nan Du,Ethan Dyer,Vladimir Feinberg,Fan Feng,Markus Freitag,Xavier Garcia,Sebastian Gehrmann,Guy Gur-Ari,Steven Hand,Hadi Hashemi,Le Hou,Joshua Howland,Anren Hu,Jeffrey Hui,Jeremy Scott Hurwitz,Michael Isard,Abe Ittycheriah,Matthew Jagielski,Wenhao Jia,Kathleen Kenealy,Maxim Krikun,Sneha Kudugunta,Katherine Lee,Benjamin N. Lee,Eric Li,Mu Li-Li,Wei Li,Yaguang Li,Jian Li,Hyeontaek Lim,Han Lin,Zhong-Zhong Liu,Frederick Liu,Marcello Maggioni,Aroma Mahendru,Joshua Maynez,Vedant Misra,Maysam Moussalem,Zachary Nado,John Nham,Eric Ni,Andrew Nystrom,Alicia Parrish,Marie Pellat,Martin Polacek,Alex Polozov,Reiner Pope,Siyuan Qiao,Emily Reif,Parker Riley,Alexandra Ros,Aurko Roy,Brennan Saeta,Rajkumar Samuel,Renee Shelby,Ambrose Jay Slone,Daniel Smilkov,David R. So,Daniela Sohn,Simon Tokumine,Vijay K. Vasudevan,Kiran Vodrahalli,Xuezhi Wang,Pidong Wang,Tao Wang,John Wieting,Yuhuai Wu,Ke Xu,Yu Yu Xu,Lin Wu Xue,Pengcheng Yin,Jia Yu,Biao Zhang,Steven X.F. Zheng,Ce Zheng,Wei Zhou,Denny Zhou,Slav Petrov,Yonghui Wu +121 more
TL;DR: The PaLM 2 model as mentioned in this paper is a Transformer-based model trained using a mixture of objectives, which has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM.
Journal ArticleDOI
A Pretrainer's Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity
Shayne Longpre,Gregory Yauney,Emily Reif,Katherine Lee,Adam Roberts,Barret Zoph,Denny Zhou,Jason Loh Seong Wei,Kevin Robinson,David Mimno,Daphne Ippolito +10 more
TL;DR: The authors show that a temporal shift between evaluation data and pretraining data leads to performance degradation, which is not overcome by finetuning, and explore the effect of quality and toxicity filters, showing a tradeoff between performance on standard benchmarks and risk of toxic generations.