K
Kathleen S. Meier-Hellstern
Researcher at Bell Labs
Publications - 7
Citations - 3042
Kathleen S. Meier-Hellstern is an academic researcher from Bell Labs. The author has contributed to research in topics: Computer science & Integrated Services Digital Network. The author has an hindex of 3, co-authored 4 publications receiving 949 citations.
Papers
More filters
Journal Article
PaLM: Scaling Language Modeling with Pathways
Aakanksha Chowdhery,Sharan Narang,Jacob Devlin,Maarten Bosma,Gaurav Mishra,Adam Roberts,Paul Barham,Hyung Won Chung,Charles Sutton,Sebastian Gehrmann,Parker Schuh,Kensen Shi,Sasha Tsvyashchenko,Joshua Maynez,Abhishek Rao,Parker Barnes,Yi Tay,Noam Shazeer,Velu Prabhakaran,Emily Reif,Nan Du,B. C. Hutchinson,Reiner Pope,James Bradbury,Jacob Austin,Michael Isard,Guy Gur-Ari,Peng Yin,Toju Duke,Anselm Levskaya,Sanjay Ghemawat,Sunipa Dev,Henryk Michalewski,Xavier Garcia,Vedant Misra,Kevin Robinson,L Fedus,Denny Zhou,Daphne Ippolito,David Luan,Hyeontaek Lim,Barret Zoph,Alexander Spiridonov,Ryan Sepassi,David Dohan,Shivani Agrawal,Mark Omernick,Andrew M. Dai,Thanumalayan Sankaranarayana Pillai,Marie Pellat,Aitor Lewkowycz,Erica Oliveira Moreira,Rewon Child,Oleksandr Polozov,Katherine Lee,Zong Tuan Zhou,Xuezhi Wang,Brennan Saeta,Mark Díaz,Orhan Firat,M. Catasta,Jason Loh Seong Wei,Kathleen S. Meier-Hellstern,Douglas Eck,Jeffrey Dean,Slav Petrov,Noah Fiedel +66 more
TL;DR: A 540-billion parameter, densely activated, Transformer language model, which is called PaLM achieves breakthrough performance, outperforming the state-of-the-art on a suite of multi-step reasoning tasks, and outperforming average human performance on the recently released BIG-bench benchmark.
Journal ArticleDOI
The Markov-modulated Poisson process (MMPP) cookbook
TL;DR: The purpose of this paper is to collect a number of useful results about Markov-modulated Poisson processes and queues with Markov -modulated input and to summary of recent developments.
Journal Article
LaMDA: Language Models for Dialog Applications
Romal Thoppilan,Daniel Adiwardana,Jamie Hall,Noam Shazeer,Apoorv Kulshreshtha,Heng-Tze Cheng,Alicia Jin,Taylor Bos,Leslie Baker,Yu Du,Yaguang Li,Hongrae Lee,Huaixiu Zheng,Amin Ghafouri,Marcelo Menegali,Yanping Huang,Maxim Krikun,Dmitry Lepikhin,James Qin,Dehao Chen,Yuanzhong Xu,Zhifeng Chen,Adam Roberts,Maarten Bosma,Yaoqi Zhou,Chung-Ching Chang,I. A. Krivokon,Willard J. Rusch,Marc Pickett,Kathleen S. Meier-Hellstern,Meredith Ringel Morris,Tulsee Doshi,Renelito Delos Santos,Toju Duke,Johnny Hartz Søraker,Bendert Zevenbergen,Velu Prabhakaran,Mark Díaz,Ben Hutchinson,Kristen Olson,Alejandra Aguirre Molina,Erin Hoffman-John,Josh Lee,Lora Aroyo,Ravindran Rajakumar,Alena Butryna,Matthew Lamm,V. O. Kuzmina,Joseph Fenton,Aaron Cohen,Rachel Bernstein,Raymond C. Kurzweil,Blaise Aguera-Arcas,Claire Cui,Marian Rogers Croak,Ed H. Chi,Quoc Hoai Le +56 more
TL;DR: The authors presented LaMDA: Language Models for Dialog Applications, a family of Transformer-based neural language models specialized for dialog, which have up to 137B parameters and are pre-trained on 1.56T words of public dialog data and web text and demonstrate that fine-tuning with annotated data and enabling the model to consult external knowledge sources can lead to significant improvements towards the two key challenges of safety and factual grounding.
Traffic models for isdn data users: office automation application
TL;DR: The results of a measurement/modeling study focused on an office automation application on an ISDN are described and the measurements are used in the construction of a canonical model of the packet arrival process generated by a single user.
Journal ArticleDOI
PaLM 2 Technical Report
Rohan Anil,Andrew M. Dai,Orhan Firat,Melvin George Johnson,Dmitry Lepikhin,Alexandre Passos,Siamak Shakeri,Emanuel Taropa,Paige Bailey,Zhi Chen,Eric Chu,Jonathan H. Clark,Laurent El Shafey,Yanping Huang,Kathleen S. Meier-Hellstern,Gaurav Mishra,Erica Oliveira Moreira,Mark Omernick,Kevin Robinson,Sebastian Ruder,Yi Pei. Tay,Kefan Xiao,Yuanzhong Xu,Yujing Zhang,Gustavo Hernandez-Abrego,Junwhan Ahn,Jacob Austin,Paul Barham,Jan A. Botha,James Bradbury,Siddhartha Brahma,Kevin Michael Brooks,M. Catasta,Yongzhou Cheng,Colin Cherry,Christopher A. Choquette-Choo,Aakanksha Chowdhery,C Crepy,Shachi Dave,Mostafa Dehghani,Sunipa Dev,Jacob Devlin,M. D'iaz,Nan Du,Ethan Dyer,Vladimir Feinberg,Fan Feng,Markus Freitag,Xavier Garcia,Sebastian Gehrmann,Guy Gur-Ari,Steven Hand,Hadi Hashemi,Le Hou,Joshua Howland,Anren Hu,Jeffrey Hui,Jeremy Scott Hurwitz,Michael Isard,Abe Ittycheriah,Matthew Jagielski,Wenhao Jia,Kathleen Kenealy,Maxim Krikun,Sneha Kudugunta,Katherine Lee,Benjamin N. Lee,Eric Li,Mu Li-Li,Wei Li,Yaguang Li,Jian Li,Hyeontaek Lim,Han Lin,Zhong-Zhong Liu,Frederick Liu,Marcello Maggioni,Aroma Mahendru,Joshua Maynez,Vedant Misra,Maysam Moussalem,Zachary Nado,John Nham,Eric Ni,Andrew Nystrom,Alicia Parrish,Marie Pellat,Martin Polacek,Alex Polozov,Reiner Pope,Siyuan Qiao,Emily Reif,Parker Riley,Alexandra Ros,Aurko Roy,Brennan Saeta,Rajkumar Samuel,Renee Shelby,Ambrose Jay Slone,Daniel Smilkov,David R. So,Daniela Sohn,Simon Tokumine,Vijay K. Vasudevan,Kiran Vodrahalli,Xuezhi Wang,Pidong Wang,Tao Wang,John Wieting,Yuhuai Wu,Ke Xu,Yu Yu Xu,Lin Wu Xue,Pengcheng Yin,Jia Yu,Biao Zhang,Steven X.F. Zheng,Ce Zheng,Wei Zhou,Denny Zhou,Slav Petrov,Yonghui Wu +121 more
TL;DR: The PaLM 2 model as mentioned in this paper is a Transformer-based model trained using a mixture of objectives, which has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM.