H
Huaixiu Zheng
Publications - 5
Citations - 719
Huaixiu Zheng is an academic researcher. The author has contributed to research in topics: Computer science. The author has an hindex of 5, co-authored 5 publications receiving 719 citations.
Papers
More filters
Journal Article
LaMDA: Language Models for Dialog Applications
Romal Thoppilan,Daniel Adiwardana,Jamie Hall,Noam Shazeer,Apoorv Kulshreshtha,Heng-Tze Cheng,Alicia Jin,Taylor Bos,Leslie Baker,Yu Du,Yaguang Li,Hongrae Lee,Huaixiu Zheng,Amin Ghafouri,Marcelo Menegali,Yanping Huang,Maxim Krikun,Dmitry Lepikhin,James Qin,Dehao Chen,Yuanzhong Xu,Zhifeng Chen,Adam Roberts,Maarten Bosma,Yaoqi Zhou,Chung-Ching Chang,I. A. Krivokon,Willard J. Rusch,Marc Pickett,Kathleen S. Meier-Hellstern,Meredith Ringel Morris,Tulsee Doshi,Renelito Delos Santos,Toju Duke,Johnny Hartz Søraker,Bendert Zevenbergen,Velu Prabhakaran,Mark Díaz,Ben Hutchinson,Kristen Olson,Alejandra Aguirre Molina,Erin Hoffman-John,Josh Lee,Lora Aroyo,Ravindran Rajakumar,Alena Butryna,Matthew Lamm,V. O. Kuzmina,Joseph Fenton,Aaron Cohen,Rachel Bernstein,Raymond C. Kurzweil,Blaise Aguera-Arcas,Claire Cui,Marian Rogers Croak,Ed H. Chi,Quoc Hoai Le +56 more
TL;DR: The authors presented LaMDA: Language Models for Dialog Applications, a family of Transformer-based neural language models specialized for dialog, which have up to 137B parameters and are pre-trained on 1.56T words of public dialog data and web text and demonstrate that fine-tuning with annotated data and enabling the model to consult external knowledge sources can lead to significant improvements towards the two key challenges of safety and factual grounding.
Journal Article
Unifying Language Learning Paradigms
Yi Tay,Mostafa Dehghani,Vinh Q. Tran,Xavier Garcia,Dara Bahri,T. Schuster,Huaixiu Zheng,Neil Houlsby,Donald Metzler +8 more
TL;DR: UL2 achieves SOTA performance on 50 well-established supervised NLP tasks ranging from language generation, language understanding, text classification, question answering, commonsense reasoning, long text reasoning, structured knowledge grounding and information retrieval.
Proceedings ArticleDOI
HyperPrompt: Prompt-based Task-Conditioning of Transformers
Yun He,Huaixiu Zheng,Yi Pei. Tay,Jai Gupta,Yu Du,Vamsi Aribandi,Zhe Zhao,Yaguang Li,Zhaoji Chen,Donald Metzler,Heng-Tze Cheng,Ed H. Chi +11 more
TL;DR: Through extensive empirical experiments, it is demonstrated that HyperPrompt can achieve superior performances over strong T5 multi-task learning baselines and parameter-efficient adapter variants including Prompt-Tuning and HyperFormer++ on Natural Language Understanding benchmarks of GLUE and SuperGLUE across many model sizes.
Proceedings Article
UL2: Unifying Language Learning Paradigms
Yi Tay,Mostafa Dehghani,Vinh Q. Tran,Xavier Garcia,Jason Loh Seong Wei,Xuezhi Wang,Hyung Won Chung,Dara Bahri,T. Schuster,Huaixiu Zheng,Denny Zhou,Neil Houlsby,Donald Metzler +12 more
TL;DR: By scaling the model up to 20B parameters, this paper achieves SOTA performance on 50 well-established supervised NLP tasks ranging from language generation, language understanding, text classification, question answering, commonsense reasoning, long text reasoning, structured knowledge grounding and information retrieval.
Journal ArticleDOI
Transcending Scaling Laws with 0.1% Extra Compute
Yi Tay,Jason Loh Seong Wei,Hyung Won Chung,Vinh Q. Tran,David R. So,Siamak Shakeri,Xavier Garcia,Huaixiu Zheng,Jinfeng Rao,Aakanksha Chowdhery,Denny Zhou,Donald Metzler,Slav Petrov,Neil Houlsby,Quoc V. Le,Mostafa Dehghani +15 more
TL;DR: U-PaLM outperforms PaLM on many few-shot setups, i.e., English NLP tasks, reasoning tasks with chain-of-thought, multilingual tasks, MMLU and challenging BIG-Bench tasks, and is able to substantially improve the scaling properties of large language models on downstream metrics.