S
Sebastian Goodman
Researcher at Google
Publications - 13
Citations - 6320
Sebastian Goodman is an academic researcher from Google. The author has contributed to research in topics: Automatic summarization & Computer science. The author has an hindex of 5, co-authored 9 publications receiving 3259 citations.
Papers
More filters
Proceedings Article
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
TL;DR: This work presents two parameter-reduction techniques to lower memory consumption and increase the training speed of BERT, and uses a self-supervised loss that focuses on modeling inter-sentence coherence.
Posted Content
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
TL;DR: The authors proposed a self-supervised loss that focuses on modeling inter-sentence coherence, and showed it consistently helps downstream tasks with multientence inputs, achieving state-of-the-art results on the GLUE, RACE, and \squad benchmarks.
Proceedings ArticleDOI
Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning
TL;DR: The Conceptual Captions dataset as discussed by the authors contains an order of magnitude more images than the MS-COCO dataset and represents a wider variety of both images and image caption styles.
Proceedings ArticleDOI
PaLI: A Jointly-Scaled Multilingual Language-Image Model
Xi Chen,Xiao Jing Wang,Soravit Changpinyo,AJ Piergiovanni,Piotr Padlewski,Daniel M Salz,Sebastian Goodman,Adam Grycner,Basil Mustafa,Lucas Beyer,Alexander Kolesnikov,Joan Puigcerver,Nan Ding,Keran Rong,Hassan Akbari,Gaurav Mishra,Linting Xue,Ashish V. Thapliyal,James Bradbury,Weicheng Kuo,Mojtaba Seyedhosseini,Chao Jia,Burcu Karagol Ayan,Carlos Riquelme,Andreas Steiner,Anelia Angelova,Xiaohua Zhai,Neil Houlsby,Radu Soricut +28 more
TL;DR: PaLI achieves state-of-the-art in multiple vision and language tasks, while retaining a simple, modular, and scalable design.
Journal ArticleDOI
Scaling Up Models and Data with t5x and seqio
Adam Roberts,Hyung Won Chung,Anselm Levskaya,Gaurav Mishra,James Bradbury,Daniel Andor,Sharan Narang,Brian Lester,Colin Gaffney,Afroz Mohiuddin,Curtis Hawthorne,Aitor Lewkowycz,Alexandru D. Sălcianu,M. van Zee,Jacob Austin,Sebastian Goodman,Livio Soares,Haitang Hu,Sasha Tsvyashchenko,Aakanksha Chowdhery,Jasmijn Bastings,Jannis Bulian,Xavier Garcia,Jianmo Ni,A. Chen,Kathleen Kenealy,Jonathan H. Clark,Stephan G. Lee,Daniel H Garrette,James P. Lee-Thorp,Colin Raffel,Noam Shazeer,Marvin Ritter,Maarten Bosma,Alexandre Passos,Jeremy Maitin-Shepard,Noah Fiedel,Mark Omernick,Brennan Saeta,Ryan Sepassi,Alexander Spiridonov,Joshua Newlan,Andrea Gesmundo +42 more
TL;DR: Two software libraries are presented: t5x simplifies the process of building and training large language models at scale while maintaining ease of use, and seqio provides a task-based API for simple creation of fast and reproducible training data and evaluation pipelines.