Y
Yutian Chen
Researcher at Google
Publications - 55
Citations - 9573
Yutian Chen is an academic researcher from Google. The author has contributed to research in topics: Computer science & Artificial neural network. The author has an hindex of 17, co-authored 45 publications receiving 6537 citations. Previous affiliations of Yutian Chen include University of California, Irvine & University of Cambridge.
Papers
More filters
Journal ArticleDOI
Mastering the game of Go without human knowledge
David Silver,Julian Schrittwieser,Karen Simonyan,Ioannis Antonoglou,Aja Huang,Arthur Guez,Thomas Hubert,Lucas Baker,Matthew Lai,Adrian Bolton,Yutian Chen,Timothy P. Lillicrap,Fan Hui,Laurent Sifre,George van den Driessche,Thore Graepel,Demis Hassabis +16 more
TL;DR: An algorithm based solely on reinforcement learning is introduced, without human data, guidance or domain knowledge beyond game rules, that achieves superhuman performance, winning 100–0 against the previously published, champion-defeating AlphaGo.
Journal ArticleDOI
A Generalist Agent
S Reed,Konrad Żołna,Emilio Parisotto,Sergio Gomez Colmenarejo,Alexander Novikov,Gabriel Barth-Maron,Mai Gimenez,Yury Sulsky,Jackie Kay,Jost Tobias Springenberg,Tom Eccles,Jake Bruce,Ali Razavi,Ashley D. Edwards,Nicolas Heess,Yutian Chen,Raia Hadsell,Oriol Vinyals,Mahyar Bordbar,Nando de Freitas +19 more
TL;DR: Gato as mentioned in this paper is a generalist agent that can play Atari, caption images, chat, stack blocks with a real robot arm and much more, deciding based on its context whether to output text, joint torques, button presses, or other tokens.
Proceedings Article
Learning to Learn without Gradient Descent by Gradient Descent
Yutian Chen,Matthew W. Hoffman,Sergio Gomez Colmenarejo,Misha Denil,Timothy P. Lillicrap,Matthew Botvinick,Nando de Freitas +6 more
TL;DR: It is shown that recurrent neural network optimizers trained on simple synthetic functions by gradient descent exhibit a remarkable degree of transfer in that they can be used to efficiently optimize a broad range of derivative-free black-box functions, including Gaussian process bandits, simple control objectives, global optimization benchmarks and hyper-parameter tuning tasks.
Proceedings Article
Austerity in MCMC Land: Cutting the Metropolis-Hastings Budget
TL;DR: In this paper, an approximate MH rule based on a sequential hypothesis test was proposed to accept or reject samples with high confidence using only a fraction of the data required for the exact MH rule.
Proceedings Article
Super-samples from kernel herding
TL;DR: The kernel herding algorithm as mentioned in this paper is an infinite memory deterministic process that learns to approximate a PDF with a collection of samples, and it has been shown that it decreases the error of expectations of functions in the Hilbert space at a rate O(1/T) which is much faster than the usual O(T) for iid random samples.