D
Dustin Morrill
Researcher at University of Alberta
Publications - 21
Citations - 1293
Dustin Morrill is an academic researcher from University of Alberta. The author has contributed to research in topics: Reinforcement learning & Regret. The author has an hindex of 9, co-authored 19 publications receiving 933 citations.
Papers
More filters
Journal ArticleDOI
DeepStack: Expert-level artificial intelligence in heads-up no-limit poker
Matej Moravčík,Matej Moravčík,Martin Schmid,Martin Schmid,Neil Burch,Viliam Lisý,Viliam Lisý,Dustin Morrill,Nolan Bard,Trevor Davis,Kevin Waugh,Michael Johanson,Michael Bowling +12 more
TL;DR: DeepStack is introduced, an algorithm for imperfect-information settings that combines recursive reasoning to handle information asymmetry, decomposition to focus computation on the relevant decision, and a form of intuition that is automatically learned from self-play using deep learning.
Journal ArticleDOI
DeepStack: Expert-Level Artificial Intelligence in No-Limit Poker
Matej Moravcik,Martin Schmid,Neil Burch,Viliam Lisý,Dustin Morrill,Nolan Bard,Trevor Davis,Kevin Waugh,Michael Johanson,Michael Bowling +9 more
TL;DR: DeepStack as discussed by the authors combines recursive reasoning to handle information asymmetry, decomposition to focus computation on the relevant decision, and a form of intuition that is automatically learned from self-play using deep learning.
Posted Content
OpenSpiel: A Framework for Reinforcement Learning in Games.
Marc Lanctot,Edward Lockhart,Jean-Baptiste Lespiau,Vinicius Zambaldi,Satyaki Upadhyay,Julien Perolat,Sriram Srinivasan,Finbarr Timbers,Karl Tuyls,Shayegan Omidshafiei,Daniel Hennes,Dustin Morrill,Paul Muller,Timo Ewalds,Ryan Faulkner,János Kramár,Bart De Vylder,Brennan Saeta,James Bradbury,David Ding,Sebastian Borgeaud,Matthew Lai,Julian Schrittwieser,Thomas Anthony,Edward Hughes,Ivo Danihelka,Jonah Ryan-Davis +26 more
TL;DR: This document serves both as an overview of the code base and an introduction to the terminology, core concepts, and algorithms across the fields of reinforcement learning, computational game theory, and search.
Proceedings Article
Solving games with functional regret estimation
TL;DR: This paper proposed an online learning method for minimizing regret in large extensive-form games, which learns a function approximator online to estimate the regret for choosing a particular action and uses these estimates in place of the true regrets to define a sequence of policies.
Proceedings ArticleDOI
Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent
Edward Lockhart,Marc Lanctot,Julien Perolat,Jean-Baptiste Lespiau,Dustin Morrill,Finbarr Timbers,Karl Tuyls +6 more
TL;DR: The exploitability descent algorithm is presented, a new algorithm to compute approximate equilibria in two-player zero-sum extensive-form games with imperfect information, by direct policy optimization against worst-case opponents, and it is proved that when following this optimization, the exploitability of a player's strategy converges asymptotically to zero.