N
Neil Burch
Researcher at University of Alberta
Publications - 62
Citations - 3913
Neil Burch is an academic researcher from University of Alberta. The author has contributed to research in topics: Perfect information & Game theory. The author has an hindex of 27, co-authored 58 publications receiving 3173 citations. Previous affiliations of Neil Burch include Google.
Papers
More filters
Proceedings ArticleDOI
Online implicit agent modelling
TL;DR: This work describes an end-to-end approach for building an implicit modelling agent, compute robust response strategies, show how to select strategies for the portfolio, and apply existing variance reduction and online learning techniques to dynamically adapt the agent's strategy to its opponent.
Proceedings Article
Block A*: database-driven search with applications in any-angle path-planning
TL;DR: A new type of database, the Local Distance Database (LDDB), that contains distances between boundary points of a local neighborhood that calculates the optimal path between start and goal locations given the local distances stored in the LDDB is introduced.
Proceedings Article
Solving checkers
Jonathan Schaeffer,Yngvi Björnsson,Neil Burch,Akihiro Kishimoto,M. M¨ uller,Robert Lake,Paul Lu,Steve Sutphen +7 more
Abstract: AI has had notable success in building high-performance game-playing programs to complete against the best human players. However, the availability of fast and plentiful machines with large memories and disks creates the possibility of solving a game. This has been done before for simple or relatively small games. In this paper, we present new ideas and algorithms for solving the game of checkers. Checkers is a popular game of skill with a search space of 1020 possible positions. This paper reports on our first result. One of the most challenging checkers openings has been solved-the White Doctor opening is a draw. Solving roughly 50 more openings will result in the game-theoretic value of checkers being determined.
Proceedings Article
A new algorithm for generating equilibria in massive zero-sum games
TL;DR: A new measure of game complexity that links existing state-of-the-art algorithms for computing approximate equilibria to a more human measure, which considers the range of skill in a game, i.e. how many different skill levels exist.
Journal ArticleDOI
Mastering the game of Stratego with model-free multiagent reinforcement learning
Julien Perolat,Bart De Vylder,Daniel Hennes,Eugene V. Tarassov,F Strub,Vincent de Boer,Paul Muller,Jerome T. Connor,Neil Burch,Thomas Anthony,Stephen McAleer,Romuald Elie,Sarah Cen,Zhe Wang,Audrunas Gruslys,Aleksandra Malysheva,Mina Khan,Sherjil Ozair,Finbarr Timbers,Toby Pohlen,Tom Eccles,Mark Rowland,Marc Lanctot,Jean-Baptiste Lespiau,Bilal Piot,Shayegan Omidshafiei,Edward Paul Lockhart,Laurent Sifre,Nathalie Beauguerlange,Rémi Munos,David Silver,Satinder Singh,Demis Hassabis,Karl Tuyls +33 more
TL;DR: DeepNash as discussed by the authors uses a game-theoretic, model-free deep reinforcement learning method, without search, that learns to master Stratego through self-play from scratch.