A video game description language for model-based or interactive learning
Tom Schaul
- pp 1-8
TLDR
It is shown how to learn competent behaviors when a model of the game dynamics is available or when it is not, when full state information is given to the agent or just subjective observations, when learning is interactive or in batch-mode, and for a number of different learning algorithms, including reinforcement learning and evolutionary search.Abstract:
We propose a powerful new tool for conducting research on computational intelligence and games. `PyVGDL' is a simple, high-level description language for 2D video games, and the accompanying software library permits parsing and instantly playing those games. The streamlined design of the language is based on defining locations and dynamics for simple building blocks, and the interaction effects when such objects collide, all of which are provided in a rich ontology. It can be used to quickly design games, without needing to deal with control structures, and the concise language is also accessible to generative approaches. We show how the dynamics of many classical games can be generated from a few lines of PyVGDL. The main objective of these generated games is to serve as diverse benchmark problems for learning and planning algorithms; so we provide a collection of interfaces for different types of learning agents, with visual or abstract observations, from a global or first-person viewpoint. To demonstrate the library's usefulness in a broad range of learning scenarios, we show how to learn competent behaviors when a model of the game dynamics is available or when it is not, when full state information is given to the agent or just subjective observations, when learning is interactive or in batch-mode, and for a number of different learning algorithms, including reinforcement learning and evolutionary search.read more
Citations
More filters
Posted Content
StarCraft II: A New Challenge for Reinforcement Learning
Oriol Vinyals,Timo Ewalds,Sergey Bartunov,Petko Georgiev,Alexander Vezhnevets,Michelle Yeo,Alireza Makhzani,Heinrich Küttler,John P. Agapiou,Julian Schrittwieser,John Quan,Stephen Gaffney,Stig Petersen,Karen Simonyan,Tom Schaul,Hado van Hasselt,David Silver,Timothy P. Lillicrap,Kevin Calderone,Paul Keet,Anthony Brunasso,David Lawrence,Anders Ekermo,Jacob Repp,Rodney Tsing +24 more
TL;DR: This paper introduces SC2LE (StarCraft II Learning Environment), a reinforcement learning environment based on the StarCraft II game that offers a new and challenging environment for exploring deep reinforcement learning algorithms and architectures and gives initial baseline results for neural networks trained from this data to predict game outcomes and player actions.
Book
Procedural Content Generation in Games
TL;DR: This book presents the most up-to-date coverage of procedural content generation (PCG) for games, specifically the procedural generation of levels, landscapes, items, rules, quests, or other types of content.
Proceedings Article
The Malmo platform for artificial intelligence experimentation
TL;DR: Project Malmo provides a sophisticated abstraction layer on top of Minecraft that supports a wide range of experimentation scenarios, ranging from navigation and survival to collaboration and problem solving tasks, to support openness and collaboration in AI research.
BookDOI
Artificial Intelligence and Games
TL;DR: This is the first textbook dedicated to explaining how artificial intelligence techniques can be used in and for games, and how to use AI to play games, to generate content for games and to model players.
Journal ArticleDOI
The 2014 General Video Game Playing Competition
Diego Perez-Liebana,Spyridon Samothrakis,Julian Togelius,Tom Schaul,Simon M. Lucas,Adrien Couëtoux,Jerry Lee,Chong-U Lim,Tommy Thompson +8 more
TL;DR: All controllers submitted to the first General Video Game Playing Competition are described, with an in-depth description of four of them by their authors, including the winner and the runner-up entries of the contest.
References
More filters
Journal ArticleDOI
Long short-term memory
TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Book
Reinforcement Learning: An Introduction
TL;DR: This book provides a clear and simple account of the key ideas and algorithms of reinforcement learning, which ranges from the history of the field's intellectual foundations to the most recent developments and applications.
Journal ArticleDOI
A Survey of Monte Carlo Tree Search Methods
Cameron Browne,Edward J. Powley,Daniel Whitehouse,Simon M. Lucas,Peter I. Cowling,Philipp Rohlfshagen,S. Tavener,Diego Perez,Spyridon Samothrakis,Simon Colton +9 more
TL;DR: A survey of the literature to date of Monte Carlo tree search, intended to provide a snapshot of the state of the art after the first five years of MCTS research, outlines the core algorithm's derivation, impart some structure on the many variations and enhancements that have been proposed, and summarizes the results from the key game and nongame domains.
Journal ArticleDOI
The arcade learning environment: an evaluation platform for general agents
TL;DR: The Arcade Learning Environment (ALE) as discussed by the authors is a platform for evaluating the development of general, domain-independent AI technology, which provides an interface to hundreds of Atari 2600 game environments, each one different, interesting, and designed to be a challenge for human players.
Journal ArticleDOI
Least-squares policy iteration
TL;DR: The new algorithm, least-squares policy iteration (LSPI), learns the state-action value function which allows for action selection without a model and for incremental policy improvement within a policy-iteration framework.