scispace - formally typeset
Open AccessProceedings Article

Successor Features for Transfer in Reinforcement Learning

Reads0
Chats0
TLDR
This work proposes a transfer framework for the scenario where the reward function changes between tasks but the environment's dynamics remain the same, and derives two theorems that set the approach in firm theoretical ground and presents experiments that show that it successfully promotes transfer in practice.
Abstract
Transfer in reinforcement learning refers to the notion that generalization should occur not only within a task but also across tasks. We propose a transfer framework for the scenario where the reward function changes between tasks but the environment's dynamics remain the same. Our approach rests on two key ideas: "successor features", a value function representation that decouples the dynamics of the environment from the rewards, and "generalized policy improvement", a generalization of dynamic programming's policy improvement operation that considers a set of policies rather than a single one. Put together, the two ideas lead to an approach that integrates seamlessly within the reinforcement learning framework and allows the free exchange of information across tasks. The proposed method also provides performance guarantees for the transferred policy even before any learning has taken place. We derive two theorems that set our approach in firm theoretical ground and present experiments that show that it successfully promotes transfer in practice, significantly outperforming alternative methods in a sequence of navigation tasks and in the control of a simulated robotic arm.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Machine learning

TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Posted Content

Deep Reinforcement Learning: An Overview

Yuxi Li
- 25 Jan 2017 - 
TL;DR: This work discusses core RL elements, including value function, in particular, Deep Q-Network (DQN), policy, reward, model, planning, and exploration, and important mechanisms for RL, including attention and memory, unsupervised learning, transfer learning, multi-agent RL, hierarchical RL, and learning to learn.
Journal ArticleDOI

Reinforcement learning for control: Performance, stability, and deep approximators

TL;DR: This review mainly covers artificial-intelligence approaches to RL, from the viewpoint of the control engineer, and explains how approximate representations of the solution make RL feasible for problems with continuous states and control actions.
Proceedings ArticleDOI

Deep reinforcement learning with successor features for navigation across similar environments

TL;DR: In this paper, a successor-feature-based deep reinforcement learning algorithm is proposed to transfer navigation knowledge from previously mastered navigation tasks to new problem instances, which substantially decreases the required learning time after the first task instance has been solved, making it easily adaptable to changing environments.
Proceedings Article

NerveNet: Learning Structured Policy with Graph Neural Networks

TL;DR: NerveNet is proposed to explicitly model the structure of an agent, which naturally takes the form of a graph, and is demonstrated to be significantly more transferable and generalizable than policies learned by other models and are able to transfer even in a zero-shot setting.
Related Papers (5)