scispace - formally typeset
Open AccessJournal ArticleDOI

A survey of multi-objective sequential decision-making

TLDR
This article surveys algorithms designed for sequential decision-making problems with multiple objectives and proposes a taxonomy that classifies multi-objective methods according to the applicable scenario, the nature of the scalarization function, and the type of policies considered.
Abstract
Sequential decision-making problems with multiple objectives arise naturally in practice and pose unique challenges for research in decision-theoretic planning and learning, which has largely focused on single-objective settings. This article surveys algorithms designed for sequential decision-making problems with multiple objectives. Though there is a growing body of literature on this subject, little of it makes explicit under what circumstances special methods are needed to solve multi-objective problems. Therefore, we identify three distinct scenarios in which converting such a problem to a single-objective one is impossible, infeasible, or undesirable. Furthermore, we propose a taxonomy that classifies multi-objective methods according to the applicable scenario, the nature of the scalarization function (which projects multi-objective values to scalar ones), and the type of policies considered. We show how these factors determine the nature of an optimal solution, which can be a single policy, a convex hull, or a Pareto front. Using this taxonomy, we survey the literature on multi-objective methods for planning and learning. Finally, we discuss key applications of such methods and outline opportunities for future work.

read more

Content maybe subject to copyright    Report

Citations
More filters
Book ChapterDOI

Multiple criteria decision making

TL;DR: In this Chapter, a decision maker (or a group of experts) trying to establish or examine fair procedures to combine opinions about alternatives related to different points of view is imagined.
Posted Content

Deep Reinforcement Learning: An Overview

Yuxi Li
- 25 Jan 2017 - 
TL;DR: This work discusses core RL elements, including value function, in particular, Deep Q-Network (DQN), policy, reward, model, planning, and exploration, and important mechanisms for RL, including attention and memory, unsupervised learning, transfer learning, multi-agent RL, hierarchical RL, and learning to learn.
Journal ArticleDOI

Deep Reinforcement Learning for Autonomous Driving: A Survey

TL;DR: This review summarises deep reinforcement learning algorithms, provides a taxonomy of automated driving tasks where (D)RL methods have been employed, highlights the key challenges algorithmically as well as in terms of deployment of real world autonomous driving agents, the role of simulators in training agents, and finally methods to evaluate, test and robustifying existing solutions in RL and imitation learning.
Posted Content

Multi-Task Learning as Multi-Objective Optimization

TL;DR: This paper cast multi-task learning as a multi-objective optimization problem, with the overall objective of finding a Pareto optimal solution, and propose an upper bound for the multiobjective loss and show that it can be optimized efficiently.
Posted Content

Challenges of Real-World Reinforcement Learning

TL;DR: A set of nine unique challenges that must be addressed to productionize RL to real world problems are presented and an example domain that has been modified to present these challenges as a testbed for practical RL research is presented.
References
More filters
Proceedings ArticleDOI

Evolving policies for multi-reward partially observable markov decision processes (MR-POMDPs)

TL;DR: This work contributes the multi-reward partially-observable Markov decision process (MR-POMDP) as a general modelling framework and presents two hybrid (memetic) multi-objective evolutionary algorithms that generate non-dominated sets of policies (in the form of stochastic finite state controllers).
Book ChapterDOI

A New Distributed Reinforcement Learning Algorithm for Multiple Objective Optimization Problems

TL;DR: A non-dominant criterion is used to construct Pareto fronts and by delaying adjustments on the rewards MDQL achieves better distributions of solutions and an extension for applying reinforcement learning to continuous functions is given.
Journal ArticleDOI

GIS and Intelligent Agents for Multiobjective Natural Resource Allocation: A Reinforcement Learning Approach

TL;DR: The model integrates agent-based modeling in a GIS environment with reinforcement learning – a heuristic method for generating, evaluating, and improving multiobjective decision making solutions and is validated to demonstrate that it can provide practical solutions to natural resource decision making.
Proceedings ArticleDOI

EDA-RL: estimation of distribution algorithms for reinforcement learning problems

TL;DR: Conditional Random Fields by Lafferty et al. is newly introduced into EDAs in this paper, which are extended to solve reinforcement learning problems which arise naturally in a framework for autonomous agents.
Journal ArticleDOI

Multiobjective dynamic programming for forest resource management

TL;DR: The MODP method is introduced in this paper to take the two issues of forest resource management problems into account simultaneously, i.e. multiple objectives and uncertainty.
Related Papers (5)