scispace - formally typeset
Open AccessPosted Content

Approximating the Stationary Hamilton-Jacobi-Bellman Equation by Hierarchical Tensor Products

Reads0
Chats0
TLDR
This work treats infinite horizon optimal control problems by solving the associated stationary Hamilton-Jacobi-Bellman (HJB) equation numerically to compute the value function and an optimal feedback law, and uses low rank hierarchical tensor product approximation/tree-based tensor formats, in particular tensor trains (TT tensors), and multi-polynomials, together with high dimensional quadrature.
Abstract
We treat infinite horizon optimal control problems by solving the associated stationary Hamilton-Jacobi-Bellman (HJB) equation numerically, for computing the value function and an optimal feedback area law. The dynamical systems under consideration are spatial discretizations of nonlinear parabolic partial differential equations (PDE), which means that the HJB is suffering from the curse of dimensions. To overcome numerical infeasability we use low-rank hierarchical tensor product approximation, or tree-based tensor formats, in particular tensor trains (TT tensors) and multi-polynomials, since the resulting value function is expected to be smooth. To this end we reformulate the Policy Iteration algorithm as a linearization of HJB equations. The resulting linear hyperbolic PDE remains the computational bottleneck due to high-dimensions. By the methods of characteristics it can be reformulated via the Koopman operator in the spirit of dynamic programming. We use a low rank tensor representation for approximation of the value function. The resulting operator equation is solved using high-dimensional quadrature, e.g. Variational Monte-Carlo methods. From the knowledge of the value function at computable samples $x_i$ we infer the function $ x \mapsto v (x)$. We investigate the convergence of this procedure. By controlling destabilized versions of viscous Burgers and Schloegl equations numerical evidences are given.

read more

Citations
More filters

Solving high-dimensional Hamilton–Jacobi–Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space

TL;DR: The potential of iterative diffusion optimisation techniques is investigated, in particular considering applications in importance sampling and rare event simulation, and focusing on problems without diffusion control, with linearly controlled drift and running costs that depend quadratically on the control.
Posted Content

Solving high-dimensional Hamilton-Jacobi-Bellman PDEs using neural networks: perspectives from the theory of controlled diffusions and measures on path space

TL;DR: In this paper, the authors investigated the potential of iterative diffusion optimisation techniques, in particular considering applications in importance sampling and rare event simulation, and developed a principled framework based on divergences between path measures.
Posted Content

Tensor Decomposition Methods for High-dimensional Hamilton-Jacobi-Bellman Equations

TL;DR: A tensor decomposition approach for the solution of high-dimensional, fully nonlinear Hamilton-Jacobi-Bellman equations arising in optimal feedback control of nonlinear dynamics is presented in this article.
Posted Content

Approximative Policy Iteration for Exit Time Feedback Control Problems driven by Stochastic Differential Equations using Tensor Train format

TL;DR: In this paper, the authors considered a stochastic optimal exit time feedback control problem, where the Bellman equation is solved approximatively via the policy iteration algorithm on a polynomial ansatz space by a sequence of linear equations.
Posted Content

Actor-Critic Method for High Dimensional Static Hamilton-Jacobi-Bellman Partial Differential Equations based on Neural Networks.

TL;DR: In this paper, an actor-critic framework inspired by reinforcement learning is proposed for high-dimensional elliptic partial differential equations (PDEs) with high dimensional value functions, where the authors employ a policy gradient approach to improve the control and derive a variance reduced least square temporal difference method (VR-LSTD) using stochastic calculus.
References
More filters
Book

Deep Learning

TL;DR: Deep learning as mentioned in this paper is a form of machine learning that enables computers to learn from experience and understand the world in terms of a hierarchy of concepts, and it is used in many applications such as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames.
Book

Dynamic Programming and Optimal Control

TL;DR: The leading and most up-to-date textbook on the far-ranging algorithmic methododogy of Dynamic Programming, which can be used for optimal control, Markovian decision problems, planning and sequential decision making under uncertainty, and discrete/combinatorial optimization.
Book

Support Vector Machines

TL;DR: This book explains the principles that make support vector machines (SVMs) a successful modelling and prediction tool for a variety of applications and provides a unique in-depth treatment of both fundamental and recent material on SVMs that so far has been scattered in the literature.
Book

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

TL;DR: In this paper, the main ideas on a model problem with continuous viscosity solutions of Hamilton-Jacobi equations are discussed. But the main idea of the main solutions is not discussed.
Related Papers (5)