scispace - formally typeset
Open AccessBook

Markov Decision Processes: Discrete Stochastic Dynamic Programming

Reads0
Chats0
TLDR
Puterman as discussed by the authors provides a uniquely up-to-date, unified, and rigorous treatment of the theoretical, computational, and applied research on Markov decision process models, focusing primarily on infinite horizon discrete time models and models with discrete time spaces while also examining models with arbitrary state spaces, finite horizon models, and continuous time discrete state models.
Abstract
From the Publisher: The past decade has seen considerable theoretical and applied research on Markov decision processes, as well as the growing use of these models in ecology, economics, communications engineering, and other fields where outcomes are uncertain and sequential decision-making processes are needed. A timely response to this increased activity, Martin L. Puterman's new work provides a uniquely up-to-date, unified, and rigorous treatment of the theoretical, computational, and applied research on Markov decision process models. It discusses all major research directions in the field, highlights many significant applications of Markov decision processes models, and explores numerous important topics that have previously been neglected or given cursory coverage in the literature. Markov Decision Processes focuses primarily on infinite horizon discrete time models and models with discrete time spaces while also examining models with arbitrary state spaces, finite horizon models, and continuous-time discrete state models. The book is organized around optimality criteria, using a common framework centered on the optimality (Bellman) equation for presenting results. The results are presented in a "theorem-proof" format and elaborated on through both discussion and examples, including results that are not available in any other book. A two-state Markov decision process model, presented in Chapter 3, is analyzed repeatedly throughout the book and demonstrates many results and algorithms. Markov Decision Processes covers recent research advances in such areas as countable state space models with average reward criterion, constrained models, and models with risk sensitive optimality criteria. It also explores several topics that have received little or no attention in other books, including modified policy iteration, multichain models with average reward criterion, and sensitive optimality. In addition, a Bibliographic Remarks section in each chapter comments on relevant historic

read more

Citations
More filters
Journal ArticleDOI

Anti-Jamming Games in Multi-Channel Cognitive Radio Networks

TL;DR: This paper derives a channel hopping defense strategy using the Markov decision process approach with the assumption of perfect knowledge, and proposes two learning schemes for secondary users to gain knowledge of adversaries to handle cases without perfect knowledge.
Proceedings ArticleDOI

Age of information: Design and analysis of optimal scheduling algorithms

TL;DR: This paper forms a Markov decision process (MDP) to find dynamic transmission scheduling schemes, with the purpose of minimizing the long-run average age, and proposes both optimal off-line and online scheduling algorithms for the finite-approximate MDPs, depending on knowledge of time-varying arrivals.
Proceedings Article

Decision-Theoretic, High-Level Agent Programming in the Situation Calculus

TL;DR: The DTGologmodel allows one to partially specify a control program in a highlevel, logical language, and provides an interpreter that will determine the optimal completion of that program (viewed as a Markov decision process).
Journal ArticleDOI

Percentile Optimization for Markov Decision Processes with Parameter Uncertainty

TL;DR: A set of percentile criteria that are conceptually natural and representative of the trade-off between optimistic and pessimistic views of the question are presented and the use of these criteria under different forms of uncertainty for both the rewards and the transitions is studied.
Journal ArticleDOI

Modeling secrecy and deception in a multiple-period attacker–defender signaling game

TL;DR: This multiple-period model provides insights into the balance between capital and expense for defensive investments (and the effects of defender private information, such as defense effectiveness, target valuations, and costs), and shows that defenders can achieve more cost-effective security through secrecy and deception (possibly lasting more than one period), in a multiple- period game.
References
More filters
Book

Functional analysis

Walter Rudin
Book

Dynamic Programming

TL;DR: The more the authors study the information processing aspects of the mind, the more perplexed and impressed they become, and it will be a very long time before they understand these processes sufficiently to reproduce them.
Journal ArticleDOI

Finding Optimal (s, S) Policies Is About As Simple As Evaluating a Single Policy

TL;DR: A new algorithm for computing optimal ( s , S ) policies is derived based upon a number of new properties of the infinite horizon cost function c as well as a new upper bound for optimal order-up-to levels S * and a new lower bound for ideal reorder levels s *.
Related Papers (5)