scispace - formally typeset
L

Leslie Pack Kaelbling

Researcher at Massachusetts Institute of Technology

Publications -  375
Citations -  36830

Leslie Pack Kaelbling is an academic researcher from Massachusetts Institute of Technology. The author has contributed to research in topics: Motion planning & Markov decision process. The author has an hindex of 65, co-authored 354 publications receiving 32988 citations. Previous affiliations of Leslie Pack Kaelbling include Aalborg University & Artificial Intelligence Center.

Papers
More filters
Journal ArticleDOI

Reinforcement learning: a survey

TL;DR: Central issues of reinforcement learning are discussed, including trading off exploration and exploitation, establishing the foundations of the field via Markov decision theory, learning from delayed reinforcement, constructing empirical models to accelerate learning, making use of generalization and hierarchy, and coping with hidden state.
Posted Content

Reinforcement Learning: A Survey

TL;DR: A survey of reinforcement learning from a computer science perspective can be found in this article, where the authors discuss the central issues of RL, including trading off exploration and exploitation, establishing the foundations of RL via Markov decision theory, learning from delayed reinforcement, constructing empirical models to accelerate learning, making use of generalization and hierarchy, and coping with hidden state.
Journal ArticleDOI

Planning and Acting in Partially Observable Stochastic Domains

TL;DR: A novel algorithm for solving pomdps off line and how, in some cases, a finite-memory controller can be extracted from the solution to a POMDP is outlined.
Proceedings Article

Acting Optimally in Partially Observable Stochastic Domains

TL;DR: The existing algorithms for computing optimal control strategies for partially observable stochastic environments are found to be highly computationally inefficient and a new algorithm is developed that is empirically more efficient.
Book

Learning in Embedded Systems

TL;DR: This dissertation addresses the problem of designing algorithms for learning in embedded systems using Sutton's techniques for linear association and reinforcement comparison, while the interval estimation algorithm uses the statistical notion of confidence intervals to guide its generation of actions.