scispace - formally typeset
N

Nicolas Meuleau

Researcher at Ames Research Center

Publications -  31
Citations -  3052

Nicolas Meuleau is an academic researcher from Ames Research Center. The author has contributed to research in topics: Markov decision process & Partially observable Markov decision process. The author has an hindex of 21, co-authored 31 publications receiving 2983 citations. Previous affiliations of Nicolas Meuleau include Massachusetts Institute of Technology & Brown University.

Papers
More filters
Proceedings Article

Learning to cooperate via policy search

TL;DR: This paper provides a gradient-based distributed policy-search method for cooperative games and compares the notion of local optimum to that of Nash equilibrium, and demonstrates the effectiveness of this method experimentally in a small, partially observable simulated soccer domain.
Journal ArticleDOI

Model-based search for combinatorial optimization: A critical survey

TL;DR: This paper introduces model-based search as a unifying framework accommodating some recently proposed metaheuristics for combinatorial optimization such as ant colony optimization, stochastic gradient ascent, cross-entropy and estimation of distribution methods.
Posted Content

Learning to Cooperate via Policy Search

TL;DR: In this article, a gradient-based distributed policy search method for cooperative games is proposed and compared to the notion of local optimum to that of Nash equilibrium, which is a reasonable alternative to value-based methods for partially observable environments.
Posted Content

Hierarchical Solution of Markov Decision Processes using Macro-actions

TL;DR: In this article, a hierarchical MDP with macro-actions is proposed to reduce the size of the state space by treating macroactions as local policies that act in certain regions of state space, and restricting states in the abstract MDP to those at the boundaries of regions.
Proceedings Article

Hierarchical solution of Markov decision processes using macro-actions

TL;DR: A hierarchical model is proposed (using an abstract MDP) that works with macro-actions only, and that significantly reduces the size of the state space, and is shown to justify the computational overhead of macro-action generation.