scispace - formally typeset
G

Geoffrey J. Gordon

Researcher at Carnegie Mellon University

Publications -  176
Citations -  15435

Geoffrey J. Gordon is an academic researcher from Carnegie Mellon University. The author has contributed to research in topics: Markov decision process & Artificial neural network. The author has an hindex of 54, co-authored 176 publications receiving 13660 citations. Previous affiliations of Geoffrey J. Gordon include Microsoft & University of Pittsburgh.

Papers
More filters
Proceedings Article

A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning

TL;DR: In this article, a no-regret algorithm is proposed to find a policy with good performance under the distribution of observations it induces in such sequential settings, which can be seen as a no regret algorithm in an online learning setting.
Proceedings ArticleDOI

Relational learning via collective matrix factorization

TL;DR: This model generalizes several existing matrix factorization methods, and therefore yields new large-scale optimization algorithms for these problems, which can handle any pairwise relational schema and a wide variety of error models.
Posted Content

A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning

TL;DR: In this article, a no-regret algorithm is proposed to train a stationary deterministic policy with good performance under the distribution of observations it induces in such sequential settings, and it outperforms previous approaches on two challenging imitation learning problems and a benchmark sequence labeling problem.
Proceedings Article

ARA*: Anytime A* with Provable Bounds on Sub-Optimality

TL;DR: An anytime heuristic search, ARA*, is proposed, which tunes its performance bound based on available search time, and starts by finding a suboptimal solution quickly using a loose bound, then tightens the bound progressively as time allows.
Book ChapterDOI

Stable function approximation in dynamic programming

TL;DR: A proof of convergence is provided for a wide class of temporal difference methods involving function approximators such as k-nearest-neighbor, and it is shown experimentally that these methods can be useful.