scispace - formally typeset
O

Ofir Nachum

Researcher at Google

Publications -  97
Citations -  5554

Ofir Nachum is an academic researcher from Google. The author has contributed to research in topics: Reinforcement learning & Computer science. The author has an hindex of 27, co-authored 76 publications receiving 3203 citations. Previous affiliations of Ofir Nachum include Massachusetts Institute of Technology.

Papers
More filters
Journal Article

D4RL: Datasets for Deep Data-Driven Reinforcement Learning

TL;DR: This work introduces benchmarks specifically designed for the offline setting, guided by key properties of datasets relevant to real-world applications of offline RL, and releases benchmark tasks and datasets with a comprehensive evaluation of existing algorithms and an evaluation protocol together with an open-source codebase.
Proceedings Article

Data-Efficient Hierarchical Reinforcement Learning

TL;DR: This paper studies how to develop HRL algorithms that are general, in that they do not make onerous additional assumptions beyond standard RL algorithms, and efficient, in the sense that they can be used with modest numbers of interaction samples, making them suitable for real-world problems such as robotic control.
Posted Content

Behavior Regularized Offline Reinforcement Learning

TL;DR: A general framework, behavior regularized actor critic (BRAC), is introduced to empirically evaluate recently proposed methods as well as a number of simple baselines across a variety of offline continuous control tasks.
Proceedings ArticleDOI

MorphNet: Fast & Simple Resource-Constrained Structure Learning of Deep Networks

TL;DR: MorphNet iteratively shrinks and expands a network, shrinking via a resource-weighted sparsifying regularizer on activations and expanding via a uniform multiplicative factor on all layers, which is scalable to large networks, adaptable to specific resource constraints, and capable of increasing the network's performance.
Proceedings Article

Bridging the Gap Between Value and Policy Based Reinforcement Learning

TL;DR: A new RL algorithm, Path Consistency Learning (PCL), is developed that minimizes a notion of soft consistency error along multi-step action sequences extracted from both on- and off-policy traces and significantly outperforms strong actor-critic and Q-learning baselines across several benchmarks.