scispace - formally typeset
D

Dhiraj Gandhi

Researcher at Carnegie Mellon University

Publications -  25
Citations -  1656

Dhiraj Gandhi is an academic researcher from Carnegie Mellon University. The author has contributed to research in topics: Robot learning & Modular design. The author has an hindex of 15, co-authored 25 publications receiving 1013 citations. Previous affiliations of Dhiraj Gandhi include Facebook.

Papers
More filters
Proceedings Article

Learning to Explore using Active Neural SLAM

TL;DR: This work presents a modular and hierarchical approach to learn policies for exploring 3D environments, called `Active Neural SLAM', which leverages the strengths of both classical and learning-based methods, by using analytical path planners with learned SLAM module, and global and local policies.
Proceedings ArticleDOI

Learning to fly by crashing

TL;DR: This paper builds a drone whose sole purpose is to crash into objects: it samples naive trajectories and crashes into random objects to create one of the biggest UAV crash dataset.
Posted Content

Object Goal Navigation using Goal-Oriented Semantic Exploration

TL;DR: A modular system called, `Goal-Oriented Semantic Exploration' which builds an episodic semantic map and uses it to explore the environment efficiently based on the goal object category and outperforms a wide range of baselines including end-to-end learning-based methods as well as modular map- based methods.
Posted Content

Self-Supervised Exploration via Disagreement

TL;DR: In this article, an ensemble of dynamics models is used to incentivize the agent to explore such that the disagreement of those ensembles is maximized, which results in a sample-efficient exploration.
Proceedings Article

Self-Supervised Exploration via Disagreement

TL;DR: This paper proposes a formulation for exploration inspired by the work in active learning literature and trains an ensemble of dynamics models and incentivizes the agent to explore such that the disagreement of those ensembles is maximized, which results in a sample-efficient exploration.