scispace - formally typeset
Open AccessProceedings ArticleDOI

Socially aware motion planning with deep reinforcement learning

TLDR
Using deep reinforcement learning, this work develops a time-efficient navigation policy that respects common social norms and is shown to enable fully autonomous navigation of a robotic vehicle moving at human walking speed in an environment with many pedestrians.
Abstract
For robotic vehicles to navigate safely and efficiently in pedestrian-rich environments, it is important to model subtle human behaviors and navigation rules (e.g., passing on the right). However, while instinctive to humans, socially compliant navigation is still difficult to quantify due to the stochasticity in people's behaviors. Existing works are mostly focused on using feature-matching techniques to describe and imitate human paths, but often do not generalize well since the feature values can vary from person to person, and even run to run. This work notes that while it is challenging to directly specify the details of what to do (precise mechanisms of human navigation), it is straightforward to specify what not to do (violations of social norms). Specifically, using deep reinforcement learning, this work develops a time-efficient navigation policy that respects common social norms. The proposed method is shown to enable fully autonomous navigation of a robotic vehicle moving at human walking speed in an environment with many pedestrians.

read more

Citations
More filters
Journal ArticleDOI

Unmasking Clever Hans Predictors and Assessing What Machines Really Learn

TL;DR: The authors investigate how these methods approach learning in order to assess the dependability of their decision making and propose a semi-automated Spectral Relevance Analysis that provides a practically effective way of characterizing and validating the behavior of nonlinear learning machines.
Proceedings ArticleDOI

Motion Planning Among Dynamic, Decision-Making Agents with Deep Reinforcement Learning

TL;DR: This work extends the previous approach to develop an algorithm that learns collision avoidance among a variety of types of dynamic agents without assuming they follow any particular behavior rules, and introduces a strategy using LSTM that enables the algorithm to use observations of an arbitrary number of other agents, instead of previous methods that have a fixed observation size.
Journal ArticleDOI

Unmasking Clever Hans predictors and assessing what machines really learn.

TL;DR: In this article, the authors apply recent techniques for explaining decisions of state-of-the-art learning machines and analyze various tasks from computer vision and arcade games, and propose a semi-automated Spectral Relevance Analysis that provides a practically effective way of characterizing and validating the behavior of nonlinear learning machines.
Proceedings ArticleDOI

Crowd-Robot Interaction: Crowd-Aware Robot Navigation With Attention-Based Deep Reinforcement Learning

TL;DR: This work proposes to rethink pairwise interactions with a self-attention mechanism, and jointly model Human-Robot as well as Human-Human interactions in the deep reinforcement learning framework, and captures the Human- human interactions occurring in dense crowds that indirectly affects the robot’s anticipation capability.
Proceedings ArticleDOI

Safe Reinforcement Learning With Model Uncertainty Estimates

TL;DR: MC-Dropout and Bootstrapping are used to give computationally tractable and parallelizable uncertainty estimates and are embedded in a Safe Reinforcement Learning framework to form uncertainty-aware navigation around pedestrians, resulting in a collision avoidance policy that knows what it does not know and cautiously avoids pedestrians that exhibit unseen behavior.
References
More filters
Journal ArticleDOI

Human-level control through deep reinforcement learning

TL;DR: This work bridges the divide between high-dimensional sensory inputs and actions, resulting in the first artificial agent that is capable of learning to excel at a diverse array of challenging tasks.
Journal ArticleDOI

Mastering the game of Go with deep neural networks and tree search

TL;DR: Using this search algorithm, the program AlphaGo achieved a 99.8% winning rate against other Go programs, and defeated the human European Go champion by 5 games to 0.5, the first time that a computer program has defeated a human professional player in the full-sized game of Go.
Proceedings Article

Asynchronous methods for deep reinforcement learning

TL;DR: A conceptually simple and lightweight framework for deep reinforcement learning that uses asynchronous gradient descent for optimization of deep neural network controllers and shows that asynchronous actor-critic succeeds on a wide variety of continuous motor control problems as well as on a new task of navigating random 3D mazes using a visual input.
Journal ArticleDOI

Social Force Model for Pedestrian Dynamics

TL;DR: Computer simulations of crowds of interacting pedestrians show that the social force model is capable of describing the self-organization of several observed collective effects of pedestrian behavior very realistically.
Proceedings ArticleDOI

Apprenticeship learning via inverse reinforcement learning

TL;DR: This work thinks of the expert as trying to maximize a reward function that is expressible as a linear combination of known features, and gives an algorithm for learning the task demonstrated by the expert, based on using "inverse reinforcement learning" to try to recover the unknown reward function.
Related Papers (5)