scispace - formally typeset
Open AccessBook ChapterDOI

Vulnerability of Deep Reinforcement Learning to Policy Induction Attacks

TLDR
In this article, the transferability of adversarial examples is verified across different DQN models, and a novel class of attacks based on this vulnerability is presented to enable policy manipulation and induction in the learning process of DQNs.
Abstract
Deep learning classifiers are known to be inherently vulnerable to manipulation by intentionally perturbed inputs, named adversarial examples. In this work, we establish that reinforcement learning techniques based on Deep Q-Networks (DQNs) are also vulnerable to adversarial input perturbations, and verify the transferability of adversarial examples across different DQN models. Furthermore, we present a novel class of attacks based on this vulnerability that enable policy manipulation and induction in the learning process of DQNs. We propose an attack mechanism that exploits the transferability of adversarial examples to implement policy induction attacks on DQNs, and demonstrate its efficacy and impact through experimental study of a game-learning scenario.

read more

Citations
More filters
Journal ArticleDOI

Deep Learning in Mobile and Wireless Networking: A Survey

TL;DR: This paper bridges the gap between deep learning and mobile and wireless networking research, by presenting a comprehensive survey of the crossovers between the two areas, and provides an encyclopedic review of mobile and Wireless networking research based on deep learning, which is categorize by different domains.
Proceedings ArticleDOI

Audio Adversarial Examples: Targeted Attacks on Speech-to-Text

TL;DR: A white-box iterative optimization-based attack to Mozilla's implementation DeepSpeech end-to-end has a 100% success rate, and the feasibility of this attack introduce a new domain to study adversarial examples.
Posted Content

The Space of Transferable Adversarial Examples

TL;DR: It is found that adversarial examples span a contiguous subspace of large (~25) dimensionality, which indicates that it may be possible to design defenses against transfer-based attacks, even for models that are vulnerable to direct attacks.
Posted Content

Adversarially Robust Generalization Requires More Data

TL;DR: In this paper, the authors study adversarially robust learning from the viewpoint of generalization and show that the sample complexity of robust learning can be significantly larger than that of "standard" learning.
Journal ArticleDOI

Adversarial Attacks and Defenses in Deep Learning

TL;DR: The theoretical foundations, algorithms, and applications of adversarial attack techniques are introduced and a few research efforts on the defense techniques are described, which cover the broad frontier in the field.
References
More filters
Journal ArticleDOI

Human-level control through deep reinforcement learning

TL;DR: This work bridges the divide between high-dimensional sensory inputs and actions, resulting in the first artificial agent that is capable of learning to excel at a diverse array of challenging tasks.
Proceedings Article

Intriguing properties of neural networks

TL;DR: It is found that there is no distinction between individual highlevel units and random linear combinations of high level units, according to various methods of unit analysis, and it is suggested that it is the space, rather than the individual units, that contains of the semantic information in the high layers of neural networks.
Posted Content

Playing Atari with Deep Reinforcement Learning

TL;DR: This work presents the first deep learning model to successfully learn control policies directly from high-dimensional sensory input using reinforcement learning, which outperforms all previous approaches on six of the games and surpasses a human expert on three of them.
Proceedings Article

Explaining and Harnessing Adversarial Examples

TL;DR: It is argued that the primary cause of neural networks' vulnerability to adversarial perturbation is their linear nature, supported by new quantitative results while giving the first explanation of the most intriguing fact about them: their generalization across architectures and training sets.
Related Papers (5)