scispace - formally typeset
Open AccessJournal ArticleDOI

A Survey of Deep Learning Methods for Cyber Security

TLDR
This survey paper describes a literature review of deep learning methods for cyber security applications, including deep autoencoders, restricted Boltzmann machines, recurrent neural networks, generative adversarial networks, and several others.
Abstract
This survey paper describes a literature review of deep learning (DL) methods for cyber security applications. A short tutorial-style description of each DL method is provided, including deep autoencoders, restricted Boltzmann machines, recurrent neural networks, generative adversarial networks, and several others. Then we discuss how each of the DL methods is used for security applications. We cover a broad array of attack types including malware, spam, insider threats, network intrusions, false data injection, and malicious domain names used by botnets.

read more

Citations
More filters
Journal ArticleDOI

Deep learning for cyber security intrusion detection: Approaches, datasets, and comparative study

TL;DR: A survey of deep learning approaches for cyber security intrusion detection, the datasets used, and a comparative study to evaluate the efficiency of several methods are presented.
Journal ArticleDOI

Cybersecurity data science: an overview from machine learning perspective

TL;DR: This paper focuses and briefly discusses on cybersecurity data science, where the data is being gathered from relevant cybersecurity sources, and the analytics complement the latest data-driven patterns for providing more effective security solutions.
Journal ArticleDOI

Comprehensive Review of Artificial Neural Network Applications to Pattern Recognition

TL;DR: There is a need for state-of-the-art in neural networks application to PR to urgently address the above-highlights problems and the research focus on current models and the development of new models concurrently for more successes in the field.
Journal ArticleDOI

Scientific Machine Learning Through Physics–Informed Neural Networks: Where we are and What’s Next

TL;DR: A comprehensive review of the literature on physics-informed neural networks can be found in this article , where the primary goal of the study was to characterize these networks and their related advantages and disadvantages, as well as incorporate publications on a broader range of collocation-based physics informed neural networks.
Journal ArticleDOI

Deep learning and big data technologies for IoT security

TL;DR: A comprehensive survey on state-of-the-art deep learning, IoT security, and big data technologies is conducted and a thematic taxonomy is derived from the comparative analysis of technical studies of the three aforementioned domains.
References
More filters
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Journal ArticleDOI

Generative Adversarial Nets

TL;DR: A new framework for estimating generative models via an adversarial process, in which two models are simultaneously train: a generative model G that captures the data distribution and a discriminative model D that estimates the probability that a sample came from the training data rather than G.
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Related Papers (5)