scispace - formally typeset
Open AccessProceedings ArticleDOI

EIGEN: Ecologically-Inspired GENetic Approach for Neural Network Structure Searching From Scratch

TLDR
Zhang et al. as discussed by the authors proposed an ecologically-inspired GENetic (EIGEN) approach that uses the concept of succession, extinction, mimicry and gene duplication to search neural network structure from scratch with poorly initialized simple network and few constraints forced during the evolution, as they assume no prior knowledge about the task domain.
Abstract
Designing the structure of neural networks is considered one of the most challenging tasks in deep learning, especially when there is few prior knowledge about the task domain. In this paper, we propose an Ecologically-Inspired GENetic (EIGEN) approach that uses the concept of succession, extinction, mimicry, and gene duplication to search neural network structure from scratch with poorly initialized simple network and few constraints forced during the evolution, as we assume no prior knowledge about the task domain. Specifically, we first use primary succession to rapidly evolve a population of poorly initialized neural network structures into a more diverse population, followed by a secondary succession stage for fine-grained searching based on the networks from the primary succession. Extinction is applied in both stages to reduce computational cost. Mimicry is employed during the entire evolution process to help the inferior networks imitate the behavior of a superior network and gene duplication is utilized to duplicate the learned blocks of novel structures, both of which help to find better network structures. Experimental results show that our proposed approach can achieve similar or better performance compared to the existing genetic approaches with dramatically reduced computation cost. For example, the network discovered by our approach on CIFAR-100 dataset achieves 78.1% test accuracy under 120 GPU hours, compared to 77.0% test accuracy in more than 65, 536 GPU hours in [35].

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

A Survey on Evolutionary Neural Architecture Search

TL;DR: This article reviews over 200 articles of most recent EC-based NAS methods in light of the core components, to systematically discuss their design principles and justifications on the design.
Journal ArticleDOI

A Survey on Evolutionary Neural Architecture Search

TL;DR: In this article , the authors reviewed over 200 papers of most recent Evolutionary Computation-based Neural Architecture Search (NAS) methods in light of the core components, to systematically discuss their design principles as well as justifications on the design.
Proceedings ArticleDOI

Teachers Do More Than Teach: Compressing Image-to-Image Models

TL;DR: CAT as mentioned in this paper introduces a teacher network that provides a search space in which efficient network architectures can be found, in addition to performing knowledge distillation, which achieves similar or even better image fidelity than the original models with much reduced computational cost, e.g., MACs.
Journal ArticleDOI

Lights and shadows in Evolutionary Deep Learning: Taxonomy, critical methodological analysis, cases of study, learned lessons, recommendations and challenges

TL;DR: This work comprehensively review and critically examine contributions made so far based on three axes - optimization and taxonomy, critical analysis, and challenges - which outline a complete vision of a merger of two technologies drawing up an exciting future for this area of fusion research.
Journal ArticleDOI

A Review on Convolutional Neural Network Encodings for Neuroevolution

TL;DR: A comprehensive review on the state-of-the-art encodings for CNNs can be found in this paper , where the authors present a comprehensive review of the most widely used encoding methods.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings ArticleDOI

Densely Connected Convolutional Networks

TL;DR: DenseNet as mentioned in this paper proposes to connect each layer to every other layer in a feed-forward fashion, which can alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
Posted Content

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Batch Normalization as mentioned in this paper normalizes layer inputs for each training mini-batch to reduce the internal covariate shift in deep neural networks, and achieves state-of-the-art performance on ImageNet.
Related Papers (5)