scispace - formally typeset
Open AccessPosted Content

FNA++: Fast Network Adaptation via Parameter Remapping and Architecture Search

TLDR
This paper proposes a Fast Network Adaptation (FNA++) method, which can adapt both the architecture and parameters of a seed network to become a network with different depths, widths, or kernel sizes via a parameter remapping technique, making it possible to use NAS for segmentation and detection tasks a lot more efficiently.
Abstract
Deep neural networks achieve remarkable performance in many computer vision tasks. Most state-of-the-art (SOTA) semantic segmentation and object detection approaches reuse neural network architectures designed for image classification as the backbone, commonly pre-trained on ImageNet. However, performance gains can be achieved by designing network architectures specifically for detection and segmentation, as shown by recent neural architecture search (NAS) research for detection and segmentation. One major challenge though is that ImageNet pre-training of the search space representation (a.k.a. super network) or the searched networks incurs huge computational cost. In this paper, we propose a Fast Network Adaptation (FNA++) method, which can adapt both the architecture and parameters of a seed network (e.g. an ImageNet pre-trained network) to become a network with different depths, widths, or kernel sizes via a parameter remapping technique, making it possible to use NAS for segmentation/detection tasks a lot more efficiently. In our experiments, we conduct FNA++ on MobileNetV2 to obtain new networks for semantic segmentation, object detection, and human pose estimation that clearly outperform existing networks designed both manually and by NAS. We also implement FNA++ on ResNets and NAS networks, which demonstrates a great generalization ability. The total computation cost of FNA++ is significantly less than SOTA segmentation/detection NAS approaches: 1737x less than DPC, 6.8x less than Auto-DeepLab, and 8.0x less than DetNAS. The code will be released at this https URL.

read more

Citations
More filters
Journal ArticleDOI

EfficientPose: Efficient human pose estimation with neural architecture search

TL;DR: This paper proposes an efficient framework targeted at human pose estimation including two parts, the efficient backbone and the efficient head, by implementing the differentiable neural architecture search method and customize the backbone network design for pose estimation and reduce the computation cost with negligible accuracy degradation.
Posted Content

Cyclic Differentiable Architecture Search

TL;DR: New joint optimization objectives are proposed and a novel Cyclic Differentiable ARchiTecture Search framework is proposed, dubbed CyDAS, which enables the evolution of the architecture to fit the final evaluation network.
Journal ArticleDOI

A Survey on Surrogate-assisted Efficient Neural Architecture Search

TL;DR: This paper begins with a brief introduction to the general framework of NAS, followed by a description of surrogate-assisted NAS, which is divided into three different categories, namely Bayesian optimization for NAS, surrogate- assisted evolutionary algorithms forNAS, and MOP for NAS.
Journal ArticleDOI

Neural Architecture Search Survey: A Computer Vision Perspective

TL;DR: In this article , the basic concepts of automated neural architecture search (NAS) are summarized and an overview of recent studies on the applications of NAS is provided. But, to the best knowledge of the present authors, this study is the first to look at NAS from a computer vision perspective, and recent trends found in each study on NAS were analyzed in detail.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Related Papers (5)