scispace - formally typeset
Proceedings ArticleDOI

Challenges in Energy-Efficient Deep Neural Network Training With FPGA

TLDR
A performance metric and evaluation workflow are proposed to compare the FPGA-based systems for DNN training in terms of usage of on-chip resources, training efficiency, energy efficiency, and model performance for specific computer vision tasks.
Abstract
In recent years, it is highly demanding to deploy Deep Neural Networks (DNNs) on edge devices, such as mobile phones, drones, robotics, and wearable devices, to process visual data collected by the cameras embedded in these systems. In addition to the model inference, training DNNs locally can benefit model customization and data privacy protection. Since many edge systems are powered by batteries or have limited energy budgets, Field-Programmable Gate Array (FPGA) is commonly used as the primary processing engine to satisfy both demands in performance and energy-efficiency. Although many recent research papers have been published on the topic of DNN inference with FPGAs, training a DNN with FPGAs has not been well exploited by the community. This paper summarizes the current status of adopting FPGA for DNN computation and identifies the main challenges in deploying DNN training on FPGAs. Moreover, a performance metric and evaluation workflow are proposed to compare the FPGA-based systems for DNN training in terms of (1) usage of on-chip resources, (2) training efficiency, (3) energy efficiency, and (4) model performance for specific computer vision tasks.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

FPGA Implementation for CNN-Based Optical Remote Sensing Object Detection

TL;DR: This paper optimize the CNN-based model for hardware implementation, which establishes a foundation for efficiently mapping the network on a field-programmable gate array (FPGA), and proposes a hardware architecture for the CNN -based remote sensing object detection model.
Journal ArticleDOI

A systematic review of Green AI

TL;DR: In this article , the authors present a systematic review of the Green AI literature, which includes position papers, observational studies, and solution papers and conclude that the time is suitable to adopt other Green AI research strategies, and port the numerous promising academic results to industrial practice.
Journal ArticleDOI

EF-Train: Enable Efficient On-device CNN Training on FPGA through Data Reshaping for Online Adaptation or Personalization

TL;DR: EF-Train is designed, an efficient DNN training accelerator with a unified channel-level parallelism-based convolution kernel that can achieve end-to-end training on resource-limited low-power edge-level FPGAs and develops a data reshaping approach with intra-tile continuous memory allocation and weight reuse.
Journal ArticleDOI

FitNN: A Low-Resource FPGA-Based CNN Accelerator for Drones

TL;DR: A field-programmable gate array (FPGA)-based convolutional neural network (CNN) accelerator, named FitNN, is presented, which improves the speed and power efficiency of CNN inference by reducing data movements.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Related Papers (5)