scispace - formally typeset
Open AccessBook ChapterDOI

LightLayers: Parameter Efficient Dense and Convolutional Layers for Image Classification

TLDR
LightLayers as discussed by the authors is a method for reducing the number of trainable parameters in DNNs, which consists of LightDense and LightConv2D layers that are as efficient as regular Conv2D and Dense layers but uses less parameters.
Abstract
Deep Neural Networks (DNNs) have become the de-facto standard in computer vision, as well as in many other pattern recognition tasks. A key drawback of DNNs is that the training phase can be very computationally expensive. Organizations or individuals that cannot afford purchasing state-of-the-art hardware or tapping into cloud hosted infrastructures may face a long waiting time before the training completes or might not be able to train a model at all. Investigating novel ways to reduce the training time could be a potential solution to alleviate this drawback, and thus enabling more rapid development of new algorithms and models. In this paper, we propose LightLayers, a method for reducing the number of trainable parameters in DNNs. The proposed LightLayers consists of LightDense and LightConv2D layers that are as efficient as regular Conv2D and Dense layers but uses less parameters. We resort to Matrix Factorization to reduce the complexity of the DNN models resulting in lightweight DNN models that require less computational power, without much loss in the accuracy. We have tested LightLayers on MNIST, Fashion MNIST, CIFAR 10, and CIFAR 100 datasets. Promising results are obtained for MNIST, Fashion MNIST, and CIFAR-10 datasets whereas CIFAR 100 shows acceptable performance by using fewer parameters.

read more

Citations
More filters
Journal ArticleDOI

Explainable neural networks that simulate reasoning

TL;DR: In this article, the authors show how neural circuits can directly encode cognitive processes via simple neurobiological principles, and demonstrate how neural systems can encode cognitive functions, and use the proposed model to train robust, scalable deep neural networks that are explainable and capable of symbolic reasoning and domain generalization.
Journal ArticleDOI

New convolutional neural network models for efficient object recognition with humanoid robots

TL;DR: In this article, the object recognition capability of a humanoid robot is discussed and the robot can manipulate the objects they have not previously seen in real-life environments, thus, it is important that the robot has the ability to recognize objects.
Book ChapterDOI

Dynamic Neural Diversification: Path to Computationally Sustainable Neural Networks

TL;DR: In this article, the authors explore the diversity of the neurons within the hidden layer during the learning process, and analyze how the diversity affects predictions of the model, and introduce several techniques to dynamically reinforce diversity between neurons during the training.
Journal ArticleDOI

Konvolüsyonel Sinir Ağları Tabanlı Türkçe Metin Sınıflandırma

TL;DR: Ayrıca tüm algoritmaların birbiri ile karşılaştırmasını içeren tablolar oluşturularak sonuçlar analiz edilmiştir.
Book ChapterDOI

Dynamic Neural Diversification: Path to Computationally Sustainable Neural Networks

TL;DR: In this paper, the authors explore the diversity of the neurons within the hidden layer during the learning process, and analyze how the diversity affects predictions of the model, and introduce several techniques to dynamically reinforce diversity between neurons during the training.
References
More filters
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Dissertation

Learning Multiple Layers of Features from Tiny Images

TL;DR: In this paper, the authors describe how to train a multi-layer generative model of natural images, using a dataset of millions of tiny colour images, described in the next section.
Posted Content

Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms

TL;DR: Fashion-MNIST is intended to serve as a direct drop-in replacement for the original MNIST dataset for benchmarking machine learning algorithms, as it shares the same image size, data format and the structure of training and testing splits.
Related Papers (5)