scispace - formally typeset
Open AccessProceedings ArticleDOI

Learning Structure and Strength of CNN Filters for Small Sample Size Training

TLDR
SSF-CNN as discussed by the authors uses dictionary-based filter learning to learn the structure and strength of the filter for small sample size problems such as newborn face recognition and Omniglot.
Abstract
Convolutional Neural Networks have provided state-of-the-art results in several computer vision problems. However, due to a large number of parameters in CNNs, they require a large number of training samples which is a limiting factor for small sample size problems. To address this limitation, we propose SSF-CNN which focuses on learning the "structure" and "strength" of filters. The structure of the filter is initialized using a dictionary based filter learning algorithm and the strength of the filter is learned using the small sample training data. The architecture provides the flexibility of training with both small and large training databases, and yields good accuracies even with small size training data. The effectiveness of the algorithm is first demonstrated on MNIST, CIFAR10, and NORB databases, with varying number of training samples. The results show that SSF-CNN significantly reduces the number of parameters required for training while providing high accuracies on the test databases. On small sample size problems such as newborn face recognition and Omniglot, it yields state-of-the-art results. Specifically, on the IIITD Newborn Face Database, the results demonstrate improvement in rank-1 identification accuracy by at least 10%.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Generalizing from a Few Examples: A Survey on Few-Shot Learning

TL;DR: A thorough survey to fully understand Few-Shot Learning (FSL), and categorizes FSL methods from three perspectives: data, which uses prior knowledge to augment the supervised experience; model, which used to reduce the size of the hypothesis space; and algorithm, which using prior knowledgeto alter the search for the best hypothesis in the given hypothesis space.
Proceedings ArticleDOI

Meta-Transfer Learning for Few-Shot Learning

TL;DR: In this paper, the authors proposed a meta-transfer learning approach to adapt a base-learner to a new task for which only a few labeled samples are available, which learns scaling and shifting functions of DNN weights for each task.
Proceedings ArticleDOI

DeepEMD: Few-Shot Image Classification With Differentiable Earth Mover’s Distance and Structured Classifiers

TL;DR: Zhang et al. as discussed by the authors adopt the Earth Mover's distance (EMD) as a metric to compute a structural distance between dense image representations to determine image relevance, which is used to represent the image distance for classification.
Posted Content

Meta-Transfer Learning for Few-Shot Learning.

TL;DR: A novel few-shot learning method called meta-transfer learning (MTL) which learns to adapt a deep NN for few shot learning tasks and introduces the hard task (HT) meta-batch scheme as an effective learning curriculum for MTL.
Posted Content

Few-shot Learning: A Survey

Yaqing Wang, +1 more
TL;DR: A comprehensive survey of the core issues of Few-Shot Learning, and existing works from the birth of FSL to the most recent published ones are categorized in a unified taxonomy, with thorough discussion of the pros and cons for different categories.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Trending Questions (1)
How many parameters should I have in my CNN for a fixed amount of training datapoints?

For the proposed SSF-CNN, the total number of strength parameters to be learned for the CIFAR-10 dataset is 26,928, which is significantly less than the 242,352 parameters in traditional ResNet architecture.