scispace - formally typeset
Book ChapterDOI

GradientBased Learning Applied to Document Recognition

TLDR
Various methods applied to handwritten character recognition are reviewed and compared and Convolutional Neural Networks, that are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques.
Abstract
Multilayer Neural Networks trained with the backpropagation algorithm constitute the best example of a successful Gradient-Based Learning technique. Given an appropriate network architecture, Gradient-Based Learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional Neural Networks, that are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation, recognition, and language modeling. A new learning paradigm, called Graph Transformer Networks (GTN), allows such multi-module systems to be trained globally using Gradient-Based methods so as to minimize an overall performance measure. Two systems for on-line handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of Graph Transformer Networks. A Graph Transformer Network for reading bank check is also described. It uses Convolutional Neural Network character recognizers combined with global training techniques to provides record accuracy on business and personal checks. It is deployed commercially and reads several million checks per day.

read more

Citations
More filters
Posted Content

Learning Sparse High Dimensional Filters: Image Filtering, Dense CRFs and Bilateral Neural Networks

TL;DR: In this paper, a gradient descent algorithm is proposed to learn high-dimensional linear filters that operate in sparsely populated feature spaces, which can be used to learn the pairwise potentials in densely connected conditional random fields.
Proceedings ArticleDOI

AutoFocus: Efficient Multi-Scale Inference

TL;DR: AutoFocus as discussed by the authors predicts category agnostic segmentation maps for small objects at coarser scales, called FocusPixels, which can be predicted with high recall, and in many cases, they only cover a small fraction of the entire image.
Proceedings Article

Spectral rotation versus K-means in spectral clustering

TL;DR: This paper proposes to impose an additional orthonormal constraint to better approximate the optimal continuous solution to the graph cut objective functions, and provides efficient algorithm to solve the new problem rigorously, which is not significantly more costly than K-Means.
Proceedings ArticleDOI

Making Memristive Neural Network Accelerators Reliable

TL;DR: A new error correction scheme for analog neural network accelerators based on arithmetic codes that reduces the respective misclassification rates by 1.5x and 1.1x and encodes the data through multiplication by an integer, which preserves addition operations through the distributive property.
Proceedings ArticleDOI

Everything You Wanted to Know about Deep Learning for Computer Vision but Were Afraid to Ask

TL;DR: The most fundamental concepts of Deep Learning for Computer Vision in particular CNNs, AEs and GANs are introduced, including architectures, inner workings and optimization.
Related Papers (5)