scispace - formally typeset
Book ChapterDOI

GradientBased Learning Applied to Document Recognition

TLDR
Various methods applied to handwritten character recognition are reviewed and compared and Convolutional Neural Networks, that are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques.
Abstract
Multilayer Neural Networks trained with the backpropagation algorithm constitute the best example of a successful Gradient-Based Learning technique. Given an appropriate network architecture, Gradient-Based Learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional Neural Networks, that are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation, recognition, and language modeling. A new learning paradigm, called Graph Transformer Networks (GTN), allows such multi-module systems to be trained globally using Gradient-Based methods so as to minimize an overall performance measure. Two systems for on-line handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of Graph Transformer Networks. A Graph Transformer Network for reading bank check is also described. It uses Convolutional Neural Network character recognizers combined with global training techniques to provides record accuracy on business and personal checks. It is deployed commercially and reads several million checks per day.

read more

Citations
More filters
Proceedings Article

A Two-Step Computation of the Exact GAN Wasserstein Distance

TL;DR: This approach optimizes the exact Wasserstein distance, obviating the need for weight clipping previously used in WGANs, and theoretically proves that the proposed formulation is equivalent to the discrete MongeKantorovich dual formulation.
Proceedings ArticleDOI

ShieldNets: Defending Against Adversarial Attacks Using Probabilistic Adversarial Robustness

TL;DR: Experimental results show that the Probabilistic adversarial robustness approach is generalizable, robust against adversarial transferability and resistant to a wide variety of attacks on the Fashion-MNIST and CIFAR10 datasets, respectively.
Proceedings ArticleDOI

Policy-GNN: Aggregation Optimization for Graph Neural Networks

TL;DR: In this article, a meta-policy framework is proposed to model the sampling procedure and message passing of GNNs into a combined learning process, which is trained with deep reinforcement learning by exploiting the feedback from the model.
Proceedings ArticleDOI

Support Vector Machines on GPU with Sparse Matrix Format

TL;DR: Sparse matrix format is introduced into parallel SVM to achieve better performance and Experimental results show that the speedup of 55x–133.8x over LIBSVM can be achieved in training process on NVIDIA GeForce GTX470.
Proceedings ArticleDOI

Learning lexical features of programming languages from imagery using convolutional neural networks

TL;DR: The results indicate that not only can computer vision models based on deep architectures be taught to differentiate among programming languages with over 98% accuracy, but can learn language-specific lexical features in the process.
Related Papers (5)