scispace - formally typeset
Book ChapterDOI

Eigenvector Orientation Corrected LeNet for Digit Recognition

Reads0
Chats0
TLDR
The approach ‘Eigenvector Orientation Corrected LeNet (EOCL)’ presents a simple method to make ordinary LeNet capable of detecting rotated digits, and also to predict the relative angle of orientation of digits with unknown orientation.
Abstract
Convolutional Neural Networks (CNNs) are being used popularly for detecting and classifying objects. Rotational invariance is not guaranteed by many of the existing CNN architectures. Many attempts have been made to acquire rotational invariance in CNNs. Our approach ‘Eigenvector Orientation Corrected LeNet (EOCL)’ presents a simple method to make ordinary LeNet [1] capable of detecting rotated digits, and also to predict the relative angle of orientation of digits with unknown orientation. The proposed method does not demand any modification in the existing LeNet architecture, and requires training with digits having only single orientation. EOCL incorporates an ‘orientation estimation and correction’ step prior to the testing phase. Using Principal Component Analysis, we find the maximum spread direction (Principal Component) of each test sample and then align it vertically. We demonstrate the improvement in classification accuracy and reduction in test time achieved by our approach, on rotated-MNIST [2] and MNIST_rot_12k test datasets, compared to other existing methods.

read more

Citations
More filters
Journal ArticleDOI

Review on One-Stage Object Detection Based on Deep Learning

TL;DR: The current mainstream one-stage object detection model is summarized, and based on YOLOv1, it is continuously optimized, and the improvements and shortcomings are summarized in detail.
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Book ChapterDOI

SURF: speeded up robust features

TL;DR: A novel scale- and rotation-invariant interest point detector and descriptor, coined SURF (Speeded Up Robust Features), which approximates or even outperforms previously proposed schemes with respect to repeatability, distinctiveness, and robustness, yet can be computed and compared much faster.
Proceedings Article

Spatial transformer networks

TL;DR: This work introduces a new learnable module, the Spatial Transformer, which explicitly allows the spatial manipulation of data within the network, and can be inserted into existing convolutional architectures, giving neural networks the ability to actively spatially transform feature maps.
Journal ArticleDOI

The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web]

TL;DR: “Best of the Web” presents the modified National Institute of Standards and Technology (MNIST) resources, consisting of a collection of handwritten digit images used extensively in optical character recognition and machine learning research.
Related Papers (5)