scispace - formally typeset
Search or ask a question
Author

Jun Sun

Other affiliations: Kyushu University
Bio: Jun Sun is an academic researcher from Fujitsu. The author has contributed to research in topics: Convolutional neural network & Binary image. The author has an hindex of 16, co-authored 163 publications receiving 1227 citations. Previous affiliations of Jun Sun include Kyushu University.


Papers
More filters
Proceedings ArticleDOI
Li Chen1, Song Wang1, Wei Fan1, Jun Sun1, Satoshi Naoi1 
01 Nov 2015
TL;DR: In the experiments, the proposed CNN-based handwritten character recognition framework performed even better than human on handwritten digit (MNIST) and Chinese character (CASIA) recognition.
Abstract: Because of the various appearance (different writers, writing styles, noise, etc.), the handwritten character recognition is one of the most challenging task in pattern recognition. Through decades of research, the traditional method has reached its limit while the emergence of deep learning provides a new way to break this limit. In this paper, a CNN-based handwritten character recognition framework is proposed. In this framework, proper sample generation, training scheme and CNN network structure are employed according to the properties of handwritten characters. In the experiments, the proposed framework performed even better than human on handwritten digit (MNIST) and Chinese character (CASIA) recognition. The advantage of this framework is proved by these experimental results.

117 citations

Proceedings ArticleDOI
Chunpeng Wu1, Wei Fan1, Yuan He1, Jun Sun1, Satoshi Naoi1 
15 Dec 2014
TL;DR: The relaxation convolution layer adopted in the R-CNN, unlike traditional convolutional layer, does not require neurons within a feature map to share the same Convolutional kernel, endowing the neural network with more expressive power.
Abstract: Deep learning methods have recently achieved impressive performance in the area of visual recognition and speech recognition. In this paper, we propose a hand- writing recognition method based on relaxation convolutional neural network (R-CNN) and alternately trained relaxation convolutional neural network (ATR-CNN). Previous methods regularize CNN at full-connected layer or spatial-pooling layer, however, we focus on convolutional layer. The relaxation convolution layer adopted in our R-CNN, unlike traditional convolutional layer, does not require neurons within a feature map to share the same convolutional kernel, endowing the neural network with more expressive power. As relaxation convolution sharply increase the total number of parameters, we adopt alternate training in ATR-CNN to regularize the neural network during training procedure. Our previous C- NN took the 1st place in ICDAR'13 Chinese Handwriting Character Recognition Competition, while our latest ATR-CNN outperforms our previous one and achieves the state-of-the-art accuracy with an error rate of 3.94%, further narrowing the gap between machine and human observers (3.87%).

105 citations

Journal ArticleDOI
TL;DR: A novel document skew detection algorithm based on wavelet decompositions and projection profile analysis is proposed and shows that this algorithm performs well on document images of various layouts and is also robust to different languages.

80 citations

Journal ArticleDOI
TL;DR: A sparse learning algorithm for Support Vector Classification, called SSVC, which leads to sparse solutions by automatically setting the irrelevant parameters exactly to zero, and it is shown that SSVC offers competitive performance to SVC but needs significantly fewer Support Vectors.

58 citations

Patent
Chunpeng Wu1, Wei Fan1, Yuan He1, Jun Sun1
26 Jun 2014
TL;DR: In this paper, a convolutional-neural-network-based classifier is proposed, which consists of a plurality of feature map layers, at least one feature map in each layer, and a plurality-of-convolutional templates corresponding to the plurality of regions.
Abstract: The present invention relates to a convolutional-neural-network-based classifier, a classifying method by using a convolutional-neural-network-based classifier and a method for training the convolutional-neural-network-based classifier The convolutional-neural-network-based classifier comprises: a plurality of feature map layers, at least one feature map in at least one of the plurality of feature map layers being divided into a plurality of regions; and a plurality of convolutional templates corresponding to the plurality of regions respectively, each of the convolutional templates being used for obtaining a response value of a neuron in the corresponding region

54 citations


Cited by
More filters
Patent
17 Sep 2008
TL;DR: The type of information in metadata for one type of file differs from the type of metadata for another type of a file as discussed by the authors, and the metadata from files created by several different software applications are captured and the captured metadata is searched.
Abstract: Systems and methods for managing data, such as metadata. In one exemplary method, metadata from files created by several different software applications are captured, and the captured metadata is searched. The type of information in metadata for one type of file differs from the type of information in metadata for another type of file. Other methods are described and data processing systems and machine readable media are also described.

947 citations

Journal ArticleDOI
TL;DR: This review provides a fundamental comparison and analysis of the remaining problems in the field and summarizes the fundamental problems and enumerates factors that should be considered when addressing these problems.
Abstract: This paper analyzes, compares, and contrasts technical challenges, methods, and the performance of text detection and recognition research in color imagery It summarizes the fundamental problems and enumerates factors that should be considered when addressing these problems Existing techniques are categorized as either stepwise or integrated and sub-problems are highlighted including text localization, verification, segmentation and recognition Special issues associated with the enhancement of degraded text and the processing of video text, multi-oriented, perspectively distorted and multilingual text are also addressed The categories and sub-categories of text are illustrated, benchmark datasets are enumerated, and the performance of the most representative approaches is compared This review provides a fundamental comparison and analysis of the remaining problems in the field

709 citations

Journal ArticleDOI
TL;DR: An accurate and robust method for detecting texts in natural scene images using a fast and effective pruning algorithm to extract Maximally Stable Extremal Regions (MSERs) as character candidates using the strategy of minimizing regularized variations is proposed.
Abstract: Text detection in natural scene images is an important prerequisite for many content-based image analysis tasks. In this paper, we propose an accurate and robust method for detecting texts in natural scene images. A fast and effective pruning algorithm is designed to extract Maximally Stable Extremal Regions (MSERs) as character candidates using the strategy of minimizing regularized variations. Character candidates are grouped into text candidates by the single-link clustering algorithm, where distance weights and clustering threshold are learned automatically by a novel self-training distance metric learning algorithm. The posterior probabilities of text candidates corresponding to non-text are estimated with a character classifier; text candidates with high non-text probabilities are eliminated and texts are identified with a text classifier. The proposed system is evaluated on the ICDAR 2011 Robust Reading Competition database; the f-measure is over 76%, much better than the state-of-the-art performance of 71%. Experiments on multilingual, street view, multi-orientation and even born-digital databases also demonstrate the effectiveness of the proposed method.

616 citations

Journal ArticleDOI
TL;DR: The experimental results obtained on real hyperspectral data sets including airport, beach, and urban scenes demonstrate that the performance of the proposed method is quite competitive in terms of computing time and detection accuracy.
Abstract: A novel method for anomaly detection in hyperspectral images is proposed. The method is based on two ideas. First, compared with the surrounding background, objects with anomalies usually appear with small areas and distinct spectral signatures. Second, for both the background and the objects with anomalies, pixels in the same class are usually highly correlated in the spatial domain. In this paper, the pixels with specific area property and distinct spectral signatures are first detected with attribute filtering and a Boolean map-based fusion approach in order to obtain an initial pixel-wise detection result. Then, the initial detection result is refined with edge-preserving filtering to make full use of the spatial correlations among adjacent pixels. Compared with other widely used anomaly detection methods, the experimental results obtained on real hyperspectral data sets including airport, beach, and urban scenes demonstrate that the performance of the proposed method is quite competitive in terms of computing time and detection accuracy.

298 citations