scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Graph fourier transform based descriptor for gesture classification

TL;DR: Simulation results show that the proposed GFT based descriptor gives a comparable results with Karhunen-Loeve transform (K-LT) and Discrete Cosine transform (DCT) based descriptors for gesture classification.
Abstract: This paper proposes a method for gesture classification based on Graph Fourier transform (GFT) coefficients. GFT coefficients are the projection of image pixel block onto the eigenvectors of a Laplacian matrix. This Laplacian matrix is generated from undirected graph, representing a spatial connectedness between each pixel within an image block. This work proposes a method for generating an undirected graph by using edge information of the image. Edge information of the image is obtained by average sum of absolute difference between the current pixel and its neighboring pixels by using an appropriate threshold. The resulting GFT based feature vector is formed by concatenating GFT coefficients of each block. The resultant feature vector is applied to linear Support Vector Machine (SVM) classifier to predict the gesture class. For NTU and Massey hand gesture datasets, threshold value 30 gives maximum prediction accuracy. We compare the results of the proposed GFT based descriptor approach with Karhunen-Loeve transform (K-LT) and Discrete Cosine transform (DCT) based descriptors on three different gesture datasets: NTU, Cambridge and Massey. Simulation results show that the proposed GFT based descriptor gives a comparable results with Karhunen-Loeve transform (K-LT) and Discrete Cosine transform (DCT) based descriptors for gesture classification.
Citations
More filters
Proceedings ArticleDOI
01 Feb 2020
TL;DR: Simulation results show that the proposed GDFT based features from Gaussian Weighted Visibility Graph (VG) can detect epileptic seizure with 100 % accuracy.
Abstract: Epileptic Seizure is a chronic nervous system disorder which is analyzed using Electroencephalogram (EEG) signals. This paper proposes a Graph Signal Processing technique called Graph Discrete Fourier Transform (GDFT) for the detection of epilepsy. EEG data points are projected on the Eigen space of Laplacian matrix of graph to produce GDFT coefficients. The Laplacian matrix is generated from weighted visibility graph constructed from EEG signals. It proposes Gaussian kernel based edge weights between the nodes. The proposed GDFT based feature vectors are then used to detect the seizure class from the given EEG signal using a crisp rule based classification. Simulation results show that the proposed GDFT based features from Gaussian Weighted Visibility Graph (VG) can detect epileptic seizure with 100 % accuracy.

15 citations

References
More filters
Journal ArticleDOI
TL;DR: Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.
Abstract: LIBSVM is a library for Support Vector Machines (SVMs). We have been actively developing this package since the year 2000. The goal is to help users to easily apply SVM to their applications. LIBSVM has gained wide popularity in machine learning and many other areas. In this article, we present all implementation details of LIBSVM. Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.

40,826 citations


Additional excerpts

  • ...LIBSVM library [25] is used for classification....

    [...]

Proceedings ArticleDOI
20 Jun 2005
TL;DR: It is shown experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection, and the influence of each stage of the computation on performance is studied.
Abstract: We study the question of feature sets for robust visual object recognition; adopting linear SVM based human detection as a test case. After reviewing existing edge and gradient based descriptors, we show experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection. We study the influence of each stage of the computation on performance, concluding that fine-scale gradients, fine orientation binning, relatively coarse spatial binning, and high-quality local contrast normalization in overlapping descriptor blocks are all important for good results. The new approach gives near-perfect separation on the original MIT pedestrian database, so we introduce a more challenging dataset containing over 1800 annotated human images with a large range of pose variations and backgrounds.

31,952 citations


"Graph fourier transform based descr..." refers methods in this paper

  • ...We tested the proposed algorithm on the INRIA [17] dataset which is divided into pos and neg folders....

    [...]

  • ...[17] used Histogram of Oriented Gradient (HOG) for human detection....

    [...]

Journal ArticleDOI
TL;DR: A near-real-time computer system that can locate and track a subject's head, and then recognize the person by comparing characteristics of the face to those of known individuals, and that is easy to implement using a neural network architecture.
Abstract: We have developed a near-real-time computer system that can locate and track a subject's head, and then recognize the person by comparing characteristics of the face to those of known individuals. The computational approach taken in this system is motivated by both physiology and information theory, as well as by the practical requirements of near-real-time performance and accuracy. Our approach treats the face recognition problem as an intrinsically two-dimensional (2-D) recognition problem rather than requiring recovery of three-dimensional geometry, taking advantage of the fact that faces are normally upright and thus may be described by a small set of 2-D characteristic views. The system functions by projecting face images onto a feature space that spans the significant variations among known face images. The significant features are known as "eigenfaces," because they are the eigenvectors (principal components) of the set of faces; they do not necessarily correspond to features such as eyes, ears, and noses. The projection operation characterizes an individual face by a weighted sum of the eigenface features, and so to recognize a particular face it is necessary only to compare these weights to those of known individuals. Some particular advantages of our approach are that it provides for the ability to learn and later recognize new faces in an unsupervised manner, and that it is easy to implement using a neural network architecture.

14,562 citations


"Graph fourier transform based descr..." refers methods in this paper

  • ...In the field of image processing, K-LT and DCT was initially used for image compression and reconstruction [26], [27] but they can also be employed for image recognition [28], [29]....

    [...]

Journal ArticleDOI
TL;DR: In this article, the authors present the most common spectral clustering algorithms, and derive those algorithms from scratch by several different approaches, and discuss the advantages and disadvantages of these algorithms.
Abstract: In recent years, spectral clustering has become one of the most popular modern clustering algorithms. It is simple to implement, can be solved efficiently by standard linear algebra software, and very often outperforms traditional clustering algorithms such as the k-means algorithm. On the first glance spectral clustering appears slightly mysterious, and it is not obvious to see why it works at all and what it really does. The goal of this tutorial is to give some intuition on those questions. We describe different graph Laplacians and their basic properties, present the most common spectral clustering algorithms, and derive those algorithms from scratch by several different approaches. Advantages and disadvantages of the different spectral clustering algorithms are discussed.

9,141 citations

Journal ArticleDOI
TL;DR: This paper presents work on computing shape models that are computationally fast and invariant basic transformations like translation, scaling and rotation, and proposes shape detection using a feature called shape context, which is descriptive of the shape of the object.
Abstract: We present a novel approach to measuring similarity between shapes and exploit it for object recognition. In our framework, the measurement of similarity is preceded by: (1) solving for correspondences between points on the two shapes; (2) using the correspondences to estimate an aligning transform. In order to solve the correspondence problem, we attach a descriptor, the shape context, to each point. The shape context at a reference point captures the distribution of the remaining points relative to it, thus offering a globally discriminative characterization. Corresponding points on two similar shapes will have similar shape contexts, enabling us to solve for correspondences as an optimal assignment problem. Given the point correspondences, we estimate the transformation that best aligns the two shapes; regularized thin-plate splines provide a flexible class of transformation maps for this purpose. The dissimilarity between the two shapes is computed as a sum of matching errors between corresponding points, together with a term measuring the magnitude of the aligning transform. We treat recognition in a nearest-neighbor classification framework as the problem of finding the stored prototype shape that is maximally similar to that in the image. Results are presented for silhouettes, trademarks, handwritten digits, and the COIL data set.

6,693 citations


"Graph fourier transform based descr..." refers methods in this paper

  • ...[13] presented a method that measured similarity between the shapes and used this descriptor for object recognition....

    [...]