scispace - formally typeset
Open AccessPosted Content

Building Deep Networks on Grassmann Manifolds

Reads0
Chats0
TLDR
This paper proposes a deep network architecture by generalizing the Euclidean network paradigm to Grassmann manifolds and designs full rank mapping layers to transform input Grassmannian data to more desirable ones, and exploits re-orthonormalization layers to normalize the resulting matrices.
Abstract
Learning representations on Grassmann manifolds is popular in quite a few visual recognition tasks. In order to enable deep learning on Grassmann manifolds, this paper proposes a deep network architecture by generalizing the Euclidean network paradigm to Grassmann manifolds. In particular, we design full rank mapping layers to transform input Grassmannian data to more desirable ones, exploit re-orthonormalization layers to normalize the resulting matrices, study projection pooling layers to reduce the model complexity in the Grassmannian context, and devise projection mapping layers to respect Grassmannian geometry and meanwhile achieve Euclidean forms for regular output layers. To train the Grassmann networks, we exploit a stochastic gradient descent setting on manifolds of the connection weights, and study a matrix generalization of backpropagation to update the structured data. The evaluations on three visual recognition tasks show that our Grassmann networks have clear advantages over existing Grassmann learning methods, and achieve results comparable with state-of-the-art approaches.

read more

Citations
More filters
Proceedings ArticleDOI

Attention-Aware Deep Reinforcement Learning for Video Face Recognition

TL;DR: An attention-aware deep reinforcement learning (ADRL) method for video face recognition, which aims to discard the misleading and confounding frames and find the focuses of attentions in face videos for person recognition.
Posted Content

Deep Learning on Lie Groups for Skeleton-based Action Recognition

TL;DR: Li et al. as mentioned in this paper incorporated the Lie group structure into a deep network architecture to learn more appropriate Lie group features for skeleton-based action recognition, and designed rotation mapping layers to transform the input Lie group feature into desirable ones, which are aligned better in the temporal domain.
Proceedings ArticleDOI

A Neural Network Based on SPD Manifold Learning for Skeleton-Based Hand Gesture Recognition

TL;DR: Li et al. as discussed by the authors proposed a new neural network based on SPD manifold learning for skeleton-based hand gesture recognition, given the stream of hand's joint positions, their approach combines two aggregation processes on respectively spatial and temporal domains.
Journal ArticleDOI

Fuzzy Integral-Based CNN Classifier Fusion for 3D Skeleton Action Recognition

TL;DR: Fuzzy fusion through the Choquet integral leverages the degree of uncertainty of decision scores obtained from four CNNs to adaptively generate final decision score based upon confidence of each information source.
Proceedings ArticleDOI

Learning Discriminative Aggregation Network for Video-Based Face Recognition

TL;DR: Experimental results show that the DAN method for video face recognition can generate discriminative images from video clips and improve the overall recognition performance in both the speed and accuracy on three widely used datasets.
References
More filters
Proceedings Article

Rectified Linear Units Improve Restricted Boltzmann Machines

TL;DR: Restricted Boltzmann machines were developed using binary stochastic hidden units that learn features that are better for object recognition on the NORB dataset and face verification on the Labeled Faces in the Wild dataset.
Posted Content

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

TL;DR: This work proposes a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit and derives a robust initialization method that particularly considers the rectifier nonlinearities.
Proceedings ArticleDOI

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

TL;DR: In this paper, a Parametric Rectified Linear Unit (PReLU) was proposed to improve model fitting with nearly zero extra computational cost and little overfitting risk, which achieved a 4.94% top-5 test error on ImageNet 2012 classification dataset.
Book ChapterDOI

Large-Scale Machine Learning with Stochastic Gradient Descent

Léon Bottou
TL;DR: A more precise analysis uncovers qualitatively different tradeoffs for the case of small-scale and large-scale learning problems.
Proceedings ArticleDOI

Deep face recognition

TL;DR: It is shown how a very large scale dataset can be assembled by a combination of automation and human in the loop, and the trade off between data purity and time is discussed.
Related Papers (5)