scispace - formally typeset
Proceedings ArticleDOI

Disentangled Representation Learning GAN for Pose-Invariant Face Recognition

TLDR
Quantitative and qualitative evaluation on both controlled and in-the-wild databases demonstrate the superiority of DR-GAN over the state of the art.
Abstract
The large pose discrepancy between two face images is one of the key challenges in face recognition. Conventional approaches for pose-invariant face recognition either perform face frontalization on, or learn a pose-invariant representation from, a non-frontal face image. We argue that it is more desirable to perform both tasks jointly to allow them to leverage each other. To this end, this paper proposes Disentangled Representation learning-Generative Adversarial Network (DR-GAN) with three distinct novelties. First, the encoder-decoder structure of the generator allows DR-GAN to learn a generative and discriminative representation, in addition to image synthesis. Second, this representation is explicitly disentangled from other face variations such as pose, through the pose code provided to the decoder and pose estimation in the discriminator. Third, DR-GAN can take one or multiple images as the input, and generate one unified representation along with an arbitrary number of synthetic images. Quantitative and qualitative evaluation on both controlled and in-the-wild databases demonstrate the superiority of DR-GAN over the state of the art.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Disentangled and Side-Aware Unsupervised Domain Adaptation for Cross-Dataset Subjective Tinnitus Diagnosis

TL;DR: Wang et al. as discussed by the authors proposed Disentangled and Side-aware Unsupervised Domain Adaptation (DSUDA) for cross-dataset tinnitus diagnosis, where a disentangled auto-encoder is developed to decouple class-irrelevant information from the EEG signals.
Posted Content

DSRGAN: Explicitly Learning Disentangled Representation of Underlying Structure and Rendering for Image Generation without Tuple Supervision.

TL;DR: Comparison to the state-of-the-art methods shows that DSRGAN can significantly outperform them in disentanglability, and proposes a quantitative criterion (the Normalized Disentangled DisentangLability) to quantify disentangleability.
Journal ArticleDOI

A Review of Facial Expression Recognition

TL;DR: Wang et al. as discussed by the authors summarized some widely used public data sets for facial expression recognition and analyzed some existing deep learning methods, especially deep convolutional neural network (DCNN), and compared the performance of four classical CNNs (AlexNet, GoogleNet, VGGNet and ResNet).
Proceedings ArticleDOI

Multi-task and Multi-scale Face Recognition Based on CNN

TL;DR: Wang et al. as mentioned in this paper proposed a multi-scale feature fusion convolutional neural network combined with multi-task learning, where the face recognition main task is decomposed into pose estimation, illumination classification and occlusion classification subtasks, which are jointly used to promote the optimization of the main task.
Posted Content

Novel View Synthesis from a Single Image via Unsupervised learning.

TL;DR: In this paper, a token transformation module (TTM) was proposed to transform the features extracted from a source viewpoint image into an intrinsic representation with respect to a pre-defined reference pose and a view generation module was used to synthesize an arbitrary view from the representation.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Generative Adversarial Nets

TL;DR: A new framework for estimating generative models via an adversarial process, in which two models are simultaneously train: a generative model G that captures the data distribution and a discriminative model D that estimates the probability that a sample came from the training data rather than G.
Journal ArticleDOI

Representation Learning: A Review and New Perspectives

TL;DR: Recent work in the area of unsupervised feature learning and deep learning is reviewed, covering advances in probabilistic models, autoencoders, manifold learning, and deep networks.
Proceedings ArticleDOI

FaceNet: A unified embedding for face recognition and clustering

TL;DR: A system that directly learns a mapping from face images to a compact Euclidean space where distances directly correspond to a measure offace similarity, and achieves state-of-the-art face recognition performance using only 128-bytes perface.
Posted Content

Conditional Generative Adversarial Nets

Mehdi Mirza, +1 more
- 06 Nov 2014 - 
TL;DR: The conditional version of generative adversarial nets is introduced, which can be constructed by simply feeding the data, y, to the generator and discriminator, and it is shown that this model can generate MNIST digits conditioned on class labels.
Related Papers (5)