scispace - formally typeset
Open AccessPosted Content

Prototype Memory for Large-scale Face Representation Learning.

TLDR
Prototype Memory as discussed by the authors uses a limited-size memory module for storing recent class prototypes and employs a set of algorithms to update it in appropriate way, which can be used with various loss functions, hard example mining algorithms and encoder architectures.
Abstract
Face representation learning using datasets with massive number of identities requires appropriate training methods. Softmax-based approach, currently the state-of-the-art in face recognition, in its usual "full softmax" form is not suitable for datasets with millions of persons. Several methods, based on the "sampled softmax" approach, were proposed to remove this limitation. These methods, however, have a set of disadvantages. One of them is a problem of "prototype obsolescence": classifier weights (prototypes) of the rarely sampled classes, receive too scarce gradients and become outdated and detached from the current encoder state, resulting in an incorrect training signals. This problem is especially serious in ultra-large-scale datasets. In this paper, we propose a novel face representation learning model called Prototype Memory, which alleviates this problem and allows training on a dataset of any size. Prototype Memory consists of the limited-size memory module for storing recent class prototypes and employs a set of algorithms to update it in appropriate way. New class prototypes are generated on the fly using exemplar embeddings in the current mini-batch. These prototypes are enqueued to the memory and used in a role of classifier weights for usual softmax classification-based training. To prevent obsolescence and keep the memory in close connection with encoder, prototypes are regularly refreshed, and oldest ones are dequeued and disposed. Prototype Memory is computationally efficient and independent of dataset size. It can be used with various loss functions, hard example mining algorithms and encoder architectures. We prove the effectiveness of the proposed model by extensive experiments on popular face recognition benchmarks.

read more

Citations
More filters
Posted Content

STC speaker recognition systems for the NIST SRE 2021.

TL;DR: In this paper, a number of diverse subsystems based on using deep neural networks as feature extractors were used for speaker verification filed in the NIST 2021 Speaker Recognition Evaluation for both fixed and open training conditions.
Book ChapterDOI

FaceMix: Transferring Local Regions for Data Augmentation in Face Recognition

TL;DR: FaceMix as mentioned in this paper is a flexible face-specific data augmentation technique that transfers a local area of an image to another image, and it can generate new images for a class, using face data from other classes, and these two modes also could be combined.
References
More filters
Book ChapterDOI

RBF-Softmax: Learning Deep Representative Prototypes with Radial Basis Function Softmax

TL;DR: A novel Radial Basis Function (RBF) distances are introduced to replace the commonly used inner products in the softmax loss function, such that it can adaptively assign losses to regularize the intra-class and inter-class distances by reshaping the relative differences, and thus creating more representative prototypes of classes to improve optimization.
Posted Content

Learning to Support: Exploiting Structure Information in Support Sets for One-Shot Learning.

TL;DR: A novel meta-learner is proposed which shows state-of-the-art performance on common benchmarks for one/few shot classification and can be viewed as an approximation to an ensemble, which saves the factor of in training and test times and in the storage of the final model.
Proceedings ArticleDOI

Margin based knowledge distillation for mobile face recognition

TL;DR: A new method for learning fast and compact face recognition model which has a similar performance to a much more complex model used for transferring its knowledge is proposed and it is shown that both these models can be used for verification in a single face recognition system.
Journal ArticleDOI

LinCos-Softmax: Learning Angle-Discriminative Face Representations With Linearity-Enhanced Cosine Logits

TL;DR: This work proposes a Linear-Cosine Softmax Loss (LinCos-Softmax) to more effectively learn angle-discriminative facial features and proposes an automatic scale parameter selection scheme, which can conveniently provide an appropriate scale for different logits without the need for exhaustive parameter search to improve performance.
Proceedings Article

Extreme Classification via Adversarial Softmax Approximation

TL;DR: This paper proposed an adversarial sampling mechanism that produces negative samples at a cost only logarithmic in the number of classes, thus still resulting in cheap gradient updates, and a mathematical proof that this method minimizes the gradient variance while any bias due to non-uniform sampling can be removed.
Related Papers (5)