scispace - formally typeset
Open AccessProceedings ArticleDOI

Class Rectification Hard Mining for Imbalanced Deep Learning

Reads0
Chats0
TLDR
This work develops an end-to-end deep learning framework capable of avoiding the dominant effect of majority classes by discovering sparsely sampled boundaries of minority classes by introducing a Class Rectification Loss (CRL) regularising algorithm.
Abstract
Recognising detailed facial or clothing attributes in images of people is a challenging task for computer vision, especially when the training data are both in very large scale and extremely imbalanced among different attribute classes. To address this problem, we formulate a novel scheme for batch incremental hard sample mining of minority attribute classes from imbalanced large scale training data. We develop an end-to-end deep learning framework capable of avoiding the dominant effect of majority classes by discovering sparsely sampled boundaries of minority classes. This is made possible by introducing a Class Rectification Loss (CRL) regularising algorithm. We demonstrate the advantages and scalability of CRL over existing state-of-the-art attribute recognition and imbalanced data learning models on two large scale imbalanced benchmark datasets, the CelebA facial attribute dataset and the X-Domain clothing attribute dataset.

read more

Citations
More filters
Posted Content

SelectNet: Learning to Sample from the Wild for Imbalanced Data Training

TL;DR: SelectNet as discussed by the authors adopts a semi-supervised learning paradigm by training a deep neural network, referred to as SelectNet, to selectively add unlabeled data together with their predicted labels to the training dataset, thus gradually mitigating the imbalance in the data used for training the classifier.
Journal ArticleDOI

Improving Imbalanced Land Cover Classification with K-Means SMOTE: Detecting and Oversampling Distinctive Minority Spectral Signatures

TL;DR: In this article, the authors used K-means and the Synthetic Minority Oversampling Technique (SMOTE) as an improved oversampling algorithm to improve the quality of newly created artificial data by addressing both the between-class imbalance, as traditional oversamplers do, but also the withinclass imbalance.
Posted Content

Transferability and Hardness of Supervised Classification Tasks

TL;DR: In this article, the authors estimate the difficulty and transferability of supervised classification tasks using an information theoretic approach, treating training labels as random variables and exploring their statistics, and show that the conditional entropy between two such variables is related to the loss of the transferred model.
Proceedings ArticleDOI

Learning to Learn across Diverse Data Biases in Deep Face Recognition

TL;DR: In this paper , a multi-variation cosine margin (MvCoM) is proposed to simultaneously consider the multiple variation factors, which orthogonally enhances the face recognition losses to incorporate the importance of training samples.
Proceedings ArticleDOI

Metricbert: Text Representation Learning Via Self-Supervised Triplet Training

TL;DR: MetricBERT as mentioned in this paper is a BERT-based model that learns to embed text under a well-defined similarity metric while simultaneously adhering to the traditional masked-language task.
References
More filters
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Journal ArticleDOI

ImageNet Large Scale Visual Recognition Challenge

TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Book ChapterDOI

Microsoft COCO: Common Objects in Context

TL;DR: A new dataset with the goal of advancing the state-of-the-art in object recognition by placing the question of object recognition in the context of the broader question of scene understanding by gathering images of complex everyday scenes containing common objects in their natural context.
Related Papers (5)