scispace - formally typeset
Open AccessProceedings Article

Selective Classification for Deep Neural Networks

Reads0
Chats0
TLDR
A method to construct a selective classifier given a trained neural network, which allows a user to set a desired risk level and the classifier rejects instances as needed, to grant the desired risk (with high probability).
Abstract
Selective classification techniques (also known as reject option) have not yet been considered in the context of deep neural networks (DNNs). These techniques can potentially significantly improve DNNs prediction performance by trading-off coverage. In this paper we propose a method to construct a selective classifier given a trained neural network. Our method allows a user to set a desired risk level. At test time, the classifier rejects instances as needed, to grant the desired risk (with high probability). Empirical results over CIFAR and ImageNet convincingly demonstrate the viability of our method, which opens up possibilities to operate DNNs in mission-critical applications. For example, using our method an unprecedented 2% error in top-5 ImageNet classification can be guaranteed with probability 99.9%, with almost 60% test coverage.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Who Needs Decoders? Efficient Estimation of Sequence-level Attributes

TL;DR: This paper proposed Non-Autoregressive Proxy (NAP) models that can predict general scalar-valued sequence-level attributes directly from the encodings, avoiding the expensive autoregressive decoding stage.

L everaging u nlabeled d ata to p redict o ut - of -d istribution p erformance

TL;DR: This work proposes Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model’s confidence, predicting accuracy as the fraction of unlabeled examples for which model confidence exceeds that threshold.
Proceedings ArticleDOI

Human-Centered Deferred Inference: Measuring User Interactions and Setting Deferral Criteria for Human-AI Teams

TL;DR: The authors focus on the role of users in deferred inference and argue that the deferral criteria should be a function of the user and model as a team, not simply the model itself.
Journal ArticleDOI

Partial Matrix Completion

TL;DR: Interestingly, even though the distribution is unknown and arbitrarily complex, the efficient algorithm is able to guarantee: (a) high accuracy over all completed entries, and (b) high coverage, meaning that it covers at least as much of the matrix as the distribution of observations.

Learning Large Margin Sparse Embeddings for Open Set Medical Diagnosis

TL;DR: Wang et al. as mentioned in this paper proposed margin loss with adaptive scale (MLAS) and open margin cosine loss (OMCL) for medical open set recognition (OSR).
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Journal ArticleDOI

ImageNet Large Scale Visual Recognition Challenge

TL;DR: The ImageNet Large Scale Visual Recognition Challenge (ILSVRC) as mentioned in this paper is a benchmark in object category classification and detection on hundreds of object categories and millions of images, which has been run annually from 2010 to present, attracting participation from more than fifty institutions.
Proceedings ArticleDOI

Very deep convolutional neural network based image classification using small training sample size

TL;DR: In this article, a modified VGG-16 network was used to fit CIFAR-10 without severe overfitting and achieved 8.45% error rate on the dataset.
Journal ArticleDOI

An optimum character recognition system using decision functions

TL;DR: The character recognition problem, usually resulting from characters being corrupted by printing deterioration and/or inherent noise of the devices, is considered from the viewpoint of statistical decision theory and the optimum recogition is obtained.