scispace - formally typeset
Open AccessJournal ArticleDOI

Classifier design for computer-aided diagnosis: effects of finite sample size on the mean performance of classical and neural network classifiers.

Reads0
Chats0
TLDR
The understanding of the performance of the classifiers under the constraint of a finite design sample size is expected to facilitate the selection of a proper classifier for a given classification task and the design of an efficient resampling scheme.
Abstract
Classifier design is one of the key steps in the development of computer-aided diagnosis(CAD) algorithms. A classifier is designed with case samples drawn from the patient population. Generally, the sample size available for classifier design is limited, which introduces variance and bias into the performance of the trained classifier, relative to that obtained with an infinite sample size. For CAD applications, a commonly used performance index for a classifier is the area, A z , under the receiver operating characteristic (ROC) curve. We have conducted a computer simulation study to investigate the dependence of the mean performance, in terms of A z , on design sample size for a linear discriminant and two nonlinear classifiers, the quadratic discriminant and the backpropagation neural network (ANN). The performances of the classifiers were compared for four types of class distributions that have specific properties: multivariate normal distributions with equal covariance matrices and unequal means, unequal covariance matrices and unequal means, and unequal covariance matrices and equal means, and a feature space where the two classes were uniformly distributed in disjoint checkerboard regions. We evaluated the performances of the classifiers in feature spaces of dimensionality ranging from 3 to 15, and design sample sizes from 20 to 800 per class. The dependence of the resubstitution and hold-out performance on design (training) sample size (N t ) was investigated. For multivariate normal class distributions with equal covariance matrices, the linear discriminant is the optimal classifier. It was found that its A z - versus -1/N t curves can be closely approximated by linear dependences over the range of sample sizes studied. In the feature spaces with unequal covariance matrices where the quadratic discriminant is optimal, the linear discriminant is inferior to the quadratic discriminant or the ANN when the design sample size is large. However, when the design sample is small, a relatively simple classifier, such as the linear discriminant or an ANN with very few hidden nodes, may be preferred because performance bias increases with the complexity of the classifier. In the regime where the classifier performance is dominated by the 1/N t term, the performance in the limit of infinite sample size can be estimated as the intercept (1/N t =0) of a linear regression of A z versus 1/N t . The understanding of the performance of the classifiers under the constraint of a finite design sample size is expected to facilitate the selection of a proper classifier for a given classification task and the design of an efficient resampling scheme.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Deep learning in medical imaging and radiation therapy.

TL;DR: The general principles of DL and convolutional neural networks are introduced, five major areas of application of DL in medical imaging and radiation therapy are surveyed, common themes are identified, methods for dataset expansion are discussed, and lessons learned, remaining challenges, and future directions are summarized.
Journal ArticleDOI

A review of evidence of health benefit from artificial neural networks in medical intervention

TL;DR: The purpose of this review is to assess the evidence of healthcare benefits involving the application of artificial neural networks to the clinical functions of diagnosis, prognosis and survival analysis, in the medical domains of oncology, critical care and cardiovascular medicine.
Journal ArticleDOI

Machine Learning Testing: Survey, Landscapes and Horizons

TL;DR: A comprehensive survey of machine learning testing can be found in this article, which covers 138 papers on testing properties (e.g., correctness, robustness, and fairness), testing components (i.e., the data, learning program, and framework), testing workflow, and application scenarios.
Journal ArticleDOI

A study on several Machine-learning methods for classification of Malignant and benign clustered microcalcifications

TL;DR: This paper investigated several state-of-the-art machine-learning methods for automated classification of clustered microcalcifications (MCs), and formulated differentiation of malignant from benign MCs as a supervised learning problem, and applied these learning methods to develop the classification algorithm.
Journal ArticleDOI

Anniversary paper: History and status of CAD and quantitative image analysis: the role of Medical Physics and AAPM.

TL;DR: The future includes the combined research efforts from physicists working in CAD with those working on quantitative imaging systems to readily yield information on morphology, function, molecular structure, and more-from animal imaging research to clinical patient care.
References
More filters
Book

Introduction to Statistical Pattern Recognition

TL;DR: This completely revised second edition presents an introduction to statistical pattern recognition, which is appropriate as a text for introductory courses in pattern recognition and as a reference book for workers in the field.
Journal ArticleDOI

Effects of sample size in classifier design

TL;DR: The effect of finite sample-size on parameter estimates and their subsequent use in a family of functions are discussed, and an empirical approach is presented to enable asymptotic performance to be accurately estimated using a very small number of samples.
Journal ArticleDOI

On Dimensionality, Sample Size, Classification Error, and Complexity of Classification Algorithm in Pattern Recognition

TL;DR: Four classification algorithms-discriminant functions when classifying individuals into two multivariate populations are compared and it is shown that the classification error EPN depends on the structure of a classification algorithm, asymptotic probability of misclassification P¿, and the ratio of learning sample size N to dimensionality p:N/p.

On Dimensionality, SampleSize, Classification Error, andComplexity ofClassification Algorithm in Pattern Recognition

TL;DR: In this article, the authors compared four classification algorithms-discriminant functions when classifying individuals into two multivariate populations and derived analytical formulas for the expected probability of misclassification EPN.
Journal ArticleDOI

Variance-component modeling in the analysis of receiver operating characteristic index estimates

TL;DR: The variance-component approach is a useful statistical tool for modeling different sources of variation that contribute to the overall variance of ROC data and index estimates derived from those data.
Related Papers (5)