scispace - formally typeset
Journal ArticleDOI

On Comparing Classifiers: Pitfalls toAvoid and a Recommended Approach

Steven L. Salzberg
- 31 Jan 1997 - 
- Vol. 1, Iss: 3, pp 317-328
Reads0
Chats0
TLDR
Several phenomena that can, if ignored, invalidate an experimental comparison and the conclusions that follow apply not only to classification, but to computational experiments in almost any aspect of data mining.

Content maybe subject to copyright    Report

Citations
More filters
Journal Article

Statistical Comparisons of Classifiers over Multiple Data Sets

TL;DR: A set of simple, yet safe and robust non-parametric tests for statistical comparisons of classifiers is recommended: the Wilcoxon signed ranks test for comparison of two classifiers and the Friedman test with the corresponding post-hoc tests for comparisons of more classifiers over multiple data sets.
Posted Content

Principles of data mining

TL;DR: This paper gives a lightning overview of data mining and its relation to statistics, with particular emphasis on tools for the detection of adverse drug reactions.
Journal ArticleDOI

Neural networks for classification: a survey

TL;DR: The issues of posterior probability estimation, the link between neural and conventional classifiers, learning and generalization tradeoff in classification, the feature variable selection, as well as the effect of misclassification costs are examined.
Journal ArticleDOI

Logistic regression and artificial neural network classification models: a methodology review

TL;DR: In this paper, the differences and similarities of these models from a technical point of view, and compare them with other machine learning algorithms are summarized and compared using a set of quality criteria for logistic regression and artificial neural networks.
Journal ArticleDOI

Querying and mining of time series data: experimental comparison of representations and distance measures

TL;DR: An extensive set of time series experiments are conducted re-implementing 8 different representation methods and 9 similarity measures and their variants and testing their effectiveness on 38 time series data sets from a wide variety of application domains to provide a unified validation of some of the existing achievements.
References
More filters
Journal ArticleDOI

Bayesian Model Selection in Social Research

TL;DR: In this article, a Bayesian approach to hypothesis testing, model selection, and accounting for model uncertainty is presented, which is straightforward through the use of the simple and accurate BIC approximation, and it can be done using the output from standard software.
Proceedings Article

Multi-Interval Discretization of Continuous-Valued Attributes for Classification Learning

TL;DR: This paper addresses the use of the entropy minimization heuristic for discretizing the range of a continuous-valued attribute into multiple intervals.
Journal ArticleDOI

Very Simple Classification Rules Perform Well on Most Commonly Used Datasets

TL;DR: On most datasets studied, the best of very simple rules that classify examples on the basis of a single attribute is as accurate as the rules induced by the majority of machine learning systems.