scispace - formally typeset
Open AccessJournal ArticleDOI

Instance-Based Learning Algorithms

TLDR
This paper describes how storage requirements can be significantly reduced with, at most, minor sacrifices in learning rate and classification accuracy and extends the nearest neighbor algorithm, which has large storage requirements.
Abstract
Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances to solve incremental learning tasks. In this paper, we describe a framework and methodology, called instance-based learning, that generates classification predictions using only specific instances. Instance-based learning algorithms do not maintain a set of abstractions derived from specific instances. This approach extends the nearest neighbor algorithm, which has large storage requirements. We describe how storage requirements can be significantly reduced with, at most, minor sacrifices in learning rate and classification accuracy. While the storage-reducing algorithm performs well on several real-world databases, its performance degrades rapidly with the level of attribute noise in training instances. Therefore, we extended it with a significance test to distinguish noisy instances. This extended algorithm's performance degrades gracefully with increasing noise levels and compares favorably with a noise-tolerant decision tree algorithm.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Software measurement data reduction using ensemble techniques

TL;DR: This study utilized 16 real-world software measurement data sets of different sizes and built 54,400 classification models using four well known classifiers and the main conclusion is that ensembles of very few rankers are very effective and even better than ensembled of many or all rankers.
Journal Article

Learning efficient disambiguation

TL;DR: The dissertation addresses "grammar and model specialization" and presents a new framework, the Ambiguity-Reduction Specialization (ARS) framework, that formulates the necessary and sufficient conditions for successful specialization.
Book ChapterDOI

Keep It Simple: A Case-Base Maintenance Policy Based on Clustering and Information Theory

TL;DR: A case-base maintenance method that avoids building sophisticated structures around a case base or perform complex operations on a casebase, and is based on a decision forest built with the attributes that are obtained through an innovative modification of the ID3 algorithm.
Proceedings ArticleDOI

Machine Learning-Based Runtime Scheduler for Mobile Offloading Framework

TL;DR: This study considers 19 different machine learning algorithms and four workloads, with a dataset obtained through the deployment of an Android-based remote offloading framework prototype on actual mobile and cloud resources, and uses Instance-Based Learning to evaluate an online adaptive scheduler for mobile offloading.

Predicting phrase breaks with memory-based learning

TL;DR: It is shown that a simple memory-based learning algorithm that uses only minimal context and information outperforms the HMM approach, in terms of precision and recall.
References
More filters
Journal ArticleDOI

Induction of Decision Trees

J. R. Quinlan
- 25 Mar 1986 - 
TL;DR: In this paper, an approach to synthesizing decision trees that has been used in a variety of systems, and it describes one such system, ID3, in detail, is described, and a reported shortcoming of the basic algorithm is discussed.
Book

Classification and regression trees

Leo Breiman
TL;DR: The methodology used to construct tree structured rules is the focus of a monograph as mentioned in this paper, covering the use of trees as a data analysis method, and in a more mathematical framework, proving some of their fundamental properties.
Journal ArticleDOI

Nearest neighbor pattern classification

TL;DR: The nearest neighbor decision rule assigns to an unclassified sample point the classification of the nearest of a set of previously classified points, so it may be said that half the classification information in an infinite sample set is contained in the nearest neighbor.