scispace - formally typeset
Open Access

Kernel based learning methods for pattern and feature analysis

Zhili Wu
TLDR
The Rival Penalized Competitive Learning (RPCL) is transformed from data space to feature space for automatic clustering and spectral analysis of kernel matrices is used to address the seeds initialization problem associated with RPCL.
Abstract
Kernel-based learning methods (kernel methods) have significant influences on recent development of machine learning research. This thesis is on devising and improving kernel methods, and on applying them to pattern and feature analysis. Part of our research focuses on improving Support Vector Machines (SVMs). In solving SVMs, we find that some proposed cache policies for sequential minimal optimization (SMO) result in low efficiency. A better strategy is to cache gradients for all vectors frequently checked. Moreover, we propose a strategy that utilizes the nearest neighboring vectors to speed up the convergence of SMO. We also suggest the use of Hadamard codes for multiclass label prediction by SVMs. We prove that the Hadamard codes are optimal in correcting the wrong labels predicted by base classifiers. Furthermore, we design a newsummation of exponential (SoE) kernel for solving regression tasks with missing values. We show SoE kernels are admissible to kernel conditions and insensitive to missing values. This thesis also deals with unsupervised and semi-supervised kernel methods. Specifically, we transform the Rival Penalized Competitive Learning (RPCL) from data space to feature space for automatic clustering. In addition, we use spectral analysis of kernel matrices to address the seeds initialization problem associated with RPCL. We also improve the SVM-based feature selection in a semi-supervised manner by utilizing both labeled and unlabeled data. The new feature selection method exhibits good performance in solving feature selection benchmark problems.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Optimization of SVM MultiClass by Particle Swarm (PSO-SVM)

TL;DR: The results obtained show that approach PSO-SVM gives a better classification in terms of accuracy even though the execution time is increased, which makes it possible to optimize the performance of classifier SVM (Separating with Vast Margin).
Book ChapterDOI

Semi-supervised SVM-based Feature Selection for Cancer Classification using Microarray Gene Expression Data

TL;DR: This paper aims to propose a semi-supervised SVM-based feature selection S$$^3$$VM-FS, which simultaneously exploit the knowledge from unlabelled and labelled data, which achieves the higher accuracy yet requires shorter processing time compared with the well-known supervised method.
Journal Article

Adaptive dimension reduction for clustering high dimensional data

TL;DR: Clustering analysis performed on highly overlapped Gaussians, DNA gene expression profiles and Internet newsgroups demonstrate the effectiveness of the proposed algorithm by repeated dimension reductions such that K-means or EM are performed only in very low dimensions.
Book ChapterDOI

Improving support vector machine using a stochastic local search for classification in datamining

TL;DR: The computational experiments show that the proposed SVM+SLS provides competitive results and finds high quality solutions to the classification problem in Datamining.
Journal Article

Kernel matrix completion by semidefinite programming

TL;DR: It is argued that semidefinite programming provides an interesting convex optimisation framework for machine learning in general and for kernel-machines in particular.
References
More filters
Journal ArticleDOI

LIBSVM: A library for support vector machines

TL;DR: Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.

Statistical learning theory

TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Journal ArticleDOI

A Tutorial on Support Vector Machines for Pattern Recognition

TL;DR: There are several arguments which support the observed high accuracy of SVMs, which are reviewed and numerous examples and proofs of most of the key theorems are given.
Journal ArticleDOI

Nonlinear dimensionality reduction by locally linear embedding.

TL;DR: Locally linear embedding (LLE) is introduced, an unsupervised learning algorithm that computes low-dimensional, neighborhood-preserving embeddings of high-dimensional inputs that learns the global structure of nonlinear manifolds.