New Support Vector Algorithms
Reads0
Chats0
TLDR
A new class of support vector algorithms for regression and classification that eliminates one of the other free parameters of the algorithm: the accuracy parameter in the regression case, and the regularization constant C in the classification case.Abstract:
We propose a new class of support vector algorithms for regression and classification. In these algorithms, a parameter ν lets one effectively control the number of support vectors. While this can be useful in its own right, the parameterization has the additional benefit of enabling us to eliminate one of the other free parameters of the algorithm: the accuracy parameter epsilon in the regression case, and the regularization constant C in the classification case. We describe the algorithms, give some theoretical results concerning the meaning and the choice of ν, and report experimental results.read more
Citations
More filters
Journal ArticleDOI
Experimentally optimal ν in support vector regression for different noise models and parameter settings
TL;DR: Good agreement is found with the values that had previously been predicted by a theoretical argument based on a the asymptotic efficiency of a simplified model of SV regression of Support Vector regression.
Journal ArticleDOI
GIS-based groundwater potential mapping using artificial neural network and support vector machine models: the case of Boryeong city in Korea
TL;DR: Groundwater productivity-potential was analysed using the data mining models of an artificial neural network and a support vector machine in Boryeong city, Korea and it proves that the ANN and SVM models will be highly conducive to developing useful groundwater resources.
Posted Content
Composite Binary Losses
TL;DR: This paper characterised when margin losses can be proper composite losses, explicitly showed how to determine a symmetric loss in full from half of one of its partial losses, and gave a complete characterisation of the relationship between proper losses and ''classification calibrated'' losses.
Journal ArticleDOI
Soft Margin Multiple Kernel Learning
Xinxing Xu,Ivor W. Tsang,Dong Xu +2 more
TL;DR: A novel soft margin perspective for MKL is presented, which introduces an additional slack variable called kernel slack variable to each quadratic constraint of MKL, which corresponds to one support vector machine model using a single base kernel.
Proceedings Article
A geometric interpretation of ν-SVM classifiers
TL;DR: In this article, the authors show that the recently proposed variant of the Support Vector machine (SVM) algorithm, known as ν-SVM, can be interpreted as a maximal separation between subsets of the convex hulls of the data.
References
More filters
Book
The Nature of Statistical Learning Theory
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Journal ArticleDOI
Support-Vector Networks
Corinna Cortes,Vladimir Vapnik +1 more
TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Book
Matrix Analysis
Roger A. Horn,Charles R. Johnson +1 more
TL;DR: In this article, the authors present results of both classic and recent matrix analyses using canonical forms as a unifying theme, and demonstrate their importance in a variety of applications, such as linear algebra and matrix theory.
Journal ArticleDOI
A Tutorial on Support Vector Machines for Pattern Recognition
TL;DR: There are several arguments which support the observed high accuracy of SVMs, which are reviewed and numerous examples and proofs of most of the key theorems are given.