Structural Minimax Probability Machine
Reads0
Chats0
TLDR
This paper uses two finite mixture models to capture the structural information of the data from binary classification and proposes a structural MPM, which can be interpreted as a large margin classifier and can be transformed to support vector machine and maxi–min margin machine under certain special conditions.Abstract:
Minimax probability machine (MPM) is an interesting discriminative classifier based on generative prior knowledge. It can directly estimate the probabilistic accuracy bound by minimizing the maximum probability of misclassification. The structural information of data is an effective way to represent prior knowledge, and has been found to be vital for designing classifiers in real-world problems. However, MPM only considers the prior probability distribution of each class with a given mean and covariance matrix, which does not efficiently exploit the structural information of data. In this paper, we use two finite mixture models to capture the structural information of the data from binary classification. For each subdistribution in a finite mixture model, only its mean and covariance matrix are assumed to be known. Based on the finite mixture models, we propose a structural MPM (SMPM). SMPM can be solved effectively by a sequence of the second-order cone programming problems. Moreover, we extend a linear model of SMPM to a nonlinear model by exploiting kernelization techniques. We also show that the SMPM can be interpreted as a large margin classifier and can be transformed to support vector machine and maxi–min margin machine under certain special conditions. Experimental results on both synthetic and real-world data sets demonstrate the effectiveness of SMPM.read more
Citations
More filters
Journal ArticleDOI
The Modified HZ Conjugate Gradient Algorithm for Large-Scale Nonsmooth Optimization.
TL;DR: Numerical results show that the presented methods can be better efficiency for large-scale nonsmooth problems, and several problems are tested.
Journal ArticleDOI
Double feature selection algorithm based on low-rank sparse non-negative matrix factorization
TL;DR: Experimental results show that NMF-LRSR is more effective than the other six feature selection algorithms and the double feature selection theory is used to this paper, which makes the result of feature selection more accurate.
Journal ArticleDOI
The Bibliometric Analysis on Finite Mixture Model
TL;DR: In this article , a review of finite mixture model literature via bibliometric analysis, focusing on the trend and link between finite mixture models studies is presented, and the results show that there is an increasing trend of annual publication on FMM studies.
Journal ArticleDOI
Visual tracking using global sparse coding and local convolutional features
TL;DR: A robust tracking algorithm by representing the target at two levels: global and local levels is presented, which outperforms several state-of-the-art models.
Journal ArticleDOI
A Grouping Particle Swarm Optimizer with Personal-Best-Position Guidance for Large Scale Optimization
TL;DR: A variant of PSO named Grouping PSO with Personal-Best-Position with Guidance which maintains the population diversity by preserving the diversity of exemplars and exhibits a competitive performance to maintain population diversity.
References
More filters
Journal ArticleDOI
A tutorial on hidden Markov models and selected applications in speech recognition
TL;DR: In this paper, the authors provide an overview of the basic theory of hidden Markov models (HMMs) as originated by L.E. Baum and T. Petrie (1966) and give practical details on methods of implementation of the theory along with a description of selected applications of HMMs to distinct problems in speech recognition.
Journal ArticleDOI
Hierarchical Grouping to Optimize an Objective Function
TL;DR: In this paper, a procedure for forming hierarchical groups of mutually exclusive subsets, each of which has members that are maximally similar with respect to specified characteristics, is suggested for use in large-scale (n > 100) studies when a precise optimal solution for a specified number of groups is not practical.
Book ChapterDOI
Neural Networks for Pattern Recognition
Suresh Kothari,Heekuck Oh +1 more
TL;DR: The chapter discusses two important directions of research to improve learning algorithms: the dynamic node generation, which is used by the cascade correlation algorithm; and designing learning algorithms where the choice of parameters is not an issue.
Journal ArticleDOI
Using SeDuMi 1.02, a MATLAB toolbox for optimization over symmetric cones
TL;DR: This paper describes how to work with SeDuMi, an add-on for MATLAB, which lets you solve optimization problems with linear, quadratic and semidefiniteness constraints by exploiting sparsity.
Journal ArticleDOI
A comparison of methods for multiclass support vector machines
Hsu Chih-Wei,Chih-Jen Lin +1 more
TL;DR: Decomposition implementations for two "all-together" multiclass SVM methods are given and it is shown that for large problems methods by considering all data at once in general need fewer support vectors.
Related Papers (5)
A Robust Regularization Path Algorithm for $\nu $ -Support Vector Classification
Bin Gu,Victor S. Sheng +1 more