scispace - formally typeset
Search or ask a question
Topic

Statistical learning theory

About: Statistical learning theory is a research topic. Over the lifetime, 1618 publications have been published within this topic receiving 158033 citations.


Papers
More filters
Proceedings ArticleDOI
29 Jul 2010
TL;DR: The learning rates of the estimator in reconstructing such class of functions also exploiting recent advances in statistical learning theory are characterized and Monte Carlo studies are used to illustrate the definite advantages of this new nonparametric approach over classical parametric prediction error methods in terms of accuracy in impulse responses reconstruction.
Abstract: A new nonparametric paradigm to model identification has been recently introduced in the literature. Instead of adopting finite-dimensional models of the system transfer function, the system impulse response is searched for within an infinite dimensional space using regularization theory. The method exploits the so called stable spline kernels which are associated with hypothesis spaces embedding information on both regularity and stability of the impulse response. In this paper, the potentiality of this approach is studied with respect to the reconstruction of sums of exponentials. In particular, first, we characterize the learning rates of our estimator in reconstructing such class of functions also exploiting recent advances in statistical learning theory. Then, we use Monte Carlo studies to illustrate the definite advantages of this new nonparametric approach over classical parametric prediction error methods in terms of accuracy in impulse responses reconstruction.

54 citations

Journal ArticleDOI
TL;DR: The SVM‐based reconstruction is used to develop time series forecasts for multiple lead times ranging from 2 weeks to several months and is able to extract the dynamics using only a few past observed data points out of the training examples.
Abstract: [1] The reconstruction of low-order nonlinear dynamics from the time series of a state variable has been an active area of research in the last decade. The 154 year long, biweekly time series of the Great Salt Lake volume has been analyzed by many researchers from this perspective. In this study, we present the application of a powerful state space reconstruction methodology using the method of support vector machines (SVM) to this data set. SVM are machine learning systems that use a hypothesis space of linear functions in a kernel-induced higher-dimensional feature space. SVM are optimized by minimizing a bound on a generalized error (risk) measure rather than just the mean square error over a training set. Under Mercer's conditions on the kernels the corresponding optimization problems are convex; hence global optimal solutions can be readily computed. The SVM-based reconstruction is used to develop time series forecasts for multiple lead times ranging from 2 weeks to several months. Unlike previously reported methodologies, SVM are able to extract the dynamics using only a few past observed data points out of the training examples. The reliability of the algorithm in learning and forecasting the dynamics is tested using split sample sensitivity analysis, with a particular interest in forecasting extreme states. Efforts are also made to assess variations in predictability as a function of initial conditions and as a function of the degree of extrapolation from the state space used for learning the model.

53 citations

Journal ArticleDOI
TL;DR: In this article, a robust learning method developed based on statistical learning theory namely least squares support vector machine (LSSVM) has been employed for calculating the freezing point depression (FPD) of different electrolyte solutions.

53 citations

Proceedings ArticleDOI
14 Jan 2016
TL;DR: Concepts from statistical and online learning theory are adapted to reason about application-specific algorithm selection, and dimension notions from statistical learning theory, historically used to measure the complexity of classes of binary- and real-valued functions, are relevant in a much broader algorithmic context.
Abstract: The best algorithm for a computational problem generally depends on the "relevant inputs," a concept that depends on the application domain and often defies formal articulation. While there is a large literature on empirical approaches to selecting the best algorithm for a given application domain, there has been surprisingly little theoretical analysis of the problem.This paper adapts concepts from statistical and online learning theory to reason about application-specific algorithm selection. Our models capture several state-of-the-art empirical and theoretical approaches to the problem, ranging from self-improving algorithms to empirical performance models, and our results identify conditions under which these approaches are guaranteed to perform well. We present one framework that models algorithm selection as a statistical learning problem, and our work here shows that dimension notions from statistical learning theory, historically used to measure the complexity of classes of binary- and real-valued functions, are relevant in a much broader algorithmic context. We also study the online version of the algorithm selection problem, and give possibility and impossibility results for the existence of no-regret learning algorithms.

53 citations

Journal ArticleDOI
TL;DR: The Algorithmic Stability framework is relied on to prove learning bounds for the unsupervised concept drift detection on data streams, and the Plover algorithm is designed to detect drifts using different measure functions, such as Statistical Moments and the Power Spectrum.
Abstract: Motivated by the Statistical Learning Theory (SLT), which provides a theoretical framework to ensure when supervised learning algorithms generalize input data, this manuscript relies on the Algorithmic Stability framework to prove learning bounds for the unsupervised concept drift detection on data streams. Based on such proof, we also designed the Plover algorithm to detect drifts using different measure functions, such as Statistical Moments and the Power Spectrum. In this way, the criterion for issuing data changes can also be adapted to better address the target task. From synthetic and real-world scenarios, we observed that each data stream may require a different measure function to identify concept drifts, according to the underlying characteristics of the corresponding application domain. In addition, we discussed about the differences of our approach against others from literature, and showed illustrative results confirming the usefulness of our proposal.

53 citations


Network Information
Related Topics (5)
Artificial neural network
207K papers, 4.5M citations
86% related
Cluster analysis
146.5K papers, 2.9M citations
82% related
Feature extraction
111.8K papers, 2.1M citations
81% related
Optimization problem
96.4K papers, 2.1M citations
80% related
Fuzzy logic
151.2K papers, 2.3M citations
79% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20239
202219
202159
202069
201972
201847