scispace - formally typeset
Open AccessJournal ArticleDOI

Implementation of algorithms for tuning parameters in regularized least squares problems in system identification

Tianshi Chen, +1 more
- 01 Jul 2013 - 
- Vol. 49, Iss: 7, pp 2213-2220
Reads0
Chats0
TLDR
This work investigates implementation of algorithms for solving the hyper-parameter estimation problem that can deal with both large data sets and possibly ill-conditioned computations and proposes a QR factorization based matrix-inversion-free algorithm to evaluate the cost function in an efficient and accurate way.
About
This article is published in Automatica.The article was published on 2013-07-01 and is currently open access. It has received 126 citations till now. The article focuses on the topics: QR decomposition & System identification.

read more

Citations
More filters
Journal ArticleDOI

Survey Kernel methods in system identification, machine learning and function estimation: A survey

TL;DR: A survey of kernel-based regularization and its connections with reproducing kernel Hilbert spaces and Bayesian estimation of Gaussian processes to demonstrate that learning techniques tailored to the specific features of dynamic systems may outperform conventional parametric approaches for identification of stable linear systems.
Journal ArticleDOI

System Identification Via Sparse Multiple Kernel-Based Regularization Using Sequential Convex Optimization Techniques

TL;DR: A multiple kernel-based regularization method is proposed to handle model estimation and structure detection with short data records and it is shown that the locally optimal solutions lead to good performance for randomly generated starting points.
Journal ArticleDOI

A shift in paradigm for system identification

TL;DR: The purpose of this contribution is to provide an accessible account of the main ideas and results of kernel-based regularisation methods for system identification.
Journal ArticleDOI

On kernel design for regularized LTI system identification

TL;DR: This paper proposes two methods to design kernels: one is from a machine learning perspective and the other one is a system theory perspective, and provides analysis results for both methods, which enhances the understanding for the existing kernels but also directs the design of new kernels.
Journal ArticleDOI

Maximum entropy properties of discrete-time first-order stable spline kernel

TL;DR: In this article, the authors formulate the exact maximum entropy problem solved by the first order stable spline (SS-1) kernel without Gaussian and uniform sampling assumptions and derive the special structure of the SS-1 kernel under general sampling assumption.
References
More filters
Book

System Identification: Theory for the User

Lennart Ljung
TL;DR: Das Buch behandelt die Systemidentifizierung in dem theoretischen Bereich, der direkte Auswirkungen auf Verstaendnis and praktische Anwendung der verschiedenen Verfahren zur IdentifIZierung hat.
Book

Applied Regression Analysis

TL;DR: In this article, the Straight Line Case is used to fit a straight line by least squares, and the Durbin-Watson Test is used for checking the straight line fit.
Book

Gaussian Processes for Machine Learning

TL;DR: The treatment is comprehensive and self-contained, targeted at researchers and students in machine learning and applied statistics, and deals with the supervised learning problem for both regression and classification.
Book

Linear statistical inference and its applications

TL;DR: Algebra of Vectors and Matrices, Probability Theory, Tools and Techniques, and Continuous Probability Models.
Book

System identification

Frequently Asked Questions (13)
Q1. What have the authors contributed in "Implementation of algorithms for tuning parameters in regularized least squares problems in system identification" ?

There is recently a trend to study linear system identification with high order finite impulse response ( FIR ) models using the regularized least-squares approach. 

When seeking efficient algorithms to compute the costfunction in (6) for very large N and N À n, the numerical accuracy depends on the conditioning and the magnitude of the matrices P(α) and ΦNΦTN . 

For the impulse response estimation problem, the matrixAT A = σ2In +LT ΦNΦTNL (17)can be ill-conditioned due to the following two problems:• 

The command fmincon is used here to solve the nonconvex optimization problem (6) with the trust region reflective algorithm selected. 

For each data set, the authors aim to estimate FIR model (3) with n = 125 using the regularized least squares (5) including the empirical Bayes method (6). 

For the “DC” kernel (31c), further assume 0.72≤ λ < 1 and−0.99≤ ρ ≤ 0.99 so that the condition number of the DC kernel is smaller than 2.0×1020. • 

It is well-known, e.g. (Golub & Van Loan, 1996, Section 5), that the least squares problem (15) can be solved more accurately with the QR factorization method than the Cholesky factorization method, when the condition number of AT A defined in (17) is ill-conditioned. 

The linear least squares problem to estimate linear regressions is one of the most basic estimation problems, and there is an extensive literature around it, e.g. (Rao, 1973; Daniel & Wood, 1980; Draper & Smith, 1981). 

So one may question if the extra constraints can eventually cause performance loss in the regularized least squares estimate (5b). 

The magnitude of LT ΦNΦTNL can be very large if the element of α that controls the magnitude of P(α) is large, which is often the case for the stable spline kernel (Pillonetto & Nicolao, 2010; Pillonetto et al., 2011). 

It is the maximum likelihood method to estimate α from (4) under the (Bayesian) assumptions that θ is Gaussian with zero mean and covariance matrix P(α) and VN is Gaussian with zero mean and covariance matrix σ2IN . 

In this case, as discussed in (Chen, Ohlsson & Ljung, 2012), the authors can first estimate, with the Maximum Likelihood/Prediction Error Method e.g. (Ljung, 1999), a low-order “base-line model” that can take care of the dominating part of the impulse response. 

The authors then use regularized least squares (based on Algorithm 2) to estimate an FIR model with reasonably large n, which should capture the residual (fast decaying) dynamics (Chen, Ohlsson & Ljung, 2012).