A Modified Principal Component Technique Based on the LASSO
read more
Citations
Principal component analysis: a review and recent developments
Applied Predictive Modeling
Sparse Principal Component Analysis
Regression shrinkage and selection via the lasso: a retrospective
Online Learning for Matrix Factorization and Sparse Coding
References
Regression Shrinkage and Selection via the Lasso
Principal Component Analysis
Global Optimization of Statistical Functions with Simulated Annealing
Penalized Regressions: The Bridge versus the Lasso
Better subset regression using the nonnegative garrote
Related Papers (5)
Frequently Asked Questions (14)
Q2. What is the explanation of this anomaly?
The explanation of this anomaly is in the projected gradient method used for numerical solution of the problem, which approximates the LASSO constraint with a certain smooth function and thus the zero-loadings produced may be also approximate.
Q3. What is the extra constraint in the technique?
In their technique the extra constraint is in the form of a bound on the sum of the absolute values of the loadings in that component.
Q4. What is the simplest way to construct a correlation matrix?
Given a vector l of positive real numbers and an orthogonal matrix A, the authors can attempt to nd a covariance matrix or correlation matrix whose eigenvalues are the elements of l, and whose eigenvectorsare the column of A.
Q5. How is the solution of the modi ed maximization problem found?
The solution of this modi ed maximization problem is then found as an ascent gradient vector ow onto the p-dimensional unit sphere following the standard projected gradient formalism (Chu and Trenda lov 2001; Helmke and Moore 1994).
Q6. What are the other criteria used in the simulations?
Other rotation criteria, such as quartimax, can in theory nd uniform vectors of loadings, but they were tried and also found to be unsuccessful in their simulations.
Q7. How many parameters of the projected gradient method need to be dened?
To achieve an appropriate solution, a number of parameters of the projected gradient method (e.g., starting points, absolute and relative tolerances) also need to be de ned.
Q8. What is the equivalent way of deriving LASSO estimates?
An equivalent way of deriving LASSO estimates is to minimize the residual sum of squares with the addition of a penalty function based on pj = 1 j jj.
Q9. Why is the PCA algorithm slower than SCoTLASS?
This is because SCoTLASS is implemented subject to an extra restriction on PCA and the authors lose the advantage of calculation via the singular value decomposition which makes the PCA algorithm fast.
Q10. How many zeros are in the solution with t = 1:50?
One can see that the solution with t = 1:50 contains a total of 56 loadings with less than 0.005 magnitude, compared to 42 in the case t = 1:75.
Q11. Why is it preferred to rotated principal components?
It is preferred in many respects to rotated principal components, as a means of simplifying interpretation compared to principal component analysis.
Q12. What is the alternative approach to the simulation study?
An alternative approach to the simulation study would be to replace the near-zero loadings by exact zeros and the nearly equal loadingsby exact equalities.
Q13. Why is the problem in the regression tree?
These problems may occur due to the instability of the regression coefcients in the presence of collinearity, or simply because of the large number of variables included in the regression equation.
Q14. What is the difference between the PC and the SCoTLASS?
SCoTLASS looks for simple sources of variation and, likePCA, aims for highvariance,but becauseof simplicityconsiderationsthesimpli ed components can, in theory, be moderately different from the PCs.