scispace - formally typeset
Open AccessJournal ArticleDOI

Reconstructing Kernel-based Machine Learning Force Fields with Super-linear Convergence

Reads0
Chats0
TLDR
In this article , the authors consider the broad class of Nyström-type methods to construct preconditioners based on successively more sophisticated low-rank approximations of the original kernel matrix, each of which provides a different set of computational tradeoffs.
Abstract
Kernel machines have sustained continuous progress in the field of quantum chemistry. In particular, they have proven to be successful in the low-data regime of force field reconstruction. This is because many equivariances and invariances due to physical symmetries can be incorporated into the kernel function to compensate for much larger data sets. So far, the scalability of kernel machines has however been hindered by its quadratic memory and cubical runtime complexity in the number of training points. While it is known that iterative Krylov subspace solvers can overcome these burdens, their convergence crucially relies on effective preconditioners, which are elusive in practice. Effective preconditioners need to partially presolve the learning problem in a computationally cheap and numerically robust manner. Here, we consider the broad class of Nyström-type methods to construct preconditioners based on successively more sophisticated low-rank approximations of the original kernel matrix, each of which provides a different set of computational trade-offs. All considered methods aim to identify a representative subset of inducing (kernel) columns to approximate the dominant kernel spectrum.

read more

Citations
More filters
Journal ArticleDOI

Robust, randomized preconditioning for kernel ridge regression

TL;DR: In this article , two randomized preconditioning techniques for robustly solving kernel ridge regression (KRR) problems with a medium to large number of data points were introduced. But their performance was limited to a restricted version of the KRR problem with a cost of O((N + k^2) k \log k) ).
References
More filters
Journal ArticleDOI

Generalized Gradient Approximation Made Simple

TL;DR: A simple derivation of a simple GGA is presented, in which all parameters (other than those in LSD) are fundamental constants, and only general features of the detailed construction underlying the Perdew-Wang 1991 (PW91) GGA are invoked.
Book

Iterative Methods for Sparse Linear Systems

Yousef Saad
TL;DR: This chapter discusses methods related to the normal equations of linear algebra, and some of the techniques used in this chapter were derived from previous chapters of this book.
Journal ArticleDOI

Methods of Conjugate Gradients for Solving Linear Systems

TL;DR: An iterative algorithm is given for solving a system Ax=k of n linear equations in n unknowns and it is shown that this method is a special case of a very general method which also includes Gaussian elimination.
Journal ArticleDOI

Generalized neural-network representation of high-dimensional potential-energy surfaces.

TL;DR: A new kind of neural-network representation of DFT potential-energy surfaces is introduced, which provides the energy and forces as a function of all atomic positions in systems of arbitrary size and is several orders of magnitude faster than DFT.
Related Papers (5)