论文标题
用超级线性收敛重建基于内核的机器学习力场
Reconstructing Kernel-based Machine Learning Force Fields with Super-linear Convergence
论文作者
论文摘要
内核机在量子化学领域持续持续进展。特别是,事实证明,它们在力场重建的低数据表中取得了成功。这是因为可以将由于物理对称性引起的许多均等和不稳定性纳入内核函数,以补偿更大的数据集。到目前为止,内核机器的可伸缩性受到训练点数中的二次记忆和立方运行时的复杂性的阻碍。众所周知,迭代的Krylov子空间求解器可以克服这些负担,但它们的收敛性至关重要地依赖于有效的预处理,而实际上难以捉摸。有效的预处理需要以计算上便宜且数值稳健的方式部分预见学习问题。在这里,我们考虑了基于原始内核矩阵的更复杂的低级别近似值的Nyström-type方法的广泛类别,每种方法都提供了一组不同的计算权衡。所有考虑的方法旨在识别诱导(内核)列的代表性子集以近似主要的内核光谱。
Kernel machines have sustained continuous progress in the field of quantum chemistry. In particular, they have proven to be successful in the low-data regime of force field reconstruction. This is because many equivariances and invariances due to physical symmetries can be incorporated into the kernel function to compensate for much larger datasets. So far, the scalability of kernel machines has however been hindered by its quadratic memory and cubical runtime complexity in the number of training points. While it is known, that iterative Krylov subspace solvers can overcome these burdens, their convergence crucially relies on effective preconditioners, which are elusive in practice. Effective preconditioners need to partially pre-solve the learning problem in a computationally cheap and numerically robust manner. Here, we consider the broad class of Nyström-type methods to construct preconditioners based on successively more sophisticated low-rank approximations of the original kernel matrix, each of which provides a different set of computational trade-offs. All considered methods aim to identify a representative subset of inducing (kernel) columns to approximate the dominant kernel spectrum.