1 Introduction
Kernel methods are a way to embed the input space into a high dimensional feature space, which enables learning over a richer and more complex hypothesis class. One of the most elementary applications of kernel methods is Kernel Ridge Regression (KRR). The problem setup can be defined in the following manner  let the training data be defined as where is the input domain and is the output domain, let , and a regularization parameter
, with the response for a given input x estimated as:
(1) 
where is the solution of the equation
(2) 
Remark 1.1.
is the kernel matrix defined by
KRR, while simple, is a powerful technique that achieves empirically good results. However, KRR comes with the drawback of having to compute the KRR estimator. The straightforward solution for the estimator requires inverting the matrix, which requires time, and is very restrictive for large datasets. Furthermore, holding the original kernel in memory requires storage, which by itself is prohibitive for large n. These drawbacks have driven research to focus on scaling up kernel based methods, mostly by approximation, yielding methods such as the Nyström method for lowrank kernel approximation, or the randomized construction of feature maps such as Random Fourier Features (RFF), originally proposed by Rahimi and Recht [31]. Another approach for the kernel ridge regression problem is finding a preconditioner [34, 11]. It was recently suggested to use RFF as a preconditioner for the kernel matrix [17] and then to use conjugate gradient to solve the regression problem. In this paper, a different approach for the preconditioner is presented, based on randomized matrix factorizations [18, 22, 33, 7]
. The suggested preconditioner combines between the interpolative decomposition, Nyströmapproximation and fast Gauss transform to achieve an efficient preconditioner that can be applied in
operations, i.e. linear in the number of data points.1.1 Related work
Building scalable and computationally efficient algorithms for kernel methods is an active research area. Some methods use low rank approximations to approximate the kernel directly, such as Nyström approximation or random Fourier features. This approach is very popular and has many variants where the main idea is to reduce the computational cost by low rank approximation while retaining reasonable model quality. Recent work on these approaches can be found in [20, 9, 1, 26]
. A different approach uses fast matrix vector multiplications, e.g Fast Gauss Transform or Fast Multipole methods to solve the regression using Krylov iterations
[32, 11, 34]. These methods reduce the computational cost of each Krylov iteration significantly, from to or even . However, when the kernel matrix is illconditioned, the number of iterations required might be huge. Another approach is to use a preconditioner to reduce the condition number of the kernel matrix and in so doing reduce the required number of Krylov iterations. These approaches are discussed in [17, 35, 8].1.2 Contributions
The paper proposes the use of randomized matrix decompositions [18] (specifically, the interpolative decomposition [16]), for constructing a preconditioner for the kernel matrix. The preconditioner can be used to solve the kernel ridge regression problem (2) using a substantially smaller number of iterations while maintaining high accuracy. Moreover, the preconditioner is designed in its structure to be both efficient in its application to the kernel matrix and strict in memory consumption. The paper presents a theoretical analysis of the condition number of the preconditioned matrix. Furthermore, for a Gaussian kernel, the paper provides a theoretical lower bound for the low rank approximation required to reach the desired condition number, this bound can be calculated directly from the data.
2 Preliminaries
2.1 Notation
Throughout the paper, matrices are indicated by capital letters and vectors and scalars by small letters. The norm when applied to vectors, refers to the standard Euclidean norm, i.e. and to the spectral norm when applied to matrices, i.e. . The norm applied to a vector with respect to a positive semidefinite matrix is defined
. Singular values are indicated in descending order, by
and eigenvalues by
also in descending order when they are real. on a square matrix indicates that is positive semidefinite and means .2.2 Interpolative and CUR decompositions
Interpolative decomposition (ID) [16] is a matrix factorization that can be applied to any matrix and defined as the following:
(3) 
where is a subset of the columns of and is an interpolation matrix with certain properties (such as small norm) that is used for reconstructing from . The columns of are computed by rank revealing factorization [15, 28, 24, 6, 19] in order to get an error bounded by a factor proportional to the the singular value of . In addition to the deterministic algorithm for computing the ID that is described in [16], there is a randomized version that starts by projecting
onto a low dimensional space using a random normally distributed matrix
[23, 18]. This property is used implicitly in Algorithm 1. The CUR decomposition [21, 10], is a pseudoskeleton decomposition [12, 27] that factors a matrix into three matrices, where and are subsets of the columns and rows of respectively, and is defined as the inverse matrix of the overlap between and . Most of the matrices discussed in the paper are symmetric, so the Nyström approximation will be used as a special case of the CUR decomposition. The following Lemma gives an error bound for the CUR decomposition using columns and rows selected by the interpolative decomposition:Lemma 2.1.
[36] Let , that satisfies and , where C,R are the k columns and k rows of A, respectively, and W, are the interpolation matrices from the interpolative decomposition. Suppose further that R is full rank, and that U is defined as . Then
(4) 
2.3 Fast Gauss Transform
Fast Gauss transform (FGT) is a variant of the fast multipole method (FMM) [13]. FMM was originally formulated as an efficient method for complex physical calculations, efficiently performing matrixvector products. FGT deals with the evaluation of the discrete Gauss transform:
(5) 
where represent the centers of the Gaussians, each of which has bandwidth , and are the weight coefficients for each Gaussian. Direct evaluation of the summation for a set of points, , is computationally costly in large scale scenarios, as it requires operations. FGT allows setting a desired degree of precision for the approximation of the Gaussian function, and reduces the complexity of the summation (5) to , accompanied by a constant factor, which grows exponentially with the dimension [14]
. To overcome this disadvantage, which makes plain FGT inapplicable to highdimensional data, the Fast Improved Gauss (FIG) transform
[37] uses tree data structures and an adaptive space subdivision technique, leading to reduction of the constant factor to asymptotically polynomial order. It is worth noting, that the computational complexity in is dependent on the data as well as the selection of [25].Definition 2.1.
Let represent the application of the FIG transform to compute equation (5), where and represent matrices of data points for which the kernel is generated, weighted by the vector . According to this notation, it holds that .
The algorithm discussed in the paper can be applied to any other kernels that can utilize fast matrixvector product such as other FMM functions. Another approach for fast matrixvector product for any kernel and based on nearest neighbors is described in [34].
3 Description of the algorithm
The proposed algorithm is based on conjugate gradient with a preconditioner designed not only to be effective in its ability to reduce the condition number, but also in its efficiency to be applied in terms of computational cost and memory consumption. In order to keep the time and storage complexity to a minimum , the proposed scheme uses the Fast Improved Gauss transform [37, 38] (“FIG transform”). If the kernel is Gaussian, then the FIG transform can be used for applying the kernel to any vector in , but the preconditioner has to be applied using the FIG transform as well, to keep the advantage of using a Gaussian kernel. In order to achieve this, the kernel is approximated using a Nyström decomposition, i.e. , where is a subset of columns of and is the inverse of a submatrix of . Adding the ridge regression parameter, and using the Woodbury matrix identity:
(6) 
Since is a submatrix of , then the application of both and to a matrix, can be done using FIG transform in . is also a subset of (unlike itself, which involves matrix inversion) and can therefore be applied in the same way, in . The Woodbury matrix identity requires computing the inverse of , however this is a small matrix of size . Constructing the matrix
can be done by the application of the FIG transform to the identity matrix,
(of size ) in operations. Each application of Eq. 6 involves solving the following linear system(7) 
It is better to store the Cholesky factor of the matrix in Eq. 7, but this matrix tends to be illconditioned in large scales. Instead, Eq. 7 is modified to:
(8) 
where the matrix is generally more stable, and its Cholesky decomposition can be applied to solve Eq. 8 for and then restoring using .
(9) 
where is the Cholesky factor of size . The algorithm can be viewed as two steps:

Preparation stage (or “setup” stage), which selects anchor data points from the dataset and also performs some computations to be used later in the iterative stage, such as building the matrix of the Cholesky decomposition and the matrix .

Iterative stage, which applies conjugate gradient using the preconditioner that was computed in the preparation stage.
Selection of the anchor points can be done in various ways, such as random sampling or farthest point sampling (FPS) [5]. In this work, the suggested method is based on the interpolative decomposition (or equivalently, on rank revealing decompositions, [15, 28, 24], which has the following advantages:

It is strongly related to the CUR decomposition [36], and therefore also to the Nyström approximation. The interpolative decomposition is also integral to the method, yielding theoretical spectral bounds that can give theoretical guarantees for the overall performance of the algorithm.

It selects columns of the input matrix, which enables the usage of the FIGtransform later on in the algorithm

It has a randomized version, which can be computed (using the FIGtransform) in linear time and the most demanding computational part can be parallelized.
Remark 3.1.
Algorithm 1 uses sampling technique based on interpolative decomposition. When using pivoted QR, the computational complexity is . In practice, the performance is very similar to RRQR.
Remark 3.2.
The computational complexity of Algorithm 2 is
Remark 3.3.
Applying the preconditioner (Algorithm 3) in each iteration takes operations.
Remark 3.4.
Remark 3.5.
Anchor points selection can be done differently, for example by random sampling or FPS. In this case, the theoretical bounds will not hold, but may still produce good results in practice and reduce the computational cost of Algorithm 1. For example, from using pivoted QR, to using random sampling.
4 Theoretical Analysis
Lemma 4.1.
Let be a symmetric matrix, and let be its Nyström approximation, then where p(k,n) is a function bounded by a low degree polynomial in and .
Proof.
By Lemma 2.1
(10) 
Note however that since K is symmetric, meaning
(11) 
hence . Therefore it follows immediately that
(12) 
We also know from by definition of RRQR, that the decomposition must satisfy
(13) 
We therefore combine (7) and (8), yielding
(14) 
∎
Lemma 4.2.
Let be a positive semidefinite matrix, and let its Nyström approximation, then .
Proof.
The lemma infers directly from the Schur’s compliment of . ∎
Lemma 4.3 ([4] pp. 673).
For any two matrices the following holds:
(15) 
Definition 4.1.
[2] The numerical rank of the Gaussian kernel up to precision is
Theorem 4.4.
The following theorem gives an upper bound to the condition number:
Theorem 4.5.
Let be a rank Nyström approximation for a positive semidefinite matrix , such that , for a positive constant (that may depend on and ) and a ridge parameter , then
(17) 
Proof.
Combining Lemma 4.2 and Lemma 4.3 with the fact that and are positive semidefinite, gives
(18) 
Modifying Eq. 18 and adding gives
(19) 
Applying to both sides of the equation and using the fact that is symmetric positive semidefinite gives
(20) 
Clearly, , which yields
(21) 
and finally,
(22) 
which completes the proof. ∎
Corollary 4.6.
Let be a Gaussian kernel matrix over dataset , i.e. . Let be a rank Nyström approximation, such that and let be a ridge parameter. Then, for a maximal condition number,
(23) 
where are the lengths of the intervals of the bounding box of the dataset .
Proof.
Remark 4.7.
Corollary 4.6 enables to determine the required rank Nyström approximation and illustrates the tradeoff between the condition number and the low rank, which has implications over the memory consumption and computational load of the algorithm. Smaller condition number yields less iterations on one hand, but on the other hand requires processing larger matrix .
Remark 4.8.
As a simple example, for the deterministic interpolative decomposition, , which yields
Remark 4.9.
Corollary 4.10.
For a Nyström preconditioner built by choosing , where are columns of chosen by the interpolative decomposition, then
(26) 
Remark 4.11.
A similar bound can be developed immediately for other matrix decompositions.
5 Numerical Results
In this section, empirical evaluation of the algorithm is presented and compared against the naive implementation (no preconditioning) and RFF preconditioning. The datasets used are described in the supplementary appendix[30]. The number of data points is , rank , noise level and
. The most costly computation in the presented flow is the QR decomposition in Algorithm
1 which is , compared to where the number of Gaussian components, therefore in order to maintain similar computational complexity, the number of Gaussian components was set to , and therefore . Figure 1 shows the comparison between the distance , where is the exact solution, and is the solution at iteration .Remark 5.1.
It is important to note that each iteration of the proposed algorithm is , while every iteration of RFF is for . For , each RFF iteration is asymptotically more costly in terms of runtime complexity by a factor of .
Remark 5.2.
When the spectral decay is fast, a small will be sufficient for satisfying results.
Figure 2 compares between the norm of the residual error, , where refers to the iteration number and is the solution at iteration of the CG algorithm. Also presented is the distance
It’s worth noting that the graph is slightly misleading, as the number of Gaussian RV’s taken for the RFF preconditioner is very high (in order to achieve the same computational complexity), however per iteration the proposed method is significantly faster, therefore in reality the convergence is still far faster for the proposed method.
5.1 Synthetic Electromagnetic Field Simulator
A 3dimensional synthetic dataset generated by an electric field simulator was used to test the performance of the algorithm. The simulation was composed of the following parts:

setting the support of the points to

choosing points from a uniform distribution over the defined support at random, where:

5 points represent the positive pointcharges

30,000 points represent arbitrary sampling locations of the electric field

Following the superposition rule, at each sampling point , each field component was summed as follows:
(27) 
(28) 
(29) 
where is the distance between charge and the point , and and are the spherical angles with respect to axes and respectively. Additionally, an electric potential was calculated at each point, to be used as the problem’s label as follows:
(30) 
6 Conclusions
The paper shows that the use of randomized matrix decompositions (specifically RID with CUR) for constructing a preconditioner for the matrix kernel is very successful in solving the scalability drawback intrinsic to kernel matrix methods. The method is strict in memory consumption (no need to hold the matrix), maintains accuracy, allows efficient application of the kernel matrix and reduces its condition number. Our results show fast convergence of the CG algorithm while maintaining accuracy, outperforming similar SOTA methods. The paper also provides theoretical insights and bounds for the lowrank approximation needed to reach the desired condition number.
References
 [1] Haim Avron and Vikas Sindhwani. Highperformance kernel machines with implicit distributed optimization and randomization. Technometrics, 58(3):341–349, 2016.
 [2] Amit Bermanis, Amir Averbuch, and Ronald R Coifman. Multiscale data sampling and function extension. Applied and Computational Harmonic Analysis, 34(1):15–29, 2013.
 [3] Amit Bermanis, Guy Wolf, and Amir Averbuch. Coverbased bounds on the numerical rank of Gaussian kernels. Applied and Computational Harmonic Analysis, 36(2):302–315, 2014.
 [4] Dennis S Bernstein. Matrix Mathematics: Theory, Facts, and Formulas. Princeton University Press, 2009.
 [5] Alexander M Bronstein, Michael M Bronstein, and Ron Kimmel. Numerical geometry of nonrigid shapes. Springer Science & Business Media, 2008.
 [6] S. Chandrasekaran and I. Ipsen. On rankrevealing QR factorisations.
 [7] Kenneth L Clarkson and David P Woodruff. Lowrank approximation and regression in input sparsity time. Journal of the ACM (JACM), 63(6):54, 2017.

[8]
Kurt Cutajar, Michael Osborne, John Cunningham, and Maurizio Filippone.
Preconditioning kernel matrices.
In
International Conference on Machine Learning
, pages 2529–2538, 2016.  [9] Bo Dai, Bo Xie, Niao He, Yingyu Liang, Anant Raj, MariaFlorina F Balcan, and Le Song. Scalable kernel methods via doubly stochastic gradients. In Advances in Neural Information Processing Systems, pages 3041–3049, 2014.
 [10] Petros Drineas, Michael W Mahoney, and S Muthukrishnan. Relativeerror CUR matrix decompositions. SIAM Journal on Matrix Analysis and Applications, 30(2):844–881, 2008.
 [11] Nando D Freitas, Yang Wang, Maryam Mahdaviani, and Dustin Lang. Fast Krylov methods for Nbody learning. In Advances in neural information processing systems, pages 251–258, 2006.
 [12] Sergei A Goreinov, Eugene E Tyrtyshnikov, and Nickolai L Zamarashkin. A theory of pseudoskeleton approximations. Linear algebra and its applications, 261(13):1–21, 1997.
 [13] Leslie Greengard and Vladimir Rokhlin. A fast algorithm for particle simulations. Journal of computational physics, 73(2):325–348, 1987.
 [14] Leslie Greengard and John Strain. The fast Gauss transform. SIAM Journal on Scientific and Statistical Computing, 12(1):79–94, 1991.
 [15] Ming Gu and Stanley C Eisenstat. Efficient algorithms for computing a strong rankrevealing QR factorization. SIAM Journal on Scientific Computing, 17(4):848–869, 1996.
 [16] P.G Martinsson V. Rokhlin H. Cheng, Z. Gimbutas. On the compression of low rank matrices. SIAM, 26.
 [17] Kenneth L. Clarkson Haim Avron and David P. Woodruff. Faster kernel ridge regression using sketching and preconditioning. SIAM, 38:1116–1138.
 [18] Nathan Halko, PerGunnar Martinsson, and Joel A Tropp. Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions. SIAM review, 53(2):217–288, 2011.

[19]
Y.P Hong and C.T. Pan.
Rankrevealing QR factorizations and the singular value decomposition.
Math. Comp.  [20] Zhiyun Lu, Avner May, Kuan Liu, Alireza Bagheri Garakani, Dong Guo, Aurélien Bellet, Linxi Fan, Michael Collins, Brian Kingsbury, Michael Picheny, et al. How to scale up kernel methods to be as good as deep neural nets. arXiv preprint arXiv:1411.4000, 2014.
 [21] Michael W Mahoney and Petros Drineas. CUR matrix decompositions for improved data analysis. Proceedings of the National Academy of Sciences, 106(3):697–702, 2009.
 [22] Michael W Mahoney et al. Randomized algorithms for matrices and data. Foundations and Trends® in Machine Learning, 3(2):123–224, 2011.
 [23] PerGunnar Martinsson, Vladimir Rokhlin, and Mark Tygert. A randomized algorithm for the decomposition of matrices. Applied and Computational Harmonic Analysis, 30(1):47–68, 2011.
 [24] L Miranian and Ming Gu. Strong rank revealing lu factorizations. Linear algebra and its applications, 367:1–16, 2003.
 [25] Vlad I Morariu, Balaji V Srinivasan, Vikas C Raykar, Ramani Duraiswami, and Larry S Davis. Automatic online tuning for fast Gaussian summation. In Advances in neural information processing systems, pages 1113–1120, 2009.
 [26] Cameron Musco and Christopher Musco. Recursive sampling for the Nystrom method. In Advances in Neural Information Processing Systems, pages 3833–3845, 2017.
 [27] AI Osinsky and NL Zamarashkin. Pseudoskeleton approximations with better accuracy estimates. Linear Algebra and its Applications, 537:221–249, 2018.
 [28] CT Pan. On the existence and computation of rankrevealing LU factorizations. Linear Algebra and its Applications, 316(13):199–222, 2000.
 [29] Vladimir Rokhlin P.G Martinsson and Mark Tygert. A randomized algorithm for the decomposition of matrices. Applied and Computational Harmonic Analysis, 30.
 [30] Adam Coates Pieter Abbeel and Andrew Y. Ng. Autonomous helicopter aerobatics through apprenticeship learning. IJRR, 2010.
 [31] Ali Rahimi and Benjamin Recht. Random features for largescale kernel machines. Advances in neural information processing systems, pages 1177–1184, 2008.
 [32] Vikas C Raykar and Ramani Duraiswami. Fast large scale Gaussian process regression using approximate matrixvector products. In Learning workshop, 2007.
 [33] Gil Shabat, Yaniv Shmueli, Yariv Aizenbud, and Amir Averbuch. Randomized LU decomposition. Applied and Computational Harmonic Analysis, 44(2):246–272, 2018.
 [34] Yirong Shen, Matthias Seeger, and Andrew Y Ng. Fast Gaussian process regression using kdtrees. In Advances in neural information processing systems, pages 1225–1232, 2006.
 [35] Balaji Vasan Srinivasan, Qi Hu, Nail A Gumerov, Raghu Murtugudde, and Ramani Duraiswami. Preconditioned Krylov solvers for kernel regression. arXiv preprint arXiv:1408.1237, 2014.
 [36] Sergey Voronin and PerGunnar Martinsson. Efficient algorithms for CUR and interpolative matrix decompositions. Advances in Computational Mathematics, 43(3):495–516, 2017.

[37]
C Yang, R Duraiswami, NA Gumerov, and L Davis.
Improved fast Gauss transform and efficient kernel density estimation.
InProceedings Ninth IEEE International Conference on Computer Vision
.  [38] Changjiang Yang, Ramani Duraiswami, and Larry S Davis. Efficient kernel machines using the improved fast Gauss transform. In Advances in neural information processing systems, pages 1561–1568, 2005.
Comments
There are no comments yet.