Matrix-free Krylov iteration for implicit convolution of numerically low-rank data

Evaluating the response of a linear shift-invariant system is a problem that occurs frequently in a wide variety of science and engineering problems. Calculating the system response via a convolution may be done efficiently with Fourier transforms. When one must compute the response of one system to m input signals, or the response of m systems to one signal, it may be the case that one may approximate all system responses without having to compute all m Fourier transforms. This can lead to substantial computational savings. Rather than process each point individually, one may only process basis vectors that span the output data space. However, to get a low-error approximation, it is necessary that the output vectors have low numerical rank if they were assembled into a matrix. We develop theory that shows how the singular value decay of a matrix ź A that is a product of a convolution operator ź and an arbitrary matrix A depends in a linear fashion on the singular value decays of ź and A . We propose gap-rank, a measure of the relative numerical rank of a matrix. We show that convolution cannot destroy the numerical low-rank-ness of ź A data with only modest assumptions. We then develop a new method that exploits low-rank problems with block Golub-Kahan iteration in a Krylov subspace to approximate the low-rank problem. Our method can exploit parallelism in both the individual convolutions and the linear algebra operations in the block Golub-Kahan algorithm. We present numerical examples from signal and image processing that show the low error and scalability of our method.

[1]  Andrew Lumsdaine,et al.  Minimal krylov subspaces for dimension reduction , 2013 .

[2]  Richard R. Underwood An iterative block Lanczos method for the solution of large sparse symmetric eigenproblems , 1975 .

[3]  Qian Wang,et al.  AUGEM: Automatically generate high performance Dense Linear Algebra kernels on x86 CPUs , 2013, 2013 SC - International Conference for High Performance Computing, Networking, Storage and Analysis (SC).

[4]  J. Cullum,et al.  A block Lanczos algorithm for computing the q algebraically largest eigenvalues and a corresponding eigenspace of large, sparse, real symmetric matrices , 1974, CDC 1974.

[5]  H. P. Lee,et al.  PROPER ORTHOGONAL DECOMPOSITION AND ITS APPLICATIONS—PART I: THEORY , 2002 .

[6]  Yousef Saad,et al.  Fast Approximate kNN Graph Construction for High Dimensional Data via Recursive Lanczos Bisection , 2009, J. Mach. Learn. Res..

[7]  Qiang Ye,et al.  An adaptive block Lanczos algorithm , 1996, Numerical Algorithms.

[8]  H. Simon Analysis of the symmetric Lanczos algorithm with reorthogonalization methods , 1984 .

[9]  David Skillicorn,et al.  Using Matrix Decompositions for Data Mining (Chapman & Hall/Crc Data Mining and Knowledge Discovery Series) , 2007 .

[10]  J. Cullum,et al.  Lanczos Algorithms for Large Symmetric Eigenvalue Computations, Vol. 1 , 2002 .

[11]  I. Jolliffe Principal Component Analysis , 2002 .

[12]  J. Cullum,et al.  Lanczos algorithms for large symmetric eigenvalue computations , 1985 .

[13]  Gene H. Golub,et al.  Calculating the singular values and pseudo-inverse of a matrix , 2007, Milestones in Matrix Computation.

[14]  Yousef Saad,et al.  Lanczos Vectors versus Singular Vectors for Effective Dimension Reduction , 2009, IEEE Transactions on Knowledge and Data Engineering.

[15]  Andrew V. Knyazev,et al.  Angles between subspaces and their tangents , 2012, J. Num. Math..

[16]  Dao-Qing Dai,et al.  Bilinear Lanczos components for fast dimensionality reduction and feature extraction , 2010, Pattern Recognit..

[17]  A. Swindlehurst,et al.  Subspace-based signal analysis using singular value decomposition , 1993, Proc. IEEE.

[18]  Y. Saad On the Rates of Convergence of the Lanczos and the Block-Lanczos Methods , 1980 .

[19]  Axel Ruhe,et al.  A Krylov Subspace Method for Information Retrieval , 2005, SIAM J. Matrix Anal. Appl..

[20]  B. Parlett,et al.  The Lanczos algorithm with selective orthogonalization , 1979 .

[21]  T. Landauer,et al.  Indexing by Latent Semantic Analysis , 1990 .

[22]  Pablo Laguna,et al.  Principal Component Analysis in ECG Signal Processing , 2007, EURASIP J. Adv. Signal Process..

[23]  D. O’Leary The block conjugate gradient algorithm and related methods , 1980 .

[24]  R. Bracewell The Fourier Transform and Its Applications , 1966 .

[25]  Hongyuan Zha,et al.  Low-Rank Matrix Approximation Using the Lanczos Bidiagonalization Process with Applications , 1999, SIAM J. Sci. Comput..

[26]  Fuzhen Zhang Matrix Theory: Basic Results and Techniques , 1999 .

[27]  Franklin T. Luk,et al.  A Block Lanczos Method for Computing the Singular Values and Corresponding Singular Vectors of a Matrix , 1981, TOMS.

[28]  Antoine Petitet,et al.  Minimizing development and maintenance costs in supporting persistently optimized BLAS , 2005 .

[29]  H. Simon The Lanczos algorithm with partial reorthogonalization , 1984 .

[30]  Vijay K. Madisetti,et al.  The Digital Signal Processing Handbook , 1997 .

[31]  J. Mendel Lessons in Estimation Theory for Signal Processing, Communications, and Control , 1995 .

[32]  Jack Dongarra,et al.  Templates for the Solution of Algebraic Eigenvalue Problems , 2000, Software, environments, tools.

[33]  M. Turk,et al.  Eigenfaces for Recognition , 1991, Journal of Cognitive Neuroscience.

[34]  David B. Skillicorn,et al.  Understanding Complex Datasets: Data Mining with Matrix Decompositions , 2007 .

[35]  James Demmel,et al.  Applied Numerical Linear Algebra , 1997 .

[36]  C. Lanczos An iteration method for the solution of the eigenvalue problem of linear differential and integral operators , 1950 .