Known polylog parallel algorithms for the solution of linear systems and related problems require computation of the characteristic polynomial or related forms, which are known to be highly unstable in practice. However, matrix factorizations of various types, bypassing computation of the characteristic polynomial, are used extensively in sequential numerical computations and are essential in many applications.
This paper gives new parallel methods for various exact factorizations of several classes of matrices. We assume the input matrices are <italic>n</italic> × <italic>n</italic> with either integer entries of size ≤ 2<supscrpt><italic>n</italic><supscrpt>&ogr;(1)</supscrpt></supscrpt>. We make no other assumption on the input. We assume the arithmetic PRAM model of parallel computation. Our main result is a reduction of the known parallel time bounds for these factorizations from <italic>O</italic>(log<supscrpt>3</supscrpt><italic>n</italic>) to <italic>O</italic>(log<supscrpt>2</supscrpt><italic>n</italic>). Our results are work efficient; we match the best known work bounds of parallel algorithms with polylog time bounds, and are within a log <italic>n</italic> factor of the work bounds for the best known sequential algorithms for the same problems.
The exact factorizations we compute for symmetric positive definite matrices include:
<list><item>1. recursive factorization sequences and trees,
</item><item>2. LU factorizations,
</item><item>3. QR factorizations, and
</item><item>4. reduction to upper Hessenberg form.
</item></list>The classes of matrices for which we can efficiently compute these factorizations include:
<list><item>1. dense matrices, in time <italic>O</italic>(log<supscrpt>2</supscrpt><italic>n</italic>) with processor bound <italic>P</italic>(<italic>n</italic>) (the number of processors needed to multiply two <italic>n</italic> × <italic>n</italic> matrices in <italic>O</italic>(log <italic>n</italic> time),
</item>,<item>2. block diagonal matrices, in time <italic>O</italic>(log<supscrpt>2</supscrpt><italic>b</italic> with <italic>P(b)n/b</italic> processors,
</item><item>3. sparse matrices which are <italic>s(n)</italic>-separable (recursive factorizations only), in time <italic>O</italic>(log<supscrpt>2</supscrpt><italic>n</italic>) with <italic>P(s(n))</italic> processors where <italic>s(n)</italic> is of the form <italic>n</italic><supscrpt>&ngr;</supscrpt> for 0 < &ngr; < 1, and
</item><item>4. banded matrices, in parallel time <italic>O</italic>((log<italic>n</italic>) log <italic>b</italic>) with <italic>P(b)n/b</italic> processors.
</item></list>Our factorizations also provide us similarly efficient algorithms for exact computation (given arbitrary rational matrices that need not be symmetric positive definite) of the following:
<list><item>1. solution of the corresponding linear systems,
</item><item>2. the determinant,
</item><item>3. the inverse.
</item></list>Thus our results provide the first known efficient parallel algorithms for exact solution of these matrix problems, that avoids computation of the characteristic polynomial or related forms. Instead we use a construction which modifies the input matrix, which may initially have arbitrary condition, so as to have condition nearly 1, and then applies a multilevel, pipelined Newton iteration, followed by a similar multilevel, pipelined Hensel Lifting.
[1]
B. Cuny,et al.
Fast and Efficient Parallel Solution of Dense Linear Systems
,
1989
.
[2]
Erich Kaltofen,et al.
On Wiedemann's Method of Solving Sparse Linear Systems
,
1991,
AAECC.
[3]
Victor Y. Pan,et al.
Processor efficient parallel solution of linear systems over an abstract field
,
1991,
SPAA '91.
[4]
Robert T. Moenck,et al.
Approximate algorithms to derive exact solutions to systems of linear equations
,
1979,
EUROSAM.
[5]
Victor Y. Pan,et al.
Fast and Efficient Parallel Solution of Sparse Linear Systems
,
1993,
SIAM J. Comput..
[6]
H. Hotelling.
Some New Methods in Matrix Calculation
,
1943
.
[7]
Victor Y. Pan,et al.
Efficient parallel solution of linear systems
,
1985,
STOC '85.
[8]
Victor Y. Pan,et al.
Fast and Efficient Parallel Algorithms for the Exact Inversion of Integer Matrices
,
1985,
FSTTCS.
[9]
R. Tarjan,et al.
A Separator Theorem for Planar Graphs
,
1977
.
[10]
Adi Ben-Israel,et al.
On Iterative Computation of Generalized Inverses and Associated Projections
,
1966
.
[11]
D. Rose,et al.
Generalized nested dissection
,
1977
.
[12]
D Kozen,et al.
Fast parallel orthogonalization
,
1986,
SIGA.
[13]
John H. Reif,et al.
Space and time efficient implementations of parallel nested dissection
,
1992,
SPAA '92.
[14]
Joseph JáJá,et al.
An Introduction to Parallel Algorithms
,
1992
.
[15]
Allan Borodin,et al.
The computational complexity of algebraic and numeric problems
,
1975,
Elsevier computer science library.
[16]
Donald E. Knuth,et al.
The art of computer programming. Vol.2: Seminumerical algorithms
,
1981
.
[17]
Victor Y. Pan,et al.
Processor-efficient parallel solution of linear systems. II. The positive characteristic and singular cases
,
1992,
Proceedings., 33rd Annual Symposium on Foundations of Computer Science.
[18]
Victor Y. Pan,et al.
The Parallel Computation of Minimum Cost Paths in Graphs by Stream Contraction
,
1991,
Inf. Process. Lett..
[19]
Adi Ben-Israel,et al.
A note on an iterative method for generalized inversion of matrices
,
1966
.
[20]
Victor Y. Pan,et al.
Complexity of Parallel Matrix Computations
,
1987,
Theor. Comput. Sci..
[21]
John H. Reif,et al.
Synthesis of Parallel Algorithms
,
1993
.
[22]
Jacob T. Schwartz,et al.
Fast Probabilistic Algorithms for Verification of Polynomial Identities
,
1980,
J. ACM.
[23]
Victor Y. Pan,et al.
Fast and efficient parallel solution of dense linear systems
,
1989
.