Supereffective slow-down of parallel computations

Brent’s scheduling principle provides a general simulation scheme when fewer processors are available than specified by the fastest parallel algorithm. Such a scheme preserves the actual number of executed operations, and when applicable, it provides a processor balancing technique that significantly reduces the work, expressed as the number of ezecutcdde operators. In this paper we discuss a new technique, called supereffective slow-down, that yields quite fast an algorithm with work significantly smaller than that of the fastest algorithm for the same problem. This technique can be viewed as a work-preserving acceleration of an existing recursive sequential algorithm for the considered problem. The presented examples include the computation of path algebras in graphs and digraphs and various computations in linear albegra. Some of the new algorithms may have practical value; for instance, we substantially improve the performance of the known parallel algorithms for triangular linear systems of equations.

[1]  Franco P. Preparata,et al.  An Improved Parallel Processor Bound in Fast Matrix Inversion , 1978, Inf. Process. Lett..

[2]  Mihalis Yannakakis,et al.  High-probability parallel transitive closure algorithms , 1990, SPAA '90.

[3]  Gene H. Golub,et al.  Matrix computations , 1983 .

[4]  Alfred V. Aho,et al.  The Design and Analysis of Computer Algorithms , 1974 .

[5]  Richard P. Brent,et al.  The Parallel Evaluation of General Arithmetic Expressions , 1974, JACM.

[6]  Victor Y. Pan,et al.  Parallel Evaluation of the Determinant and of the Inverse of a Matrix , 1989, Inf. Process. Lett..

[7]  Victor Y. Pan,et al.  Processor efficient parallel solution of linear systems over an abstract field , 1991, SPAA '91.

[8]  Richard M. Karp,et al.  A Survey of Parallel Algorithms for Shared-Memory Machines , 1988 .

[9]  Victor Y. Pan,et al.  Processor-efficient parallel solution of linear systems. II. The positive characteristic and singular cases , 1992, Proceedings., 33rd Annual Symposium on Foundations of Computer Science.

[10]  Victor Y. Pan,et al.  Improved Parallel Polynomial Division , 1993, SIAM J. Comput..

[11]  Victor Y. Pan,et al.  Parallel polynomial computations by recursive processes , 1990, ISSAC '90.

[12]  Thomas H. Spencer More time-work tradeoffs for parallel graph algorithms , 1991, SPAA '91.

[13]  Mihalis Yannakakis,et al.  High-Probability Parallel Transitive-Closure Algorithms , 1991, SIAM J. Comput..

[14]  Victor Y. Pan,et al.  Fast and Efficient Solution of Path Algebra Problems , 1989, J. Comput. Syst. Sci..

[15]  V. Pan On computations with dense structured matrices , 1990 .

[16]  Don Coppersmith,et al.  Matrix multiplication via arithmetic progressions , 1987, STOC.

[17]  Alexander L. Chistov,et al.  Fast parallel calculation of the rank of matrices over a field of arbitrary characteristic , 1985, FCT.

[18]  Thomas H. Spencer Time-work tradeoffs for parallel graph algorithms , 1991, SODA '91.

[19]  L. Csanky,et al.  Fast Parallel Matrix Inversion Algorithms , 1976, SIAM J. Comput..

[20]  T. Kailath,et al.  Fast Parallel Algorithms for QR and Triangular Factorization , 1987 .

[21]  Julian D. Laderman,et al.  On practical algorithms for accelerated matrix multiplication , 1992 .

[22]  V. Pan PARAMETRIZATION OF NEWTON'S ITERATION FOR COMPUTATIONS WITH STRUCTURED MATRICES AND APPLICATIONS , 1992 .

[23]  Victor Y. Pan,et al.  Complexity of Parallel Matrix Computations , 1987, Theor. Comput. Sci..

[24]  Stuart J. Berkowitz,et al.  On Computing the Determinant in Small Parallel Time Using a Small Number of Processors , 1984, Inf. Process. Lett..