Warp experience: we can map computations onto a parallel computer efficiently

Warp is a programmable, systolic array computer developed by Carnegie Mellon and produced by GE. A 10-cell Warp machine can perform 100 million floating-point operations per second (10 MFLOPS). A variety of applications have been mapped onto Warp. The experience has been that the mapping is not a real problem; in fact, usually near-optimal mapping is relatively easy to obtain, and the actual implementation of the mapping on the machine can often be automated. This paper explains why this is the case by examining some computational models which are frequently used on Warp. Carnegie Mellon and Intel are jointly developing a VLSI version of Warp, called iWarp. It is expected that many applications can be efficiently mapped onto low-cost iWarp arrays to achieve an effective computation bandwidth of about one GigaFLOPS.

[1]  H. T. Kung,et al.  Using warp as a supercomputer in signal processing , 1986, ICASSP '86. IEEE International Conference on Acoustics, Speech, and Signal Processing.

[2]  H. T. Kung,et al.  Warp as a machine for low-level vision , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[3]  H. T. Kung Systolic communication , 1988, [1988] Proceedings. International Conference on Systolic Arrays.

[4]  Louis A. Hageman,et al.  Iterative Solution of Large Linear Systems. , 1971 .

[5]  Gérard M. Baudet,et al.  Optimal Sorting Algorithms for Parallel Computers , 1978, IEEE Transactions on Computers.

[6]  H. T. Kung Systolic algorithms for the CMU warp processor , 1984 .

[7]  J. A. Webb,et al.  Low-level vision on Warp and the apply programming model. Technical report , 1987 .

[8]  Monica Sin-Ling Lam,et al.  A Systolic Array Optimizing Compiler , 1989 .

[9]  Azriel Rosenfeld,et al.  Scene Labeling by Relaxation Operations , 1976, IEEE Transactions on Systems, Man, and Cybernetics.

[10]  H. T. Kung Why systolic architectures? , 1982, Computer.

[11]  H. T. Kung,et al.  Path Planning On The Warp Computer: Using A Linear Systolic Array In Dynamic Programming , 1988, Optics & Photonics.

[12]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[13]  Azriel Rosenfeld,et al.  Iterative methods in image analysis , 1978, Pattern Recognit..

[14]  D. S. Touretzky,et al.  Neural network simulation at Warp speed: how we got 17 million connections per second , 1988, IEEE 1988 International Conference on Neural Networks.

[15]  Thomas R. Gross,et al.  Compilation for a high-performance systolic array , 1986, SIGPLAN '86.