We study networks that can sort n items even when a large number of the comparators in the network are faulty. We restrict attention to networks that consist of registers, comparators, and replicators. (Replicators are used to copy an item from one register to another, and they are assumed to be fault-free.) We consider the scenario of bot h random and worst-case comparator faults, and we follow the general model of destructive comparator failure proposed by Assaf and Upfal [2] in which the two outputs of a faulty comparator can fail independently of each other. In the case of random faults, Assaf and Upfal [2] showed how to construct a network with O(n log2 n) comparators that (with high probability) can sort n items even if a constant fraction of the comparators are faulty, Whether or not the bound on the number of comparators can be improved (to, say, O(n log n)) for sorting (or merging) has remained an interesting open question. We resolve this question in the paper by prov“This research is supported by AF05R Contract F49620-92-J-0125, DARPA Contract NOO014-91-J-1698, and DARPA Contract NOO01492-J-1799. Permission to copy without fee all or part of this material is granted provided that the copies are not made or distributed for direct commercial advantage, the ACM copyright notice and the title of the publication and ita date appear, and notice ia given that copying is by permission of the Association for Computing Machinery. To copy otherwise, or to republish, requires a fee and/or specific permission. ACM-SPAA’93-6193 /Velen,Germany. 01993 ACM 0-89791 -59921931000610030 . ..$1 .50 ing that any n-item sorting or merging network which can tolerate a constant fraction of random failures must have Q(n log2 n) comparators. In the case of worst-case faults, we show that $l(kn log n) comparators are necessary to construct a sorting or merging network that can tolerate up to k worst-case faults. We also show that this bound is tight for k = O(log n). The lower bound is particularly significant since it formally proves that the cost of being tolerant to worstcase failures is very high. Both the lower bound for random faults and the lower bound for worst-case faults are the first nontrivial lower bounds on the size of a fault-tolerant sorting or merging network.
[1]
F. Leighton,et al.
Introduction to Parallel Algorithms and Architectures: Arrays, Trees, Hypercubes
,
1991
.
[2]
Kenneth E. Batcher,et al.
Sorting networks and their applications
,
1968,
AFIPS Spring Joint Computing Conference.
[3]
Nicholas Pippenger,et al.
On networks of noisy gates
,
1985,
26th Annual Symposium on Foundations of Computer Science (sfcs 1985).
[4]
C. Greg Plaxton,et al.
Highly fault-tolerant sorting circuits
,
1991,
[1991] Proceedings 32nd Annual Symposium of Foundations of Computer Science.
[5]
E. Szemerédi,et al.
O(n LOG n) SORTING NETWORK.
,
1983
.
[6]
M. V. Wilkes,et al.
The Art of Computer Programming, Volume 3, Sorting and Searching
,
1974
.
[7]
Andrew Chi-Chih Yao,et al.
On Fault-Tolerant Networks for Sorting
,
1985,
SIAM J. Comput..
[8]
Manfred Schimmler,et al.
A Correction Network for N-Sorters
,
1989,
SIAM J. Comput..
[9]
E. Szemerédi,et al.
Sorting inc logn parallel steps
,
1983
.
[10]
Larry Rudolph,et al.
A Robust Sorting Network
,
1985,
IEEE Transactions on Computers.
[11]
Eli Upfal,et al.
Fault tolerant sorting network
,
1990,
Proceedings [1990] 31st Annual Symposium on Foundations of Computer Science.
[12]
Frank Thomson Leighton,et al.
Tight Bounds on the Complexity of Parallel Sorting
,
1984,
IEEE Transactions on Computers.