We study certain combinatorial aspects of listdecoding, motivated by the exponential gap between the known upper bound (of O(1/γ)) and lower bound (of Ω<sub>p</sub>(log(1/γ))) for the list size needed to list decode up to error fraction p with rate γ away from capacity, i.e., 1 - h(p) - γ [here p E (0, 1/2) and γ > 0]. Our main result is that we prove that in any binary code C ⊆ (0, 1)<sup>n</sup> of rate 1 - h(p) - γ, there must exist a set l ⊂ C of p(1/√γ) codewords such that the average distance of the points in L from their centroid is at most pn. In other words, there must exist Ω<sub>p</sub>(1/√γ) codewords with low average radius. The standard notion of list decoding corresponds to working with the maximum distance of a collection of codewords from a center instead of average distance. The average radius form is in itself quite natural; for instance, the classical Johnson bound in fact implies average-radius list-decodability. The remaining results concern the standard notion of list-decoding, and help clarify the current state of affairs regarding combinatorial bounds for list-decoding as follows. First, we give a short simple proof, over all fixed alphabets, of the above-mentioned Ω<sub>p</sub>(log(1/γ)) lower bound. Earlier, this bound followed from a complicated, more general result of Blinovsky. Second, we show that one cannot improve the Ω<sub>p</sub>(log(1/γ)) lower bound via techniques based on identifying the zero-rate regime for list-decoding of constantweight codes [this is a typical approach for negative results in coding theory, including the Ω<sub>p</sub>(log(1/γ)) list-size lower bound]. On a positive note, our Ω<sub>p</sub>(1/√γ) lower bound for average radius list-decoding circumvents this barrier. Third, we exhibit a reverse connection between the existence of constant-weight and general codes for list-decoding, showing that the best possible list-size, as a function of the gap γ of the rate to the capacity limit, is the same up to constant factors for both constant-weight codes (with weight bounded away from p) and general codes. Fourth, we give simple second moment-based proofs that w.h.p. a list-size of Ω<sub>p</sub>(1/γ) is needed for list-decoding random codes from errors as well as erasures. For random linear codes, the corresponding list-size bounds are Ω<sub>p</sub>(1/γ) for errors and expΩ<sub>p</sub>(log(1/γ)) for erasures.
[1]
Vladimir M. Blinovsky,et al.
List decoding
,
1992,
Discret. Math..
[2]
Atri Rudra,et al.
Limits to List Decoding of Random Codes
,
2009,
IEEE Transactions on Information Theory.
[3]
Stasys Jukna,et al.
Extremal Combinatorics - With Applications in Computer Science
,
2001,
Texts in Theoretical Computer Science. An EATCS Series.
[4]
Peter Elias,et al.
List decoding for noisy channels
,
1957
.
[5]
Peter Elias,et al.
Error-correcting codes for list decoding
,
1991,
IEEE Trans. Inf. Theory.
[6]
Venkatesan Guruswami,et al.
Linear-Algebraic List Decoding of Folded Reed-Solomon Codes
,
2011,
2011 IEEE 26th Annual Conference on Computational Complexity.
[7]
Venkatesan Guruswami,et al.
Combinatorial Limitations of Average-Radius List Decoding
,
2013,
APPROX-RANDOM.
[8]
Venkatesan Guruswami,et al.
Explicit Codes Achieving List Decoding Capacity: Error-Correction With Optimal Redundancy
,
2005,
IEEE Transactions on Information Theory.
[9]
Venkatesan Guruswami,et al.
Extensions to the Johnson bound
,
2001
.
[10]
Venkatesan Guruswami,et al.
Linear-Algebraic List Decoding for Variants of Reed–Solomon Codes
,
2013,
IEEE Transactions on Information Theory.
[11]
Vladimir M. Blinovsky,et al.
Code bounds for multiple packings over a nonbinary finite alphabet
,
2005,
Probl. Inf. Transm..
[12]
Venkatesan Guruswami,et al.
A Lower Bound on List Size for List Decoding
,
2005,
IEEE Trans. Inf. Theory.
[13]
Venkatesan Guruswami,et al.
On the List-Decodability of Random Linear Codes
,
2010,
IEEE Transactions on Information Theory.
[14]
Vladimir M. Blinovsky,et al.
On the convexity of one coding-theory function
,
2008,
Probl. Inf. Transm..
[15]
V. Blinovsky,et al.
Asymptotic Combinatorial Coding Theory
,
1997
.