Bounds for list-decoding and list-recovery of random linear codes

Venkatesan Guruswami, Ray Li, Jonathan Mosheiff, Nicolas Resch, Shashwat Silas, Mary Wootters

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

A family of error-correcting codes is list-decodable from error fraction p if, for every code in the family, the number of codewords in any Hamming ball of fractional radius p is less than some integer L that is independent of the code length. It is said to be list-recoverable for input list size ℓ if for every sufficiently large subset of codewords (of size L or more), there is a coordinate where the codewords take more than ℓ values. The parameter L is said to be the “list size” in either case. The capacity, i.e., the largest possible rate for these notions as the list size L → ∞, is known to be 1 − hq(p) for list-decoding, and 1 − logq ℓ for list-recovery, where q is the alphabet size of the code family. In this work, we study the list size of random linear codes for both list-decoding and list-recovery as the rate approaches capacity. We show the following claims hold with high probability over the choice of the code (below q is the alphabet size, and ε > 0 is the gap to capacity). A random linear code of rate 1 − logq(ℓ) − ε requires list size L ≥ ℓΩ(1) for list-recovery from input list size ℓ. This is surprisingly in contrast to completely random codes, where L = O(ℓ/ε) suffices w.h.p. A random linear code of rate 1−hq(p)−ε requires list size L ≥ bhq(p)/ε + 0.99c for list-decoding from error fraction p, when ε is sufficiently small. A random binary linear code of rate 1 − h2(p) − ε is list-decodable from average error fraction p with list size with L ≤ bh2(p)/εc+ 2. (The average error version measures the average Hamming distance of the codewords from the center of the Hamming ball, instead of the maximum distance as in list-decoding.) The second and third results together precisely pin down the list sizes for binary random linear codes for both list-decoding and average-radius list-decoding to three possible values. Our lower bounds follow by exhibiting an explicit subset of codewords so that this subset - or some symbol-wise permutation of it - lies in a random linear code with high probability. This uses a recent characterization of (Mosheiff, Resch, Ron-Zewi, Silas, Wootters, 2019) of configurations of codewords that are contained in random linear codes. Our upper bound follows from a refinement of the techniques of (Guruswami, Håstad, Sudan, Zuckerman, 2002) and strengthens a previous result of (Li, Wootters, 2018), which applied to list-decoding rather than average-radius list-decoding.

Original languageAmerican English
Title of host publicationApproximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, APPROX/RANDOM 2020
EditorsJaroslaw Byrka, Raghu Meka
PublisherSchloss Dagstuhl- Leibniz-Zentrum fur Informatik GmbH, Dagstuhl Publishing
ISBN (Electronic)9783959771641
DOIs
StatePublished - 1 Aug 2020
Externally publishedYes
Event23rd International Conference on Approximation Algorithms for Combinatorial Optimization Problems and 24th International Conference on Randomization and Computation, APPROX/RANDOM 2020 - Virtual, Online, United States
Duration: 17 Aug 202019 Aug 2020

Publication series

NameLeibniz International Proceedings in Informatics, LIPIcs
Volume176

Conference

Conference23rd International Conference on Approximation Algorithms for Combinatorial Optimization Problems and 24th International Conference on Randomization and Computation, APPROX/RANDOM 2020
Country/TerritoryUnited States
CityVirtual, Online
Period17/08/2019/08/20

Keywords

  • Coding theory
  • List-decoding
  • List-recovery
  • Random linear codes

All Science Journal Classification (ASJC) codes

  • Software

Fingerprint

Dive into the research topics of 'Bounds for list-decoding and list-recovery of random linear codes'. Together they form a unique fingerprint.

Cite this