HOME

TheInfoList



OR:

In
coding theory Coding theory is the study of the properties of codes and their respective fitness for specific applications. Codes are used for data compression, cryptography, error detection and correction, data transmission and data storage. Codes are studied ...
, folded Reed–Solomon codes are like Reed–Solomon codes, which are obtained by mapping m Reed–Solomon codewords over a larger alphabet by careful bundling of codeword symbols. Folded Reed–Solomon codes are also a special case of Parvaresh–Vardy codes. Using optimal parameters one can decode with a rate of ''R'', and achieve a decoding radius of 1 − ''R''. The term "folded Reed–Solomon codes" was coined in a paper by V.Y. Krachkovsky with an algorithm that presented Reed–Solomon codes with many random "phased burst" error

The list-decoding algorithm for folded RS codes corrects beyond the 1-\sqrt bound for Reed–Solomon codes achieved by the Guruswami
Sudan Sudan ( or ; ar, السودان, as-Sūdān, officially the Republic of the Sudan ( ar, جمهورية السودان, link=no, Jumhūriyyat as-Sūdān), is a country in Northeast Africa. It shares borders with the Central African Republic ...
algorithm for such phased burst errors.


History

One of the ongoing challenges in Coding Theory is to have error correcting codes achieve an optimal trade-off between (Coding) Rate and Error-Correction Radius. Though this may not be possible to achieve practically (due to Noisy Channel Coding Theory issues), quasi optimal tradeoffs can be achieved theoretically. Prior to Folded Reed–Solomon codes being devised, the best Error-Correction Radius achieved was 1- \sqrt , by Reed–Solomon codes for all rates R . An improvement upon this 1- \sqrt bound was achieved by Parvaresh and Vardy for rates R <\tfrac. For R\to 0 the Parvaresh–Vardy algorithm can decode a fraction 1- O(R \log(1/R)) of errors. Folded Reed–Solomon Codes improve on these previous constructions, and can be list decoded in polynomial time for a fraction (1-R-\varepsilon) of errors for any constant \varepsilon > 0.


Definition

:f(X)\mapsto\beginf(1)\\f(\gamma)\\\vdots\\f(\gamma^)\end,\beginf(\gamma^m)\\f(\gamma^)\\\vdots\\f(\gamma^)\end,\ldots,\beginf(\gamma^)\\f(\gamma^)\\\vdots\\f(\gamma^)\end Consider a Reed–Solomon =q-1,kq code of length n and
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coordin ...
k and a folding parameter m \ge 1 . Assume that m divides n. Mapping for Reed–Solomon codes like this: :f \mapsto \left \langle f(1), f \left (\gamma^1 \right ), f \left (\gamma^2 \right ), \ldots , f \left (\gamma^ \right ) \right \rangle where \gamma \in \mathbb_q is a primitive element in :\mathbb_q = \left\ . The m folded version of Reed Solomon code C, denoted FRS_ is a code of block length N = n/m over \mathbb^m. FRS_ are just -1,k Reed Solomon codes with m consecutive symbols from RS codewords grouped together.


Graphic description

The above definition is made more clear by means of the diagram with m=3, where m is the folding parameter. The message is denoted by f(X) , which when encoded using Reed–Solomon encoding, consists of values of f at x_0, x_1, x_2, \ldots,x_ , where x_i =\gamma^i . Then bundling is performed in groups of 3 elements, to give a codeword of length n/3 over the alphabet \mathbb_q^3 . Something to be observed here is that the folding operation demonstrated does not change the rate R of the original Reed–Solomon code. To prove this, consider a linear
, k, d The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline o ...
q code, of length n ,
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coordin ...
k and
distance Distance is a numerical or occasionally qualitative measurement of how far apart objects or points are. In physics or everyday usage, distance may refer to a physical length or an estimation based on other criteria (e.g. "two counties over"). ...
d . The m folding operation will make it a \left tfrac, \tfrac,\tfrac\right code. By this, the rate R = \tfrac will be the same.


Folded Reed–Solomon codes and the singleton bound

According to the asymptotic version of the
singleton bound In coding theory, the Singleton bound, named after Richard Collom Singleton, is a relatively crude upper bound on the size of an arbitrary block code C with block length n, size M and minimum distance d. It is also known as the Joshibound. proved b ...
, it is known that the relative distance \delta , of a code must satisfy R \leqslant 1-\delta + o(1) where R is the rate of the code. As proved earlier, since the rate R is maintained, the relative distance \delta \leqslant 1-R also meets the Singleton bound.


Why folding might help?

Folded Reed–Solomon codes are basically the same as Reed Solomon codes, just viewed over a larger alphabet. To show how this might help, consider a folded Reed–Solomon code with m=3. Decoding a Reed–Solomon code and folded Reed–Solomon code for the same fraction of errors \rho are tasks of almost the same computational intensity: one can unfold the received word of the folded Reed–Solomon code, treat it as a received word of the original Reed–Solomon code, and run the Reed–Solomon list decoding algorithm on it. Obviously, this list will contain all the folded Reed–Solomon codewords within distance \rho of the received word, along with some extras, which we can expurgate. Also, decoding a folded Reed–Solomon code is an easier task. Suppose we want to correct a third of the errors. The decoding algorithm chosen must correct an error pattern that corrects every third symbol in the Reed–Solomon encoding. But after folding, this error pattern will corrupt all symbols over \mathbb_q^3 and will eliminate the need for error correction. This propagation of errors is indicated by the blue color in the graphical description. This proves that for a fixed fraction of errors \rho, the folding operation reduces the channel's flexibility to distribute errors, which in turn leads to a reduction in the number of error patterns that need to be corrected.


How folded Reed–Solomon (FRS) codes and Parvaresh Vardy (PV) codes are related

We can relate Folded Reed Solomon codes wit
Parvaresh Vardy
codes which encodes a polynomial f of degree k with polynomials f_0=f, f_1 , \ldots, f_ (s \geqslant 2) where f_i(X)=f_(X)^d \mod E(X) where E(X) is an
irreducible polynomial In mathematics, an irreducible polynomial is, roughly speaking, a polynomial that cannot be factored into the product of two non-constant polynomials. The property of irreducibility depends on the nature of the coefficients that are accepted f ...
. While choosing irreducible polynomial E(X)=X^q -\gamma and parameter d we should check if every polynomial f of degree at most k satisfies f(\gamma X)=f(X)^d \mod E(X) since f(\gamma X) is just the shifted counterpart of f(X) where \gamma is the primitive element in \mathbb_q. Thus folded RS code with bundling together code symbols is PV code of order s=m for the set of evaluation points :\left \. If we compare the folded RS code to a PV code of order 2 for the set of evaluation points : \left \ we can see that in PV encoding of f, for every 0 \leq i \leq n/m-1 and every 0 appears at f(\gamma^) and f_1(\gamma^\gamma^), unlike in the folded FRS encoding in which it appears only once. Thus, the PV and folded RS codes have same information but only the rate of FRS is bigger by a factor of 2(m-1)/m and hence the
list decoding In coding theory, list decoding is an alternative to unique decoding of error-correcting codes for large error rates. The notion was proposed by Elias in the 1950s. The main idea behind list decoding is that the decoding algorithm instead of outp ...
radius trade-off is better for folded RS code by just using the list decodability of the PV codes. The plus point is in choosing FRS code in a way that they are compressed forms of suitable PV code with similar error correction performance with better rate than corresponding PV code. One can use this idea to construct a folded RS codes of rate R that are list decodable up to radius approximately 1-R^ for s\geq 1.


Brief overview of list-decoding folded Reed–Solomon codes

A
list decoding In coding theory, list decoding is an alternative to unique decoding of error-correcting codes for large error rates. The notion was proposed by Elias in the 1950s. The main idea behind list decoding is that the decoding algorithm instead of outp ...
algorithm which runs in quadratic time to decode FRS code up to radius 1-R-\varepsilon is presented by Guruswami. The algorithm essentially has three steps namely the interpolation step in which welch berlekamp style interpolation is used to interpolate the non-zero polynomial :Q(X,Y_1,Y_2,\ldots,Y_s)=A_0(X) + A_1(X)Y_1 + A_2(X)Y_2 + \cdots + A_s(X)Y_s, after which all the polynomials f\in\mathbb_q /math> with degree k-1 satisfying the equation derived in interpolation are found. In the third step the actual list of close-by codewords are known by pruning the solution subspace which takes q^stime.


Linear-algebraic list decoding algorithm

Guruswami presents a n^ time list decoding algorithm based on linear-algebra, which can decode folded Reed–Solomon code up to radius 1-R-\varepsilon with a list-size of . There are three steps in this algorithm: Interpolation Step, Root Finding Step and Prune Step. In the Interpolation step it will try to find the candidate message polynomial f(x) by solving a linear system. In the Root Finding step, it will try to find the solution subspace by solving another linear system. The last step will try to prune the solution subspace gained in the second step. We will introduce each step in details in the following.


Step 1: The interpolation step

It is a Welch–Berlekamp-style interpolation (because it can be viewed as the higher-dimensional generalization of the Welch–Berlekamp algorithm). Suppose we received a codeword y of the m-folded Reed–Solomon code as shown below :\left(\beginy_0\\y_1\\y_2\\\cdots\\y_\end,\beginy_m\\y_\\y_\\\cdots\\y_\end,\ldots,\beginy_\\y_\\y_\\\cdots\\y_\end\right) We interpolate the nonzero polynomial : Q(X,Y_1,\ldots,Y_s)=A_0(X) + A_1(X)Y_1 + \cdots + A_s(X)Y_s, \qquad \begin \deg(A_i) \leqslant D & 1 \leqslant i \leqslant s \\ \deg(A_0) \leqslant D + k -1 & \end by using a carefully chosen degree parameter D. : D= \left \lfloor \frac \right \rfloor So the interpolation requirements will be : Q \left (\gamma^,y_,y_,\ldots,y_ \right )=0, \quad \text \quad i=0,1,\ldots, \tfrac - 1, j=0,1,\ldots,m-s. Then the number of monomials in Q(X,Y_1,\ldots,Y_s) is : (D + 1)s + D + k = (D+1)(s+1) + k -1 > N(m - s + 1) Because the number of monomials in Q(X,Y_1,\ldots,Y_s) is greater than the number of interpolation conditions. We have below lemma :Lemma 1. 0 \neq Q \in \mathbb_q ,Y_1,\ldots,Y_s/math> satisfying the above interpolation condition can be found by solving a homogeneous linear system over \mathbb_q with at most Nm constraints and variables. Moreover this interpolation can be performed in O(Nm\log^2(Nm)\log\log(Nm)) operations over \mathbb_q. This lemma shows us that the interpolation step can be done in near-linear time. For now, we have talked about everything we need for the multivariate polynomial Q(X,Y_1,\ldots,Y_s). The remaining task is to focus on the message polynomials f(X). :Lemma 2. If a candidate message polynomial f(X) \in \mathbb /math> is a polynomial of degree at most k-1 whose Folded Reed-Solomon encoding agrees with the received word y in at least t columns with ::t>, :then Q(X,f(X),f(\gamma X),\ldots,f(\gamma_X))=0. Here "agree" means that all the m values in a column should match the corresponding values in codeword y. This lemma shows us that any such polynomial Q(X,Y_1,\ldots,Y_s) presents an algebraic condition that must be satisfied for those message polynomials f(x) that we are interested in list decoding. Combining Lemma 2 and parameter D, we have :t(m-s+1)>\frac Further we can get the decoding bound :t\geqslant \frac+\frac\cdot\frac=N\left(\frac+\frac\cdot\frac\right) We notice that the fractional agreement is :\dfrac + \dfrac \cdot \dfrac


Step 2: The root-finding step

During this step, our task focus on how to find all polynomials f\in with degree no more than k-1 and satisfy the equation we get from Step 1, namely :A_0(X) + A_1(X)f(X) + A_2(X)f(\gamma X) + \cdots + A_s(X)f(\gamma^X)=0 Since the above equation forms a linear system equations over \mathbb_q in the coefficients f_0,f_1,\ldots,f_ of the polynomial :f(X) = f_0 + f_1X + \cdots + f_X^, the solutions to the above equation is an
affine subspace In mathematics, an affine space is a geometric structure that generalizes some of the properties of Euclidean spaces in such a way that these are independent of the concepts of distance and measure of angles, keeping only the properties related ...
of \mathbb^k_q. This fact is the key point that gives rise to an efficient algorithm - we can solve the linear system. It is natural to ask how large is the dimension of the solution? Is there any upper bound on the dimension? Having an upper bound is very important in constructing an efficient list decoding algorithm because one can simply output all the codewords for any given decoding problem. Actually it indeed has an upper bound as below lemma argues. :Lemma 3. If the order of \gamma is at least k (in particular when \gamma is primitive), then the dimension of the solution is at most s-1.For the details of the proof, please refer to Guruswami's paper. This lemma shows us the upper bound of the dimension for the solution space. Finally, based on the above analysis, we have below theorem :Theorem 1. For the folded Reed–Solomon code FRS^_q ,k/math> of block length N=\tfrac and rate R=\tfrac, the following holds for all integers s, 1 \leqslant s \leqslant m. Given a received word y \in (\mathbb_q^m)^N, in O((Nm\log q)^2) time, one can find a basis for a subspace of dimension at most s-1 that contains all message polynomials f \in \mathbb_q /math> of degree less than k whose FRS encoding differs from y in at most a fraction ::\frac\left(1-\frac\right) :of the N codeword positions. When s = m = 1, we notice that this reduces to a unique decoding algorithm with up to a fraction (1-R)/2 of errors. In other words, we can treat unique decoding algorithm as a specialty of list decoding algorithm. The quantity is about n^ for the parameter choices that achieve a list decoding radius of 1-R-\varepsilon. Theorem 1 tells us exactly how large the error radius would be. Now we finally get the solution subspace. However, there is still one problem standing. The list size in the worst case is n^. But the actual list of close-by codewords is only a small set within that subspace. So we need some process to prune the subspace to narrow it down. This prune process takes q^s time in the worst case. Unfortunately it is not known how to improve the running time because we do not know how to improve the bound of the list size for folded Reed-Solomon code. Things get better if we change the code by carefully choosing a subset of all possible degree k -1 polynomials as messages, the list size shows to be much smaller while only losing a little bit in the rate. We will talk about this briefly in next step.


Step 3: The prune step

By converting the problem of decoding a folded Reed–Solomon code into two linear systems, one linear system that is used for the interpolation step and another linear system to find the candidate solution subspace, the complexity of the decoding problem is successfully reduced to quadratic. However, in the worst case, the bound of list size of the output is pretty bad. It was mentioned in Step 2 that if one carefully chooses only a subset of all possible degree k -1 polynomials as messages, the list size can be much reduced. Here we will expand our discussion. To achieve this goal, the idea is to limit the coefficient vector (f_0,f_1,\ldots,f_) to a special subset \nu \subseteq \mathbb_q^k, which satisfies below two conditions: :Condition 1. The set \nu must be large enough (, \nu, \geq q^). This is to make sure that the rate will be at most reduced by factor of (1-\varepsilon). :Condition 2. The set \nu should have low intersection with any subspace S of dimension s satisfying S \subset \mathbb_q^k and , S \cap \nu, \leqslant L. Such a subset is called subspace-evasive subset. The bound for the list size at worst case is n^, and it can be reduced to a relative small bound O(1/\varepsilon^2) by using subspace-evasive subsets. During this step, as it has to check each element of the solution subspace that we get from Step 2, it takes q^s time in the worst case (s is the dimension of the solution subspace). Dvir and Lovett improved the result based on the work of Guruswami, which can reduce the list size to a constant. Here is only presented the idea that is used to prune the solution subspace. For the details of the prune process, please refer to papers by Guruswami, Dvir and Lovett, which are listed in the reference.


Summary

If we don't consider the Step 3, this algorithm can run in quadratic time. A summary for this algorithm is listed below.


See also

*
Reed–Solomon error correction Reed–Solomon codes are a group of error-correcting codes that were introduced by Irving S. Reed and Gustave Solomon in 1960. They have many applications, the most prominent of which include consumer technologies such as MiniDiscs, CDs, DVDs, ...
*
Singleton bound In coding theory, the Singleton bound, named after Richard Collom Singleton, is a relatively crude upper bound on the size of an arbitrary block code C with block length n, size M and minimum distance d. It is also known as the Joshibound. proved b ...
*
Coding theory Coding theory is the study of the properties of codes and their respective fitness for specific applications. Codes are used for data compression, cryptography, error detection and correction, data transmission and data storage. Codes are studied ...
*
List decoding In coding theory, list decoding is an alternative to unique decoding of error-correcting codes for large error rates. The notion was proposed by Elias in the 1950s. The main idea behind list decoding is that the decoding algorithm instead of outp ...


References


Atri Rudra
s Lecture Notes
Folded Reed–Solomon Codes
# Atri Rudra's Lecture Notes
Bounds
# A paper b
Atri Rudra
an
Venkatesan GuruswamiDecoding Folded Reed–Solomon Codes
# A chapter on List Decoding of folded Reed–Solomon codes
List Decoding of Folded Reed–Solomon Codes
# Venkatesan Guruswami's lecture notes
Elementary bounds on codes
# Venkatesan Guruswami's lecture notes
List Decoding Folded Reed–Solomon Code
# # # PHD Thesis b
Kristian BranderInterpolation and List Decoding of Algebraic Codes
# {{DEFAULTSORT:Folded Reed-Solomon code Coding theory Error detection and correction Computational complexity theory