scispace - formally typeset
Search or ask a question

Showing papers on "List decoding published in 2013"


Journal ArticleDOI
TL;DR: Unified descriptions of the SC, SCL, and SCS decoding algorithms are given as path search procedures on the code tree of polar codes and a new decoding algorithm called the successive cancellation hybrid (SCH) is proposed to provide a flexible configuration when the time and space complexities are limited.
Abstract: As improved versions of the successive cancellation (SC) decoding algorithm, the successive cancellation list (SCL) decoding and the successive cancellation stack (SCS) decoding are used to improve the finite-length performance of polar codes. In this paper, unified descriptions of the SC, SCL, and SCS decoding algorithms are given as path search procedures on the code tree of polar codes. Combining the principles of SCL and SCS, a new decoding algorithm called the successive cancellation hybrid (SCH) is proposed. This proposed algorithm can provide a flexible configuration when the time and space complexities are limited. Furthermore, a pruning technique is also proposed to lower the complexity by reducing unnecessary path searching operations. Performance and complexity analysis based on simulations shows that under proper configurations, all the three improved successive cancellation (ISC) decoding algorithms can approach the performance of the maximum likelihood (ML) decoding but with acceptable complexity. With the help of the proposed pruning technique, the time and space complexities of ISC decoders can be significantly reduced and be made very close to those of the SC decoder in the high signal-to-noise ratio regime.

212 citations


Proceedings ArticleDOI
09 Jun 2013
TL;DR: A simple quasi-uniform puncturing algorithm to generate the puncturing table is proposed and it is proved that this method has better row-weight property than that of the random puncturing.
Abstract: CRC (cyclic redundancy check) concatenated polar codes are superior to the turbo codes under the successive cancellation list (SCL) or successive cancellation stack (SCS) decoding algorithms. But the code length of polar codes is limited to the power of two. In this paper, a family of rate-compatible punctured polar (RCPP) codes is proposed to satisfy the construction with arbitrary code length. We propose a simple quasi-uniform puncturing algorithm to generate the puncturing table. And we prove that this method has better row-weight property than that of the random puncturing. Simulation results under the binary input additive white Gaussian noise channels (BI-AWGNs) show that these RCPP codes outperform the performance of turbo codes in WCDMA (Wideband Code Division Multiple Access) or LTE (Long Term Evolution) wireless communication systems in the large range of code lengths. Especially, the RCPP code with CRC-aided SCL/SCS algorithm can provide over 0.7dB performance gain at the block error rate (BLER) of 10-4 with short code length M = 512 and code rate R = 0.5.

199 citations


Patent
03 Dec 2013
TL;DR: In this article, a method for decoding block and concatenated codes based on belief propagation algorithms, with particular advantages when applied to codes having higher density parity check matrices, is presented.
Abstract: Systems and methods for decoding block and concatenated codes are provided. These include advanced iterative decoding techniques based on belief propagation algorithms, with particular advantages when applied to codes having higher density parity check matrices. Improvements are also provided for performing channel state information estimation including the use of optimum filter lengths based on channel selectivity and adaptive decision-directed channel estimation. These improvements enhance the performance of various communication systems and consumer electronics. Particular improvements are also provided for decoding HD Radio signals, including enhanced decoding of reference subcarriers based on soft-diversity combining, joint enhanced channel state information estimation, as well as iterative soft-input soft-output and list decoding of convolutional codes and Reed-Solomon codes. These and other improvements enhance the decoding of different logical channels in HD Radio systems.

132 citations


Journal ArticleDOI
TL;DR: An improved version of the simplified successive-cancellation decoding algorithm that increases decoding throughput without degrading the error-correction performance is presented.
Abstract: The serial nature of successive-cancellation decoding results in low polar decoder throughput. In this letter we present an improved version of the simplified successive-cancellation decoding algorithm that increases decoding throughput without degrading the error-correction performance. We show that the proposed algorithm has up to three times the throughput of the simplified successive-cancellation decoding algorithm and up to twenty-nine times the throughput of a standard successive-cancellation decoder while using the same number of processing elements.

119 citations


Journal ArticleDOI
TL;DR: Numerical results show that length-compatible polar codes designed by the proposed method provide a performance gain of about 1.0 - 5.0 dB over those obtained by random puncturing when successive cancellation decoding is employed.
Abstract: Length-compatible polar codes are a class of polar codes which can support a wide range of lengths with a single pair of encoder and decoder. In this paper we propose a method to construct length-compatible polar codes by employing the reduction of the 2n × 2n polarizing matrix proposed by Arikan. The conditions under which a reduced matrix becomes a polarizing matrix supporting a polar code of a given length are first analyzed. Based on these conditions, length-compatible polar codes are constructed in a suboptimal way by codeword-puncturing and information-refreezing processes. They have low encoding and decoding complexity since they can be encoded and decoded in a similar way as a polar code of length 2n. Numerical results show that length-compatible polar codes designed by the proposed method provide a performance gain of about 1.0 - 5.0 dB over those obtained by random puncturing when successive cancellation decoding is employed.

99 citations


Journal ArticleDOI
TL;DR: In this paper, a two-slice characterization of the parity polytope is presented, which simplifies the representation of points in the parity space and allows the decoding of large-scale error-correcting codes efficiently.
Abstract: When binary linear error-correcting codes are used over symmetric channels, a relaxed version of the maximum likelihood decoding problem can be stated as a linear program (LP). This LP decoder can be used to decode error-correcting codes at bit-error-rates comparable to state-of-the-art belief propagation (BP) decoders, but with significantly stronger theoretical guarantees. However, LP decoding when implemented with standard LP solvers does not easily scale to the block lengths of modern error correcting codes. In this paper, we draw on decomposition methods from optimization theory, specifically the alternating direction method of multipliers (ADMM), to develop efficient distributed algorithms for LP decoding. The key enabling technical result is a “two-slice” characterization of the parity polytope, the polytope formed by taking the convex hull of all codewords of the single parity check code. This new characterization simplifies the representation of points in the polytope. Using this simplification, we develop an efficient algorithm for Euclidean norm projection onto the parity polytope. This projection is required by the ADMM decoder and its solution allows us to use LP decoding, with all its theoretical guarantees, to decode large-scale error correcting codes efficiently. We present numerical results for LDPC codes of lengths more than 1000. The waterfall region of LP decoding is seen to initiate at a slightly higher SNR than for sum-product BP, however an error floor is not observed for LP decoding, which is not the case for BP. Our implementation of LP decoding using the ADMM executes as fast as our baseline sum-product BP decoder, is fully parallelizable, and can be seen to implement a type of message-passing with a particularly simple schedule.

98 citations


Patent
14 Jun 2013
TL;DR: In this article, an early decoding termination detection for QC-LDPC decoders is discussed, where the controller terminates decoding the data unit in response to determining that the decoded data units from more than one layer decoding operation satisfy a parity check equation.
Abstract: Embodiments of decoders having early decoding termination detection are disclosed. The decoders can provide for flexible and scalable decoding and early termination detection, particularly when quasi-cyclic low-density parity-check code (QC-LDPC) decoding is used. In one embodiment, a controller iteratively decodes a data unit using a coding matrix comprising a plurality of layers. The controller terminates decoding the data unit in response to determining that the decoded data units from more than one layer decoding operation satisfy a parity check equation and that the decoded data units from more than one layer decoding operation are the same. Advantageously, the termination of decoding of the data unit can reduce a number of iterations performed to decode the data unit.

96 citations


Journal ArticleDOI
TL;DR: It is demonstrated that in the case of multiple relays, there is no improvement on the achievable rate by joint decoding either, and it is discovered that any compressions not supporting successive decoding will actually lead to strictly lower achievable rates for the original message.
Abstract: In the classical compress-and-forward relay scheme developed by Cover and El Gamal, the decoding process operates in a successive way: the destination first decodes the compression of the relay's observation and then decodes the original message of the source. Recently, several modified compress-and-forward relay schemes were proposed, where the destination jointly decodes the compression and the message, instead of successively. Such a modification on the decoding process was motivated by realizing that it is generally easier to decode the compression jointly with the original message, and more importantly, the original message can be decoded even without completely decoding the compression. Thus, joint decoding provides more freedom in choosing the compression at the relay. However, the question remains in these modified compress-and-forward relay schemes-whether this freedom of selecting the compression necessarily improves the achievable rate of the original message. It has been shown by El Gamal and Kim in 2010 that the answer is negative in the single-relay case. In this paper, it is further demonstrated that in the case of multiple relays, there is no improvement on the achievable rate by joint decoding either. More interestingly, it is discovered that any compressions not supporting successive decoding will actually lead to strictly lower achievable rates for the original message. Therefore, to maximize the achievable rate for the original message, the compressions should always be chosen to support successive decoding. Furthermore, it is shown that any compressions not completely decodable even with joint decoding will not provide any contribution to the decoding of the original message. The above phenomenon is also shown to exist under the repetitive encoding framework recently proposed by Lim , which improved the achievable rate in the case of multiple relays. Here, another interesting discovery is that the improvement is not a result of repetitive encoding, but the benefit of delayed decoding after all the blocks have been finished. The same rate is shown to be achievable with the simpler classical encoding process of Cover and El Gamal with a block-by-block backward decoding process.

94 citations


Patent
10 Jul 2013
TL;DR: In this paper, the decoding paths are successively duplicated and selectively pruned to generate a list of potential decoding paths and a single decoding path among the list of possible decoding paths is selected as the output and a candidate codeword is thereby identified.
Abstract: A method of decoding data encoded with a polar code and devices that encode data with a polar code. A received word of polar encoded data is decoded following several distinct decoding paths to generate a list of codeword candidates. The decoding paths are successively duplicated and selectively pruned to generate a list of potential decoding paths. A single decoding path among the list of potential decoding paths is selected as the output and a single candidate codeword is thereby identified. In another preferred embodiment, the polar encoded data includes redundancy values in its unfrozen bits. The redundancy values aid the selection of the single decoding path. A preferred device of the invention is a cellular network device, (e.g., a handset) that conducts decoding in accordance with the methods of the invention.

94 citations


Journal ArticleDOI
TL;DR: It is proved that a random linear code over $\mathbb{F}_q, with probability arbitrarily close to 1, is list decodable at radius $1-1/q-\epsilon$ with list size and rate and the desired average-distance guarantees hold.
Abstract: We prove that a random linear code over $\mathbb{F}_q$, with probability arbitrarily close to 1, is list decodable at radius $1-1/q-\epsilon$ with list size $L=O(1/\epsilon^2)$ and rate $R=\Omega_q(\epsilon^2/(\log^3(1/\epsilon)))$. Up to the polylogarithmic factor in $1/\epsilon$ and constant factors depending on $q$, this matches the lower bound $L=\Omega_q(1/\epsilon^2)$ for the list size and upper bound $R=O_q(\epsilon^2)$ for the rate. Previously only existence (and not abundance) of such codes was known for the special case $q=2$ (Guruswami et al., 2002). In order to obtain our result, we employ a relaxed version of the well-known Johnson bound on list decoding that translates the average Hamming distance between codewords to list decoding guarantees. We furthermore prove that the desired average-distance guarantees hold for a code provided that a natural complex matrix encoding the codewords satisfies the restricted isometry property with respect to the Euclidean norm. For the case of random binary...

90 citations


Proceedings ArticleDOI
01 Dec 2013
TL;DR: Analysis and simulation of the iterative HDD of tightly-braided block codes with BCH component codes for high-speed optical communication shows that these codes are competitive with the best schemes based on HDD.
Abstract: Designing error-correcting codes for optical communication is challenging mainly because of the high data rates (e.g., 100 Gbps) required and the expectation of low latency, low overhead (e.g., 7% redundancy), and large coding gain (e.g., >9dB). Although soft-decision decoding (SDD) of low-density parity-check (LDPC) codes is an active area of research, the mainstay of optical transport systems is still the iterative hard-decision decoding (HDD) of generalized product codes with algebraic syndrome decoding of the component codes. This is because iterative HDD allows many simplifications and SDD of LDPC codes results in much higher implementation complexity. In this paper, we use analysis and simulation to evaluate tightly-braided block codes with BCH component codes for high-speed optical communication. Simulation of the iterative HDD shows that these codes are competitive with the best schemes based on HDD. Finally, we suggest a specific design that is compatible with the G.709 framing structure and exhibits a coding gain of >9.35 dB at 7% redundancy under iterative HDD with a latency of approximately 1 million bits.

Journal ArticleDOI
TL;DR: In this article, the problem of finding the shortest length index code with a prescribed error-correcting capability was studied, and the Singleton bound and two other bounds, referred to as the α-bound and the κ -bound, were established.
Abstract: A problem of index coding with side information was first considered by Birk and Kol in 1998. In this study, a generalization of index coding scheme, where transmitted symbols are subject to errors, is studied. Error-correcting methods for such a scheme, and their parameters, are investigated. In particular, the following question is discussed: given the side information hypergraph of index coding scheme and the maximal number of erroneous symbols δ , what is the shortest length of a linear index code, such that every receiver is able to recover the required information? This question turns out to be a generalization of the problem of finding a shortest length error-correcting code with a prescribed error-correcting capability in the classical coding theory. The Singleton bound and two other bounds, referred to as the α-bound and the κ -bound, for the optimal length of a linear error-correcting index code (ECIC) are established. For large alphabets, a construction based on concatenation of an optimal index code with a maximum distance separable classical code is shown to attain the Singleton bound. For smaller alphabets, however, this construction may not be optimal. A random construction is also analyzed. It yields another inexplicit bound on the length of an optimal linear ECIC. Further, the problem of error-correcting decoding by a linear ECIC is studied. It is shown that in order to decode correctly the desired symbol, the decoder is required to find one of the vectors, belonging to an affine space containing the actual error vector. The syndrome decoding is shown to produce the correct output if the weight of the error pattern is less or equal to the error-correcting capability of the corresponding ECIC. Finally, the notion of static ECIC, which is suitable for use with a family of instances of an index coding problem, is introduced. Several bounds on the length of static ECICs are derived, and constructions for static ECICs are discussed. Connections of these codes to weakly resilient Boolean functions are established.

Journal ArticleDOI
TL;DR: This work considers the decoding of spatially coupled codes through a windowed decoder that aims to retain many of the attractive features of belief propagation, while trying to reduce complexity further, by defining thresholds on channel erasure rates that guarantee a target erasure rate.
Abstract: Spatially coupled codes have been of interest recently owing to their superior performance over memoryless binary-input channels. The performance is good both asymptotically, since the belief propagation thresholds approach the Shannon limit, as well as for finite lengths, since degree-2 variable nodes that result in high error floors can be completely avoided. However, to realize the promised good performance, one needs large blocklengths. This in turn implies a large latency and decoding complexity. For the memoryless binary erasure channel, we consider the decoding of spatially coupled codes through a windowed decoder that aims to retain many of the attractive features of belief propagation, while trying to reduce complexity further. We characterize the performance of this scheme by defining thresholds on channel erasure rates that guarantee a target erasure rate. We give analytical lower bounds on these thresholds and show that the performance approaches that of belief propagation exponentially fast in the window size. We give numerical results including the thresholds computed using density evolution and the erasure rate curves for finite-length spatially coupled codes.

Journal ArticleDOI
TL;DR: In this article, lattice codes are used to achieve the same performance as known i.i.d. Gaussian random codes for the Gaussian relay channel, and several examples of how this may be combined with the linearity of lattices codes in multisource relay networks.
Abstract: Lattice codes are known to achieve capacity in the Gaussian point-to-point channel, achieving the same rates as i.i.d. random Gaussian codebooks. Lattice codes are also known to outperform random codes for certain channel models that are able to exploit their linearity. In this paper, we show that lattice codes may be used to achieve the same performance as known i.i.d. Gaussian random coding techniques for the Gaussian relay channel, and show several examples of how this may be combined with the linearity of lattices codes in multisource relay networks. In particular, we present a nested lattice list decoding technique in which lattice codes are shown to achieve the decode-and-forward (DF) rate of single source, single destination Gaussian relay channels with one or more relays. We next present two examples of how this DF scheme may be combined with the linearity of lattice codes to achieve new rate regions which for some channel conditions outperform analogous known Gaussian random coding techniques in multisource relay channels. That is, we derive a new achievable rate region for the two-way relay channel with direct links and compare it to existing schemes, and derive a new achievable rate region for the multiple access relay channel. We furthermore present a lattice compress-and-forward (CF) scheme for the Gaussian relay channel which exploits a lattice Wyner-Ziv binning scheme and achieves the same rate as the Cover-El Gamal CF rate evaluated for Gaussian random codes. These results suggest that structured/lattice codes may be used to mimic, and sometimes outperform, random Gaussian codes in general Gaussian networks.

Journal ArticleDOI
TL;DR: In this article, the rank-metric equivalent of Gabidulin codes is studied and three bounds on the list size of rank-matric codes are derived, and it is shown that polynomial-time list decoding is possible only with exponential time complexity.
Abstract: So far, there is no polynomial-time list decoding algorithm (beyond half the minimum distance) for Gabidulin codes These codes can be seen as the rank-metric equivalent of Reed-Solomon codes In this paper, we provide bounds on the list size of rank-metric codes in order to understand whether polynomial-time list decoding is possible or whether it works only with exponential time complexity Three bounds on the list size are proven The first one is a lower exponential bound for Gabidulin codes and shows that for these codes no polynomial-time list decoding beyond the Johnson radius exists Second, an exponential upper bound is derived, which holds for any rank-metric code of length n and minimum rank distance d The third bound proves that there exists a rank-metric code over \BBFqm of length n ≤ m such that the list size is exponential in the length for any radius greater than half the minimum rank distance This implies that there cannot exist a polynomial upper bound depending only on n and d similar to the Johnson bound in Hamming metric All three rank-metric bounds reveal significant differences to bounds for codes in Hamming metric

Journal ArticleDOI
TL;DR: A new family of channel codes, called ISI-free codes, are introduced, which improve the communication reliability while keeping the decoding complexity fairly low in the diffusion environment modeled by the Brownian motion.
Abstract: Molecular communications emerges as a promising scheme for communications between nanoscale devices. In diffusion-based molecular communications, molecules as information symbols diffusing in the fluid environments suffer from molecule crossovers, i.e., the arriving order of molecules is different from their transmission order, leading to intersymbol interference (ISI). In this paper, we introduce a new family of channel codes, called ISI-free codes, which improve the communication reliability while keeping the decoding complexity fairly low in the diffusion environment modeled by the Brownian motion. We propose general encoding/decoding schemes for the ISI-free codes, working upon the modulation schemes of transmitting a fixed number of identical molecules at a time. In addition, the bit error rate (BER) approximation function of the ISI-free codes is derived mathematically as an analytical tool to decide key factors in the BER performance. Compared with the uncoded systems, the proposed ISI-free codes offer good performance with reasonably low complexity for diffusion-based molecular communication systems.

Journal ArticleDOI
14 Nov 2013
TL;DR: This work analyzes the multivariate generalization of Howgrave-Graham's algorithm for the approximate common divisor problem and develops a corresponding lattice-based list decoding algorithm for Parvaresh-Vardy and Guruswami-Rudra codes, which are multivariate extensions of Reed-Solomon codes.
Abstract: We analyze the multivariate generalization of Howgrave-Graham’s algorithm for the approximate common divisor problem. In the m-variable case with modulus N and approximate common divisor of size Nβ , this improves the size of the error tolerated from Nβ 2 to Nβ (m+1)/m , under a commonly used heuristic assumption. This gives a more detailed analysis of the hardness assumption underlying the recent fully homomorphic cryptosystem of van Dijk, Gentry, Halevi, and Vaikuntanathan. While these results do not challenge the suggested parameters, a 2n e approximation algorithm with e < 2/3 for lattice basis reduction in n dimensions could be used to break these parameters. We have implemented the algorithm, and it performs better in practice than the theoretical analysis suggests. Our results fit into a broader context of analogies between cryptanalysis and coding theory. The multivariate approximate common divisor problem is the number-theoretic analogue of multivariate polynomial reconstruction, and we develop a corresponding lattice-based algorithm for the latter problem. In particular, it specializes to a lattice-based list decoding algorithm for ParvareshVardy and Guruswami-Rudra codes, which are multivariate extensions of Reed-Solomon codes. This yields a new proof of the list decoding radii for these codes.

Journal ArticleDOI
TL;DR: In this paper, a simple linear-algebra-based analysis of folded Reed-Solomon (RS) codes is presented, which eliminates the need for the computationally expensive root-finding step over extension fields.
Abstract: Folded Reed-Solomon (RS) codes are an explicit family of codes that achieve the optimal tradeoff between rate and list error-correction capability: specifically, for any e > 0, Guruswami and Rudra presented an nO(1/ e) time algorithm to list decode appropriate folded RS codes of rate R from a fraction 1-R-e of errors. The algorithm is based on multivariate polynomial interpolation and root-finding over extension fields. It was noted by Vadhan that interpolating a linear polynomial suffices for a statement of the above form. Here, we give a simple linear-algebra-based analysis of this variant that eliminates the need for the computationally expensive root-finding step over extension fields (and indeed any mention of extension fields). The entire list-decoding algorithm is linear-algebraic, solving one linear system for the interpolation step, and another linear system to find a small subspace of candidate solutions. Except for the step of pruning this subspace, the algorithm can be implemented to run in quadratic time. We also consider a closely related family of codes, called (order m) derivative codes and defined over fields of large characteristic, which consist of the evaluations of f as well as its first m-1 formal derivatives at N distinct field elements. We show how our linear-algebraic methods for folded RS codes can be used to show that derivative codes can also achieve the above optimal tradeoff. The theoretical drawback of our analysis for folded RS codes and derivative codes is that both the decoding complexity and proven worst-case list-size bound are nΩ(1/ e). By combining the above idea with a pseudorandom subset of all polynomials as messages, we get a Monte Carlo construction achieving a list-size bound of O(1/ e2) which is quite close to the existential O(1/ e) bound (however, the decoding complexity remains nΩ(1/ e)). Our work highlights that constructing an explicit subspace-evasive subset that has small intersection with low-dimensional subspaces-an interesting problem in pseudorandomness in its own right-could lead to explicit codes with better list-decoding guarantees.

Journal ArticleDOI
TL;DR: This paper proposes a linear-complexity algorithm for the projection onto a parity polytope (having a computational complexity of small O(d), where small d is the check-node degree), as compared to recent work .
Abstract: Linear program (LP) decoding has become increasingly popular for error-correcting codes due to its simplicity and promising performance. Low-complexity and efficient iterative algorithms for LP decoding are of great importance for practical applications. In this paper we focus on solving the binary LP decoding problem by using the alternating direction method of multipliers (ADMM). Our main contribution is that we propose a linear-complexity algorithm for the projection onto a parity polytope (having a computational complexity of small O(d), where small d is the check-node degree), as compared to recent work , which has a computational complexity of small O(d log d). In particular, we show that the projection onto the parity polytope can be transformed to a projection onto a simplex.

Journal ArticleDOI
TL;DR: This work shows a number of ways in which conventional error-correcting codes can be modified to correct errors in the Kendall space and presents several general constructions of codes in permutations that cover a broad range of code parameters.
Abstract: Rank modulation is a way of encoding information to correct errors in flash memory devices as well as impulse noise in transmission lines. Modeling rank modulation involves construction of packings of the space of permutations equipped with the Kendall tau distance. As our main set of results, we present several general constructions of codes in permutations that cover a broad range of code parameters. In particular, we show a number of ways in which conventional error-correcting codes can be modified to correct errors in the Kendall space. Our constructions are nonasymptotic and afford simple encoding and decoding algorithms of essentially the same complexity as required to correct errors in the Hamming metric. As an example, from binary Bose-Chaudhuri-Hocquenghem codes, we obtain codes correcting t Kendall errors in n memory cells that support the order of n!/(log2n!)t messages, for any constant t=1,2,.... We give many examples of rank modulation codes with specific parameters. Turning to asymptotic analysis, we construct families of rank modulation codes that correct a number of errors that grows with n at varying rates, from Θ(n) to Θ(n2). One of our constructions gives rise to a family of rank modulation codes for which the tradeoff between the number of messages and the number of correctable Kendall errors approaches the optimal scaling rate.

Proceedings ArticleDOI
14 Nov 2013
TL;DR: A real-time streaming system where messages created at regular time intervals at a source are encoded for transmission to a receiver over a packet erasure link and an upper bound on the decoding probability for any time-invariant code is derived.
Abstract: We consider a real-time streaming system where messages created at regular time intervals at a source are encoded for transmission to a receiver over a packet erasure link; the receiver must subsequently decode each message within a given delay from its creation time. We study a bursty erasure model in which all erasure patterns containing erasure bursts of a limited length are admissible. For certain classes of parameter values, we provide code constructions that asymptotically achieve the maximum message size among all codes that allow decoding under all admissible erasure patterns. We also study an i.i.d. erasure model in which each transmitted packet is erased independently with the same probability; the objective is to maximize the decoding probability for a given message size. We derive an upper bound on the decoding probability for any time-invariant code, and show that the gap between this bound and the performance of a family of time-invariant intrasession codes is small in the high reliability regime.

Journal ArticleDOI
TL;DR: It is shown that the coding scheme achieves the capacity region of noiseless WOMs when an arbitrary number of multiple writes is permitted and the results can be generalized from binary to generalized WOMs, described by an arbitrary directed acyclic graph.
Abstract: A coding scheme for write once memory (WOM) using polar codes is presented. It is shown that the scheme achieves the capacity region of noiseless WOMs when an arbitrary number of multiple writes is permitted. The encoding and decoding complexities scale as O(N log N), where N is the blocklength. For N sufficiently large, the error probability decreases subexponentially in N. The results can be generalized from binary to generalized WOMs, described by an arbitrary directed acyclic graph, using nonbinary polar codes. In the derivation, we also obtain results on the typical distortion of polar codes for lossy source coding. Some simulation results with finite length codes are presented.

Journal ArticleDOI
TL;DR: Simulation results demonstrate that the MCTC aided HARQ schemes are capable of significantly reducing the complexity of the appropriately selected benchmarkers, which is achieved without degrading the Packet Loss Ratio (PLR) and throughput.
Abstract: Hybrid Automatic Repeat reQuest (HARQ) has become an essential error control technique in communication networks, which relies on a combination of arbitrary error correction codes and retransmissions. When combining turbo codes with HARQ, the associated complexity becomes a critical issue, since conventionally iterative decoding is immediately activated after each transmission, even though the iterative decoder might fail in delivering an error-free codeword even after a high number of iterations. In this scenario, precious battery-power would be wasted. In order to reduce the associated complexity, we will present design examples based on Multiple Components Turbo Codes (MCTCs) and demonstrate that they are capable of achieving an excellent performance based on the lowest possible memory octally represented generator polynomial (2, 3)o. In addition to using low-complexity generator polynomials, we detail two further techniques conceived for reducing the complexity. Firstly, an Early Stopping (ES) strategy is invoked for curtailing iterative decoding, when its Mutual Information (MI) improvements become less than a given threshold. Secondly, a novel Deferred Iteration (DI) strategy is advocated for the sake of delaying iterative decoding, until the receiver confidently estimates that it has received sufficient information for successful decoding. Our simulation results demonstrate that the MCTC aided HARQ schemes are capable of significantly reducing the complexity of the appropriately selected benchmarkers, which is achieved without degrading the Packet Loss Ratio (PLR) and throughput.

Proceedings ArticleDOI
01 Jun 2013
TL;DR: This work considers Reed-Solomon codes whose evaluation points belong to a subfield, and gives a linear-algebraic list decoding algorithm that can correct a fraction of errors approaching the code distance, while pinning down the candidate messages to a well-structured affine space of dimension a constant factor smaller than the code dimension.
Abstract: We consider Reed-Solomon (RS) codes whose evaluation points belong to a subfield, and give a linear-algebraic list decoding algorithm that can correct a fraction of errors approaching the code distance, while pinning down the candidate messages to a well-structured affine space of dimension a constant factor smaller than the code dimension. By pre-coding the message polynomials into a subspace-evasive set, we get a Monte Carlo construction of a subcode of Reed-Solomon codes that can be list decoded from a fraction (1-R-e) of errors in polynomial time (for any fixed e > 0) with a list size of O(1/e). Our methods extend to algebraic-geometric (AG) codes, leading to a similar claim over constant-sized alphabets. This matches parameters of recent results based on folded variants of RS and AG codes. but our construction here gives subcodes of Reed-Solomon and AG codes themselves (albeit with restrictions on the evaluation points).Further, the underlying algebraic idea also extends nicely to Gabidulin's construction of rank-metric codes based on linearized polynomials. This gives the first construction of positive rate rank-metric codes list decodable beyond half the distance, and in fact gives codes of rate R list decodable up to the optimal (1-R-e) fraction of rank errors. A similar claim holds for the closely related subspace codes studied by Koetter and Kschischang.We introduce a new notion called subspace designs as another way to pre-code messages and prune the subspace of candidate solutions. Using these, we also get a deterministic construction of a polynomial time list decodable subcode of RS codes. By using a cascade of several subspace designs, we extend our approach to AG codes, which gives the first deterministic construction of an algebraic code family of rate R with efficient list decoding from 1-R-e fraction of errors over an alphabet of constant size (that depends only on e). The list size bound is almost a constant (governed by log* (block length)), and the code can be constructed in quasi-polynomial time.

Proceedings ArticleDOI
07 Jul 2013
TL;DR: In this article, the theory of rank metric and Gabidulin codes is transposed to the case of fields of characteristic zero and the Frobenius automorphism is replaced by any element of the Galois group.
Abstract: We transpose the theory of rank metric and Gabidulin codes to the case of fields of characteristic zero. The Frobenius automorphism is then replaced by any element of the Galois group. We derive some conditions on the automorphism to be able to easily transpose the results obtained by Gabidulin as well and a classical polynomial-time decoding algorithm. We also provide various definitions for the rank-metric.

Journal ArticleDOI
TL;DR: Simulation results show that the performance degradation caused by the iterative multistage decoding algorithms is relevant to the code structure and can be utilized to trade off the performance against the complexity.
Abstract: This letter is concerned with a class of nonbinary low-density parity-check (LDPC) codes, referred to as column-scaled LDPC (CS-LDPC) codes, whose parity-check matrices have a property that each column is a scaled binary vector. The CS-LDPC codes, which include algebraically constructed nonbinary LDPC codes as subclasses, admit fast encoding and decoding algorithms. Specifically, for a code over the finite field F2p, the encoder can be implemented with p parallel binary LDPC encoders followed by a series of bijective mappers, while the decoder can be implemented with an iterative decoder in which no message permutations are required during the iterations. In addition, there exist low-complexity iterative multistage decoders that can be utilized to trade off the performance against the complexity. Simulation results show that the performance degradation caused by the iterative multistage decoding algorithms is relevant to the code structure.

Journal ArticleDOI
TL;DR: This brief study studies the application of a similar technique to a class of Euclidean geometry low density parity check (EG-LDPC) codes that are one step majority logic decodable and shows that the method is also effective for EG- LDPC codes.
Abstract: In a recent paper, a method was proposed to accelerate the majority logic decoding of difference set low density parity check codes. This is useful as majority logic decoding can be implemented serially with simple hardware but requires a large decoding time. For memory applications, this increases the memory access time. The method detects whether a word has errors in the first iterations of majority logic decoding, and when there are no errors the decoding ends without completing the rest of the iterations. Since most words in a memory will be error-free, the average decoding time is greatly reduced. In this brief, we study the application of a similar technique to a class of Euclidean geometry low density parity check (EG-LDPC) codes that are one step majority logic decodable. The results obtained show that the method is also effective for EG-LDPC codes. Extensive simulation results are given to accurately estimate the probability of error detection for different code sizes and numbers of errors.

Journal ArticleDOI
TL;DR: In this article, the authors considered weighted Reed-Muller codes over point ensembles S 1 × · · · × S m where S i needs not be of the same size as S j. For m = 2, they determined optimal weights and analyzed the impact of the ratio |S 1|/|S 2| on the minimum distance.
Abstract: We consider weighted Reed---Muller codes over point ensemble S 1 × · · · × S m where S i needs not be of the same size as S j . For m = 2 we determine optimal weights and analyze in detail what is the impact of the ratio |S 1|/|S 2| on the minimum distance. In conclusion the weighted Reed---Muller code construction is much better than its reputation. For a class of affine variety codes that contains the weighted Reed---Muller codes we then present two list decoding algorithms. With a small modification one of these algorithms is able to correct up to 31 errors of the [49,11,28] Joyner code.

Proceedings ArticleDOI
07 Jul 2013
TL;DR: In this article, a scheme for concatenating binary polar codes with interleaved Reed-Solomon codes is proposed, which achieves the capacity-achieving property of polar codes, while having a significantly better errordecay rate.
Abstract: A scheme for concatenating the recently invented polar codes with interleaved block codes is considered. By concatenating binary polar codes with interleaved Reed-Solomon codes, we prove that the proposed concatenation scheme captures the capacity-achieving property of polar codes, while having a significantly better error-decay rate. We show that for any e > 0, and total frame length N, the parameters of the scheme can be set such that the frame error probability is less than 2-N 1-e, while the scheme is still capacity achieving. This improves upon 2-N 0.5-e, the frame error probability of Arikan's polar codes. We also propose decoding algorithms for concatenated polar codes, which significantly improve the error-rate performance at finite block lengths while preserving the low decoding complexity.

Proceedings ArticleDOI
07 Jul 2013
TL;DR: This paper investigates the use of instantly decodable network coding (IDNC) for minimizing the mean decoding delay in multicast cooperative data exchange systems, where the clients cooperate with each other to obtain their missing packets.
Abstract: This paper investigates the use of instantly decodable network coding (IDNC) for minimizing the mean decoding delay in multicast cooperative data exchange systems, where the clients cooperate with each other to obtain their missing packets. Here, IDNC is used to reduce the decoding delay of each transmission across all clients. We first introduce a new framework to find the optimum client and coded packet that result in the minimum mean decoding delay. However, since finding the optimum solution of the proposed framework is NP-hard, we further propose a heuristic algorithm that aims to minimize the lower bound on the expected decoding delay in each transmission. The effectiveness of the proposed algorithm is assessed through simulations.