The present application is related to and claims the benefit under 35 U.S.C. § 119(a) of a Russian patent application filed on Apr. 10, 2014 in the Russian Intellectual Property Office and assigned Serial No. 2014114215, the entire disclosure of which is hereby incorporated by reference.
The present application relates generally to noiseless coding and, more specifically, to the use of twisted polar code in data encoding and decoding.
Polar codes proposed by E. Arikan “Channel polarization: A method for constructing capacity achieving codes for symmetric binary-input memoryless channels” (IEEE Transactions On Information Theory, Volume 55, No. 7, pages 3051-3073, July 2009 (REF 1)) represent a first class of correcting codes that achieve the bandwidth of a broad class of communications channels. Polar codes are linear block codes generated by a submatrix of matrix A=BmR⊗m=R⊗mBm where Bm is a bit-organized permutation matrix
and ⊗m is m-fold Kronecker product of the matrix by itself. Thus, the code word of the polar code is c=uA, c∈GF(2)n, n=2m, where ui, i∈(0, 1, . . . , n−1)/F are coded information symbols, and ui, i∈F are assumed to be 0 (frozen symbols, such as symbols whose values never change). Classic method of decoding y obtained by transmission of c over a memory-free channel, symmetric by output, is the successive cancellation method (SC). The method comprises successively finding the most probable a posteriori values ui using values u0, . . . , ui−1, y0, . . . , yn−1. The found value ui is used in the subsequent steps of the decoding algorithm. It can be shown that the combination of a linear transformation specified by matrix A, physical data channel and SC decoder (assuming that intermediate solutions are correct) are represented as a set of n bit subchannels of various quality, such as each ui is assumed to be transmitted over i-th subchannel. A quality (such as bandwidth, probability of error per bit) of these subchannels is analyzed using the probability density evolution method consisting in computing the density of distribution of log likelihood ratios computed at various stages of the SC method (I. Tal and A. Vardy “How to construct polar codes”, IEEE Transactions On Information Theory, Volume 59, No. 10, October 2013 (REF 2)), or its Gaussian approximation (P. Trifonov “Efficient design and decoding of polar codes”, IEEE Transactions on Communications, Volume 60, No. 11, pages 3221-3227, November 2012 (REF 3)). A possible method of constructing a polar code of length n and dimension k is to select F as a plurality of indices of the worst bit subchannels.
Noiseless coding is widely used in current systems of communication and storage of information. It is a primary object to provide a new structure of error-correcting codes and respective methods of encoding and decoding data. The present codes represent a generalization of recently suggested polar codes. The present approach solves the problem of small minimum distance, because of which the error correction capability of existing polar codes is significantly worse than that of LDPC and turbo codes. As a result, the present codes provide a lower probability of error in the waterfall region and a lower level of error probability saturation than LDPC and turbo codes. The present decoding method relies on the successive cancellation stack algorithm and improves it through the use of directed search methods. It also takes into account the structure of inventive twisted polar codes. Use of this decoding method with twisted polar codes provides some reduction in the decoding complexity as compared to classical polar codes.
In a first example, a method of encoding data is provided. The method includes pre-coding, by a pre-coding module, data presented in the form of k-dimensional binary vector x. The coding includes in computing u(0)=xW, where W is a matrix of dimension κ×2m, and m is a code parameter. The method also includes performing, by a pre-coding module, m-layer twisted polarization transformation of vector u(0). The i-th transformation layer includes in partitioning vector u(i−1) into 2m−1 sub-vectors of length 2 and multiplying the subvectors by matrix
The method further includes merging the resulting subvectors in one vector of dimension 2m and permuting elements of the resulting vector to obtain vector u(i), so that vector u(m) is obtained on m-th layer. The permutations are selected from a common affinity group. The method includes obtaining, by a multiplexer of an encoding apparatus, data encoded with twisted polar code by multiplexing elements of vector u(m). Use of matrix Wallows representing some elements (symbols) of vector u(0), the number of which is equal to n−k, as linear functions of its other elements (symbols). Such symbols are referred to as dynamically frozen. Note that classical polar codes is a special case of this structure, in which all these linear functions are identically equal to 0 and respective symbols are referred to as (statically) frozen.
In further aspects of the invention, matrix W is selected so that the resulting codewords are codewords of base extended BCH (Bose-Chaudhuri-Hocquenghem) code. The permutations are specified for subvectors of vector u(i) including 2i contiguous elements by indexing elements of the subvectors with binary vectors of length i and specifying basis of i-dimensional linear spaces for each of the vectors, so that the output index is obtained as the product of input index by the matrix corresponding to this basis.
In a second example, a method of decoding encoded data is provided. The encoded data is encoded, for example, according to the above method of encoding where decoded data is obtained in a decoding apparatus using a successive cancellation method. In the successive cancellation method probabilities of paths in a code tree are computed recursively on m layers, and probabilities on layer i−1 are computed from probabilities permuted on layer i, wherein the used permutation is the same as used in the above method of encoding, and the decoding apparatus selects paths with regard to values of dynamically frozen symbols.
In further aspects of disclosure, values of dynamically frozen symbol are obtained from equation u(0)VT=0, where WVT=0. The matrix V has the form in which no more than one row starts in each column, and no more than one row ends in each column. Values of dynamically frozen symbols are found using symbolic expressions of intermediate data of the decoding apparatus. The decoding apparatus stores a plurality of paths corresponding to sequences (u0(0), . . . , uϕl−1(0)), and iteratively selects one of them for expansion to obtain one or two paths (u0(0), . . . , uϕl−1(0), uϕl(0)), uϕl(0)∈{0, 1}, so that constraints of dynamic freezing of symbols are met, wherein the selected path corresponds to the highest value P(uo(0), . . . , uϕ
In a third embodiment an encoding apparatus is provided. The encoding apparatus includes a module configured to precode data presented in the form of k-dimensional binary vector x, adapted to compute u(0)=xW, where W is a matrix of dimension κ×2m and m is a code parameter. The module is configured to perform “exclusive or” logic function. The module is also configured to transmit data without change. The data is joined to m layers to enable m-layer twisted polarization transformation of vector u(0) and obtain vector u(m), wherein i-th transformation layer consists in partitioning vector u(i−1) into 2m−1 subvectors of length 2, multiplying the subvectors by matrix
merging the resulting subvectors in one vector of dimension 2m and permuting elements of the resulting vector to obtain vector u(i), wherein the permutations are selected from a common affinity group. The apparatus also includes a multiplexer configured to produce a codeword encoded with twisted code from vector u(m) obtained on m-th transformation layer from vector u(0).
In a fourth example a decoding apparatus is provided. The decoding apparatus includes a first module configured to compute probabilities adapted to compute probabilities of intermediate symbols on m layers. The probabilities on (i−1)-th layer are obtained from permuted probabilities on layer i, the used permutation being the same that in the above encoding method. The first module is also configured to transmit the computed probability to second module for sequential processing. The apparatus also includes a second module configured to iteratively build paths in a code tree by retrieving a path with maximum metric from a stack. The second module is also configured to expand the path. The second module is further configured to store obtained paths in the stack. The second module is configured to determine values of decoded symbols. The decoding apparatus further includes a stack configured to store path identifiers and metrics. The stack is connected communicatively to the second module. The decoding apparatus includes a third module configured to update values of intermediate symbols. The third module is also configured to recursively compute values of intermediate symbols on m layers. The third module is configured to transmit the computed values of intermediate symbols to a memory device. The decoding apparatus also includes a memory device configured to store values of intermediate symbols u(i) and transmit the values of intermediate symbols to the module for sequential processing and the first module to compute probabilities.
The present invention offers a code structure and respective methods and apparatus for encoding and decoding, which provide higher correction capability and lower decoding complexity compared to classical polar codes and their improvements discussed above. Numerical results show that twisted polar codes provide a coding gain of up to 0.5 dB as compared with turbo and LDPC codes, and a better error correction capability compared with untwisted and classic polar codes with external CRC. Twisted polar codes also allow reducing the level of error probability saturation (the phenomenon observed with turbo codes and untwisted polar codes). Furthermore, the present decoding method uses heuristic function, which enables the number of iterations performed by the stack decoding algorithm to be significantly reduced.
The present invention increases the speed of encoding and/or decoding through the reduction in the number of iterations to be performed.
Before undertaking the DETAILED DESCRIPTION below, it may be advantageous to set forth definitions of certain words and phrases used throughout this patent document: the terms “include” and “comprise,” as well as derivatives thereof, mean inclusion without limitation; the term “or,” is inclusive, meaning and/or; the phrases “associated with” and “associated therewith,” as well as derivatives thereof, may mean to include, be included within, interconnect with, contain, be contained within, connect to or with, couple to or with, be communicable with, cooperate with, interleave, juxtapose, be proximate to, be bound to or with, have, have a property of, or the like; and the term “controller” means any device, system or part thereof that controls at least one operation, such a device may be implemented in hardware, firmware or software, or some combination of at least two of the same. It should be noted that the functionality associated with any particular controller may be centralized or distributed, whether locally or remotely. Definitions for certain words and phrases are provided throughout this patent document, those of ordinary skill in the art should understand that in many, if not most instances, such definitions apply to prior, as well as future uses of such defined words and phrases.
For a more complete understanding of the present disclosure and its advantages, reference is now made to the following description taken in conjunction with the accompanying drawings, in which like reference numerals represent like parts:
Polar codes proposed by E. Arikan “Channel polarization: A method for constructing capacity achieving codes for symmetric binary-input memoryless channels” (IEEE Transactions On Information Theory, Volume 55, No. 7, pages 3051-3073, July 2009 (REF 1)) represent a first class of correcting codes that achieve the bandwidth of a broad class of communications channels. Polar codes are linear block codes generated by a submatrix of matrix A=BmR⊗m=R⊗mBm where Bm is a bit-organized permutation matrix
and ⊗m is m-fold Kronecker product of the matrix by itself. Thus, the code word of the polar code is c=uA, c∈GF(2)n, n=2m where ui, i∈(0, 1, . . . , n−1)\F are coded information symbols, and ui, i∈F are assumed to be 0 (frozen symbols, such as symbols whose values never change). Classic method of decoding y obtained by transmission of c over a memory-free channel, symmetric by output, is the successive cancellation method (SC). The method comprises successively finding the most probable a posteriori values ui using values u0, . . . , ui−1, y0, . . . , yn−1. The found value ui is used in the subsequent steps of the decoding algorithm. It can be shown that the combination of a linear transformation specified by matrix A, physical data channel and SC decoder (assuming that intermediate solutions are correct) are represented as a set of n bit subchannels of various quality, such as each ui is assumed to be transmitted over i-th subchannel. A quality (such as bandwidth, probability of error per bit) of these subchannels is analyzed using the probability density evolution method consisting in computing the density of distribution of log likelihood ratios computed at various stages of the SC method (I. Tal and A. Vardy “How to construct polar codes”, IEEE Transactions On Information Theory, Volume 59, No. 10, October 2013 (REF 2)), or its Gaussian approximation (P. Trifonov “Efficient design and decoding of polar codes”, IEEE Transactions on Communications, Volume 60, No. 11, pages 3221-3227, November 2012 (REF 3)). A possible method of constructing a polar code of length n and dimension k is to select F as a plurality of indices of the worst bit subchannels.
The polar code is represented as a tree whose leaves are its codewords, and paths from the root to the leaves are defined by possible values of symbols u0, . . . , un−1. Optimal decoder find the most probable path in the tree, corresponding to received vector (y0, . . . , yn−1). The above approach is used to construct codes reaching the channel bandwidth at n→∞. However, the error correction capability of polar codes of average length when they are decoded by the SC method is quite low. I. Tal and A. Vardy “List decoding of polar codes”, IEEE International Symposium on Information Theory, 2011 (REF 4) has shown this problem is solved by list SC decoding. That is, instead of making a single decision on ui, at each step of the SC decoder two paths can be analyzed, which correspond to different values of unfrozen symbol ui. As a result, the number of paths (u0, . . . , ui) monitored by the decoder is doubled at each step i∉F. To prevent the exponential growth of complexity and memory consumption it is suggested (as shown in REF 4) to retain no more than L most probable paths at every step. Probability of path Pn(i)(u0i|y0n−1) is computed according to
where αlj=(αi, αi−1, . . . , αj), αi,ej and αi,0j are subvectors αlj with even and odd numbered components, respectively, and Pl(0)(α|y) is the probability that symbol α∈GF(2) was transmitted with the proviso that the channel outputs symbol y (see REF 4).
The described approach significantly improves the error correction capability of polar codes in the low SNR domain. However, classical polar codes exhibit the effect of error probability saturation of a sufficiently high level, since their minimum Hamming distance is min i∉F wt(Ai), where Ai is i-th row of matrix A, and wt(x) is the Hamming weight of vector x. Data is encoded with CRC error detection code before it is encoded with polar code (see REF 4). Then among L paths in the code tree, found by the above algorithm, paths that have correct CRC checksum are chosen. It was shown experimentally that this approach provides a fairly good error correction capability, although no non-trivial estimates of the minimum distance have been found.
To reduce the complexity of decoding, K. Niu and K. Chen “CRC-aided decoding of polar codes”, IEEE Communications Letters, Volume 16, No. 10, October 2012 (REF (5), suggests that at every step of the SC decoder the most probable path, such as the path with the highest P{u0, . . . , ui−1|y}, is to be chosen for further disclosure. The decoder maintains a stack (such as a priority queue) that contains paths of different lengths, so to effectively select the most probable of them. This approach (stack SC decoding) ensures 2-3-fold reduction in complexity compared to the list SC decoding without any loss in the error correction capability.
Another implementation of this approach in case of polar codes with external CRC is presented by B. Li, H. Shen, D. Tse “An Adaptive Successive Cancellation List Decoder for Polar Codes with Cyclic Redundancy Check”, IEEE Communications Letters, 16 (12), December 2012 [6], and B. Li, H. Shen “Decoding method and decoding apparatus for polar code concatenated with cyclic redundancy check, patent application” (WO2013107140) (REF 7). The implementation comprises application of the Tala-Vardi algorithm of list decoding with a certain size of list L, searching among the found codewords for words with correct CRC and, in the absence thereof, re-starting the list decoder with double size of the list. These steps are repeated until a code word with correct CRC is found or the size of the list exceeds a predetermined value. The main disadvantage of this approach is its high computational complexity that is proportional to the maximum size of the list.
The problem of decoding polar codes is also analyzed by W. Gross, G. Sarkis, A. Raymond, S. Leroux, I. Tal, A. Vardy “Methods and Systems for Decoding Polar Codes” (US20130117344) (REF 8), where some methods are suggested to reduce the complexity of a non-list SC decoder. However the methods do not increase the error correction capability. Data encoding methods according to this disclosure provide a much smaller probability of decoding error per codeword as compared to REF 8.
The present disclosure provides an apparatus and methods to provide a regular structure of encoder 010 and decoder, like many modern designs of LDPC codes, for example, RU2395902, T. Richardson, X. Jin, “Methods and Apparatus for LDPC-encoding” (REF 9). However, the error correction capability of the present code structure is much better.
Basic idea of the present approach, which is reflected in respective methods for encoding and decoding, relies on the introduction of permutation of intermediate symbols appearing in the polarization transformation before forwarding them to the next layer of polarization transformation. This allows dynamic freezing of symbols, providing a sufficiently large minimum distance of the code, where all quite bad bit subchannels are frozen. Constrains of dynamic freezing are obtained using a parent extended BCH (Bose-Chaudhuri-Hocquenghem) code, and permutations are selected from a common affinity group, which is a group of automorphisms of Reed-Muller sub-codes for this code (F. J. MacWilliams and N. J. A. Sloane “The Theory of Error-Correcting Codes”, 1977 (REF 10)). More specifically, elements of sub-blocks of intermediate vectors of polarization transformation are indexed with binary vectors which are equivalently considered as integers. For each layer and each sub-block, Ci,j,0, . . . , Ci,j,i−1 basis of corresponding 1-dimensional linear subspace is indicated, so that output index x is computed from input s=s0+2s1+ . . . +2i−2si−2, Sj∈{0, 1} as x=s0Ci,j,1⊕s1Ci,j,2 . . . ⊕si−2Ci,j,2. This defines permutation of vector elements. For example, matrix
corresponds to permutation 0→0, 2→3, 4→4, 6→7, 1→2, 3→1, 5→6, 7→. Formally, the present method of encoding for a code of length n=2m comprises the following steps:
At a first step, binary vector u(0) are constructed so that the dynamic freezing constraints, specified by a matrix V, are observed, such as u(0)VT=0.
At a second step, let i=1.
At a third step, let u2
At a fourth step, let i=i+l. If i≤m, and return to the third step.
It is seen that the encoder 010 structure described in the third step is similar to the structure used in the encoder of classical Arikan polar codes. Note that the latter are obtained by selecting Ci,j,l equal to l-th unit vector and using only statically frozen symbols. This allows the use of SC decoder and its modifications taking into account inventive modifications. In particular, the effective embodiment of a list SC decoder described in the fourth step or its stack version, illustrate in
Let A′ is n×n matrix corresponding to the transformation defined by the third and fourth steps of the above algorithm. To provide a sufficiently large minimum distance of the resulting code an additional constraint is applied, which consist in the fact that codewords u(0)A′ obtained in the aforementioned manner belong to another code (herein it is assumed that this code is an extended BCH code) with test matrix H, such as u(0)A′HT=0. Then matrix V, used in the first step of the encoding algorithm, is obtained as VT=A′HTQ, where Q is an invertible matrix. It is suggested to select matrix Q so that the complexity of recovery of intermediate symbols in processing of dynamic freezing constraints is minimal.
To reduce complexity of the stack decoding algorithm it is suggested to use values Pl,m[0, α]ψ(ϕl+1) instead of Pl,m[0, α]α∈{0, 1} as path metrics, where ψ(i) is the heuristic function, which defines the average probability of correct decisions after phase i. This allows the decoder to avoid frequent switching between short and long paths, thereby reducing the total number of iterations.
This transformation is realized by “exclusive or” unit 004, which computes the modulo-2 sum of input values, and pass-through unit 005, which outputs the input value without change. This yields vector u(1), which is further processed in a similar manner. The module configured to perform “exclusive or” function and the pass-through module (configured to transmit data unchanged) are arranged in m layers. On layer i intermediate vector u(i+1) is represented as 2m−i−1 subvectors of length 2i+1. Prior to transmitting the intermediate vector to the next layer, its components are permuted, and components of different subvectors are permuted in different manner. Vector c=u(4) thus obtained is denoted by final vector 007. Its elements are multiplexed using a multiplexor 008 into codeword 009 of twisted polar code. This codeword is further transmitted over a communications channel or stored by a recording device. Note that the rearrangement on the last layer is not obligatory. This rearrangement is shown here only to illustrate a specific permutation structure described below.
It should be understood that the twisted polar code is a linear block code. The above encoding method corresponds to computing:
c=xG (3)
where G=WA′ is a code generator k×n matrix, and A′ specifies twisted polarization transformation. However, the same code is obtained by replacing G in equation (3) by G′=QG, where Q is an invertible k×k matrix. In particular, in some applications it is advisable to use systematic encoding, such as select G′=(I|S)P, where I is a unity matrix, S is a k×(n−k) matrix and P is a permutation matrix. In this case, the resulting codeword contains useful data vector x as a subvector. Although the structure of systematic (or any other) encoder differs from that described above for example as illustrate in
Matrix V is used to provide a sufficiently large minimum distance d≥d0 of the resulting code. It is constructed as follows. Let H is a test matrix of a base code of length n with minimum distance d0. Let A′ is an n×n matrix corresponding to twisted polarization transformation. Vector u(0) should be constructed so that the respective codeword c of the twisted polar code is also a codeword of the base code, such as u(0)A′HT=0. Thus, V=QTHA′T is obtained, where Q is an arbitrary invertible matrix. To increase the error correction capability, matrix V is obtained and is supplemented with rows of weight 1. Particular added rows are defined with account of the error probability in bit subchannels defined by the polarization transformation A′. l′i=min(j|Vij=l) and l″i=max(j|Vij=l) are made to be the initial and final position of row i, respectively. With elementary operations performed on the rows of matrix HAT, V is obtained such that no more than one row begins and no more than one row ends at position i, 0≤i<n. The latter property means that
ui(0)=Σj<i:Vtj=luj(0), i∈F (4)
where t is the index of the row ending at position i, F={i|∃j:lj″=i}, and modulo-2 sum is computed. The first property generally reduces the number of terms in REF 4 and simplifies computing the expression. Symbols ui(0), i∈F are referred to as frozen. Frozen symbols for which l′i=l″i (such as ui(0)=0) are referred to as statically frozen symbols, while those for which l′i≠l″i are referred to as dynamically frozen symbols.
Preferably, extended BCH code is used as the base code in the described structure. With this, the number of positions i, for which REF 4 comprises a non-trivial sum, is small, which simplifies decoding. Furthermore, this allows the use of a specific structure of permutations, which is compactly specified. The permutations are determined by indexing elements of sub-vectors of length 2i of vectors u(i) with binary vectors of length i and specifying different basis of corresponding i-dimensional linear spaces. For each layer i and each subvector u2
At a first step it is assumed that ui(0), i∉F are equal to elements of payload vector x, and compute remaining elements ui(0) according to (2). Note that this is equivalent to computing u(0)=xW for corresponding matrix W.
At a second step, let i=1.
At a third step, u2
At a fourth step i=i+l. If i≤m, then go to the third step.
At a fifth step a codeword is obtained as u(m).
It should be appreciated that there are other equivalent encoding methods, which generate the same plurality of codewords. All these methods are obtained by replacing matrix W with W′=QW in the above algorithm, where Q is a reversible k×k matrix.
For example,
C1,j,0=(1), 0≤j<8,
C2,j,0=(1,0)2=2, C2,j,1=(0,1)2=1, 0≤j<4,
C3,0,0=(0,1,0)2=2, C3,0,1=(0,0,1)2=1, C3,0,2=(1,0,0)2=4,
C3,1,0=(1,0,0)2=4, C3,1,1=(0,0,1)2=1, C3,1,2=(0,1,0)2=2,
C4,0,0=(1,1,0,0)2=12, C4,0,1=(0,0,0,1)2=1, C4,0,2=(0,0,1,0)2=2,
C4,0,3=(0,1,0,0)2=4.
It is seen that there are two distinctions over the classical polar encoder as illustrated in
Note that it is sufficient to only indicate Ci,j,0, and specify a deterministic algorithm for finding linearly independent vectors Ci,j,s, 1≤s≤i−l. Thus, each permutation is specified by i bits only, which is significantly less than log2 21 ! bits for an arbitrary permutation.
The specific form of permutations used should be determined such that to minimize the probability of errors in resulting bit subchannels. This is accomplished, for example, by complete enumeration (or one or more randomized searches) of all possible permutations defined by nondegenerate m×m binary matrices and computation of the error probability in corresponding unfrozen bit subchannels using density evolution.
The present encoder is used in data transmission or data storage systems.
The SC decoder, its list option 4 and stack option 5 should take a decision on each symbol ui(0). In the case of dynamic frozen symbols these decisions are made based on the values of symbols found at the previous algorithm steps. However, the embodiment described in 4 does not store these values explicitly. In this regard, it is suggested to recover them from the intermediate values stored in array C filled by RecursivelyCalcC function as shown in
and β=ΣiBi2i. This expression is constructed in the symbolic form when designing the decoder, so that REF 4 is expressed in terms Cl, λ[β, τ]. It is seen that the number of terms in the obtained expression for uj(o), such as its computation complexity, increases with i−j. This fact justifies the need for matrix V, such that no more than one row starts and ends at every position, since in this case i−j is usually minimized.
Furthermore, the following method is suggested to reduce the complexity of stack decoding, which is used both with twisted and conventional polar codes. Note that the MAP decoder must find a vector u(0) such that the value:
P(u0(0), . . . ,un−1(0)|y0n−1)=P(u0(0), . . . ,uϕ−1(0)|y0n−1)Πj=ϕn−1P(uj(0)|u0(0), . . . ,uj−1(0),y0n−1) (5)
is maximum. The original stack algorithm operates locally on the code tree, choosing at each step a path which maximizes the first term in expression (3), as the remaining terms are unknown. Assuming that u0(0), . . . , uj−1(0) are correct values of information symbols, find that each of the remaining terms is the probability of correct finding of uj(0) by the SC decoder [1] for given vector y0n−1. Values average for y0n−1 of these probabilities is easily computed using the density evolution method or its approximations at the second and third steps. Then, instead of maximizing P(u0(0), . . . , uϕ−1(0)|y0n−1) at each step of the stack algorithm
E[P(u0(0), . . . ,un−1(0)|y0n−1)]=P(u0(0), . . . ,uϕ−1(0)|y0n−1)Πj=ϕn−1E[P(uj(0)|u0(0), . . . ,uj−1(0),y0n−1)] (6)
is maximized.
By introducing heuristic function ψ(i)=Πj=in−1(1−Pj), where Pj=1−E[P(uj(0), . . . , uj−1(0), yn−1)]0 is the error probability in j-th bit subchannel specified by polarization transformation, an improved stack decoding algorithm is obtained as illustrated in
Pl,m[0, α]ψ(ϕl+1) is used instead of Pl,m[0, α] as a path metric in blocks 027 and 028.
Computation of dynamically frozen symbols uϕl(0) is performed in block 026 in accordance with (2) and the fast algorithm for ϕl∈F described above, which is necessary in decoding twisted polar codes with dynamically frozen symbols. However, the present method is also used for polar codes with only statically frozen symbols.
Modified algorithms RecursivelyCalcP and RecursivelyCalcC (7, 8) should be used to account for the structure of twisted polarization transformation in blocks 025, 029, 030. Initial versions thereof as illustrated in
It should be understood that those skilled in the art implement the present method in many ways. In particular, instead of path probabilities their logarithms are used. Probabilities are computed approximately. Computations of some of them, corresponding to frozen symbols, are omitted. Furthermore, more accurate heuristic functions ψ(i) are used. For example, such functions take into account the initial part of the path in the code tree to estimate the probability of remaining symbols. In general, increasing the accuracy of estimating the probability of remaining symbols, defined by function ψ(i), enables the number of iterations in the decoder to be reduced.
The methods of encoding and decoding described above is implemented in hardware (such as in ASIC) or software, which allows constructing encoding and decoding apparatus. The present encoding apparatus includes a precoder that performs a computation u(0)=xW. The encoding apparatus also includes an “exclusive or” unit and pass-through unit as illustrated in
As illustrated in
Note that modules 035 and 034 are not obligatory. Modules 032, 033, 037 are used to implement a list decoder of twisted polar codes.
The present methods of encoding and decoding are used in wireless and wireline communication systems, magnetic and optical data storage systems.
Furthermore, as shown in
where j(A,B)=max(A,B)+ln(l+e−|A−B|).
Embodiments of the inventions are not limited to the embodiments described herein; those skilled in the art will appreciate based on the information described herein and the knowledge in the art that the invention is practiced by other than the described embodiments without departing from the spirit and scope of this disclosure.
The description does not specify hardware and software for implementing the invention, but those skilled in the art will appreciate that the essence of the invention is not limited to a particular hardware or software, and therefore the disclosure uses any hardware and software known in the art. Thus, hardware is implemented in one or more application specific integrated circuits, digital signal processors, digital signal processing devices, programmable logic devices, field programmable gate arrays, processors, controllers, microcontrollers, microprocessors, electronic devices, other electronic units capable to perform the described functions, computer, or a combination thereof.
Although not specifically mentioned, it is evident that storage of data, programs, etc., implies that there is a computer readable storage medium; examples of computer readable media include read-only memory, random access memory, register, cache memory, semiconductor memory devices, magnetic media such as internal hard disks and removable disks, magneto-optical media and optical media such as CD-ROM disks and digital versatile disks (DVD), as well as any conventional other storage media.
Although the present disclosure has been described with an exemplary embodiment, various changes and modifications is suggested to one skilled in the art. It is intended that the present disclosure encompass such changes and modifications as fall within the scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2014114215 | Apr 2014 | RU | national |
Number | Name | Date | Kind |
---|---|---|---|
6341362 | Golnabi | Jan 2002 | B1 |
7346832 | Richardson et al. | Mar 2008 | B2 |
8311144 | Sun | Nov 2012 | B1 |
8347186 | Arikan | Jan 2013 | B1 |
20050243895 | Ashikhmin | Nov 2005 | A1 |
20060228116 | Bontu | Oct 2006 | A1 |
20080063115 | Varadarajan | Mar 2008 | A1 |
20100039928 | Noh | Feb 2010 | A1 |
20130117344 | Gross et al. | May 2013 | A1 |
20130132803 | Xiao | May 2013 | A1 |
20140108748 | Lee | Apr 2014 | A1 |
20140341320 | Hua | Nov 2014 | A1 |
20140365842 | Li et al. | Dec 2014 | A1 |
20150103947 | Shen | Apr 2015 | A1 |
20150194987 | Li | Jul 2015 | A1 |
20150236715 | Alhussien | Aug 2015 | A1 |
20150263767 | Shin | Sep 2015 | A1 |
20150293716 | Jiang | Oct 2015 | A1 |
Number | Date | Country |
---|---|---|
20110060635 | Jun 2011 | KR |
20120054571 | May 2012 | KR |
20130001494 | Jan 2013 | KR |
WO 2013107140 | Jul 2013 | WO |
Entry |
---|
Erdal Arikan, Senior Member, IEEE, Channel Polarization: A Method for Constructing Capacity-Achieving Codes for Symmetric Binary-Input Memoryless Channels, Transactions on Information Theory, vol. 55, No. 7, Jul. 2009, 24 pgs. |
Arikan, E.; “Channel Polarization: A Method for Constructing Capacity-Achieving Codes for Symmetric Binary-Input Memoryless Channels”; IEEE; vol. 55, Issue 7; Jul. 2009; 24 pgs. |
Arikan, E.; “Systematic Polar Coding”; IEEE; vol. 15, Issue 8; Jun. 30, 2011; 4 pgs. |
Li, et al.; “An Adaptive Successive Cancellation List Decoder for Polar Codes with Cyclic Redundancy Check”; IEEE; vol. 16, Issue 12; Nov. 19, 2012; 5 pgs. |
Mahdavifar, et al.; “On the Construction and Decoding of Concatenated Polar Codes”; ISIT; Jul. 7-12, 2013; 6 pgs. |
Niu, et al.; “CRC-Aided Decoding of Polar Codes”; IEEE; vol. 16, Issue 10; Sep. 7, 2012; 5 pgs. |
Tal, et al.; “List Decoding of Polar Codes”; ISIT; IEEE; vol. PP, Issue 99; Mar. 5, 2015; 12 pgs. |
Tal, et al.; “How to Construct Polar Codes”; IEEE; vol. 59, Issue 10; Jul. 10, 2013; 22 pgs. |
Trifonov, P.; “Efficient Design and Decoding of Polar Codes”; IEEE: vol. 60, Issue 11; Aug. 22, 2012; 8 pgs. |
Number | Date | Country | |
---|---|---|---|
20150295593 A1 | Oct 2015 | US |