Decoding method using dynamic scaler factor

Information

  • Patent Grant
  • 9647798
  • Patent Number
    9,647,798
  • Date Filed
    Tuesday, June 28, 2016
    8 years ago
  • Date Issued
    Tuesday, May 9, 2017
    7 years ago
Abstract
A decoding method applied to a convolutionally coded signal is provided. The method includes: adjusting first input information according to a first scaling factor to generate first a-priori information; b) decoding the convolutionally coded signal according to systematic information and the first a-priori information to generate first extrinsic information; c) adjusting second input information according to a second scaling factor to generate second a-priori information, wherein the second scaling factor is generated according to the first extrinsic information and the first a-priori information; and d) decoding the convolutionally coded signal according to the systematic information and the second a-priori information to generate second extrinsic information. One of step (b) and step (d) further generates a-posteriori information as a decoding result.
Description

This application claims the benefit of Taiwan application Serial No. 104123335, filed Jul. 17, 2015, the subject matter of which is incorporated herein by reference.


BACKGROUND OF THE INVENTION

Field of the Invention


The invention relates in general to a communication system, and more particularly to a soft output decoder used in a receiver of convolutional communication.


Description of the Related Art


In the process of data transmission of a digital communication system, incorrect messages may be received at a receiver end frequently due to unpredictable interference. Without increasing the transmission power, channel coding, although effectively reduces the error rate, poses a setback of occupying the transmission bandwidth. In view of the increasing demand of data transmission and storage systems of the public, not only the transmission rate will get faster but also the quality of service (QoS) will get higher in the future. As channel coding ensures that an error of the transmission bit is controlled within a certain range, channel coding is a critical consideration in the system design.


Convolutional coding is often used in channel coding to prevent a receiver from receiving incorrect messages. At a transmitting end, a code vector or an information block transmitted may be described by a trellis diagram. The complexity of a trellis diagram is determined by a constraint length of an encoder. Although the operation complexity gets higher as the length of the constraint length gets longer, such coding relatively provides better robustness.


At a receiving end, a soft-decision decoder may be adopted to identify a maximum likelihood code vector through a Viterbi algorithm and trellis architecture to perform decoding. However, the operation complexity of the Viterbi algorithm exponentially increases as the constraint length gets longer. In other words, a Viterbi decoder may require a substantial amount of memory and consume significant power to decode the convolutional coding having a longer constraint length.


Turbo coding is proven to render better performance than common coding technologies. A turbo code is formed from processing two or more convolutional codes by a turbo interleaver. To decode turbo codes, convolutional codes are individually decoded by a soft-decision decoder using an iteration approach. A soft-decision decoder decodes a convolutional code to provide extrinsic information, which allows the soft-decision decoder to provide a more accurate result when the soft-decision decoder decodes another convolutional code. In the prior art, soft-decision decoding may adopt a maximum a-posterioriprobability (MAP) algorithm or a soft output Viterbi algorithm (SOVA), both of which requiring forward recursion and backward recursion for decoding to determine the soft output of one information block. In general, in an environment with a lower signal-to-noise ratio (SNR), turbo codes render better performance than other convolutional codes.


The computation amount of the MAP algorithm not only is huge and complicated, but also involves index calculation. Without special processing, decoding hardware requirement or time delay generated during the decoding process is often significantly increased.


SUMMARY OF THE INVENTION

According to an embodiment of the present invention, a decoding method applied to a convolutionally coded signal is provided. The method includes: adjusting first input information according to a first scaling factor to generate first a-priori information; b) decoding the convolutionally coded signal according to systematic information and the first a-priori information to generate first extrinsic information; c) adjusting second input information according to a second scaling factor to generate second a-priori information, wherein the second scaling factor is generated according to the first extrinsic information and the first a-priori information; and d) decoding the convolutionally coded signal according to the systematic information and the second a-priori information to generate second extrinsic information. One of step (b) and step (d) further generates a-posteriori information as a decoding result.


According to another embodiment of the present invention, a decoding apparatus applied to a convolutionally coded signal is provided. The decoding apparatus includes a first soft-in-soft-out (SISO) decoder, a second data converter, a second SISO decoder, a first data converter, a first scaling factor generator and a second scaling factor generator. The first SISO decoder decodes the convolutionally coded signal according to systematic information and first a-priori information to generate first extrinsic information. The second data converter converts the first extrinsic information to second a-priori information according to a second scaling factor. The second SISO decoder decodes the convolutionally coded signal according to the systematic information and the second a-priori information to generate second extrinsic information. The first scaling factor generator generates one of the first scaling factor and the second scaling factor according to the first a-priori information and the first extrinsic information. The second scaling factor generator generates the other of the first scaling factor and the second scaling factor according to the second a-priori information and the second extrinsic information. One of the first and second SISO decoders further generates a-posteriori information as a decoding result.


The above and other aspects of the invention will become better understood with regard to the following detailed description of the preferred but non-limiting embodiments. The following description is made with reference to the accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 shows a trellis;



FIG. 2 shows a turbo encoder and a turbo decoder;



FIG. 3 and FIG. 4 are two turbo decoders according to embodiments of the present invention; and



FIG. 5 and FIG. 6 are another two turbo decoders according to embodiments of the present invention.





DETAILED DESCRIPTION OF THE INVENTION

The present invention is capable of improving the converging speed of a soft-decision decoder in an iteration process. In a decoder according to embodiments of the present invention, a scaling factor is introduced for adjusting a-priori information generated from interleaving or de-interleavinga extrinsic information. The scaling factor is not a constantly value, but dynamically changes along with a result generated from previous iteration decoding and a previous value of the scaling factor.


In general, convolutional codes and turbo codes may be represented by a trellis, as shown in FIG. 1. The trellis in FIG. 1 includes 13 steps, each having four possible states to represent that the constraint length of an encoder is 2. In other words, block codes having a block code length of 12 are obtained from the trellis in FIG. 1. For illustration purposes, in the description below, k represents a block code length, which represents the number of steps that one block code includes. As known in the prior art, a MAP decoder adopts forward recursion and backward recursion on a trellis to generate a soft output. Based on received information, a MAP decoder minimizes the bit error probability of the decoding process.


In FIG. 2, the left half shows a turbo encoder, and the right half shows a turbo decoder. A turbo encoder is generally formed by two parallel concatenated recursive systematic convolutional encoders RSC 12 and RSC 14, which are connected by an interleaver INT 16 in between. According to a block code X, the recursive systematic convolutional encoder RSC 12 generates a string of parity bits xkp1 that are in overall referred to as a parity code Xp1, where k=0 to K−1. Similarly, the recursive systematic convolutional encoder RSC 14 generates a parity code Xp2(a string of parity bits xkp2) according to an interleaved block code X. The block code X is also referred as a systematic block code Xs. Bits in the systematic block code Xs, the parity code Xp1 and the parity code Xp2 may be interleavingly connected and be outputted to a communication channel through a multiplexer. To increase the code rate, a part of parity bits may be omitted and not outputted. For example, only a part of the parity bits xkp1 and xkp2 of the same step are outputted to a communication channel, such that the turbo encoder in FIG. 2 may have a higher code rate.


The turbo decoder in FIG. 2 calculates the reliability of the received information, and represents the reliability in form of log-likelihood ratios (LLRs). Each LLR represents the probability of one corresponding bit being 0 or 1. Compared to the systematic block code Xs, the parity code Xp1 and the parity code Xp2, the turbo decoder generates systematic information Ys, parity information Yp1 and parity information Yp2. For example, the systematic information Ys is formed by a string of LLRs yks, and the parity information Yp1 is formed by a string of LLRs ykp1, where k=0 to K−1. The turbo decoder in FIG. 2 includes interleavers INT 18 and 24, soft-input-soft-output (SISO) decoders SISO 20 and 22, and a de-interleaver 26. Operations and iteration details of these elements substantially follow a Bahl, Cocke, Jelinek and Raviv (BCJR) algorithm, also referred to as a MAP algorithm. According to the systematic information Ys and the parity information Yp1 as well as a-priori information La1, the SISO 20 calculates a soft output (usually representing the maximum a-posteriori probability by an LLR), which is referred to as extrinsic information Le1. After an interleaving process, the extrinsic information Le1 becomes a-priori information La2. The SISO 22 calculates extrinsic information Le2 according to the interleaved systematic information Ys, the parity information Yp2 and the a-priori information L. The extrinsic information Le2 is processed by ande-interleaving process and becomes the a-priori information La1 that is then fed back to the SISO 20. The process having been performed once by the SISO 20 or 22 is referred to as half-iteration, and the computation process having been performed once by the SISO 20 and 22 is referred to as one iteration. In general, such iteration loop is repeated for a fixed number of times, or is repeated until the number of changing symbols in the extrinsic information Le1 or Le2 in the iteration loop is as small as a predetermined level. In FIG. 2, after the iteration loop repetition ends, the SISO 22 generates a-posteriori information, which is processed by de-interleaving and hard decision to serve as a decoding result {circumflex over (X)}S. It should be noted that, FIG. 2 is not to be construed as a limitation to the present invention. For example, in another embodiment, after the iteration loop repetition ends, the SISO 20 may also generate another set of a-posteriori information, which may also serve as the decoding result {circumflex over (X)}S after having undergone hard decision.


Under the condition that the foregoing MAP algorithm calculates a received message Y, the probability of the message bit being digital 1 or 0 at a step k, or referred to as a-posteriori log-likelihood ratio L(uk|Y), is defined as below.







L


(


u
k



/


Y

)


=

ln



P


(


u
k

=


+
1



/


Y


)



P


(


u
k

=


-
1



/


Y


)








The MAP algorithm calculates L(uk|Y) of each step k through forward and backward recursive operations on the trellis. L(uk|Y) is organized and represented as:










L


(


u
k



/


Y

)


=

ln








(

n
,
m

)



u
k


=
1







α

k
-
1




(
n
)


*





γ
k



(

n
,
m

)


*




β
k



(
m
)










(

n
,
,
m

)



u
k


=

-
1








α

k
-
1




(
n
)


*





γ
k



(

n
,
m

)


*




β
k



(
m
)










(
1
)







In equation (1), the branch metric rk(n,m) represents the probability of becoming a state m at the step k under conditions that the state is n at the step k−1 and the received message is Y, the forward state metric αk−1(n) represents the probability that the state is n at the step k−1 under a condition that the received message is Y, the backward state metric βk(m) represents the probability that the state ism at the step k under a condition that the received message is Y. The alphabet zigma (Σ) of the numerator refers to a calculated total of all branches that possibly generate uk=1. Similarly, the alphabet zigma (Σ) of the denominator refers to a calculated total of all branches that possibly generate uk=−1. The forward state metric αk(m), the backward state metric βk(m) and the branch metric rk(n,m) may be respectively represented as:











α
k



(
m
)


=




n
=
1

M










α

k
-
1




(
n
)


*




γ
k



(

n
,
m

)








(
2
)








β
k



(
m
)


=




n
=
1

M










β

k
+
1




(
m
)


*




γ

k
+
1




(

n
,
m

)








(
3
)








γ
k



(

n
,
m

)


=


A
k



P


(

m
|
n

)



exp




2


/





N
0

*



(



y
k
s




x
k
s



(
i
)



+


y
k
p




x
k
p



(

i
,
n
,
m

)




)










(
4
)







In the above, Ak is a constant number, and i=+1 or −1. It is known from equation (2) that, to calculate the forward state metric αk(m), the forward state metric α prior to the step k needs to be first learned; it is known from equation (3) that, to calculate the backward state metric βk(m), the backward state metric β subsequent to the step k needs to be first learned. Thus, the forward state metric αk(m) and the backward state metric βk(m) are generally calculated and obtained through iteration, with however the directions of the iteration being opposite. It is known from equation (4) that, to calculate the branch metric rk(n,m) requires index and multiplication calculations, and consumes immense resources and costs regardless of whether it is implemented by hardware or software.


A Log-MAP algorithm simplifies the calculation process by using a logarithm quadrant operation. Further, the index and logarithm calculations may be simplified through the approximation method below.

max*(x,y)≡In(ex+ey)=max(x,y)+In(1+e−|y−x|)  (5)


The last logarithm calculation may be obtained through a look-up table. A Max-Log-MAP algorithm further removes the last item in equation (5) to approximate the operator max*, as below:

max*(x,y)=In(ex+ey)≈max(x,y)  (6)


The decoding calculation in the SISOs 20 and 22 adopts the maximum a-posteriori probability (MAP) algorithm for decoding. The MAP algorithm may be the Log-MAP or Max-Log-MAP algorithm.


When using the Log-MAP or Max-Log-MAP algorithm for processing, equation (4) may be organized as:

Γk(n,m)=In(γk(n,m))=2/N0*(yksxks(i)+ykpxkp(i,n,m))+In(P(m|n))+K  (7)


In equation (7), K is a constant value, and P(min) is input information transmitted from a first half iteration, e.g., a-priori information. Components of the input information exchanged between two half iterations may be slightly adjusted, e.g., multiplying by a scaling factor sd to increase the converging speed of the iteration loop of the turbo decoder. Wherein, d is 1 or 2, which respectively correspond to the SISO 20 and 22, i.e., the first and second half iterations.


The scaling factor sd may be set as a fixed value, e.g., 0.7. However, it usually may be challenging to select a most appropriate fixed value.



FIG. 3 shows a turbo decoder 50 according to an embodiment of the present invention. Scaling factors s1 and s2 adopted are not fixed values, but are dynamically updated as the iteration loop progresses. Distinct from the turbo decoder in FIG. 2, the turbo decoder 50 additionally includes scaling factor generators 38 and 34, and multipliers 36 and 32. The scaling factor generator 38 provides the scaling factor s1 to the multiplier 36, adjusts a-priori information La1 to generate s1xLa1 as new a-priori information of the SISO 20. The SISO 20 generates extrinsic information Le1 after completing half iteration. At this point, according to the extrinsic information Le1, the a-priori information s1xLa1 and a predetermined equation, the scaling factor generator 38 identifies the new scaling factor s1 to be adopted in subsequent iteration loop. The scaling factor generator 34 and the multiplier 32 are used in the other half iteration, and associated operations and principles may be learned from the operations and principles of the scaling factor generator 38 and the multiplier 36 and shall be omitted herein.


In one embodiment of the present invention, for example, the scaling factor generators 38 and 34 generate s1 and s2 according to equations (8a) and (8b), respectively:

s1(I+1)=s1(I)+STEP*DIF1(I)/LEN  (8a)
s2(I+1)=s2(I)+STEP*DIF2(I)/LEN  (8b)


In the equations above, STEP represents a constant number and may be pre-set according to requirements, and s1(I) represents the scaling factor adopted in the 1st half iteration in the Ith iteration loop. The variance DIF1(I) is a sum of different numbers of bits between the extrinsic information Le1 generated by the first half iteration of the Ith iteration loop and the a-priori information La1 generated by the second half iteration of the (I−1)th iteration loop. That is, The variance DIF1(I) may be learned from the a-priori information La1 and the extrinsic information Le1 of the same half iteration. For example, after the first half iteration of the 2nd iteration loop, there is a difference of 3 bits between the extrinsic information Le1 and t the a-priori information La1, and so the variance DIF1(2) is 3. LEN is a constant number associated with the code block length K. In one embodiment, LEN is directly equal to the block code length K, e.g., 2080. In another embodiment, the relationship between LEN and the block code length K is LEN=K*bit/step, where bit is information bit, and bit/step is the possible number outputted in each step. In one embodiment K=2080, bit/step=2, and LEN=4160.


s2(I) and the variance DIF2(I) respectively represent the scaling factor adopted in the second half iteration of the Ith iteration loop, and the variance generated after the decoding of the second half iteration of the Ith iteration loop.


In some embodiments, initial values s1(0) and s2(0) of the scaling factors are both set to 0.25. In some embodiments, the scaling factors s1 and s2 are limited to a maximum value 1. In other words, if the scaling factors s1 and s2 obtained from equations (8a) and (8b) exceed 1, the scaling factors s1 and s2 are set to 1.


It is discovered from equations (8a) and (8b) that, the scaling factor adopted in the specific half iteration is determined by the scaling factor and the decoding result of the same specific half iteration of a previous iteration loop. Further, if the variance DIFd(I) generated by the previous iteration loop is not 0, the scaling factor adopted in the current iteration loop is then larger than the scaling factor adopted in the previous iteration loop. It is further discovered from equations (8a) and (8b) that, the scaling factors s1 and s2 used in each iteration loop are not smaller than the scaling factors s1 and s2 used in the previous iteration loop.


It is proven through computerized simulations that, the scaling factors adopting equations (8a) and (8b) are capable of causing the iteration loop to converge quickly, i.e., quickly reducing the variance DIFd(I). Thus, the decoding efficiency of the turbo decoder can be enhanced.



FIG. 4 shows a turbo decoder 52 according to another embodiment of the present invention. In this embodiment, scaling factor generators 40 and 42 generate scaling factors s1 and s2 according to calculations of equations (9a) and (9b), respectively:

s2(I)=s1(I)+STEP*DIF1(I)/LEN  (9a)
s1(I+1)=s2(I)+STEP*DIF2(I)/LEN  (9b)


The scaling factors in FIG. 3 and FIG. 4 are for adjusting the a-priori information as a non-limiting example of the present invention. For example, referring to FIG. 5 and FIG. 6 showing two turbo decoders 54 and 56 according to embodiments of the present invention, the scaling factors are for adjusting the extrinsic information. It should be noted that, adjusting the exchange input information between two SISOs using scaling factors is encompassed within the scope of the present invention.



FIG. 5 is similar to FIG. 3, and similar or identical details are omitted herein. In FIG. 5, the scaling factor generators 38 and 34 generate s1 and s2 according to calculations of equations (8a) and (8b), respectively. In FIG. 5, the multiplier 36 multiplies the extrinsic information Le2 generated by the SISO 22 by the scaling factor s1 to generate s1xLe2 to the de-interleaver 26. Similarly, the multiplier 32 multiplies the extrinsic information Le1 generated by the SISO 20 by the scaling factor s2 to generate s2xLe1 to the interleaver 24.



FIG. 6 is similar to FIG. 4, and similar or identical details are omitted herein. In FIG. 6, the scaling factor generators 40 and 42 generate s1 and s2 according to calculations of equations (9a) and (9b), respectively. In FIG. 6, the multiplier 36 multiplies the extrinsic information Le2 generated by the SISO 22 by the scaling factor s1 to generate s1xLe2 to the de-interleaver 26. Similarly, the multiplier 32 multiplies the extrinsic information Le1 generated by the SISO 20 by the scaling factor s2 to generate s2xLe1 to the interleaver 24.


In FIG. 3 and FIG. 5, the scaling factor generators 38 and 34 generate the scaling factors s1 and s2 according to the calculations of equations (8a) and (8b), respectively. For example but not limited to, each time the scaling factor is adjusted, the scaling factor used by the same half iteration in the previous iteration loop is referred. In another embodiment, the scaling factor generators 38 and 34 generate the scaling factors s1 and s2 according to the calculations of equations (10a) and (10b), respectively. In equations (10a) and (10b), each time the scaling factor is adjusted, the scaling factor used by the previous half iteration is referred.

s1(I+1)=s2(I)+STEP*DIF1(I)/LEN  (10a)
s2(I)=s1(I)+STEP*DIF2(I)/LEN  (10b)


In FIG. 4 and FIG. 6, the scaling factors 40 and 42 generate the scaling factors s1 and s2 according to the calculations of equations (9a) and (9b), respectively. For example but not limited to, each time the scaling factor is adjusted, the scaling factor used by the previous half iterationis referred. In another embodiment, the scaling factor generators 40 and 42 in FIG. 4 and FIG. 6 generate the scaling factors s1 and s2 according to the calculations of equations (11a) and (11b), respectively. In equations (11a) and (11b), each time the scaling factor is adjusted, the scaling factor used by the same half iteration in the previous iteration loop is referred.

s2(I+1)=s2(I)+STEP*DIF1(I)/LEN  (11a)
s1(I+1)=s1(I)+STEP*DIF2(I)/LEN  (11b)


In simple, each time the scaling factor is adjusted, the scaling factor used by the same half iteration in the previous iteration loop or the scaling factor used by the previous half iteration may be referred. Each time the scaling factor is adjusted, the variance generated by the same half iteration in the previous iteration loop or the variance generated by the previous half iteration may be referred. Thus, there are at least four combinations, which are encompassed within the scope of the present invention.


After the iteration loop repetition is complete, a-posteriori information is generated by the SISO 22 in FIG. 3 to FIG. 6. The a-posteriori information is processed by de-interleaving and hard decision to serve as a decoding result {circumflex over (X)}S. In other embodiments, another set of a-posteriori information may also be generated by the SISO 20, and is processed by hard decision to serve as the decoding result {circumflex over (X)}S.


While the invention has been described by way of example and in terms of the preferred embodiments, it is to be understood that the invention is not limited thereto. On the contrary, it is intended to cover various modifications and similar arrangements and procedures, and the scope of the appended claims therefore should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements and procedures.

Claims
  • 1. A decoding method applied to a convolutionally coded signal, comprising: a) adjusting first input information according to a first scaling factor to generate first a-priori information;b) decoding the convolutionally coded signal according to systematic information and the first a-priori information to generate first extrinsic information;c) adjusting second input information according to a second scaling factor to generate second a-priori information, wherein the second scaling factor is calculated based on the first extrinsic information and the first a-priori information; andd) decoding the convolutionally coded signal according to the systematic information and the second a-priori information to generate second extrinsic information;wherein, one of step (b) and step (d) further generates a-posteriori information as a decoding result.
  • 2. The method according to claim 1, decoding the convolutionally coded signal by an iteration loop method, each iteration loop comprising: a1) performing a first half iteration; andb1) performing a second half iteration;wherein, the step of generating the first a-priori information and the step of generating the first extrinsic information are the first half iteration; andthe step of generating the second a-priori information and the step of generating the second extrinsic information are the second half iteration.
  • 3. The decoding method according to claim 2, wherein the second scaling factor is calculated based on the first scaling factor.
  • 4. The decoding method according to claim 2, wherein the second scaling factor is calculated based on a previous second scaling factor.
  • 5. The decoding method according to claim 1, decoding the convolutionally coded signal by an iteration loop method, each iteration loop comprising: a1) performing a first half iteration; andb1) performing a second half iteration;wherein, the step of generating the first a-priori information and the step of generating the first extrinsic information are the first half iteration of an Nth iteration loop; andthe step of generating the second a-priori information and the step of generating the second extrinsic information are the first half iteration of an (N+1)th iteration loop.
  • 6. The decoding method according to claim 5, wherein the second scaling factor is generated according to the first scaling factor.
  • 7. The decoding method according to claim 5, wherein the second scaling factor is generated according to a previous scaling factor of the second half iteration of the Nth iteration loop.
  • 8. The decoding method according to claim 1, wherein the step of adjusting the first input information according to the first scaling factor to generate the first a-priori information comprises: interleaving the adjusted first input information to generate the first a-priori information, wherein the first input information is extrinsic information.
  • 9. The decoding method according to claim 1, wherein the first input information is extrinsic information having undergone interleaving.
  • 10. The decoding method according to claim 1, wherein the second scaling factor is greater than or equal to the first scaling factor.
  • 11. The decoding method according to claim 1, wherein the second scaling factor is limited to be smaller than or equal to a predetermined value.
  • 12. The decoding method according to claim 1, wherein the first a-priori information and the first extrinsic information determine a variance, and the second scaling factor is generated according to the variance and the first scaling factor.
  • 13. The decoding method according to claim 1, wherein the convolutionally coded signal comprises a block code length, and the second scaling factor is generated further according to the block code length.
  • 14. A decoding apparatus applied to a convolutionally coded signal, comprising: a first soft-in-soft-out (SISO) decoder, decoding the convolutionally coded signal according to systematic information and first a-priori information to generate first extrinsic information;a second data converter, converting the first extrinsic information to second a-priori information according to a second scaling factor;a second SISO decoder, decoding the convolutionally coded signal according to the systematic information and the second a-priori information to generate second extrinsic information;a first data converter, converting the second extrinsic information to the first a-priori information according a first scaling factor;a first scaling factor generator, calculating one of the first scaling factor and the second scaling factor according to the first a-priori information and the first extrinsic information; anda second scaling factor generator, calculating the other of the first scaling factor and the second scaling factor according to the second a-priori information and the second extrinsic information;wherein, one of the first and second SISOs further generates a-posteriori information as a decoding result.
  • 15. The decoding apparatus according to claim 14, wherein the first data converter comprises a first multiplier and a de-interleaver connected in series, and the first multiplier is controlled by the first scaling factor; and the second data converter comprises a second multiplier and an interleaver connected in series, and the second multiplier is controlled by the second scaling factor.
  • 16. The decoding apparatus according to claim 15, wherein the first multiplier receives an output of the de-interleaver.
  • 17. The decoding apparatus according to claim 15, wherein the de-interleaver receives an output of the first multiplier.
  • 18. The decoding apparatus according to claim 14, wherein the first a-priori information and the first extrinsic information determine a variance, and the first scaling factor generator provides one of the first scaling factor and the second scaling factor according to the variance.
Priority Claims (1)
Number Date Country Kind
104123335 A Jul 2015 TW national
US Referenced Citations (8)
Number Name Date Kind
6732327 Heinila May 2004 B1
8799737 Varnica Aug 2014 B1
20070067703 Berkmann Mar 2007 A1
20070220394 Kim Sep 2007 A1
20070286292 Moelker Dec 2007 A1
20100208849 Grosskinsky Aug 2010 A1
20120069936 Shaver Mar 2012 A1
20160352361 Fonseka Dec 2016 A1
Non-Patent Literature Citations (2)
Entry
Taiwan Intellectual Property Office, “Office Action”, issued on Jun. 27, 2016.
Yi-Nan Lin; et al., “An Efficient Soft-Input Scaling Scheme for Turbo Decoding”, 2006, IEEE.
Related Publications (1)
Number Date Country
20170019213 A1 Jan 2017 US