The present invention relates to a unified Viterbi/Turbo decoder in wireless communication systems, and in particular, to a system for receiving and decoding voice and data streams adaptively.
Error control coding techniques involves the use of a channel encoder in the transmitter and a decoder in the receiver. The channel encoder accepts message bits and adds redundancy according to a prescribed rule, thereby producing encoded data at a higher bit rate. The channel decoder exploits the redundancy to decide which message bits were actually transmitted. The purpose of error control coding is to minimize the effect of channel noise. The error control codes are generally classified into block codes and convolutional codes. Convolutional codes are preferred for wireless voice communication systems in which the retransmission of data and its associated delay is intolerable. Block codes are capable of delivering higher throughput and are preferred for the transmission of data where latency is less of a concern.
The Viterbi algorithm is a sequential trellis search algorithm for performing maximum-likelihood (ML) sequence detection, and it is an optimum decoding algorithm for convolutional codes. A published book written by S. B. Wicker, named “Error Control System for Digital Communication and Storage, Pretice Hall, 1995.” disclosed a Viterbi decoder for use in mobile communication systems.
Turbo codes, also known as parallel concatenated codes, are a class of codes whose performance is very close to the Shannon capacity limit. Turbo encoders are implemented by connecting convolutional encoders either in parallel or series to produce concatenated outputs. Bit sequences passing from one encoder to another are permuted by a pseudo-random interleaver, and thus low-weight code words produced by a single encoder are transformed into high-weight code words. Third generation (3G) mobile wireless standards, such as Code Division Multiple Access (CDMA) 2000 and Universal Mobile Telecommunication Service (UMTS) require Turbo encoding for data streams. UK patent application number GB2352943A and “Turbo codes, principles and application, Kluwer Academic Publishers, 2000.” Written by B. Vucetic and J. Yuan disclosed embodiments of the Turbo decoder with the Maximum a Posteriori probability (MAP) algorithm and Soft Output Viterbi Algorithm (SOVA).
Third Generation (3G) communication systems typically require both convolutional coding, for example, Viterbi coding, and Turbo coding for voice and data signals respectively. It is because the transmission of voice and data provides conflicting requirements for transmission rate versus latency and propagation delay. The current mode of addressing these problems is to provide separate encoding systems: Turbo coding for high data-rate data streams and convolutional coding for voice or low data-rate data streams. The receivers of such systems thus require two independent decoders, resulting in a multiplicity of hardware platforms and increasing cost.
A unified Viterbi/Turbo decoder is provided which is capable of performing either convolutional decoding or Turbo decoding according to the input coding type. In an embodiment, the decoder comprises a Branch Metric Calculation (BMC) unit, a control circuit, a trace-back unit, an interleaver/de-interleaver, and a Turbo buffer. The BMC unit receives either a Viterbi or Turbo symbols and determines a branch metric associated with each node of the code word. The decoder is configurable to adapt the input, and the control circuit is a common component for the two coding types which fully utilize the memory and processor capacity. The control circuit computes partial path metrics at each node with its processors according to the corresponding branch metric and the partial path metrics at a preceding node stored in a first memory block. The control circuit then stores the computed partial path metrics in a second memory block. For processing the next node, the control circuit reads from the second memory block then stores in the first memory block after computation, and so on. The two memory blocks interchange path metric information with the processors. The order and address for reading and writing the partial path metrics from/to the memory blocks follow a fixed routing rule, which allows the partial path metrics to be processed in parallel to fully utilize the processing capacity. The control circuit then performs either selection or Maximum a Posteriori probability (MAP) calculation depending on the input coding type to identify a surviving path at each node.
The trace-back unit receives the output of the control circuit while performing Viterbi decoding, and it tracks the surviving paths designated by the control circuit. The interleaver/de-interleaver receives the output of the control circuit while performing Turbo decoding, and permutes or recovers the order of its input in a deterministic manner. The Turbo buffer receives the input symbols and the output of the interleaver/de-interleaver, and provides storage for the BMC unit and the interleaver/de-interleaver.
In an embodiment, the control circuit comprises an Add Compare Select Processor (ACSP) array having J ACSPs (J=2m, m belongs to an integer, and m+1 must be less than the constraint length of the encoder, n+1), a path metric calculation unit having the first and second memory blocks, a fixed routing circuit, and a selection and MAP calculation unit. The ACSP array computes the partial path metrics at each node according to the corresponding branch metric. Each ACSP comprises two Add Compare Select units (ACSU) processing the partial path metrics in parallel. The path metric calculation unit stores the partial path metrics associated with a current node and a preceding node in the two memory block. Each memory block comprises I memory banks, wherein I=2m+1. The fixed routing circuit establishes fixed connections between each ACSP and four memory banks, two from each memory block, according to the fixed routing rule. The selection and MAP calculation unit performs either selection or MAP calculation to identify the surviving path at each node. The selection and MAP calculation unit computes a Log-Likelihood Ratio (LLR) according to the branch metrics obtained from the BMC unit and the path metrics obtained from the path metric calculation mean for the Turbo code word.
An embodiment of the present invention can be more fully understood by reading the subsequent detailed description in conjunction with the examples and references made to the accompanying drawings, wherein:
a illustrates the circuitry of an exemplary Add Compare Select Unit (ACSU).
b illustrates the circuitry of another exemplary ACSU.
a shows a path metric memory bank organization according to the first embodiment of the invention.
b illustrates a detail diagram for
a shows a path metric memory bank organization for 256 states Viterbi decoding according to the third embodiment of the invention.
b illustrates the data structure corresponding to the path metric memory bank for 256 states Viterbi decoding according to the third embodiment of the invention.
As shown in
The survivor path updating unit 46 comprises a Survivor Path Memory (SPM) for storing surviving paths determined at each node in the trellis diagram for Viterbi decoding.
In the first embodiment, the constraint length of the convolutional encoder is 5 (n=4), thus there are 16 states (2n=16) in the Viterbi decoding. The number of ACSPs in the ACSP array 423 is chosen to be 2m, wherein m must be less than n, and m=2 is chosen in this embodiment. Each ACSP comprises an Add Compare Select Unit (ACSU) pair, thus is capable of processing two inputs simultaneously. Consequently, an efficient design is to provide eight memory banks (2m+1=8) to cooperate with the four ACSPs, as the four ACSPs can process eight partial path metrics simultaneously.
a and 6b illustrates two types of Add Compare Select Unit (ACSU) circuitry for the previous described ACSP. The ACSU in
a shows an exemplary path metric memory bank organization illustrating the read/write selection rule defined by the fixed routing circuit 422 of
As shown in
b shows the detail of
After processing the first memory element of each bank, ACSP (00) obtains path metric information from the memory element with state index 0001 and the memory element with state index 1001 in Block A at time point t2. Similarly, ACSP (00) generates two new partial path metrics and stores them in the memory elements with state index 0001 and state index 0011 in Block B at time point t2+p. The remaining ACSPs perform similar read/write procedures as ACSP (00) to process the second memory element of each bank simultaneously.
At the next time point, each ACSP of the ACSP array reads the path metric information from the corresponding banks in Block B for generating new path metric information, and subsequently, storing the new path metric information in Block A. This read/write operation repeats twice, one for the first memory elements and one for the second memory elements of the banks. The ACSPs uses the two blocks (Block A and Block B) iteratively like a double buffer so that the partial path metrics corresponding to the previous node will only get overwrite after obtaining all the partial path metrics of the current node. The path metric information originally stored in Block A will be calculated with a corresponding branch metric and stored in Block B, and subsequently, the path metric information of Block B will be calculated with a corresponding branch metric and stored in Block A, and so on.
The read/write selection rule, also called the routing rule defined by the fixed routing unit for accessing the memory banks is analyzed using the following steps. The number of states is relevant to the constraint length (n+1) of an input symbol for decoding, 2n, and a unique state index Sn−1 . . . S2S1S0 is assigned to each state. In the first embodiment, constraint length of the encoder is 5 (n=4), resulting sixteen (2n=16) states. The number of memory banks is 2m+1, wherein m+1 must be less than or equal to n. In the first embodiment, m is chosen to be 2, so there are eight memory banks. The bank index Sn−1Sn−2. . . Sn−m−1 and the ACSP index Sn−2Sn−3. . . Sn−m−1 are indexes indicating the corresponding memory banks and the ACSP.
In the first embodiment, where n=4 and m=2, the memory element with a state index S3S2S1S0 denotes that it is included in the memory bank with a bank index S3S2S1, for example, the memory element with state index 0010 is included in the bank 001. Each ACSP reads two partial path metrics stored in the memory elements with state index S3S2S1S0 according to its ACSP index S2S1. For example, ACSP 10 reads the partial path metrics stored in the memory elements with state index 0100 or 0101, and state index 1100 or 1101. After the ACSPs generate new partial path metrics, the new partial path metrics will be wrote in the memory elements with state index S3S2S1S0 having the two most significant bits (S3S2) identical to the ACSP index. For example, ACSP 10 writes the new partial path metrics in the memory elements with state index 1000 or 1001, and state index 1010 or 1011.
As shown in
Turbo decoding in the first embodiment of the invention utilizes the MAP decoding algorithm to calculate the LLR at each time point according to the branch metric γ, forward path metric α, and backward path metric β.
a and 10b illustrate the path metric memory bank organization and the data structure corresponding to the path metric memory bank for 256-state Viterbi decoding according to the third embodiment of the invention. In the third embodiment, constraint length of the encoder is 9, and again, four ACSPs and eight memory banks are used. Every path metric memory bank has two memory blocks (Block A and Block B), and the block index for Block A and Block B are 0 and 1 respectively. The four ACSPs processes eight states during each cycle, thus for decoding a Viterbi code word with 256 states per code word, the ACSPs spend 32 cycles (256/8 =32) for reading and writing partial path metrics of a node from/in the memory banks.
The same hardware (i.e. ACSPs and memory banks) for 256-state Viterbi decoding is appropriate for 8-state Turbo decoding in terms of processing speed and memory size. For Turbo decoding, constraint length of each encoder is 4(n=3), causing eight possible states.
The selection and MAP calculation unit 424 in
The Viterbi/Turbo unified decoder of the invention is configurable depending on the type of receiving data, and the control circuit is designed to be capable of storing and computing partial path metrics for both Viterbi and Turbo decoding. The routing rule for the ACSP array to access the memory banks is fixed for both decoding types. In an embodiment, the memory banks for storing the partial path metrics are efficiently used in both Viterbi and Turbo decoding. For example, if the decoder is designed to receive either a 256-state Viterbi code word or an 8-state Turbo code word, the window size for Turbo decoding is chosen to be 32 (256/8). The memory banks store eight states of backward path metrics corresponding to 32 time points (stages) when decoding an 8-state Turbo code word, whereas for decoding an 256-state Viterbi code word, the ACSP array processes 8 states during one cycle, thus requires 32 cycles. The memory size and the processing time for the two decoding are therefore approximately the same. In another embodiment, a plurality of Viterbi/Turbo unified decoder of the invention are employed for decoding a bunch of data streams comprising Turbo symbols or/and Viterbi symbols, each decoder is capable of decoding any type of input. The decoding time for each decoder is designed to be approximately the same regardless the decoding scheme. An advantage of such design is to reduce the number of decoder required since each decoder can be used to decode both Viterbi and Turbo code words. Another advantage is that all the decoders can finish the decoding process at approximately the same time, so that the decoders can work in parallel without waiting for the processing delay of certain decoders.
Finally, while the invention has been described by way of example and in terms of the above, it is to be understood that the invention is not limited to the disclosed embodiment. On the contrary, it is intended to cover various modifications and similar arrangements as would be apparent to those skilled in the art. Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.
This application claims priority from U.S. Provisional Patent Application No. 60/523,981 filed on Nov. 24, 2003.
Number | Name | Date | Kind |
---|---|---|---|
5272706 | Park | Dec 1993 | A |
6738941 | Todoroki | May 2004 | B1 |
6823489 | Wittig et al. | Nov 2004 | B2 |
20030018942 | Seo | Jan 2003 | A1 |
20030028845 | Plante et al. | Feb 2003 | A1 |
20030101403 | Jeon et al. | May 2003 | A1 |
Number | Date | Country |
---|---|---|
2002217748 | Aug 2002 | JP |
10-2002-0067769 | Aug 2002 | KR |
Number | Date | Country | |
---|---|---|---|
20050149838 A1 | Jul 2005 | US |
Number | Date | Country | |
---|---|---|---|
60523981 | Nov 2003 | US |