The invention relates to computer networks, more specifically to decoding data received from computer networks.
Currently, local area networks (LANs) are utilizing Gigabit Ethernet over copper medium, a protocol commonly referred to as 1000BASE-T. The next generation high-speed Ethernet is 10 Gigabit Ethernet over copper medium, a protocol commonly referred to as 10GBASE-T. The Institute of Electrical and Electronic Engineers (IEEE) 802.3 10GBASE-T study group is investigating the feasibility of transmission of 10 Gigabits per second over 4 unshielded twisted pairs.
10GBASE-T will probably use a pulse amplitude modulation (PAM) scheme, such as PAM10 combined with a four dimensional trellis code as the basis for its transmission scheme. The symbol rate of this scheme is 833 M baud with each symbol representing 3 bits of information. One of the powerful yet simple algorithms to decode the code as well as to combat inter-symbol interference is the parallel decision-feedback decoding algorithm. However, the implementation and design of a parallel decision-feedback decoder (PDFD) which operates at 833 MHz is challenging due to the long critical path in the decoder structure.
Existing literature describes high-speed PDFD designs suitable for 1000BASE-T applications. However, most of the proposed techniques may not be suitable for 10GBASE-T. For example, the decision feedback pre-filtering technique only works for channels where the postcursor ISI's energy is concentrated on the first one or two taps. Otherwise, it may result in significant performance loss. Furthermore, the complexity is exponential with channel memory length, so it is only suitable for channels with short memory length while the channel memory length of 10GBASE-T is substantially longer than that of 1000BASE-T.
In general, the invention relates to techniques for pipelining parallel decision feedback decoders (PDFDs) for high speed communication systems, such as 10 Gigabit Ethernet over copper medium (10GBASE-T). In one aspect, the decoder applies look-ahead methods to two concurrent computation paths. In another aspect of the invention, retiming and reformulation techniques are applied to a parallel computation scheme of the decoder to remove all or a portion of a decision feedback unit (DFU) from a critical path of the computations of the pipelined decoder. In addition, the decoder may apply a pre-cancellation technique to a parallel computation scheme to remove the entire DFU from the critical path.
Utilization of pipelined PDFDs may enable network providers to operate 10 Gigabit Ethernet with copper cable rather than fiber optic cable. Thus, network providers may operate existing copper cable networks at higher speeds without having to incur the expense of converting copper cables to more expensive fiber optic cables. Furthermore, the pipelined PDFD techniques may reduce hardware overhead and complexity of the decoder.
In one embodiment, a parallel decision feedback decoder (PDFD) comprises a plurality of computational units, wherein the computational units are pipelined to produce a decoded symbol for each computational iteration.
In another embodiment, a method comprises receiving a signal from a network, and processing the signal with a parallel decision feedback decoder (PDFD) having a plurality of pipelined computational units to produce a decoded symbol for each computational iteration of the PDFD.
The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.
In the example of
In one embodiment, transmitter 6, located within a first network device (not shown), may transmit data to receiver 14, which may be located within a second network device (not shown). The first network device may also include a receiver substantially similar to receiver 14. The second network device may also include a transmitter substantially similar to transmitter 6. In this way, the first and second network devices may achieve two way communication with each other or other network devices. Examples of network devices that may incorporate transmitter 6 or receiver 14 include desktop computers, laptop computers, network enabled personal digital assistants (PDAs), digital televisions, or network appliances generally.
Decoder 16 may be a high-speed decoder such as a pipelined parallel decision feedback decoder (PDFD). Utilization of pipelined PDFDs may enable network providers to operate 10 Gigabit Ethernet with copper cable. For example, network providers may operate existing copper cable networks at higher speeds without having to incur the expense of converting copper cables to more expensive media, such as fiber optic cables. Furthermore, in certain embodiments of the invention, the pipelined PDFD design may reduce hardware overhead of the decoder. Although the invention will be described with respect to PDFD decoders, it shall be understood that the present invention is not limited in this respect, and that the techniques described herein may apply to other types of decoders.
Conventional PDFD algorithms perform computations in a serial manner. At time n, the conventional PDFD first computes inter-symbol interference (ISI) estimates. Next, these ISI estimates and the received samples are used to compute one dimensional (1D) branch metrics. Then the 1D branch metrics are added up to obtain four dimensional (4D) branch metrics. Lastly, the 4D branch metrics are used to update state metrics and survivor paths. This entire process is repeated at the next iteration. In this serial process, all of the computations are on the critical path.
At time n, look-ahead DFU 32 is used to compute partial ISI estimates for code state ρn+1 due to the channel coefficients {f2,j, f3,j, . . . , fN,j} based on the already known survivor symbol sequence. Assuming there is a state transition between ρn and ρn+1, then the partial ISI estimate for ρn+1 corresponding to the transition can be calculated as:
Since there are 8 code states and 4 wires, altogether 32 look-ahead ISI estimates are needed to compute. The computation time of look-ahead DFU 32 is around 4 additions if we use carry-save adder structure.
The look-ahead 1D BMU 34 computes look-ahead 1D branch metrics for transitions departing from code states {ρn+1} Inputs to the look-ahead 1D BMU are partial ISI estimates {ûn+1,j(ρn)} due to {f2,j, f3,j, . . . , fN,j} and the received sample zn+1,j. In addition, look-ahead 1D BMU 34 needs to consider the ISI partial contribution due to the channel coefficient f1,j and the 1D symbol decision an,j(ρn→ρn+1) associated with state transitions ρn→ρn+1. A speculative ISI estimate for the state transition ρn→ρn+1 can be calculated as
Since pulse amplitude modulation ten (PAM10) is utilized, there are 10 possible choices for an,j(ρn→ρn+1) and in turn 10 possibilities for ûn+1,j(ρn→ρn+1). The high-speed PDFD architecture 30 (
λn(rn,j,an,j,ρn)=(rn,j−an,j+un,j(ρn))2 (3)
the two possibilities for an,j are only dependent on ρn. Thus, there are only two possibilities for ûn+1,j(ρn→ρn+1). Therefore, the only pre-computations needed are look-ahead 1D branch metrics for the 2 possibilities, resulting in a high hardware reduction.
As the two possible choices for an,j(ρn→ρn+1) are only dependent on the initial state ρn, the possible ISI estimates for state ρn+1 are only dependent on ρn too. For code states {ρn+1=0,1,2,3}, as they have the same predecessor states {ρn=0,2,4,6}, their LA 1D branch metrics are the same. Therefore, LA 1D branch metrics for only one of them needs to be computed. This is also true for code states {ρn+1=4,5,6,7}. For wire j and initial code state ρn four look-ahead 1D branch metrics are needed to be calculated according to:
{circumflex over (λ)}n+1,j(rn+1,j,an+1,j,ρnnan,j)=(rn+1,j−an+1+un+1,j(ρn)−f1,jan.,j)2 (4)
with two (one per 1D subset for an+1,j) for an,j=an,j(ρn, A) and two for an,j=an,j(ρn, B). As there are eight code states and four wires, altogether 8×4×4=128 1D look-ahead branch metrics are needed to compute. This is a reduction to the 640 look-ahead branch metrics which are needed to compute in straightforward implementations.
For code state ρn+1 and wire j, two real 1D metrics (one for an,jεA and one for B) need to be selected among 16 precomputed branch metrics (four from each of 4 predecessor states of ρn+1).
4D branch metrics 36 (
For each state transition ρn→ρn+1 two 4D branch metrics (one is associated with an A-type 4D symbol and the other B-type) are needed to be computed. The smaller metric (referred to as λn(rn, an, ρn→ρn+1) and its associated 4D symbol an(ρn→ρn+1) are selected to be used in ACSU 38.
ACSU 38 (
The outputs of ACSU 38 are the newly decoded 4D survivor symbol an(ρn+1) and path selection decision dn(ρn+1). The outputs are used to update the survivor sequence. The new sequence will be used to compute ISI estimates in the next iteration.
As illustrated in
where Sel(d, x0,x1,x2,x3) is a 4-to-1 multiplexing function and depending on d, Sel(d, x0,x1,x2 ,x3) selects one of xi, i=0,1,2,3 as its output.
The proposed techniques in the previous sections are also applicable to other applications and trellis coded modulation schemes other than the one described in this paper. The proposed techniques may be used for any applications where it is necessary to decode trellis encoded signals in the presence of inter-symbol interference and noise. For example, the proposed techniques may be used for 1000BASE-T which uses a 5-level PAM modulation combined with a 4D 8-state trellis code.
Various embodiments of the invention have been described. These and other embodiments are within the scope of the following claims.
This application claims the benefit of U.S. Provisional Application No. 60/609,304, to Parhi et al., entitled “PIPELINED PARALLEL DECISION FEEDBACK DECODERS FOR HIGH-SPEED COMMUNICATION SYSTEMS,” filed Sep. 13, 2004, and U.S. Provisional Application No. ______, to Parhi et al., entitled “PIPELINED PARALLEL DECISION FEEDBACK DECODERS FOR HIGH-SPEED COMMUNICATION SYSTEMS,” having attorney docket no. 1008-030USP2, filed Sep. 9, 2005, the entire contents of each being incorporated herein by reference.
The invention was made with Government support from the National Science Foundation No. CCF-0429979. The Government may have certain rights in this invention.
Number | Date | Country | |
---|---|---|---|
60609304 | Sep 2004 | US | |
60715464 | Sep 2005 | US |