The present invention relates to the field of communications, and more particularly, systems and methods for adaptively decoding transmitted frames in the presence of impulsive noise and interference.
As the world has become more reliant on computers and information exchange, the need for the reliable transmission of data has become increasingly important. One key element in information exchange is the accurate and efficient transmission and reception of data across noisy transmission channels.
Signal processing methods implemented in practical communications systems are usually designed under the assumption that any underlying noise and interference is Gaussian. Although this assumption finds strong theoretical justification in the central limit theorem, the noise and interference patterns commonly present in modem mobile communications systems are far from Gaussian. Noise and interference generally exhibit “impulsive” behavior. In typical mobile communication systems, noise and interference sources often include motor-vehicle ignition noise, switching noise from electromechanical equipment, thunderstorms, and heavy bursts of interference. Current signal processing systems are not designed to handle these non-Gaussian noise sources. Accordingly, these systems may perform poorly, and might even fail, in the presence of impulsive noise.
Channel noise and interference can be effectively modeled as the superposition of many small and independent effects. In practice, these effects do not always follow a Gaussian distribution. This situation appears to contradict the central limit theorem. For many years, engineers have been unable to explain this apparent contradiction. Consequently, many of the techniques developed to cope with impulsive noise were mainly ad hoc, largely based on signal clipping and filtering prior to application of a Gaussian-based technique.
Clipping the amplitude of an input signal is only effective if the amplitude of the input signal is above or below the specific threshold values. These threshold values are typically determined by the limits of the hardware used in a receiver in a communications system. Accordingly, the threshold values are often chosen to take advantage of the full dynamic range of the analog to digital (A/D) converter(s) of the receiver. However, if impulsive noise, added to the input signal, does not cause the amplitude of the signal to exceed a specific threshold, clipping will not remove the noise. Additionally, even when noise does cause the signal to exceed the threshold, the clipping solution only removes the noise to the extent that the magnitude of the signal plus the noise is above the threshold. Accordingly, noise is not actually removed, but its effects are somewhat reduced.
When individual signals within a sequence are contaminated by noise, the sequence may not be properly decoded and efficient communications may be difficult. In typical communication systems, decoding is used to identify potential communication errors. Additionally, decoding may be able to correct some, or even most, errors. Errors may be corrected by one of many error detection and correct schemes known to those skilled in the art. Typical coding and decoding schemes are able to correct errors by inserting controlled redundancy into the transmitted information stream. This is typically performed by adding additional bits or using an expanded channel signal set. These schemes allow the receiver to detect and possibly correct errors.
In its most simple form, one problem with noisy transmission environments is that, a certain percentage of the time, a transmitted ‘1’ is received as a ‘0’ or vice versa. There are many methods of encoding data that allow received errors to be detected or even corrected. These encoding and decoding schemes are typically optimized based on a set of underlying assumptions. Preferably, these assumptions are designed to match the conditions of a real-world communications environment. Generally, decoding systems are designed under the assumption that the underlying noise and interference is Gaussian. When these assumptions do not match real-world conditions, the performance of such schemes may no longer be optimal. In real-world environments, this effect is ever-present because conditions are constantly changing. Even those systems in existence that try to accommodate impulsive noise are based on average conditions. These systems fall short of optimal performance when conditions stray from the average. These problems are compounded in a mobile system because conditions change even more rapidly and more often than stationary systems.
When a receiver is able to detect, but not correct, errors in received information, the receiver may request that the transmitter resend the information. In a noisy environment, this may lead to highly inefficient communications.
There is a need in the art for systems and methods for processing signals to alleviate impulsive noise distortion.
Additionally, there is a need in the art for systems and methods for providing an adaptive decoding system for efficient communications in non-Gaussian, non-stationary environments.
The present invention overcomes the limitations of the existing technology by providing systems and methods for providing adaptive decoding for efficient communications in non-Guassian, non-stationary environments. An exemplary adaptive decoding system in accordance with the present invention comprises a decoder having a decoding scheme for decoding a received transmission frame, a channel assessment unit for sensing channel characteristics, and a tuning unit for adjusting the decoding scheme of the decoder at least in part on the channel characteristics sensed by the channel assessment unit.
Other objects, features, and advantages of the present invention will become apparent upon reading the following detailed description of the embodiments of the invention, when taken in conjunction with the accompanying drawings and appended claims.
Referring now to the drawings, in which like numerals refer to like techniques throughout the several views, exemplary embodiments of the present invention are described.
As discussed in the background of the present invention, noise and interference patterns are comprised of small and independent effects. Placing upper and lower boundaries on the variance of these effects may allow greater reliance on the central limit theorem. From a conceptual perspective, an unbounded or infinite variance is feasible as a model of highly dispersed or impulsive phenomena. Without the finite variance constraint, a converging sum of normalized random variables can be proven to belong to a wider class of random variables known as “α-stable”. Thus, similar to Gaussian processes, α-stable processes can appear in practice as the result of physical principles. Furthermore, all non-Gaussian α-stable processes are “heavy-tailed” processes with infinite variance, explaining the often found impulsive nature of practical signals.
“Symmetric” α-stable random variables are commonly described through their characteristic function:
Φ(ω)=e−γ|ω|
where α is the index or characteristic exponent, and γ is the dispersion. Analogous to the variance in a Gaussian process, γ is a measure of signal strength. The shape of the distribution is determined by α From the equation above, it can be shown that α is restricted to values in the interval (0,2]. Qualitatively, smaller values of α correspond to more impulsiveness in the distribution. The limiting case α=2 corresponds to a Gaussian distribution, which is the least impulsive α-stable distribution and the only one with finite variance. A value of α=1 results in a random variable with a Cauchy distribution.
A theory of estimation in α-stable environments can be derived from the tools of robust statistics. In general, let ρ(x) be a symmetric cost function or metric which is monotonically non-decreasing on [0, ∞). For a set of samples x1, x2, . . . ,xN, the M-estimator of the location parameter, β, is defined as
In the theory of M-estimators, the shape of the cost function ρ determines the characteristics of the estimate, {circumflex over (β)}. For example, if ρ(x)=x2, {circumflex over (β)} becomes the least-squares estimate (i.e. the sample mean). For ρ(x)=|x|, {circumflex over (β)} is the sample median. The cost function:
ρ(x)=log(k2+x2), (3)
where k is a constant, possesses important optimality properties along the whole range of α-stable distributions.
The importance of the cost function described in the above equation is that the value of k can be tuned to give optimal estimation performance depending on the parameters of the underlying distribution. Cost-functions may be tuned to optimize performance for a set of conditions.
k(α, γ)=k(α)γ1/α. (4)
The above expression indicates a “separability” property of the optimal value of k in terms of the parameters α and γ. This reduces the problem of the functional form of k (α,γ) to that of determining the simpler form:
k(α)=k(α, 1), 0<α≦2. (5)
This function describes “the α-k plot” of α-stable distributions. Under the “maximum likelihood” optimality criterion, it can be proven that the α-k plot touches three fundamental points:
The above points suggest the general shape of the α-k plot illustrated in
where P is in general a positive constant. Due to its relative simplicity and efficient results, the following tuning function is often preferred:
One general goal of using encoding and decoding for the transmission of data, is to minimize the probability of error. In the situation where various coded sequences are equally likely, this may be accomplished using a “maximum likelihood” decoder. For hard decision decoding, it is well known that the maximum likelihood decoder selects the codeword that is closest in Hamming distance to the received sequence.
It is also well known that soft decision decoding offers a performance advantage over hard decision decoding. Soft decision decoding preserves information contained in the received sequence and passes that information on to a decoding scheme. The task is to choose a cost function appropriate for soft decision decoding. For a channel with underlying noise and interference that is Gaussian, maximum likelihood decoding is achieved using a Euclidean distance (ρ(x)=x2) as the cost function. Furthermore, if the channel can be accurately modeled as a stationary additive-noise channel with noise density function ƒ(x), an optimal cost function can be easily designated as ρ(x)=−log(ƒ(x)). In real world applications, however, the statistical behavior of a channel is not stationary. Accordingly, finding a model for the channel noise is a significant challenge. This situation is aggravated by the fact that even a very small deviation from model channel assumptions can lead to the design of highly inefficient cost functions. Thus, the choice of an appropriate cost function that gives satisfactory robust performance is important.
The present invention envisions decoders that are adapted to be adjusted according to present channel characteristics. In an exemplary embodiment of the present invention, a Viterbi decoder is used in which a metric function is changed according to channel characteristics. The present invention is optimized for use in channels having any of the following characteristics:
The metric generator 310 is adapted to use a finite number of cost functions ρ1, ρ2, . . . ρN,. The channel assessment unit 325 is, in general, adapted to use a discrete function that, based on soft data encoding from a demodulator (not shown), decides which cost function should be applied. The channel assessment unit 325 may be adapted to make use of a pilot signal or may be further adapted to work directly on the received data sequence. It may also work on a continuous basis, or only during a percentage of the transmission time in the same way as multiplexed pilots.
In an exemplary embodiment of the present invention, the channel assessment may be implemented as the estimator of a vector of channel parameters {circumflex over (θ)}=[{circumflex over (θ)}1, {circumflex over (θ)}2, . . . , {circumflex over (θ)}p]. Then, the parameter space p is partitoned into N disjoint regions, S1, S2, . . . , SN. Selection adaptation is then performed according to the simple rule:
select ρi when {circumflex over (θ)}εSi. (9)
In an exemplary embodiment of the present invention, the operation of the channel assessment unit 325 may be further explained using the following example of Gaussian/Laplacian selective decoding or “hard” decoding. It should be understood that this is only one example and in no way should be construed as a limitation to the present invention. When performing Gaussian/Laplacian hard adaptation, the following cost functions may be used:
ρ1(x)=x2 (Optimum for Gaussian noise) (10)
ρ2(x)=|x|(Optimum for Laplacian noise). (11)
Alternatively, a channel assessment unit 325 may be adapted to perform a Gaussianity test in place of the channel assessment scheme. The Gaussianity test may have different structures. One commonly used Gaussianity test is based on Bispectral estimation. In this method, the 2-D Fourier Transform of the third order cumulant sequence yields the bispectrum of the process. Essentially the bispectrum of a Gaussian process is zero over all frequencies (zero cumulant matrix). The Gaussian test hypothesizes whether there is sufficient statistical deviation from zero to imply a non-Guassian source. The main goal of the Gaussianity test is to give information about the current impulsiveness of the channel. Typically, the test will be comprised of a real function λ, applied to the demodulated history:
λ=λ({circumflex over (β)}M, {circumflex over (β)}M-1, {circumflex over (β)}M-2, . . . {circumflex over (β)}1), (12)
where {circumflex over (β)}M is the current demodulated symbol, and {circumflex over (β)}M-1 is the output of the demodulator i symbols ago. If λ is larger than a predetermined constant C, then the test determines that the channel can be considered Gaussian, and ρ1 is chosen. On the contrary, if λ≦C, the test indicates a tail heavier than Gaussian, calling for the use of a more impulse-resistant cost function, namely ρ2.
It should be noted that the performance of the system depends strongly on the quality of the test. Also, the adaptation speed and/or responsiveness of the system is determined by the relative importance that the function λ gives to recent and old data. In an exemplary embodiment of the present invention, λ is a “weighted” test function assigning more importance or weight to the most recent data. The variation of the weight distribution determines the responsiveness of the system. Furthermore, λ can be defined to operate only on the M most recent data. The size of M, in this case, will determine the system responsiveness (or adaptivity), with small values of M corresponding to more responsive systems.
When transitions from one cost function to the other occur, the memory of the decoder is adapted to store information in terms of the old cost function. As part of the normal operation of the decoder 315, the metrics associated with the new cost function are added to old accumulated metrics. Hence, metrics may be scaled in order to guarantee comparability or compatibility of ρ1 and ρ2 when both contribute information to the accumulated metric.
Noting that ρ2 and cρ2 (for c>0) are equivalent cost functions, an optimal value of c may be designed to guarantee compatibility and optimal performance during transitions. Such values of c are referred to as the transition constants of the system. Although only one non-unit transition constant is needed for the example outlined, c1 and c2 may be referred to as the associated transition constants in the sense that the implementation of the decoder 315 uses the equivalent metrics c1ρ1 and c2ρ2.
The present invention may also be used for soft adaptation implementations. When operating using soft adaptations, a family of cost functions are defined: ρk,kεL. The tuning of the system (i.e., the selection of the cost function to be used) is performed using a tuning function:
k=k({circumflex over (θ)}). (13)
The difference between hard and soft adaptation (and associated hard and soft data) can be seen as
k is a quantized function
k is continuous
In hard adaptation, a firm (or hard) decision is made regarding each bit as it is received. In soft adaptation, decisions may be changed if other received bits (typically received later in time) indicate that the earlier decision was likely incorrect.
A third function, the transition function, (in addition to the previously introduced cost and tuning functions) is used in soft adaptations. Instead of the constants c1, c2, . . . cN used in hard adaptation, a function c(k) scales each ρk appropriately to allow an efficient comparison among the different cost functions.
The use of different cost functions may be illustrated using a Myriad-based decoder. In an exemplary embodiment of the present invention, a Myriad-based decoder is adapted to use the following cost function and parameters:
ρκ(x)=log(k2+x2), kε
θ=[α, γ] (The parameters characterizing α-stable noise)
θ=[{circumflex over (α)}, {circumflex over (γ)}] (Any specific known estimator)
k(θ)=k({circumflex over (α)},{circumflex over (γ)})=k({circumflex over (α)})γ1/α,
where k({circumflex over (α)}) is a predetermined α-k plot such as the one shown in
The optimal form of c(k) may be designed through simulations or from theoretical analysis. For instance, c(k)=k2 is an exemplary sample which generally provides good performance. Alternatively, simulations may be performed to identify a function c(k) that produces optimal results for a specified system.
Once a value of α is estimated, k(α) may be calculated in step 415 using one of the approximation functions for k(α) shown above, or k(α) may be taken directly from the α-k plot shown in
ρk(x)=log(k2+x2) (14)
After the new metric is calculated, the decoding scheme may be adjusted to the new metric in step 425.
While this invention has been described with reference to embodiments thereof, it should be understood that variations and modifications can be made without departing from the spirit and scope of the present invention as defined by the claims that follow.
Number | Name | Date | Kind |
---|---|---|---|
5745524 | Hull | Apr 1998 | A |
5867531 | Shiino et al. | Feb 1999 | A |
6504868 | Koizumi | Jan 2003 | B1 |
6748032 | Kim et al. | Jun 2004 | B1 |
Number | Date | Country | |
---|---|---|---|
20030223512 A1 | Dec 2003 | US |