Non-linear neural network equalizer for high-speed data channel

Information

  • Patent Grant
  • 12284059
  • Patent Number
    12,284,059
  • Date Filed
    Tuesday, January 25, 2022
    3 years ago
  • Date Issued
    Tuesday, April 22, 2025
    7 days ago
  • Inventors
    • Nangare; Nitin (Sunnyvale, CA, US)
    • Aboutaleb; Ahmed Medhat Ahmed Fahmi Eid (Sunnyvale, CA, US)
  • Original Assignees
  • Examiners
    • Tadese; Berhanu
Abstract
A data channel on an integrated circuit device includes a non-linear equalizer having as inputs digitized samples of signals on the data channel, decoding circuitry configured to determine from outputs of the non-linear equalizer a respective value of each of the signals, and adaptation circuitry configured to adapt parameters of the non-linear equalizer based on respective ones of the value. The non-linear equalizer includes a non-linear filter portion, and a front-end filter portion configured to reduce numbers of the inputs from the digitized samples. The non-linear equalizer may be a neural network equalizer, such as a multi-layer perceptron neural network equalizer, a reduced complexity multi-layer perceptron neural network equalizer, or a radial-basis function neural network equalizer. Alternatively, the non-linear equalizer may include a linear filter and a non-linear activation function, which may be a hyperbolic tangent function.
Description
FIELD OF USE

This disclosure relates to the use of non-linear equalizers in a high-speed data channel. More particularly, this disclosure relates to the use of non-linear neural-network equalizers in the receiver side of a high-speed SERDES (serializer-deserializer) channel on an integrated circuit device, or in the read channel of a storage device controller.


BACKGROUND

The background description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the inventors hereof, to the extent the work is described in this background section, as well as aspects of the description that may not otherwise qualify as prior art at the time of filing, are neither expressly nor impliedly admitted to be prior art against the subject matter of the present disclosure.


Many integrated circuit devices, particularly “systems-on-chip” (SoCs), include high-speed serial links between various device components (such as the individual silicon dice in an SoC). Typical high-speed serial links of that type, commonly known as “SERDES” (serializer/deserializer), may suffer from significant non-linearity or channel impairment in the signal path, as a result of, e.g., insertion loss, inter-symbol-interference (ISI), and, in an optical system, non-linearities such as dispersion loss, or, in a copper (i.e., wired) system, cross-talk, jitter, etc. Various forms of linear equalization typically are used, at the receiver end of such links, to attempt to mitigate such channel impairments. However, linear equalization may not be sufficient to compensate for such non-linearities, particularly when the data levels to be distinguished are close together.


Similarly, in magnetic recording, reading and writing are performed by a head of a hard disk drive that moves relative to the surface of a storage medium and writes data to, or reads data from, circular data tracks on a magnetic disk. In order to increase recording densities, it is desirable to shrink the bit cell, or area of disk surface in which a single bit is recorded. Shrinking the bit cell, however, increases inter-symbol interference (ISI) from data recorded on the media, thereby increasing the bit error rate (BER) and decreasing the reliability of read-back data. An increased BER also effectively reduces the rate at which data can be read back, owing to the overhead inherent in error-detection or error-correction techniques employed to compensate for the increased BER, and/or owing to repeat read-back attempts that may be necessary in order to accurately read-back data after erroneous data read-back attempts.


SUMMARY

In accordance with implementations of the subject matter of this disclosure, a data channel on an integrated circuit device includes a non-linear equalizer having as inputs digitized samples of signals on the data channel, decoding circuitry configured to determine from outputs of the non-linear equalizer a respective value of each of the signals, and adaptation circuitry configured to adapt parameters of the non-linear equalizer based on respective ones of the value. The non-linear equalizer includes a non-linear filter portion, and a front-end filter portion configured to reduce numbers of the inputs from the digitized samples.


In a first implementation of such a data channel, the non-linear equalizer may be a neural network equalizer.


According to a first aspect of that first implementation, the neural network equalizer may be a multi-layer perceptron neural network equalizer.


In one instance of that first aspect of the first implementation, the multi-layer perceptron neural network equalizer may be a reduced complexity multi-layer perceptron neural network equalizer.


According to a second aspect of the first implementation, the neural network equalizer may be a radial-basis function neural network equalizer.


In a second implementation of such a data channel, the non-linear equalizer may include a linear filter and a non-linear activation function.


According to a first aspect of that second implementation, the non-linear activation function may be a hyperbolic tangent function.


In a third implementation of such a data channel, the adaptation circuitry may adapt parameters of the non-linear equalizer based on cross-entropy.


In a fourth implementation of such a data channel, the front-end filter portion may include a FIR filter.


A fifth implementation of such a data channel may include scalable bypass circuitry for controllably outputting output of the front-end filter portion as at least a portion of output of the non-linear equalizer.


A method according to implementations of the subject matter of this disclosure for detecting data on a data channel on an integrated circuit device includes performing non-linear equalization of digitized samples of input signals on the data channel, determining from output signals of the non-linear equalization a respective value of each of the output signals, and adapting parameters of the non-linear equalization based on respective ones of the value. Performing non-linear equalization of digitized samples of input signals on the data channel includes performing front-end filtering to reduce numbers of the inputs from the digitized samples, and performing non-linear filtering on the reduced number of inputs from the digitized samples


In a first implementation of such a method, performing non-linear equalization may include performing neural network equalization.


According to a first aspect of that first implementation, performing neural network equalization may include applying a multi-layer perceptron neural network equalizer.


In a first instance of that first aspect of the first implementation, performing neural network equalization may include applying a reduced complexity multi-layer perceptron neural network equalizer.


According to a second aspect of the first implementation, performing neural network equalization may include applying a radial-basis function neural network equalizer.


In a second implementation of such a method, performing non-linear equalization may include applying a non-linear activation function and performing linear filtering on output of the non-linear activation function.


According to a first aspect of that second implementation, applying a non-linear activation function may include applying a hyperbolic tangent function.


In a third implementation of such a method, adapting parameters of the non-linear equalization may include adapting parameters of the non-linear equalization based on cross-entropy.


In a fourth implementation of such a method, performing front-end filtering may include performing FIR filtering.


A fifth implementation of such a method may further include scalably bypassing the non-linear filtering for controllably outputting an output of the front-end filtering as at least a portion of output of the non-linear equalization.





BRIEF DESCRIPTION OF THE DRAWINGS

Further features of the disclosure, its nature and various advantages, will be apparent upon consideration of the following detailed description, taken in conjunction with the accompanying drawings, in which like reference characters refer to like parts throughout, and in which:



FIG. 1 illustrates a TDMR detector channel as one example of a channel with which implementations of the subject matter of this disclosure may be used;



FIG. 2 is a plot of an exclusive-OR function in a Cartesian coordinate space illustrating a problem solved by implementations of the subject matter of this disclosure;



FIG. 3 is a plot of a transformation of the exclusive-OR function of FIG. 2 into a different coordinate space illustrating a solution based on implementations of the subject matter of this disclosure;



FIG. 4 is a schematic representation of a general implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 5 is a diagram of a first implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 6 is a diagram of a second implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 7 is a diagram of a third implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 8 is an alternative representation of the implementation of the reduced-complexity non-linear neural network filter shown in FIG. 7;



FIG. 9 is a diagram of a fourth implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 10 is a diagram of a fifth implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure;



FIG. 11 is a graphical representation of a function to be filtered;



FIG. 12 is a diagram of a sixth implementation of a reduced-complexity non-linear neural network filter in accordance with the subject matter of this disclosure; and



FIG. 13 is flow diagram illustrating a method according to implementations of the subject matter of this disclosure.





DETAILED DESCRIPTION

As noted above, integrated circuit devices may include high-speed SERDES links between various device components. Typical SERDES links may suffer from significant non-linearity or channel impairment in the signal path, as a result of, e.g., insertion loss, inter-symbol-interference (ISI), and, in an optical system, non-linearities such as dispersion loss or, in a copper (i.e., wired) system, cross-talk, jitter, etc. Various forms of linear equalization typically are used, at the receiver end of such links, to attempt to deal with such channel impairments.


Similarly, in magnetic recording, reading and writing are performed by a head of a hard disk drive that moves relative to the surface of a storage medium and writes data to, or reads data from, circular data tracks on a magnetic disk. In order to increase recording densities, it is desirable to shrink the bit cell, or area of disk surface in which a single bit is recorded. Shrinking the bit cell, however, increases inter-symbol interference (ISI) from data recorded on the media, thereby increasing the bit error rate (BER) and decreasing the reliability of read-back data. An increased BER also effectively reduces the rate at which data can be read back, owing to the overhead inherent in error-detection or error-correction techniques employed to compensate for the increased BER, and/or owing to repeat read-back attempts that may be necessary in order to accurately read-back data after erroneous data read-back attempts.


Two-dimensional magnetic recording (TDMR) is another technique that has been developed in an effort to increase storage capacity in hard disk drives. TDMR employs a read-back technique that allows for greater storage capacity by combining signals simultaneously obtained from multiple read-back heads to enhance the accuracy of reading back data from one or more data tracks. A TDMR read-back channel typically includes a linear equalizer to mitigate the negative impact that noise has on the read-back channel signal integrity, and on the accuracy and reliability in reading back digital data values from the storage medium. Some TDMR read-back channels utilize minimum-mean-squared-error (MMSE) as a cost function to adapt the equalizer to further improve BER performance.


However, linear equalization may not be sufficient to compensate for such non-linearities or interference. Linear equalization may not be enough to correctly assign received samples near the threshold between levels to the correct written bit or symbol when the signal-to-noise ratio is low.


In accordance with implementations of the subject matter of this disclosure, non-linear equalization is used to compensate for non-linearities in a high-speed data channel, such as a SERDES channel, or a disk-drive read channel, thereby reducing the bit-error rate (BER). In different implementations, different types of non-linear equalizers may be used.


Conceptually, a linear equalizer performs the separation of samples for assignment to one level or another by effectively drawing a straight line between groups of samples plotted in a two-dimensional (e.g., (x,y)) space. In channels that are insufficiently linear, or where the levels are too close together, there may not be a straight line that can be drawn between samples from different levels on such a plot. A non-linear equalizer effectively re-maps the samples into a different space in which the samples from different levels may be separated by a straight line or other smooth curve.


A non-linear equalizer in accordance with implementations of the subject matter of this disclosure may be more or less complex. For example, a non-linear equalizer may have more or fewer variables, or taps, with complexity being proportional to the number of variables. In addition, a non-linear equalizer that operates at the bit level—i.e., operates separately on the bits of each symbol (e.g., two bits/symbol for four-level signaling in a data channel) rather than on the symbol as a whole—may be less complex than a non-linear equalizer that operates at the symbol level. Either way, greater complexity yields greater performance when all other considerations are equal. However, greater complexity also may require greater device area and/or power consumption.


Types of non-linear equalizers that may be used in accordance with the subject matter of this disclosure may include forms of reduced-complexity multi-layer perceptron neural network (RC-MLPNN) equalizers and forms of reduced-complexity radial-basis-function neural network (RC-RBFNN) equalizers.


Performance of the non-linear equalizer may be affected by the cost function used for adaptation of the equalizer. Implementations of the subject matter of this disclosure reduce the complexity of a non-linear equalizer whether it uses any one of various different cost functions for adaptation, including either a minimum mean-square error (MMSE or MSE) cost function, or a cross-entropy (CE)-based cost function. A CE-based cost function may yield a better result than an MMSE cost function, but a CE-based cost function is more complex than an MMSE cost function.


According to implementations of the subject matter of this disclosure, a non-linear equalizer with reduced complexity but comparable performance is provided by appending, to a non-linear neural network equalizer, a front-end filter to reduce complexity of the inputs to the non-linear equalizer. For example, a finite-impulse-response (FIR) filter may be used as a front end to reduce complexity of the non-linear equalizer by reducing the number of input parameters or dimensions.


The subject matter of this disclosure may be better understood by reference to FIGS. 1-13.



FIG. 1 illustrates a TDMR detector channel 100 as one example of a channel with which implementations of the subject matter of this disclosure may be used. However, as noted above, implementations of the subject matter of this disclosure also may be used with other forms of high-speed data channels such as a SERDES channel (not shown).


In TDMR detector channel 100, respective analog-to-digital converter (ADC) outputs 111, 121 from two separate read heads (not shown) are input to an equalizer 101 for equalization according to implementations of the subject matter of this disclosure. Output Y of equalizer 101 is then passed to Viterbi detector 102, the output of which is then passed to a soft-output Viterbi algorithm (SOVA) detector 103. SOVA detector 103 provides log-likelihood ratios (LLRs) 113 to a decoder, such as a low-density parity check (LDPC) decoder 104 which decodes the bits. LLRs 113 also are fed back to equalizer 101 which is adapted using either a mean-square error (MSE) cost function 151 or a cross-entropy cost function 161 which compares the LLRs 113 to output bits 114 (which are expressed as non-return-to-zero data 131), to set a target 141.


The purpose of implementing equalization on channel 100 is to correct for various sources of interference referred to above and thereby effectively move samples that are on the wrong side of a detection threshold (whether for bits read from a storage medium or signal levels in a SERDES channel) to the correct side of the threshold. Linear equalization effectively takes a plot of the samples in a two-dimensional (x,y) space and draws a straight line through the samples where the threshold ought to be. However, in a channel with non-linearities, there may be no straight line that can be drawn on that two-dimensional plot that would correctly separate the samples. In such a case, non-linear equalization can be used. A non-linear equalization function may effectively remap the samples into a different space in which there does exist a straight line that correctly separates the samples.


Alternatively, the non-linear equalization function may remap the samples into a space in which there exists some smooth curve other than a straight line that correctly separates the samples. For example, non-linear equalization using a radial-basis function may remap the samples into a polar-coordinate, or radial, space in which the samples are grouped into circular or annular bands that can be separated by circles or ellipses.


The advantage of non-linear equalization over linear equalization in a non-linear channel may be seen in a simplified illustration as shown in FIGS. 2 and 3, where the signal to be equalized is characterized by the exclusive-OR (XOR or @) function. FIG. 2 is plot of y=x1⊕x2 in (x1,x2) space, where the open dots 201, 202 represent y=0 and cross-hatched dots 203, 204 represent y=1. It is apparent that there is no straight line that can be drawn separating the open dots from the cross-hatched dots.


However, a radial-basis function







φ

(

r
i

)

=


φ

(



x
-

c
i




)

=

e

-





[




x
1






x
2




]

-

c
i




2









where ci is the centroid of the ith node, can be used to transform the XOR function from the linear Cartesian (x1,x2) space to a non-linear radial (φ(r1),φ(r2)) space as follows:
















x1
x2
φ(r1)
φ(r2)
y



















0
0
0.1353
1
0


0
1
0.3678
0.3678
1


1
0
0.3678
0.3678
1


1
1
1
0.1353
0










which is diagrammed in FIG. 3. As can be seen, when mapped into the non-linear radial (φ(r1),φ(r2)) space, the values 301, 302, 303 (as can be seen, both of the two y=1 points 201, 202 in (x1,x2) space map to the same point 301 in (φ(r1),φ(r2)) space) of the XOR function 300 may be separated by straight line 304.


As discussed below, various types of non-linear equalizers are available. Whatever type of non-linear equalizer is used may be adaptive to account for changing channel conditions. Various forms of adaptation may be used.


One type of adaptation function that may be used is minimum mean-squared error (MMSE), where the mean-squared error (MSE) is defined as the square of the norm of the difference between the equalized signal (Y) and the ideal signal (Y). The equalizer may initially be adapted in a training mode in which the ideal signal values are available. Later, during run-time operation, the detected output values of the equalized channel should be close enough to the ideal values to be used for adaptation.


Another type of adaptation function that may be used is the cross-entropy (CE) between a training bit and its log-likelihood ratio (LLR). In particular, cost function circuitry may be configured to compute a cross-entropy value indicative of a difference between a probability distribution of the detected bit value (which is a function of the LLR signal) and a probability distribution of the training bit value. The cost function circuitry then adapts the equalizer by setting an equalizer parameter (e.g., one or more coefficients of filter taps of the equalizer) to a value that corresponds to a minimum cross-entropy value from among the computed cross-entropy value and one or more previously computed cross-entropy values, to decrease a bit-error rate for the channel. As in the case of MSE equalization, the equalizer may initially be adapted in a training mode in which the ideal signal values are available. Later, during run-time operation, the detected output values of the equalized channel should be close enough to the ideal values to be used for adaptation. Specifically, if any forward error correction code (FEC) decoder (e.g., a Reed Solomon (RS) decoder or Low-Density Parity Check (LDPC) decoder) is available after the equalizer, then successfully decoded frames from the FEC decoder output may be used for adaptation.


LLR may be defined as the relationship between the probability (P0) of a bit being ‘0’ and the probability (P1) of a bit being ‘1’:







LLR
=

L
=

log

(


P
1


P
0


)








P
1

+

P
0


=
1





P
0

=

1

(

1
+

e
L


)







P
1

=


e
L


(

1
+

e
L


)








The cross-entropy between a training bit and its LLR may be computed as follows:








Cross



Entropy
(

bit
,
LLR

)


=



-

P
(

bit
=
0

)


·

log

(

P
0

)


-


P
(

bit
=
1

)

·

log

(

P
1

)








Cross



Entropy
(

bit
,
LLR

)


=



-

(

1
-
bit

)


·

log

(

P
0

)


-

bit
·

log

(

P
1

)








Cross


Entropy

=

Inf


when


{







bit
=
0

,


P
0

=
0








bit
=
1

,


P
1

=
0







Cross


Entropy

=

0


when


{





bit
=
0

,


P
0

=
1








bit
=
1

,


P
1

=
1















When the true bit is a logic ‘0’ but the probability of the detected bit represented by the LLR indicates that P0=0, or the true bit is a logic ‘1’ but the probability of the detected bit represented by the LLR indicates that P1=0, then the true value is the complete opposite of the expected value, meaning that cost (cross-entropy) approaches infinity. On the other hand, when the probability of a detected bit value as indicated by the LLR agrees with the true bit value, then cross-entropy equals zero. Insofar as in most cases both probabilities P0 and P1 are higher than 0 and lower than 1, cross-entropy will be a finite non-zero value. Thus, this cost function can be used for adaptation and reflects the quality of the detected bits, with the goal being to minimize cross-entropy.


The gradient of cross-entropy with respect to the LLR may be computed by substituting for P0 and P1 in the cross-entropy equation:










(

C

E

)





(

L

L

R

)



=



P
1

-
bit

=

{




P
1





when


bit

=
0








P
1

-
1

=

-

P
0







when


bit

=
1










The LLR may be adapted to minimize cross-entropy (i.e.,










(

C

E

)





(

L

L

R

)



=
0





), as follows:

LLRt+1=LLRt−α·P1 if bit=0
LLRt+1=LLRt+α·P0 if bit=1


A negative LLR means bit=0 has a higher probability than bit=1, while a positive LLR means bit=1 has a higher probability than bit=0. In these equations, P0 and P1 are probabilities and therefore are positive values, and a is an adaptation bandwidth which also is positive. Therefore, when the true bit=0 then adaptation using cross-entropy will make a negative LLR more negative, and when the true bit=1 then adaptation using cross-entropy will make a positive LLR more positive. Therefore, cross-entropy-based adaptation maximizes the magnitude of the LLR and hence is a maximum-likelihood adaptation which reduces BER. Thus, adaptation of the equalizer to minimize cross-entropy also minimizes BER.


If one assumes that there is a general computation graph from parameter X→Y→LLR→CE such that parameter X affects the value of output Y which affects the LLR, from which the cross-entropy may be computed, then the cross-entropy gradient can be expressed in terms of other parameters:










(
CE
)





(

parameter
X

)



=





(

parameter
Y

)





(

parameter
X

)



·




(

L

L

R

)





(

parameter
Y

)



·




(

C

E

)





(

L

L

R

)









Therefore, any parameter can be adapted to minimize the cross-entropy.



FIG. 4 shows a general implementation 400 of a reduced-complexity non-linear neural network filter 401 in accordance with the subject matter of this disclosure for equalizing the two ADC outputs 111, 121 in the TDMR read channel 100 of FIG. 1. Reduced-complexity non-linear neural network filter 401 accepts inputs 111, 121 of a certain complexity, but initially filters inputs 111, 121 through a front-end filter 402 to reduce the complexity of inputs 111, 121, before filtering reduced-complexity inputs 411, 421 through non-linear filter circuitry 403. Reduction of the complexity of inputs 411, 421 allows a reduction in the complexity (as measured by dimensionality) of non-linear filter circuitry 403, therefore the complexity of non-linear neural network filter 401, without having to reduce the complexity of the inputs 111, 121 being filtered.


A first implementation of a reduced-complexity non-linear neural network filter 500, shown in FIG. 5, is based on a radial-basis function non-linear neural network filter 501, with a finite-impulse-response-(FIR)-based front-end filter 502.


In radial-basis function non-linear neural network filter 501, digital samples from two inputs 511, 521 are delayed by delay line 531 and combined in radial-basis function non-linear neural network 541. As seen in FIG. 5, radial-basis function non-linear neural network 541 includes at least one hidden layer 550 of hidden nodes 551. Each hidden node 551 operates on each delayed sample with a radial-basis function, but to avoid crowding the drawing only some delays in delay line 531 are shown as being coupled to each hidden node 551. The outputs of hidden layer 550 are combined (e.g., by addition) at 552 to provide Y output 503.


Each sample input at 511, 521 adds a parameter or dimension to radial-basis function non-linear neural network filter 501, increasing filter complexity. In order to reduce the complexity of radial-basis function non-linear neural network filter 501, reduced-complexity non-linear neural network filter 500 includes front-end filter 502, which combines some of the inputs from ADC outputs 111, 121 to provide a reduced number of inputs 511, 521 to radial-basis function non-linear neural network filter 501. As can be seen in FIG. 5, in this implementation, front-end filter 502 uses FIR filtering (each line connecting a delay 512 to sum 522 represents multiplication of a sample by a coefficient (not shown), forming a filter tap, with the taps being summed at 522) to combine, e.g., every four input samples from ADC outputs 111, 121 into one input sample 511, 521. This allows a reduction in the complexity (as measured by dimensionality) of radial-basis function non-linear neural network filter 501, and therefore the complexity of non-linear neural network filter 500, without having to reduce the complexity of the inputs 111, 121 being filtered. The unseen coefficients may be parameters that adapted with a back-propagation algorithm and, for example, may be derived from the equation set forth above in connection with the cross-entropy gradient.


In the implementation of FIG. 5, each set of input samples 111, 121 is processed in a separate portion of delay line 512, and in a separate portion of delay line 531. In this implementation, with two sets of input samples (from the two read heads of TDMR channel 100), each delay line is divided into two segments. However, more generally, the number of segments corresponds to the number of input sets. Thus, for a single input set, there would be only one segment (i.e., the delay line would not be segmented) but if there were three inputs sets, the delay line may be divided into three segments, etc.


A second implementation 600 of a reduced-complexity non-linear neural network filter, shown in FIG. 6, also is based on a radial basis filter neural network filter 601, with a finite-impulse-response-(FIR)-based front-end filter 602. As in the case of front-end filter 502, front-end filter 602 uses FIR filtering (each line connecting a delay 612 to radial basis function 611, 621 represents multiplication of a sample by a coefficient (not shown; see discussion above in connection with FIG. 5) forming a filter tap) to combine, e.g., every four input samples from ADC outputs 111, 121 into one input sample 611, 621, thereby allowing a reduction in the complexity (as measured by dimensionality) of radial-basis function non-linear neural network filter 601, therefore the complexity of non-linear neural network filter 600, without having to reduce the complexity of the inputs 111, 121 being filtered.


However, in this implementation, rather than being summed, the taps of delay line 612 are input directly to the hidden nodes 650 of radial-basis function non-linear neural network filter stage 601, which in this implementation are upstream of delay line 631.


Once again, with inputs 111, 121 from two sources, half 613 of delay line 612 of front-end filter 602 is devoted to input 111, while half 614 of delay line 612 of front-end filter 602 is devoted to input 121, with one respective hidden node 650 of radial-basis function non-linear neural network filter stage 601 for each input source 111, 121. The same is true of delay line 631 within radial-basis function non-linear neural network filter stage 601, with separate halves 632, 633 of delay line 631 devoted to inputs deriving separately from inputs 111, 121. Here too, the delays 631 form individual taps of a final FIR filter, which are combined at summation node 641 to yield the output Y.


A third implementation of a reduced-complexity non-linear neural network filter 700, shown in FIG. 7, is based on a multilayer perceptron (MLP) non-linear neural network filter 702, with a finite-impulse-response-(FIR)-based front-end filter 701.


Typically, an MLP filter includes a delay line for input samples, followed by at least one hidden layer in which the samples are summed and then passed through a non-linear activation function such as, e.g., a hyperbolic tangent function tanh (ƒ), followed by a layer including one or more summations.


In finite-impulse-response-(FIR)-based front-end filter 701, delay line 731 is divided into a first portion 732 receiving inputs 111 and a second portion 733 receiving inputs 121. Each line connecting a delay 731 to one of hidden nodes 750 represents a multiplication of a sample by a coefficient (not shown; see discussion above in connection with FIG. 5) forming a FIR filter tap. The taps are summed by the summation portion of each hidden node 750, which includes a summation function followed by a non-linear activation function which in this implementation is a tanh (ƒ) function. Although the hidden layer is shown as having only one hidden node 750 for all of the inputs in each respective set of inputs 111, 121, in other implementations (not shown) there may be multiple nodes 750 for each set of inputs 111, 121. In any event, a set of outputs 711 is generated based on front-end filtering of inputs 111, and another set of outputs 721 is generated based on front-end filtering of inputs 121.


In this implementation, the boundary between the front-end filter 701 and the MLP non-linear neural network filter 702 runs through the hidden layer of hidden nodes 750, but that is not necessarily the case in all implementations.


MLP non-linear neural network filter 702 in this implementation includes a respective tanh (ƒ) non-linear activation function as part of each respective one of hidden nodes 750 and a FIR filter formed by a delay line 712 and a summation node 722. A portion 751 of delay line 712 receives output samples 711 from front-end filter 702, while a portion 752 of delay line 712 receives output samples 721 from front-end filter 701. Each line connecting a delay 712 to sum 722 represents a multiplication of a sample by a coefficient (not shown; see discussion above in connection with FIG. 5) forming a FIR filter tap, and the taps are combined at summation node 722 to yield the output Y.


Reduced-complexity non-linear neural network filter 700 may be represented as an equivalent filter arrangement 800, shown in FIG. 8. Reduced-complexity non-linear neural network filter 800 includes four FIR filters 801, 802, 803, 804, and two non-linear activation functions 805, 806 (which may be respective tanh (ƒ) non-linear activation functions).


FIR filters 801, 802 form finite-impulse-response-(FIR)-based front-end filter 810, with FIR filter 801 receiving inputs 111 while FIR filter 802 receives inputs 121. FIR filters 803, 804 and non-linear activation functions 805, 806 form reduced-complexity non-linear neural network 820. In reduced-complexity non-linear neural network 820, activation function 805 receives the outputs of FIR filter 801 and passes those outputs, after non-linear activation, to FIR filter 803, while activation function 806 receives the outputs of FIR filter 802 and passes those outputs, after non-linear activation, to FIR filter 804. The outputs of FIR filter 803 and FIR filter 804 are combined at summation node 808 to yield the output Y.


Another implementation of a reduced-complexity non-linear neural network filter 900, shown in FIG. 9, also is based on a multilayer perceptron (MLP) non-linear neural network filter 902, with a finite-impulse-response-(FIR)-based front-end filter 901. In this implementation 900, finite-impulse-response-(FIR)-based front-end filter 901 includes two FIR filters 911, 921, each of which filters a respective set of inputs 111, 121. The respective outputs of FIR filters 911, 921 are combined by summation node 931.


The outputs 941 of finite-impulse-response-(FIR)-based front-end filter 901 are then filtered by multilayer perceptron (MLP) non-linear neural network filter 902, which includes a non-linear activation function 912 (which may be a tanh (_) non-linear activation function), followed by FIR filter 922.


In a variation 1000 of reduced-complexity non-linear neural network filter 900, shown in FIG. 10, a scalable bypass path 1001 is provided around non-linear neural network filter 902. Scalable bypass path 1001 is controlled by a scaling factor g (1011). FIR filter 922 inherently includes a similar scaling control. The provision of scalable bypass path 1001 allows several modes of operation. First, if g=0, reduced-complexity non-linear neural network filter 1000 operates identically to reduced-complexity non-linear neural network filter 900. Second, by setting g=1, and setting the scaling factor of FIR filter 922 to 0, reduced-complexity non-linear neural network filter 1000 operates as a linear filter. This linear mode may be used as a “jump start” mode while the non-linear portion of the filter is adapting.


In addition, a non-linear function 1100 (particularly one that is close to a linear function 1101) can be approximated as a series of linear functions 1102 of different slopes, as shown in FIG. 11. By varying g to vary the slopes, non-linear function 1100 can be filtered using mostly finite-impulse-response-(FIR)-based front-end filter 901, which is linear, with non-linear neural network filter 902 correcting for the difference between the segmented linear approximation and the actual non-linear function.


A similar variation 1200, based on reduced-complexity non-linear neural network filter 800, is shown in FIG. 12. A scalable bypass path 1201 is provided around non-linear neural network filter 820. Scalable bypass path 1201 is controlled by a scaling factor g (1211). FIR filters 803, 804 of non-linear neural network filter 820 inherently include a similar scaling control. By controlling g at 1211, non-linear neural network filter 1200 can be operated in various modes in a manner similar to non-linear neural network filter 1000.


It can be shown that the various implementations of a reduced-complexity non-linear neural network filter shown above provide nearly as good performance as a non-reduced-complexity non-linear neural network filter, particularly when adapted using cross-entropy. However, the reduced complexity provides substantial savings in device area and power consumption.


Although the implementations shown above receive two inputs (as in the case of a TDMR channel), implementations of the subject matter of this disclosure may include channels with only one input, or with three or more inputs. In such cases, the input delay lines may not be divided into groups (in the case of one input), or may be divided into three or more groups (in the case of three or more inputs), rather than being divided into two groups as shown), with each group receiving samples from one of the inputs.


A method 1300 according to implementations of the subject matter of this disclosure is diagrammed in FIG. 13.


Method 1300 begins at 1301 where non-linear equalization of digitized samples of input signals on the data channel is performed, including performing front-end filtering at 1311 to reduce numbers of the inputs from the digitized samples, performing non-linear filtering at 1321 on the reduced number of inputs from the digitized samples. At 1302, a respective value of each of the output signals is determined from output signals of the non-linear equalization. At 1303, parameters of the non-linear equalization are adapted based on respective ones of the value, and method 1300 ends.


Thus it is seen that a high-speed data channel using a reduced-complexity non-linear equalizer has been provided.


As used herein and in the claims which follow, the construction “one of A and B” shall mean “A or B.”


It is noted that the foregoing is only illustrative of the principles of the invention, and that the invention can be practiced by other than the described embodiments, which are presented for purposes of illustration and not of limitation, and the present invention is limited only by the claims which follow.

Claims
  • 1. A data channel on an integrated circuit device, the data channel comprising: a non-linear equalizer having as inputs digitized samples of signals on the data channel;decoding circuitry configured to determine from outputs of the non-linear equalizer a respective value of each of the signals; andadaptation circuitry configured to adapt parameters of the non-linear equalizer based on respective ones of the value; wherein:the non-linear equalizer includes:a front-end filter portion configured to combine some of the inputs from the digitized samples, to provide a reduced number of inputs; anda non-linear filter portion configured to operate on the reduced number of inputs.
  • 2. The data channel of claim 1 wherein the non-linear equalizer is a neural network equalizer.
  • 3. The data channel of claim 2 wherein the neural network equalizer is a multi-layer perceptron neural network equalizer.
  • 4. The data channel of claim 3 wherein the multi-layer perceptron neural network equalizer is a reduced complexity multi-layer perceptron neural network equalizer.
  • 5. The data channel of claim 2 wherein the neural network equalizer is a radial-basis function neural network equalizer.
  • 6. The data channel of claim 1 wherein the non-linear equalizer comprises a linear filter and a non-linear activation function.
  • 7. The data channel of claim 6 wherein the non-linear activation function is a hyperbolic tangent function.
  • 8. The data channel of claim 1 wherein the adaptation circuitry adapts parameters of the non-linear equalizer based on cross-entropy.
  • 9. The data channel of claim 1 wherein the front-end filter portion comprises a finite-impulse-response filter.
  • 10. The data channel of claim 1 further comprising scalable bypass circuitry for controllably outputting output of the front-end filter portion as at least a portion of output of the non-linear equalizer.
  • 11. A method for detecting data on a data channel on an integrated circuit device, the method comprising: performing non-linear equalization of digitized samples of input signals on the data channel;determining from output signals of the non-linear equalization a respective value of each of the output signals; andadapting parameters of the non-linear equalization based on respective ones of the value; wherein:performing non-linear equalization of digitized samples of input signals on the data channel includes:performing front-end filtering to combine some of the inputs from the digitized samples, to provide a reduced number of inputs; andperforming non-linear filtering on the reduced number of inputs from the digitized samples.
  • 12. The method of claim 11 wherein performing the non-linear equalization comprises performing neural network equalization.
  • 13. The method of claim 12 wherein performing the neural network equalization comprises applying a multi-layer perceptron neural network equalizer.
  • 14. The method of claim 13 wherein performing the neural network equalization comprises applying a reduced complexity multi-layer perceptron neural network equalizer.
  • 15. The method of claim 12 wherein performing the neural network equalization comprises applying a radial-basis function neural network equalizer.
  • 16. The method of claim 11 wherein performing the non-linear equalization comprises applying a non-linear activation function and performing linear filtering on output of the non-linear activation function.
  • 17. The method of claim 16 wherein applying the non-linear activation function comprises applying a hyperbolic tangent function.
  • 18. The method of claim 11 wherein adapting the parameters of the non-linear equalization comprises adapting the parameters of the non-linear equalization based on cross-entropy.
  • 19. The method of claim 11 wherein the performing front-end filtering comprises performing finite-impulse-response filtering.
  • 20. The method of claim 11 further comprising scalably bypassing the non-linear filtering for controllably outputting an output of the front-end filtering as at least a portion of output of the non-linear equalization.
CROSS REFERENCE TO RELATED APPLICATION

This disclosure claims the benefit of copending, commonly-assigned U.S. Provisional Patent Application No. 63/147,106, filed Feb. 8, 2021, which is hereby incorporated by reference herein in its entirety.

US Referenced Citations (112)
Number Name Date Kind
5963929 Lo Oct 1999 A
5991308 Fuhrmann Nov 1999 A
6052349 Okamoto Apr 2000 A
6158271 de Corral Dec 2000 A
6236749 Satonaka May 2001 B1
6246962 Schultz Jun 2001 B1
6304539 Okamoto Oct 2001 B1
6307868 Rakib Oct 2001 B1
6377640 Trans Apr 2002 B2
6381083 Abarbanel Apr 2002 B1
6434084 Schultz et al. Aug 2002 B1
6665308 Rakib Dec 2003 B1
6687235 Chu Feb 2004 B1
6831900 Blake Dec 2004 B2
6937617 Rakib Aug 2005 B2
6983047 Chadha Jan 2006 B1
7016440 Singer Mar 2006 B1
7020165 Rakib Mar 2006 B2
7031344 Rakib Apr 2006 B2
7095707 Rakib Aug 2006 B2
7158566 Dowling Jan 2007 B2
7239650 Rakib Jul 2007 B2
7454684 Allpress Nov 2008 B2
7764622 El-Damhougy Jul 2010 B2
7885031 Han Feb 2011 B1
8046200 Kirby Oct 2011 B2
8107329 Painter Jan 2012 B1
8363517 Painter Jan 2013 B1
8416659 Xie Apr 2013 B1
8521488 Kirby Aug 2013 B2
8611411 Hayami Dec 2013 B2
8634284 Painter Jan 2014 B1
8896682 Bressler Nov 2014 B2
9484974 Tu Nov 2016 B2
9754163 Segalovitz Sep 2017 B2
9928824 Barath Mar 2018 B2
10038575 Steffan et al. Jul 2018 B1
10147442 Panchapagesan Dec 2018 B1
10388272 Thomson Aug 2019 B1
10475471 Ebenezer Nov 2019 B2
10531415 O'Shea Jan 2020 B2
10573312 Thomson Feb 2020 B1
10672383 Thomson Jun 2020 B1
10797805 Mirfakhraei et al. Oct 2020 B1
10833785 O'Shea et al. Nov 2020 B1
10847137 Mandal Nov 2020 B1
10971153 Thomson Apr 2021 B2
10979097 Luo Apr 2021 B2
10985951 Li Apr 2021 B2
11017778 Thomson May 2021 B1
11050494 Baek et al. Jun 2021 B2
11055615 Litichever Jul 2021 B2
11074495 Zadeh Jul 2021 B2
11074925 Bryan Jul 2021 B2
11145312 Thomson Oct 2021 B2
11145331 Nangare Oct 2021 B1
11170761 Thomson Nov 2021 B2
11195057 Zadeh Dec 2021 B2
11240579 Jumbe Feb 2022 B2
11255663 Binder Feb 2022 B2
11270200 Zhang et al. Mar 2022 B2
11271699 Eyuboglu Mar 2022 B1
11310084 Shen Apr 2022 B2
11356182 Ye Jun 2022 B2
11438014 Belzer Sep 2022 B2
11451419 Li Sep 2022 B2
11495248 Musha Nov 2022 B2
11496341 Luo Nov 2022 B2
11509509 Razavi Majomard Nov 2022 B2
11516053 Ye Nov 2022 B2
11568214 Tullberg Jan 2023 B2
11570023 Nangare Jan 2023 B2
11575544 Andrews Feb 2023 B2
11677595 Razavi Majomard Jun 2023 B2
20010034216 Creigh et al. Oct 2001 A1
20020150059 Blake Oct 2002 A1
20020181633 Trans Dec 2002 A1
20040260995 Allpress Dec 2004 A1
20050220185 Dowling Oct 2005 A1
20060039550 Chadha Feb 2006 A1
20080319375 Hardy Dec 2008 A1
20140146867 Shvydun et al. May 2014 A1
20180102136 Ebenezer Apr 2018 A1
20180115824 Cassidy Apr 2018 A1
20180204111 Zadeh Jul 2018 A1
20190154439 Binder May 2019 A1
20190365342 Ghaffarzadegan Dec 2019 A1
20190385057 Litichever Dec 2019 A1
20200184278 Zadeh Jun 2020 A1
20200250497 Peng Aug 2020 A1
20200293894 Kwon Sep 2020 A1
20200294401 Kerecsen Sep 2020 A1
20200295975 Li et al. Sep 2020 A1
20200389469 Litichever Dec 2020 A1
20210091979 Ye Mar 2021 A1
20210142158 Agrawal May 2021 A1
20210218606 Shen Jul 2021 A1
20210345939 Jumbe Nov 2021 A1
20210350237 Litichever Nov 2021 A1
20210399927 Razavi Majomard Dec 2021 A1
20220014401 Zhu et al. Jan 2022 A1
20220014404 Hossain Jan 2022 A1
20220128352 Binder Apr 2022 A1
20220131724 Razavi Majomard Apr 2022 A1
20220150094 Nangare May 2022 A1
20220224361 Alic et al. Jul 2022 A1
20220239510 Nangare Jul 2022 A1
20220385374 Arikawa Dec 2022 A1
20230021633 Li Jan 2023 A1
20230216724 Sundberg Jul 2023 A1
20230275787 Vitthaladevuni Aug 2023 A1
20240064045 Dufrene Feb 2024 A1
Foreign Referenced Citations (175)
Number Date Country
2004231907 Nov 2004 AU
2004231907 Dec 2009 AU
2020101786 Sep 2020 AU
2020101786 Sep 2020 AU
116054888 May 2003 CN
1485798 Mar 2004 CN
1485798 Mar 2004 CN
107342814 Nov 2017 CN
107342814 Nov 2017 CN
108363967 Aug 2018 CN
108363967 Aug 2018 CN
108399366 Aug 2018 CN
108399366 Aug 2018 CN
108650201 Oct 2018 CN
109379120 Feb 2019 CN
109379120 Feb 2019 CN
109753978 May 2019 CN
109995449 Jul 2019 CN
110198282 Sep 2019 CN
110211611 Sep 2019 CN
110233812 Sep 2019 CN
110326308 Oct 2019 CN
110326308 Oct 2019 CN
110365613 Oct 2019 CN
110365613 Oct 2019 CN
110392006 Oct 2019 CN
110399929 Nov 2019 CN
110446160 Nov 2019 CN
110446160 Nov 2019 CN
110533027 Dec 2019 CN
110547824 Dec 2019 CN
110547824 Dec 2019 CN
110598859 Dec 2019 CN
110598859 Dec 2019 CN
108683428 Apr 2020 CN
108683428 Apr 2020 CN
111191573 May 2020 CN
111191573 May 2020 CN
110198282 Jun 2020 CN
110211611 Jun 2020 CN
111630787 Sep 2020 CN
111630787 Sep 2020 CN
111683025 Sep 2020 CN
111683025 Sep 2020 CN
111740934 Oct 2020 CN
108650201 Nov 2020 CN
109995449 Dec 2020 CN
110636020 Jan 2021 CN
110636020 Jan 2021 CN
110197218 Feb 2021 CN
111565160 Mar 2021 CN
111565160 Mar 2021 CN
112532548 Mar 2021 CN
112532548 Mar 2021 CN
110808932 Apr 2021 CN
110325929 May 2021 CN
110325929 May 2021 CN
109067688 Sep 2021 CN
109067688 Sep 2021 CN
110392006 Jan 2022 CN
111488734 Feb 2022 CN
111488734 Feb 2022 CN
110505020 Mar 2022 CN
114342272 Apr 2022 CN
114342272 Apr 2022 CN
114664288 Jun 2022 CN
114664288 Jun 2022 CN
111740934 Aug 2022 CN
110533027 Sep 2022 CN
111565051 Nov 2022 CN
111565051 Nov 2022 CN
111630787 Dec 2022 CN
109753978 Feb 2023 CN
111786915 Mar 2023 CN
111786915 Mar 2023 CN
110399929 Apr 2023 CN
116054888 May 2023 CN
110741553 Nov 2023 CN
110741553 Nov 2023 CN
112532548 Feb 2024 CN
69620781 May 2007 DE
69620781 May 2007 DE
112021005915 Aug 2023 DE
112021005915 Aug 2023 DE
1107597 Jun 2001 EP
1107597 Jun 2001 EP
1107598 Jun 2001 EP
1107598 Jun 2001 EP
1107599 Jun 2001 EP
1130918 Sep 2001 EP
1130918 Sep 2001 EP
1130919 Sep 2001 EP
1130919 Sep 2001 EP
1187360 Mar 2002 EP
1107599 Mar 2006 EP
2 148 446 Jan 2010 EP
1107598 Jul 2010 EP
3535625 Feb 2021 EP
3993275 May 2022 EP
4088385 Nov 2022 EP
3673580 Oct 2023 EP
3673580 Oct 2023 EP
4270381 Nov 2023 EP
4270381 Nov 2023 EP
3118519 Jul 2022 FR
3118519 Jul 2022 FR
2557425 Jun 2018 GB
2557425 Feb 2020 GB
10106158 Apr 1998 JP
2002141829 May 2002 JP
3507080 Mar 2004 JP
3507080 Mar 2004 JP
3764269 Apr 2006 JP
3764269 Apr 2006 JP
4314553 Aug 2009 JP
2019-530113 Oct 2019 JP
2019530113 Oct 2019 JP
2009109380 Oct 2009 KR
WO 199708861 Mar 1997 WO
WO-9708861 Mar 1997 WO
WO 2004095711 Nov 2004 WO
WO-2004095711 Nov 2004 WO
WO 2008157422 Dec 2008 WO
WO-2008157422 Dec 2008 WO
WO-2020117504 Jun 2011 WO
WO-2015198271 Dec 2015 WO
WO 2016040583 Mar 2016 WO
WO 2016040590 Mar 2016 WO
WO 2016040644 Mar 2016 WO
WO-2016040583 Mar 2016 WO
WO-2016040590 Mar 2016 WO
WO-2016040644 Mar 2016 WO
WO 2016108166 Jul 2016 WO
WO-2016108166 Jul 2016 WO
WO 2017149526 Sep 2017 WO
WO-2017149526 Oct 2017 WO
WO 2018071387 Apr 2018 WO
WO-2018071387 Apr 2018 WO
WO 2018104929 Jun 2018 WO
WO-2018104929 Jun 2018 WO
WO-2019038693 Feb 2019 WO
WO 2019043446 Mar 2019 WO
WO-2019043446 Mar 2019 WO
WO 2019082239 May 2019 WO
WO 2019165969 Sep 2019 WO
WO 2019169400 Sep 2019 WO
WO-2019043466 Sep 2019 WO
WO-2019082239 Sep 2019 WO
WO-2019165969 Sep 2019 WO
WO-2019169400 Sep 2019 WO
WO-2019191099 Oct 2019 WO
WO-2019242534 Dec 2019 WO
WO 2020092391 May 2020 WO
WO-2020092391 May 2020 WO
WO 2020117504 Jun 2020 WO
WO 2020117505 Jun 2020 WO
WO 2020117506 Jun 2020 WO
WO 2020117507 Jun 2020 WO
WO-2020117505 Jun 2020 WO
WO-2020117506 Jun 2020 WO
WO-2020117507 Jun 2020 WO
WO 2021045862 Mar 2021 WO
WO-2021045862 Mar 2021 WO
WO 2021142189 Jul 2021 WO
WO-2021142189 Jul 2021 WO
WO 2021203242 Oct 2021 WO
WO-2021203242 Oct 2021 WO
WO 2021262052 Dec 2021 WO
WO-2021262052 Dec 2021 WO
WO 2021262811 Dec 2021 WO
WO-2021262811 Apr 2022 WO
WO 2022087388 Apr 2022 WO
WO-2022087388 Apr 2022 WO
WO-2022103422 May 2022 WO
WO-2022159870 Jul 2022 WO
Non-Patent Literature Citations (9)
Entry
Z. Zerdoumi, D. Chikouche and D. Benatia, “Adaptive decision feedback equalizer based neural network for nonlinear channels,” 3rd International Conference on Systems and Control, Algiers, Algeria, 2013, pp. 850-855, doi: 10.1109/ICoSC.2013.6750956. (Year: 2013).
J. K. Satapathy, K. R. Subhashini and G. Lalitha Manohar, “A highly efficient channel equalizer for digital communication system in Neural Network paradigm,” 2009 Innovative Technologies in Intelligent Systems and Industrial Applications, Kuala Lumpur, Malaysia, 2009, pp. 11-16, (Year: 2009).
Sheeja, K.L., et al., “Decision Feedback Equalization Using RBF and MLP Networks,” 2010 Second Vaagdevi International Conference on Information Technology for Real World Problems (VCON), IEEE, pp. 96-101 (Dec. 9, 2010).
Zerdoumi, Z., et al., “Adaptive Decision Feedback Equalizer Based Neural Network for Nonlinear Channels,” Proceedings of the 3rd International Conference on Systems and Control, Oct. 29-31, 2013 (6 pages).
Zhou, Qingyi, et al., “AdaNN: Adaptive Neural Network-based Equalizer via Online Semi-supervised Learning,” ARXIV.Org, Cornell University Library, Jul. 24, 2019 (10 pages).
Shen, J., et al., “Nonlinear Equalization for TDMR Channels Using Neural Networks”, 2020 54th Annual Conference on Information Sciences and Systems (CISS), IEEE, pp. 1-6 (Mar. 18, 2020).
Santamaria, I., et al., “Entropy Minimization for Supervised Digital Communications Channel Equalization”, IEEE Transactions on Signal Processing, 50(5):1184-92 (May 2002).
Chen, S., et al., “A Clustering Technique for Digital Communications Channel Equalization Using Radial Basis Function Networks”, IEEE Transactions on Neural Networks, 4(4):570-90 (Jul. 1993).
Gonzalez-Serrano, F.J., et al., “Reduced-Complexity Decision-Feedback Equalizer for Nonlinear Channels”, 9th European Signal Processing Conference (EUSIPCO), pp. 1-4, (1998).
Provisional Applications (1)
Number Date Country
63147106 Feb 2021 US