The present invention relates generally to methods and apparatus for storage and retrieval of information on memory units and more particularly to structured storage.
Memory storage is a constantly advancing field, as there is continued demand for larger memory capacities, at lower cost, lower power consumption per unit of storage and faster storage and retrieval speeds. There are different technologies that are in common use.
Storage of digital data as capacitive charges in a dynamic memory structure is well represented in the literature. Optimized logic design, innovative physical structures, and targeted semiconductor process improvements have been applied to create high density, high performance dynamic memory components of significant performance, capacity, and economy. Thus, the particular semiconductor technology used for large memory chips, such as dynamic random access memory (DRAM) chips, has been the topic of much research and DRAM technologies have been highly refined.
However, there is also a demand for comparatively small amounts of memory within, for example, a custom or semi-custom semiconductor logic device that uses a different technology. As such devices are based on different semiconductor processes and layout rules than are used to create DRAM devices, few of the optimizations developed for dedicated DRAMs are applicable to these embedded memory designs. Instead, the memory portions of such devices typically rely on planar capacitive structures of relatively low capacitance and relatively large area. Use of such relatively low capacitance, non-optimized cells can result in reduced signal output on a read operation, since readout of the stored charge is divided between the lower cell capacitance and the relatively fixed sense line parasitic capacitance. Mitigation of this effect requires the use of larger storage elements, significantly increasing the semiconductor area required for the embedded storage array.
As a result, improvements in structural design of memory would be useful.
In embodiments of a memory circuit according to aspects of the present invention, an embedded DRAM array stores information as groups of bits or data using information coding in the storage and retrieval of stored data, instead of each bit being stored separately.
In some embodiments, data is organized into data words, wherein each data word is representable as a vector having a plurality of components. In specific embodiments, the components are bits, in that they have one of two possible values. A data word that is to be stored into memory is mapped from a write data word to a storage format word. The mapping can be a mapping defined by a Hadamard matrix. The storage format word is stored as charge levels in an addressable location of the memory. An addressable location of the memory is a selectable subset of the memory is a plurality of storage cells wherein a plurality of bits of information can be stored.
For retrieving stored data, an addressable location is selected, charge levels are read from the plurality of storage cells activated when the addressable location is selected, and the charge levels are interpreted to a valid storage format word. The interpretation can be according to a maximal likelihood decoding according to a Hadamard code to derive a read data word corresponding to a previously written write data word. The write data word is then output as the result of a read of the selected addressable location, or a portion thereof.
In various embodiments, the characteristics of groups of bits are written and read collectively and optimized through the application of information coding techniques.
In a specific embodiment, an input word of B bits is mapped to an expanded word of S bits (S>B) for storage in a dynamic memory array as stored charge on S cells. On subsequent readout, the charge levels of the S cells are interpreted by an inverse mapping component that the memory device uses to reconstruct the original B bit values from the S possibly degraded signals.
In some embodiments, the mapping is a Hadamard matrix and in some embodiments, the mapping is two or more Hadamard matrix mappings concatenated for each of a plurality of storage format words.
The addressable location can be a row selection of a two-dimensional array of storage cells, such that the storage cells that collectively store a storage format word have a row selection in common. The maximal likelihood decoding can be a Fast Walsh Transform. The Hadamard code used for mapping can correspond to the Walsh codes of individual rows in a Hadamard matrix or a superposition of a plurality of Walsh codes of individual rows in a Hadamard matrix. Mapping can incorporate a distribution operation that spreads the write data word across multiple Hadamard encodings comprising the larger storage format word, and reading includes combining the result of multiple Hadamard decodings to produce the read data word. The write data word and/or the storage format word can incorporate additional error detection or error correction bits.
The following detailed description together with the accompanying drawings will provide a better understanding of the nature and advantages of the present invention.
In embodiments of a memory circuit described herein, data is stored into memory, such as an embedded dynamic random access memory (“DRAM”) device. Since the device is random access, it has a plurality of addressable storage locations and thus a write operation involves providing the device with a selection of an addressable storage location and write data, whereas a read operation involves providing the device with a selection of an addressable storage location and obtaining therefrom the read data.
Because this DRAM read operation has drained charge from the DRAM array storage capacitors 106, a refresh operation must be performed to restore the stored data. This is done by enabling write data column drivers 110 after the sense amplifiers 108 produce their result. This drives the column data 109 back onto the column sense lines 107, recharging the DRAM array storage capacitors 106. That operation complete, the row selection line 104 is disabled, returning all cell select transistors 105 to an off state and again trapping charge in the capacitive storage elements 106.
A DRAM write cycle is performed similarly to the described read cycle, except that external write data is caused to flow from data line 113 back through the column data selection multiplexor 112 to force the sense amplifier 108 state to the desired value, causing the new data to be written to the DRAM cells as part of the concluding refresh operation. As stored charge is subject to leakage currents, it is necessary to periodically refresh (that is, perform the described read/restore operation on) each cell in the array before its charge decays to an unreadable level.
Common embodiment variations from the example of
In an ideal DRAM design, the capacitance of each storage element is identical and relatively large compared to the parasitic capacitance of the row line and sense amplifier input, and the capacitive leakage current is extremely small. These characteristics translate into a high and uniform signal level when a charged cell is read, and a long retention time between necessary refresh operations. Dedicated DRAM devices are fabricated using semiconductor processes optimized for these characteristics, capable of creating compact, high capacitance storage cells with extremely low leakage characteristics. Such process characteristics are not generally available for dynamic memory designs that are embedded within a custom or semi-custom digital logic device, as the processes used there are typically optimized for logic speed and density rather than low leakage and the ability to fabricate deep trench structures as are used to create optimized DRAM capacitive cells. Instead, designers of embedded dynamic memory arrays typically adapt available design library elements to this purpose, such as using the parasitic capacitance of a standard transistor element as a storage capacitor. Obtaining the necessary storage capacitance with such a structure requires significant planar area, so that the overall space required for the memory system is constrained by the total size of the storage array elements, rather than by minimum feature size or array signal line density as in a dedicated DRAM device.
In the example of
In novel approaches described herein, when a memory device is presented with an address and data to be written, the data to be written is formed into a write data word, the memory device (such as a memory portion embedded in a circuit or chip that has other processing functions) maps the write data word into a storage format word and stores the storage format word. The write data word and the storage format word are both representable as vectors, i.e., they have a plurality of components, typically those components being bits, wherein a bit is a memory, data and/or information element that can have one of two possible logical values with physical representations that may be approximations of their logical value. Thus, in a specific embodiment, the write data word is a multi-bit value and the storage format word is also a multi-bit value. Herein, the length of the write data word is represented by an integer B, whereas the length of the storage format word is represented by an integer S. Where the length B write data word can take on up to 2B different possibilities, S will be greater than B. As explained below, the S bits of a storage format word are not independent, but are defined by a codeword mapping between write data words and storage format words.
Whereas storing data involves providing the memory device with an address and a write data word, which is converted to a length S storage format word that is then stored into S memory cells, reading involves retrieving, according to the selected addressable location, charge levels from the S storage cells activated when the addressable location is selected, determining the length S storage format word and inverse mapping, preferably using a maximal likelihood decoding according to a Hadamard code to derive a read data word corresponding to a previously written write data word. The write data word is then output as the result of a read of the selected addressable location, or a portion thereof.
Of course, it may be that fewer than B bits are written at a given write operation or fewer than B bits are read. In such cases, it may be that the memory device fills in some of the bits during a write with bits of the write data word that was mapped and stored (which may involve an inverse mapping to figure out those bits) and for reading, some of the steps are skipped for some bits or some of the resulting read data word are discarded before outputting fewer than B bits. In various embodiments, the characteristics of groups of bits are written and read collectively and optimized through the application of information coding techniques. In a specific embodiment, an input word of B bits is mapped to an expanded word of S bits (S>B) for storage in a dynamic memory array as stored charge on S cells. On subsequent readout, the charge levels of the S cells are interpreted by an inverse mapping component that the memory device uses to reconstruct the original B bit values from the S possibly degraded signals.
As explained in more detail elsewhere herein, a unique approach to memory design can improve embedded DRAM arrays. Rather than attempting to improve the signal characteristics of a single bit, e.g., through increased storage capacitance, the characteristics of groups of bits written and read collectively are optimized through the application of information coding techniques. Thus, an input word of B bits is mapped to an expanded word of S bits for storage in a dynamic memory array as stored charge on S cells. On subsequent readout, the charge levels of the S cells are interpreted by an inverse mapping component that reconstructs the original B bit value from the S degraded signals. For practical memory array capacities and word lengths, the increased array density available through reduction in the size of individual storage elements needed to obtain the same system bit error rate with use of the described coding technique more than compensates for the added storage elements required.
In one specific example of a memory device, storage is constructed such that the storage format words are S=N−1 bits, where N is an integer power of two, the mapping is according to a size N Hadamard matrix size, and the write data word size is B=log2(N)−1 bits.
Hadamard matrices of size 2N+4 are also known, and are equally applicable in embodiments based on the described method. For example, a 20×20 Hadamard matrix created using Paley's construction may be used to store four bits of data in a storage group of 20 cells.
To encode a data value, K, in the range 0≦K<2B, a numerical vector v is created of length N with the value “1” in the first position (first component), the value “−1” in the K+1-th position, and all other positions zero. For purposes of illustration, the values N=8 and B=2 will be used, allowing possible data values of {0, 1, 2, 3}, which are encoded as the numeric vectors as indicated in Table 2. It should be apparent upon reading this disclosure how to use other values for N and B.
This is illustrated in
Given the described construction of the numerical vector, the initial bit of the resulting binary results sequence will always be zero, and thus need not be stored but instead may be represented by the fixed value “0” on readout. The remaining seven bits of the sequence are stored in a group of seven storage cells, preferably accessed simultaneously and in parallel, such as seven array elements selected by the same array row address. For example, the sequence 333 will have its initial zero suppressed and may then be stored in seven storage cells as the sequence (1, 1, 0, 0, 1, 1, 0) to represent the data value “2” as encoded above.
In general, where a value K with 0≦K<2B is chosen, a vector v is created that has a “+1” at its first position, a “−1” at position K+1, and zeroes everywhere else, and the vector w=½*HN+1*v is formed that consists of zeroes and ones only. The entries of the vector w, except possibly the first which is always zero, can be stored on N−1 cells. In general, it is also possible to use values of K between 0 and N−2, leading to N−1 choices for K. In that case, a mapping is needed that maps bits to the N−1 possible values of K. Some techniques to this end are described herein below.
Decoding Data Values from Readout Charge Levels
On subsequent readout, the signals obtained from the same storage group of cells will be degraded by cell leakage and noise. Continuing the previous example, a vector v of relative values may be obtained by taking the fixed value of “0” (described above) in the first position, followed by the values obtained by reading the seven (or more generally, N−1) storage cells written with the previous sequence. An example result is shown as 334, with individual values representing both leakage decay of stored “1” values, which may vary from cell to cell, and noise-induced variation of all signal levels.
The Hadamard transform of this vector 334 is taken at 335, producing the result 336. It may be noted that the overall form of this result vector 336 is the same as that used to generate the numerical vector 330 used for storage; that is, a significantly positive value, followed by some arrangement of one significantly negative value 337 and six values near zero. In the ideal case having neither signal level degradation nor noise, the sequence read from the storage cells would be a series of “1” and “0” values, and the result vector obtained after the matrix multiplication would exactly match the original numeric vector used in the encoding process.
Given knowledge of this expected form, conventional techniques may be applied to identify which of the possible variations of that form (i.e., which of the possible numeric vectors that could have been used for the storage operation) is the greatest likelihood match of 336 to one of the possible numeric vectors which could have produced it. One embodiment performs this match by finding the most negative value in 336, here −0.366 in the fourth position 337, which corresponds to the encoding method's representation of the value K=2 as a “4” value 331 in the fourth position of the original numerical vector 330 of this example.
Some embodiments may use an equivalent operation to the Hadamard transform, such as a Fast Walsh transform. Other embodiments may incorporate additional consistency checks of the readout transformed sequence to further verify it is of the same overall form expected for an encoded sequence, such as having its most positive value in the first position, that the values approximating zero have amplitudes below a set threshold, etc.
Storage Cell Scaling Reduction of Array Size
The above example utilizes seven storage cells to encode a two bit value, which appears to be a significant storage inefficiency. Indeed, for higher values of N, this apparent cells/bit inefficiency only appears to grow. However, systems with larger matrix size (that is, larger N) also display significantly improved decoding gain, in other words an ability to tolerate lower signal to noise ratios on readout.
A study and error analysis of the overall system using a second moment analysis over random perturbations showed that the necessary array capacitance (or equivalently, the required area of the array storage cell) may be scaled down by a factor of around 2/N while maintaining the same overall bit error rate. Depending on the topology used for detection of the signals, actual scale downs of between 2/N and 4/N may be suitable. Including this scale improvement factor, the overall array size reduction obtainable over a conventional one bit/cell array for some example matrix sizes is as shown in Table 4.
Scaling of array storage elements will affect the required refresh rate if standard sense amplifier signal level refresh is used. Some embodiments instead implement refresh at the data bit level rather than the array bit level, performing the full Hadamard decode and encode for each refresh cycle. Other embodiments may simplify that behavior by decoding only to the point of determining which matrix row (e.g., Walsh function) best matches the readout data pattern, followed by a write back of that row's Walsh function bit pattern. Embodiments utilizing more complex data encoding patterns may extend that method to a best match of superpositions of Walsh functions.
Integration of Storage Group Coding into a DRAM Subsystem
Concurrently, data to be written 449 is encoded by a write data mapping operation 450 into a vector of encoded values, which undergoes a Hadamard encoding operation 451 to create a Hadamard encoded storage word 452 that can be placed on a set of wires or a bus as illustrated. It should be understood that the “operation” elements shown in
The individual values of the Hadamard encoded storage word 452 are driven by encoded write data drivers 453 onto the column sense lines 447, creating a pattern of charged cells in the selected DRAM element storage capacitors 446. All row selection lines 444 are then disabled, trapping the stored charges in the array cells.
Subsequently, the same row address 442 is presented to row address decoder 443, enabling one of the row selection lines 444. When enabled by the selected row selection line 444, each of the attached DRAM element pass transistors 445 for that row connects its associated DRAM element storage capacitor 446 to a column sense line 447. The charge stored on the DRAM element storage capacitor 446 presents a voltage on the column sense line 447, which is detected by the sense amplifier 448. The detected signals pass through a Hadamard decoder 454 (as might be implemented to match the logic described above for inverse mapping) and the results provided to a read data decoder 455 that then performs greatest likelihood matching and outputs results as a decoded read data word 456. Finally, information from the read operation may be returned along a channel 455 to a Hadamard encoder operation 451 to produce Hadamard encoded storage format words to be written back through encoded write data drivers 453 to DRAM array 441.
In some embodiments, Hadamard encoding operation 451 is implemented as a circuit that performs a Hadamard transform, i.e., it transforms its input according to a Hadamard transform, such as one described herein, and provides the result as its output. In a specific circuit implementation, a resistor network encodes for the Hadamard transform and the transform can be performed quickly and without undue overhead. In some embodiments, Hadamard decoding operation 545 is also done with a resistor network and decoder 455 is a circuit that performs peak detection to find the position of a vector that has the lowest entry.
Improved Storage Encoding Efficiency
A practical storage system is often desired, to handle data significantly wider than the two bit data width of the previous example. For example, the size of the write data words might be bytes (8 bits) or words (16 bits) rather than two bits. In one approach that addresses this need, such systems incorporate multiple instances of the mapper. For example, four instances of the N=8/B=2 Hadamard mapper might be used in parallel, each handling two bits to support an overall storage interface width of eight bits.
Simply paralleling multiple embodiments in this fashion does not provide the most efficient utilization of storage resources and further improvements are possible. Consider the previous N=8 encoding example. Additional encoded values are possible within the described numerical vector format. For example, a numerical vector formatted as a single “1” followed by some combination of a single “−1” and six zeroes can encode seven different values, where the example only employed four of those. To take advantage of these additional states in a multiple-instance system, an additional information spreading procedure may be applied.
As a simple example, two N=8 subsystems, N1 and N2, as described in the earlier example can actually represent 7*7=49 distinct sets of states, which is more than the 32 states needed to represent any five bit value. But, taken individually, the subsystems N1 and N2 cannot take advantage of those additional states, as by itself neither can represent enough states to encode an additional bit.
Consider now that a five bit vector, V=V3, V2, V1, V0), is to be mapped to the states of the two N=8 subsystems, N1 and N2. One simple information spreading procedure to map the five bit binary value of V to the states 0 to 5 on each of N1 and N2 uses the following procedure:
1) If V4=0, encode states N1={0, 1, 2, 3} and N2={0, 1, 2, 3} as:
N1=2*V4+V3, and N2=2*V2+V1
If V4=1 and V3=0, encode states N1={0, 1, 2, 3} and N2={4, 5} as
N1=2*V2+V1, and N2=4+V0
If V4=1 and V3=1, encode states N1={4, 5} and N2={0, 1, 2, 3}
N1=4+V0, and N2=2*V2±V1
As previously described, the resulting states N1 and N2 are then encoded as numeric vectors ranging from 0=(1, −1, 0, 0, 0, 0, 0, 0) to 5=(1, 0, 0, 0, 0, 0, −1, 0) and those numeric vectors are each separately multiplied by the Hadamard matrix H3 and the result scaled by ½ to obtain the possible binary results to be stored in the two seven-bit storage groups of cells. Extending this example, a Boolean logic layer or lookup table can be used to map an 8 bit value to three 1-of-7 state values (since 7*7*7=343 and 343>256, so all eight bit numbers can be distinctly represented) which are then encoded for storage in three N=8 subsystems.
On readout, the voltage levels obtained from the seven-bit storage groups of cells are decoded as previously described to obtain state values, which are then applied to an inverse lookup table or Boolean logic layer to reconstruct the original binary value (five bits from two storage groups of cells in the first example, eight bits from three storage groups of cells in the extended example.)
The particular mapping processes, circuits and/or structures used to first spread and later combine the data provide a unique mapping between input value to state values to output value, and should do so with relatively uniform distribution over the available set of state values, but otherwise may be chosen based on implementation convenience. On encoding, an embodiment may combine the mapping or information spreading and the encoding operations, directly producing multiple Hadamard-coded storage values based on an input value. Similarly, an embodiment of the decoding operation may combine Hadamard decoding and the inverse mapping or combining action in a single operation, or may selectively implement some portions as parallel operations.
With this enhanced encoding, the delivered storage density for some example stored data widths and matrix complexities are as shown in Table 5.
As may be seen from this disclosure, even a fairly simple implementation based on 8×8 Hadamard matrices or the equivalent Fast Walsh-Hadamard transform may require as little as 66% of the semiconductor area of an embedded dynamic memory subsystem designed using conventional techniques.
More Complex Encoding Methods
The previous examples utilize an encoding of data bits in a numeric vector which, when transformed by the Hadamard matrix, produces a binary storage pattern corresponding to a single Walsh function. This selection of patterns provides certain benefits with respect to high correlation within the decoding process and a result vector having a simple to analyze form. However, higher storage efficiency might sometimes be desired, using fewer storage bits to store a given number of data bits.
Other tradeoffs are possible; one embodiment utilizes numeric vectors of the form “2” followed by a sequence containing two “−1” values and all other values zero. When multiplied by the Hadamard matrix in the encoding process, and divided by 4, such a pattern produces a result that is the equivalent of the additive superposition of two Walsh functions, rather than the binary representation of a single Walsh function of the earlier example. That is, a range of charge levels is stored in particular cells within the selected storage group to represent the encoded sequence, rather than the binary selection of a single charge level or no charge per cell.
On decoding, analysis of the result vector obtained after the Hadamard matrix operation is similarly interpreted relative to this more complex expected pattern having significantly negative values in two positions. The additional degree of freedom allows significantly more states to be encoded, thus permitting higher storage efficiency at the cost of more complex processing functions and somewhat lower noise immunity.
The Hadamard transform of this vector 634 is taken at 635, producing the result 636. It may be noted that the overall form of this result vector 636 is the same as that used to generate the numeric vector 630 used for storage; that is, a significantly positive value, followed by some arrangement of two significantly negative values 637 and five values near zero. These two values could be determined using circuitry that can detect the positions of the two smallest entries in the vector.
Further embodiments extend this concept to numeric vectors of the form (w, . . . ) where w is the weight of the vector, and “ . . . ” is a sequence of entries containing w values of −1 and the remainder values of 0. Such embodiments identify the w most negative values in the result vector after the Hadamard transform within the decoding process.
In the examples described herein, Hadamard matrices are used to form the basis for an encoder and decoder. In other embodiments, other orthogonal matrices might be used instead, however initial analysis may indicate that Hadamard matrices are preferred.
Hybrid Error Detection and Correction
The described methods may also be used in combination with known techniques for error correction and error identification, including single and multi-dimensional parity codes, low density parity codes, and forward error correction codes including Reed-Solomon, Golay, BCH, and Hamming code. These techniques, collectively called ECC, may be applied over blocks or words of data presented to the external interface of the storage subsystem, or to storage code groups written and read from sets of physical memory cells. One example embodiment computes ECC across seven bit storage code groups and stores the storage code groups with the computed ECC.
As has now been explained, systems and methods that combine Hadamard coding of storage formatted words with a dynamic memory storage array provide improved signal integrity and error reduction. The systems and methods are applicable to embodiments that are constrained by the planar area required for each capacitive component of the dynamic memory array, as the improved read detection ability can compensate for reduced cell capacitance, permitting significant array size reduction. Further embodiments are described that aggregate multiple instances of the described system and method into larger configurations, and which combine those methods with other error detection and error correction methods.
This application is a continuation of U.S. application Ser. No. 15/051,344, filed Feb. 23, 2016, which is a continuation of U.S. application Ser. No. 13/844,032, filed Mar. 15, 2013, which claims the benefit of U.S. Provisional Patent Application No. 61/646,790, filed May 14, 2012, entitled “STORAGE METHOD AND APPARATUS FOR RANDOM ACCESS MEMORY USING CODEWORD STORAGE”, the entire disclosure of which is incorporated by reference herein for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
3196351 | Slepian | Jul 1965 | A |
3636463 | Ongkiehong | Jan 1972 | A |
3939468 | Mastin | Feb 1976 | A |
4163258 | Ebihara | Jul 1979 | A |
4181967 | Nash | Jan 1980 | A |
4206316 | Burnsweig | Jun 1980 | A |
4276543 | Miller | Jun 1981 | A |
4486739 | Franaszek | Dec 1984 | A |
4499550 | Ray, III | Feb 1985 | A |
4722084 | Morton | Jan 1988 | A |
4772845 | Scott | Sep 1988 | A |
4774498 | Traa | Sep 1988 | A |
4864303 | Ofek | Sep 1989 | A |
4897657 | Brubaker | Jan 1990 | A |
4974211 | Corl | Nov 1990 | A |
5053974 | Penz | Oct 1991 | A |
5166956 | Baltus | Nov 1992 | A |
5168509 | Nakamura | Dec 1992 | A |
5283761 | Gillingham | Feb 1994 | A |
5287305 | Yoshida | Feb 1994 | A |
5311516 | Kuznicki | May 1994 | A |
5412689 | Chan | May 1995 | A |
5449895 | Hecht | Sep 1995 | A |
5459465 | Kagey | Oct 1995 | A |
5511119 | Lechleider | Apr 1996 | A |
5553097 | Dagher | Sep 1996 | A |
5566193 | Cloonan | Oct 1996 | A |
5599550 | Kohlruss | Feb 1997 | A |
5659353 | Kostreski | Aug 1997 | A |
5727006 | Dreyer | Mar 1998 | A |
5802356 | Gaskins | Sep 1998 | A |
5825808 | Hershey | Oct 1998 | A |
5856935 | Moy | Jan 1999 | A |
5875202 | Venters | Feb 1999 | A |
5945935 | Kusumoto | Aug 1999 | A |
5949060 | Schattschneider | Sep 1999 | A |
5995016 | Perino | Nov 1999 | A |
5999016 | McClintock | Dec 1999 | A |
6005895 | Perino | Dec 1999 | A |
6084883 | Norrell | Jul 2000 | A |
6119263 | Mowbray | Sep 2000 | A |
6172634 | Leonowich | Jan 2001 | B1 |
6175230 | Hamblin | Jan 2001 | B1 |
6232908 | Nakaigawa | May 2001 | B1 |
6278740 | Nordyke | Aug 2001 | B1 |
6346907 | Dacy | Feb 2002 | B1 |
6359931 | Perino | Mar 2002 | B1 |
6378073 | Davis | Apr 2002 | B1 |
6398359 | Silverbrook | Jun 2002 | B1 |
6404820 | Postol | Jun 2002 | B1 |
6417737 | Moloudi | Jul 2002 | B1 |
6433800 | Holtz | Aug 2002 | B1 |
6452420 | Wong | Sep 2002 | B1 |
6473877 | Sharma | Oct 2002 | B1 |
6483828 | Balachandran | Nov 2002 | B1 |
6504875 | Perino | Jan 2003 | B2 |
6509773 | Buchwald | Jan 2003 | B2 |
6556628 | Poulton | Apr 2003 | B1 |
6563382 | Yang | May 2003 | B1 |
6621427 | Greenstreet | Sep 2003 | B2 |
6624699 | Yin | Sep 2003 | B2 |
6650638 | Walker | Nov 2003 | B1 |
6661355 | Cornelius | Dec 2003 | B2 |
6664355 | Kim | Dec 2003 | B2 |
6686879 | Shattil | Feb 2004 | B2 |
6766342 | Kechriotis | Jul 2004 | B2 |
6839429 | Gaikwad | Jan 2005 | B1 |
6839587 | Yonce | Jan 2005 | B2 |
6854030 | Perino | Feb 2005 | B2 |
6865234 | Agazzi | Mar 2005 | B1 |
6865236 | Terry | Mar 2005 | B1 |
6898724 | Chang | May 2005 | B2 |
6927709 | Kiehl | Aug 2005 | B2 |
6954492 | Williams | Oct 2005 | B1 |
6963622 | Eroz | Nov 2005 | B2 |
6972701 | Jansson | Dec 2005 | B2 |
6973613 | Cypher | Dec 2005 | B2 |
6976194 | Cypher | Dec 2005 | B2 |
6982954 | Dhong | Jan 2006 | B2 |
6990138 | Bejjani | Jan 2006 | B2 |
6999516 | Rajan | Feb 2006 | B1 |
7023817 | Kuffner | Apr 2006 | B2 |
7039136 | Olson | May 2006 | B2 |
7053802 | Cornelius | May 2006 | B2 |
7075996 | Simon | Jul 2006 | B2 |
7085153 | Ferrant | Aug 2006 | B2 |
7085336 | Lee | Aug 2006 | B2 |
7127003 | Rajan | Oct 2006 | B2 |
7130944 | Perino | Oct 2006 | B2 |
7142612 | Horowitz | Nov 2006 | B2 |
7142865 | Tsai | Nov 2006 | B2 |
7164631 | Tateishi | Jan 2007 | B2 |
7167019 | Broyde | Jan 2007 | B2 |
7180949 | Kleveland | Feb 2007 | B2 |
7184483 | Rajan | Feb 2007 | B2 |
7269212 | Chau | Sep 2007 | B1 |
7335976 | Chen | Feb 2008 | B2 |
7339990 | Hidaka | Mar 2008 | B2 |
7348989 | Stevens | Mar 2008 | B2 |
7349484 | Stojanovic | Mar 2008 | B2 |
7356213 | Cunningham | Apr 2008 | B1 |
7358869 | Chiarulli | Apr 2008 | B1 |
7362130 | Broyde | Apr 2008 | B2 |
7370264 | Worley | May 2008 | B2 |
7372390 | Yamada | May 2008 | B2 |
7389333 | Moore | Jun 2008 | B2 |
7400276 | Sotiriadis | Jul 2008 | B1 |
7428273 | Foster | Sep 2008 | B2 |
7456778 | Werner | Nov 2008 | B2 |
7462956 | Lan | Dec 2008 | B2 |
7496162 | Srebranig | Feb 2009 | B2 |
7535957 | Ozawa | May 2009 | B2 |
7570704 | Nagarajan | Aug 2009 | B2 |
7599390 | Pamarti | Oct 2009 | B2 |
7616075 | Kushiyama | Nov 2009 | B2 |
7620116 | Bessios | Nov 2009 | B2 |
7633850 | Ahn | Dec 2009 | B2 |
7639596 | Cioffi | Dec 2009 | B2 |
7643588 | Visalli | Jan 2010 | B2 |
7656321 | Wang | Feb 2010 | B2 |
7694204 | Schmidt | Apr 2010 | B2 |
7697915 | Behzad | Apr 2010 | B2 |
7706456 | Laroia | Apr 2010 | B2 |
7706524 | Zerbe | Apr 2010 | B2 |
7746764 | Rawlins | Jun 2010 | B2 |
7787572 | Scharf | Aug 2010 | B2 |
7804361 | Lim | Sep 2010 | B2 |
7808456 | Chen | Oct 2010 | B2 |
7808883 | Green | Oct 2010 | B2 |
7841909 | Murray | Nov 2010 | B2 |
7869497 | Benvenuto | Jan 2011 | B2 |
7869546 | Tsai | Jan 2011 | B2 |
7882413 | Chen | Feb 2011 | B2 |
7899653 | Hollis | Mar 2011 | B2 |
7907676 | Stojanovic | Mar 2011 | B2 |
7933770 | Kruger | Apr 2011 | B2 |
8036300 | Evans | Oct 2011 | B2 |
8050332 | Chung | Nov 2011 | B2 |
8055095 | Palotai | Nov 2011 | B2 |
8064535 | Wiley | Nov 2011 | B2 |
8085172 | Li | Dec 2011 | B2 |
8091006 | Prasad | Jan 2012 | B2 |
8106806 | Toyomura | Jan 2012 | B2 |
8149906 | Saito | Apr 2012 | B2 |
8159375 | Abbasfar | Apr 2012 | B2 |
8159376 | Abbasfar | Apr 2012 | B2 |
8180931 | Lee | May 2012 | B2 |
8185807 | Oh | May 2012 | B2 |
8199849 | Oh | Jun 2012 | B2 |
8199863 | Chen | Jun 2012 | B2 |
8218670 | AbouRjeily | Jul 2012 | B2 |
8245094 | Jiang | Aug 2012 | B2 |
8253454 | Lin | Aug 2012 | B2 |
8279094 | Abbasfar | Oct 2012 | B2 |
8289914 | Li | Oct 2012 | B2 |
8295250 | Gorokhov | Oct 2012 | B2 |
8310389 | Chui | Nov 2012 | B1 |
8365035 | Hara | Jan 2013 | B2 |
8406315 | Tsai | Mar 2013 | B2 |
8406316 | Sugita | Mar 2013 | B2 |
8429495 | Przybylski | Apr 2013 | B2 |
8437440 | Zhang | May 2013 | B1 |
8442099 | Sederat | May 2013 | B1 |
8442210 | Zerbe | May 2013 | B2 |
8443223 | Abbasfar | May 2013 | B2 |
8451913 | Oh | May 2013 | B2 |
8462891 | Kizer | Jun 2013 | B2 |
8472513 | Malipatil | Jun 2013 | B2 |
8498344 | Wilson | Jul 2013 | B2 |
8498368 | Husted | Jul 2013 | B1 |
8520348 | Dong | Aug 2013 | B2 |
8520493 | Goulahsen | Aug 2013 | B2 |
8539318 | Cronie | Sep 2013 | B2 |
8547272 | Nestler | Oct 2013 | B2 |
8577284 | Seo | Nov 2013 | B2 |
8578246 | Mittelholzer | Nov 2013 | B2 |
8588254 | Diab | Nov 2013 | B2 |
8588280 | Oh | Nov 2013 | B2 |
8593305 | Tajalli | Nov 2013 | B1 |
8638241 | Sudhakaran | Jan 2014 | B2 |
8649445 | Cronie | Feb 2014 | B2 |
8649460 | Ware | Feb 2014 | B2 |
8649556 | Wedge | Feb 2014 | B2 |
8649840 | Sheppard, Jr. | Feb 2014 | B2 |
8711919 | Kumar | Apr 2014 | B2 |
8718184 | Cronie | May 2014 | B1 |
8755426 | Cronie | Jun 2014 | B1 |
8773964 | Hsueh | Jul 2014 | B2 |
8780687 | Clausen | Jul 2014 | B2 |
8782578 | Tell | Jul 2014 | B2 |
8879660 | Peng | Nov 2014 | B1 |
8897134 | Kern | Nov 2014 | B2 |
8949693 | Ordentlich | Feb 2015 | B2 |
8951072 | Hashim | Feb 2015 | B2 |
8975948 | GonzalezDiaz | Mar 2015 | B2 |
8989317 | Holden | Mar 2015 | B1 |
9015566 | Cronie | Apr 2015 | B2 |
9020049 | Schwager | Apr 2015 | B2 |
9036764 | Hossain | May 2015 | B1 |
9069995 | Cronie | Jun 2015 | B1 |
9077386 | Holden | Jul 2015 | B1 |
9093791 | Liang | Jul 2015 | B2 |
9100232 | Hormati | Aug 2015 | B1 |
9106465 | Walter | Aug 2015 | B2 |
9124557 | Fox | Sep 2015 | B2 |
9165615 | Amirkhany | Oct 2015 | B2 |
9172412 | Kim | Oct 2015 | B2 |
9197470 | Okunev | Nov 2015 | B2 |
9281785 | Sjoland | Mar 2016 | B2 |
9288082 | Ulrich | Mar 2016 | B1 |
9288089 | Cronie | Mar 2016 | B2 |
9292716 | Winoto | Mar 2016 | B2 |
9306621 | Zhang | Apr 2016 | B2 |
9331962 | Lida | May 2016 | B2 |
9362974 | Fox | Jun 2016 | B2 |
9374250 | Musah | Jun 2016 | B1 |
20020044316 | Myers | Apr 2002 | A1 |
20020057592 | Robb | May 2002 | A1 |
20020154633 | Shin | Oct 2002 | A1 |
20030086366 | Branlund | May 2003 | A1 |
20030146783 | Bandy | Aug 2003 | A1 |
20050174841 | Ho | Aug 2005 | A1 |
20050213686 | Love | Sep 2005 | A1 |
20060067413 | Tsai | Mar 2006 | A1 |
20070263711 | Kramer | Nov 2007 | A1 |
20070265533 | Tran | Nov 2007 | A1 |
20080104374 | Mohamed | May 2008 | A1 |
20080159448 | Anim-Appiah | Jul 2008 | A1 |
20090059782 | Cole | Mar 2009 | A1 |
20100023838 | Shen | Jan 2010 | A1 |
20100046644 | Mazet | Feb 2010 | A1 |
20100180143 | Ware | Jul 2010 | A1 |
20100296556 | Rave | Nov 2010 | A1 |
20110072330 | Kolze | Mar 2011 | A1 |
20110268225 | Cronie | Nov 2011 | A1 |
20110299555 | Cronie | Dec 2011 | A1 |
20120152901 | Nagorny | Jun 2012 | A1 |
20120161945 | Single | Jun 2012 | A1 |
20130010892 | Cronie | Jan 2013 | A1 |
20140226455 | Schumacher | Aug 2014 | A1 |
20150078479 | Whitby-Strevens | Mar 2015 | A1 |
20150333940 | Shokrollahi | Nov 2015 | A1 |
20150381232 | Ulrich | Dec 2015 | A1 |
20160020796 | Hormati | Jan 2016 | A1 |
20160020824 | Ulrich | Jan 2016 | A1 |
20160036616 | Holden | Feb 2016 | A1 |
Number | Date | Country |
---|---|---|
101478286 | Jul 2009 | CN |
2039221 | Feb 2013 | EP |
2003163612 | Jun 2003 | JP |
2009084121 | Jul 2009 | WO |
2010031824 | Mar 2010 | WO |
2011119359 | Sep 2011 | WO |
Entry |
---|
Abbasfar, A., “Generalized Differential Vector Signaling”, IEEE International Conference on Communications, ICC '09, (Jun. 14, 2009), pp. 1-5. |
Brown, L., et al., “V.92: The Last Dial-Up Modem?”, IEEE Transactions on Communications, IEEE Service Center, Piscataway, NJ., USA, vol. 52, No. 1, Jan. 1, 2004, pp. 54-61. XP011106836, ISSN: 0090-6779, DOI: 10.1109/tcomm.2003.822168, pp. 55-59. |
Burr, “Spherical Codes for M-ARY Code Shift Keying”, University of York, Apr. 2, 1989, pp. 67-72, United Kingdom. |
Cheng, W., “Memory Bus Encoding for Low Power: A Tutorial”, Quality Electronic Design, IEEE, International Symposium on Mar. 26-28, 2001, pp. 199-204, Piscataway, NJ. |
Clayton, P., “Introduction to Electromagnetic Compatibility”, Wiley-Interscience, 2006. |
Dasilva et al., “Multicarrier Orthogonal CDMA Signals for Quasi-Synchronous Communication Systems”, IEEE Journal on Selected Areas in Communications, vol. 12, No. 5 (Jun. 1, 1994), pp. 842-852. |
Ericson, T., et al., “Spherical Codes Generated by Binary Partitions of Symmetric Pointsets”, IEEE Transactions on Information Theory, vol. 41, No. 1, Jan. 1995, pp. 107-129. |
Farzan, K., et al., “Coding Schemes for Chip-to-Chip Interconnect Applications”, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 14, No. 4, Apr. 2006, pp. 393-406. |
Healey, A., et al., “A Comparison of 25 Gbps NRZ & PAM-4 Modulation used in Legacy & Premium Backplane Channels”, DesignCon 2012, 16 pages. |
International Search Report and Written Opinion for PCT/EP2011/059279 mailed Sep. 22, 2011. |
International Search Report and Written Opinion for PCT/EP2011/074219 mailed Jul. 4, 2012. |
International Search Report and Written Opinion for PCT/EP2012/052767 mailed May 11, 2012. |
International Search Report and Written Opinion for PCT/US14/052986 mailed Nov. 24, 2014. |
International Search Report and Written Opinion from PCT/US2014/034220 mailed Aug. 21, 2014. |
International Search Report and Written Opinion of the International Searching Authority, mailed Jul. 14, 2011 in International Patent Application S.N. PCT/EP2011/002170, 10 pages. |
International Search Report and Written Opinion of the International Searching Authority, mailed Nov. 5, 2012, in International Patent Application S.N. PCT/EP2012/052767, 7 pages. |
International Search Report for PCT/US2014/053563, dated Nov. 11, 2014, 2 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration for PCT/EP2013/002681, dated Feb. 25, 2014, 15 pages. |
Poulton, et al., “Multiwire Differential Signaling”, UNC-CH Department of Computer Science Version 1.1, Aug. 6, 2003. |
She et al., “A Framework of Cross-Layer Superposition Coded Multicast for Robust IPTV Services over WiMAX,” IEEE Communications Society subject matter experts for publication in the WCNC 2008 proceedings, Mar. 31, 2008-Apr. 3, 2008, pp. 3139-3144. |
Skliar et al., A Method for the Analysis of Signals: the Square-Wave Method, Mar. 2008, Revista de Matematica: Teoria y Aplicationes, pp. 109-129. |
Slepian, D., “Premutation Modulation”, IEEE, vol. 52, No. 3, Mar. 1965, pp. 228-236. |
Stan, M., et al., “Bus-Invert Coding for Low-Power I/O, IEEE Transactions on Very Large Scale Integration (VLSI) Systems”, vol. 3, No. 1, Mar. 1995, pp. 49-58. |
Tallini, L., et al., “Transmission Time Analysis for the Parallel Asynchronous Communication Scheme”, IEEE Transactions on Computers, vol. 52, No. 5, May 2003, pp. 558-571. |
Wang et al., “Applying CDMA Technique to Network-on-Chip”, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 15, No. 10 (Oct. 1, 2007), pp. 1091-1100. |
“Introduction to: Analog Computers and the DSPACE System,” Course Material ECE 5230 Spring 2008, Utah State University, www.coursehero.com, 12 pages. |
Counts, L., et al., “One-Chip Slide Rule Works with Logs, Antilogs for Real-Time Processing,” Analog Devices Computational Products 6, Reprinted from Electronic Design, May 2, 1985, 7 pages. |
Design Brief 208 Using the Anadigm Multiplier CAM, Copyright 2002 Anadigm, 6 pages. |
Grahame, J., “Vintage Analog Computer Kits,” posted on Aug. 25, 2006 in Classic Computing, 2 pages, http.//www.retrothing.com/2006/08/classic—analog—.html. |
Jiang, A., et al., “Rank Modulation for Flash Memories”, IEEE Transactions of Information Theory, Jun. 2006, vol. 55, No. 6, pp. 2659-2673. |
Loh, M., et al., “A 3×9 Gb/s Shared, All-Digital CDR for High-Speed, High-Density I/O” , Matthew Loh, IEEE Journal of Solid-State Circuits, Vo. 47, No. 3, Mar. 2012. |
Notification of Transmittal of International Search Report and the Written Opinion of the International Searching Authority, for PCT/US2015/018363, mailed Jun. 18, 2015, 13 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, dated Mar. 3, 2015, for PCT/US2014/066893, 9 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2014/015840, dated May 20, 2014. 11 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2014/043965, dated Oct. 22, 2014, 10 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2015/037466, dated Nov. 19, 2015. |
Oh, et al., Pseudo-Differential Vector Signaling for Noise Reduction in Single-Ended Ended Signaling, DesignCon 2009. |
Schneider, J., et al., “ELEC301 Project: Building an Analog Computer,” Dec. 19, 1999, 8 pages, http://www.clear.rice.edu/elec301/Projects99/anlgcomp/. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2015/043463, dated Oct. 16, 2015, 8 pages. |
Zouhair Ben-Neticha et al, “The 'streTched-Golay and other codes for high-SNR finite-delay quantization of the Gaussian source at 1/2 Bit per sample”, IEEE Transactions on Communications, vol. 38, No. 12 Dec. 1, 1990, pp. 2089-2093, XP000203339, ISSN: 0090-6678, DOI: 10.1109/26.64647. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2015/039952, dated Sep. 23, 2015, 8 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, for PCT/US2015/041161, dated Oct. 7, 2015, 8 pages. |
Number | Date | Country | |
---|---|---|---|
61646790 | May 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15051344 | Feb 2016 | US |
Child | 15176081 | US | |
Parent | 13844032 | Mar 2013 | US |
Child | 15051344 | US |