Memory arbitration system and method having an arbitration packet protocol

Information

  • Patent Grant
  • 8555006
  • Patent Number
    8,555,006
  • Date Filed
    Monday, November 21, 2011
    13 years ago
  • Date Issued
    Tuesday, October 8, 2013
    11 years ago
Abstract
A memory hub and method for transmitting a read response on a data path of a memory hub interposed between a transmitting memory hub and a receiving memory hub. An arbitration packet including data indicative of a data path configuration for an associated read response is received at the memory hub. The arbitration packet is decoded, and the data path is configured in accordance with the data of the arbitration packet. The associated read response is received at the memory hub and the associated read response is coupled to the configured data path for transmitting the same to the receiving memory hub.
Description
TECHNICAL FIELD

This present invention is related generally to a memory system for a processor-based computing system, and more particularly, to a hub-based memory system having an arbitration system and method for managing memory responses therein.


BACKGROUND OF THE INVENTION

Computer systems use memory devices, such as dynamic random access memory (“DRAM”) devices, to store data that are accessed by a processor. These memory devices are normally used as system memory in a computer system. In a typical computer system, the processor communicates with the system memory through a processor bus and a memory controller. The memory devices of the system memory, typically arranged in memory modules having multiple memory devices, are coupled through a memory bus to the memory controller. The processor issues a memory request, which includes a memory command, such as a read command, and an address designating the location from which data or instructions are to be read. The memory controller uses the command and address to generate appropriate command signals as well as row and column addresses, which are applied to the system memory through the memory bus. In response to the commands and addresses, data are transferred between the system memory and the processor. The memory controller is often part of a system controller, which also includes bus bridge circuitry for coupling the processor bus to an expansion bus, such as a PCI bus.


In memory systems, high data bandwidth is desirable. Generally, bandwidth limitations are not related to the memory controllers since the memory controllers sequence data to and from the system memory as fast as the memory devices allow. One approach that has been taken to increase bandwidth is to increase the speed of the memory data bus coupling the memory controller to the memory devices. Thus, the same amount of information can be moved over the memory data bus in less time. However, despite increasing memory data bus speeds, a corresponding increase in bandwidth does not result. One reason for the non-linear relationship between data bus speed and bandwidth is the hardware limitations within the memory devices themselves. That is, the memory controller has to schedule all memory commands to the memory devices such that the hardware limitations are honored. Although these hardware limitations can be reduced to some degree through the design of the memory device, a compromise must be made because reducing the hardware limitations typically adds cost, power, and/or size to the memory devices, all of which are undesirable alternatives. Thus, given these constraints, although it is easy for memory devices to move “well-behaved” traffic at ever increasing rates, for example, sequel traffic to the same page of a memory device, it is much more difficult for the memory devices to resolve “badly-behaved traffic,” such as bouncing between different pages or banks of the memory device. As a result, the increase in memory data bus bandwidth does not yield a corresponding increase in information bandwidth.


In addition to the limited bandwidth between processors and memory devices, the performance of computer systems is also limited by latency problems that increase the time required to read data from system memory devices. More specifically, when a memory device read command is coupled to a system memory device, such as a synchronous DRAM (“SDRAM”) device, the read data are output from the SDRAM device only after a delay of several clock periods. Therefore, although SDRAM devices can synchronously output burst data at a high data rate, the delay in initially providing the data can significantly slow the operating speed of a computer system using such SDRAM devices. Increasing the memory data bus speed can be used to help alleviate the latency issue. However, as with bandwidth, the increase in memory data bus speeds do not yield a linear reduction of latency, for essentially the same reasons previously discussed.


Although increasing memory data bus speed has, to some degree, been successful in increasing bandwidth and reducing latency, other issues are raised by this approach. For example, as the speed of the memory data bus increases, loading on the memory bus needs to be decreased in order to maintain signal integrity since traditionally, there has only been wire between the memory controller and the memory slots into which the memory modules are plugged. Several approaches have been taken to accommodate the increase in memory data bus speed. For example, reducing the number of memory slots, adding buffer circuits on a memory module in order to provide sufficient fanout of control signals to the memory devices on the memory module, and providing multiple memory device interfaces on the memory module since there are too few memory module connectors on a single memory device interface. The effectiveness of these conventional approaches are, however, limited. A reason why these techniques were used in the past is that it was cost-effective to do so. However, when only one memory module can be plugged in per interface, it becomes too costly to add a separate memory interface for each required memory slot. In other words, it pushes the system controllers package out of the commodity range and into the boutique range, thereby, greatly adding cost.


One recent approach that allows for increased memory data bus speed in a cost effective manner is the use of multiple memory devices coupled to the processor through a memory hub. In a memory hub architecture, or a hub-based memory sub-system, a system controller or memory controller is coupled over a high speed bi-directional or unidirectional memory controller/hub interface to several memory modules. Typically, the memory modules are coupled in a point-to-point or daisy chain architecture such that the memory modules are connected one to another in series. Thus, the memory controller is coupled to a first memory module, with the first memory module connected to a second memory module, and the second memory module coupled to a third memory module, and so on in a daisy chain fashion.


Each memory module includes a memory hub that is coupled to the memory controller/hub interface and a number of memory devices on the module, with the memory hubs efficiently routing memory requests and responses between the controller and the memory devices over the memory controller/hub interface. Computer systems employing this architecture can use a high-speed memory data bus since signal integrity can be maintained on the memory data bus. Moreover, this architecture also provides for easy expansion of the system memory without concern for degradation in signal quality as more memory modules are added, such as occurs in conventional memory bus architectures.


Although computer systems using memory hubs can provide superior performance, various factors may affect the performance of the memory system. For example, the manner in which the flow of read data upstream (i.e., back to the memory hub controller in the computer system) from one memory hub to another is managed will affect read latency. The management of the flow of read data by a memory hub may be generally referred to as arbitration, with each memory hub arbitrating between local memory read responses and upstream memory read responses. That is, each memory hub determines whether to send local memory read responses first or to forward memory read responses from downstream (i.e., further away from the memory hub controller) memory hubs first. Although the determination of which memory read response has lower priority will only affect the latency of that specific memory read response, the additive effect of the memory read responses having increased latency will affect the overall latency of the memory system. Consequently, the arbitration technique employed by a memory hub directly affects the performance of the overall memory system. Additionally, the implementation of the arbitration scheme will affect the overall read latency as well, since inefficient implementation will negatively impact system memory performance despite utilizing a desirable arbitration scheme. Therefore, there is a need for a system and method for implementing an arbitration scheme for managing memory responses in a system memory having a memory hub architecture.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a partial block diagram of a computer system having a memory hub based system memory in which embodiments of the present invention can be implemented.



FIG. 2 is a functional block diagram of a arbitration control component according to an embodiment of the present invention that can be utilized in the memory hubs of FIG. 1.



FIG. 3 is a data structure diagram of a arbitration packet and memory response according to an embodiment of the present invention.



FIG. 4 is a flow diagram of the operation of the arbitration control component of FIG. 3 according to an embodiment of the present invention





DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS


FIG. 1 illustrates a computer system 100 having a memory hub architecture in which embodiments of the present invention can be utilized. The computer system 100 includes a processor 104 for performing various computing functions, such as executing specific software to perform specific calculations or tasks. The processor 104 includes a processor bus 106 that normally includes an address bus, a control bus, and a data bus. The processor bus 106 is typically coupled to cache memory 108, which, is typically static random access memory (“SRAM”). The processor bus 106 is further coupled to a system controller 110, which is also referred to as a bus bridge.


The system controller 110 also serves as a communications path to the processor 104 for a variety of other components. More specifically, the system controller 110 includes a graphics port that is typically coupled to a graphics controller 112, which is, in turn, coupled to a video terminal 114. The system controller 110 is also coupled to one or more input devices 118, such as a keyboard or a mouse, to allow an operator to interface with the computer system 100. Typically, the computer system 100 also includes one or more output devices 120, such as a printer, coupled to the processor 104 through the system controller 110. One or more data storage devices 124 are also typically coupled to the processor 104 through the system controller 110 to allow the processor 104 to store data or retrieve data from internal or external storage media (not shown). Examples of typical storage devices 124 include hard and floppy disks, tape cassettes, and compact disk read-only memories (CD-ROMs).


The system controller 110 contains a memory hub controller 128 coupled to several memory modules 130a-n through a bus system 154, 156. Each of the memory modules 130a-n includes a memory hub 140 coupled to several memory devices 148 through command, address and data buses, collectively shown as bus 150. The memory hub 140 efficiently routes memory requests and responses between the controller 128 and the memory devices 148. Each of the memory hubs 140 includes write buffers and read data buffers. Computer systems employing this architecture allow for the processor 104 to access one memory module 130a-n while another memory module 130a-n is responding to a prior memory request. For example, the processor 104 can output write data to one of the memory modules 130a-n in the system while another memory module 130a-n in the system is preparing to provide read data to the processor 104. Additionally, a memory hub architecture can also provide greatly increased memory capacity in computer systems.



FIG. 2 is a functional block diagram illustrating an arbitration control component 200 according to one embodiment of the present invention. The arbitration control component 200 can be included in the memory hubs 140 of FIG. 1. As shown in FIG. 2, the arbitration control component 200 includes two queues for storing associated memory responses. A local response queue 202 receives and stores local memory responses LMR from the memory devices 148 on the associated memory module 130. A remote response queue 206 receives and stores downstream memory responses which cannot be immediately forwarded upstream through a bypass path 204. An arbitration control circuit 210 is coupled to the queues 202, 206 through a control/status bus 136, which allows the arbitration control circuit 210 to monitor the contents of each of the queues 202, 206, and utilizes this information in controlling a multiplexer 208 to thereby control the overall arbitration process executed by the memory hub 140. The control/status bus 136 also allows “handshaking” signals to be coupled from the queues 202, 206 to the arbitration control circuit 210 to coordinate the transfer of control signals from the arbitration control circuit 210 to the queues 202, 206.


The arbitration control circuit 210 is further coupled to the high-speed link 134 to receive arbitration packets from downstream memory hubs. As will be explained in more detail below, arbitration packets are provided in advance of an associated memory response, and provide the arbitration control circuit 210 of an upstream memory hub with information to enable the appropriate path through the receiving memory hub in anticipation of receiving the associated memory response. Additionally, the arbitration control circuit 210 generates an arbitration packet to be provided prior to an associated LMR to serve as an early indication of the associated memory response when data is read from the memory devices 148 (FIG. 1) in response to a read request. As previously discussed, the arbitration packet will provide upstream memory hubs with appropriate information and give the respective arbitration control circuits 210 time to make decisions regarding enablement of the appropriate data paths before the memory response arrives. The arbitration control circuit 210 prepares the arbitration packet while read data for the memory response is being retrieved from memory devices 148. The arbitration packet is provided through a switch 212 to either the multiplexer 208 or the local response queue 202, depending on whether if the upstream memory hub is idle or busy. The multiplexer 208, under the control of the arbitration control circuit, couples the high-speed link 134 to receive memory responses from the remote response queue 206 or the bypass path 204, arbitration packets from the arbitration control circuit 210, or arbitration packets and memory responses from the local response queue 202. In an alternative embodiment of the present invention, the arbitration packets are generated in an arbitration packet circuit, rather than in the arbitration control circuit 210, as shown in FIG. 2. Additionally, although shown in FIG. 2 as providing the arbitration packet to the multiplexer 208 to be injected into the stream of data, the arbitration packet can alternatively be provided to the local response queue 202 and placed before the associated read response packet to be injected into the data stream. It will be appreciated by those ordinarily skilled in the art that modifications to the embodiments of the present invention, such as the location at which the arbitration packet is generated or the manner in which the arbitration packet is placed into the data stream prior to the associated read packet, can be made without departing from the scope of the present invention.



FIG. 3 illustrates a data structure 300 for arbitration packets and memory responses according to an embodiment of the present invention. The data structure 300 is divided into 8-bit bytes of information, with each byte of information corresponding to a sequential bit-time. Each bit-time represents an increment of time in which new data can be provided. A response header field 302 includes two bytes of data that indicate the response is either an arbitration packet or a memory response. An address field 304 includes data that is used to identify the particular hub to which the arbitration packet or memory response is directed. A command code field 306 will have a value to identify the data structure 300 as an arbitration packet, and not as a memory response. Arbitration packets and memory responses are similar, except that the data payload of data fields 308 are “don't cares” for arbitration packets. In the data structure 300, all 16 bits of size fields 310 carry the same value to indicate the size of the data payload carried by the memory response. For example, a “0” indicates that 32 bytes of data are included, and a “1” indicates that 64 bytes of data are included. It will be appreciated by one ordinarily skilled in the art that the embodiment of the data structure 300 shown in FIG. 3 has been provided by way of example, and that modifications to the data structure 300 can be made without deviating from the scope of the present invention. For example, the number and type of data fields of the data structure 300 can be changed or the number of bits for each bit time can be changed and still remain within the scope of the present invention.


Operation of the arbitration control component 200 (FIG. 2) will be described with reference to the flow diagram of FIG. 4. Following the receipt of a read data command, at a step 402 the memory hub initiates a read operation to retrieve the requested read data from the memory devices 148 (FIG. 1) for the memory response that will be provided to the requesting target. At a step 404, the arbitration control circuit 210 of the memory hub determines whether the local data path is idle by checking the status of the local response queue 202. If the local data path is idle, an arbitration packet is generated by the arbitrations control circuit 210 during the retrieval of the read data from the memory devices 148 at a step 406. When the arbitration packet and the memory response have been prepared, and are ready for transmission, at a step 408 an upstream memory hub is queried to determine if it is busy. Where the upstream memory hub is idle, the arbitration packet is sent to the upstream memory hub, followed by the memory response at steps 410, 412. However, if the upstream memory hub is busy, the arbitration packet is discarded at a step 414 and the memory response is stored in a local response queue 202 at a step 416. Similarly, in the event that at the step 404 it was determined that the local data path is busy, the memory response is also stored in the local response queue at the step 416. At a step 418 the memory response is stored in the local response queue 202 until it is selected for transmission to the upstream memory hub in accordance with an arbitration scheme implemented by the memory hub. At a step 420, the memory response is transmitted through each upstream memory hub in accordance with the arbitration scheme until the memory response reaches the target destination. Suitable arbitration schemes are well known in the art, and will not be described in detail herein. An example of an arbitration scheme that is also suitable for use is described in more detail in commonly assigned, co-pending U.S. patent application Ser. No. 10/690,810, entitled ARBITRATION SYSTEM AND METHOD FOR MEMORY RESPONSES IN A HUB-BASED MEMORY SYSTEM to James W. Meyer and Cory Kanski, filed on Oct. 20, 2003, which is incorporated herein by reference.


As described therein, the local and remote response queues 202, 206 and the bypass path 204 are utilized to implement various response arbitration schemes. For example, in one embodiment, the arbitration control circuit executes an arbitration scheme that gives downstream responses, or remote responses, priority over local responses. Alternatively, in another embodiment described, the arbitration control circuit executes an arbitration scheme that gives priority to local responses over downstream responses. In another embodiment, the arbitration control circuit alternates between a predetermined number of responses from local and downstream memory, for example, local and remote responses can be alternately forwarded, or two local responses are forwarded followed by two remote responses, and so on. Another embodiment described therein utilizes an oldest first algorithm in arbitrating between local and downstream memory responses. That is, in operation, the arbitration control circuit 210 monitors response identifier portions of the memory responses stored in the local response queue and the remote response queue and selects the oldest response contained in either of these queues as the next response to be forwarded upstream. Thus, independent of the response queue in which a memory response is stored, the arbitration control circuit forwards the oldest responses first.


It will be appreciated by those ordinarily skilled in the art that other arbitration methods and schemes can be utilized without departing from the scope of the present invention.


Returning to the steps 410, 412 where the arbitration packet is first transmitted to an upstream memory hub and then followed by the memory response, the arbitration control circuit 210 of the upstream memory hub receives the arbitration packet at a step 422. The arbitration packet is decoded, and the appropriate data path is enabled by the arbitration control circuit 210 based on the information decoded at steps 424, 426. By the time the memory response is received at a step 430, the appropriate data path is enabled by the arbitration control circuit 210. At a step 428, the next upstream memory hub is queried to determine if it is busy. If not, the arbitration packet and then the memory response are transmitted to the next upstream memory hub in a bypass fashion at a step 432. The transmission of the arbitration packet and the memory response in the bypass fashion is facilitated by enabling the appropriate data path through the memory hub based on the decoded information of the arbitration packet that is sent at the step 410 before the associated memory response is sent at the step 412.


Returning to the step 428, if it is determined that the next upstream memory hub is busy, the arbitration packet is discarded at the step 440, and the memory response is stored in the remote response queue 206 until the memory response is selected for transmission to the next upstream memory hub according to the arbitration scheme employed at a step 442. At the step 420, the memory response will make its way upstream through the memory hubs in accordance with the arbitration scheme until reaching its target destination.


From the foregoing it will be appreciated that, although specific embodiments of the invention have been described herein for purposes of illustration, various modifications may be made without deviating from the spirit and scope of the invention. For example, embodiments of the present invention have been described herein with respect to a memory hub-based system memory used in a computer system. However, it will be appreciated that embodiments of the present invention can be used in memory systems other than hub-based memory systems, where appropriate. Moreover, embodiments of the present invention can also be used in memory hub-based systems that are utilized in processor based systems, as known in the art, other than computer systems. Accordingly, the invention is not limited except as by the appended claims.

Claims
  • 1. A method of responding to a read request in a system memory having a responding hub and at least one interposing hub through which a read response is transmitted on a data path of the interposing hub, the method comprising: retrieving read data from a memory device coupled to the responding hub and preparing a read response including the read data;generating an arbitration packet including data indicative of a data path configuration for the read response, the arbitration packet having a plurality of 8-bit bytes including at least one byte including data used by the interposing hub to distinguish the arbitration packet from a read response;providing the arbitration packet and the read response to the interposing hub, the arbitration packet provided prior to the read response; andenabling a data path for the read response in the interposing hub in accordance with the data of the arbitration packet.
  • 2. The method of claim 1 wherein enabling the data path for the read response comprises enabling a bypass data path in the interposing hub to couple the arbitration packet and read response through the interposing hub.
  • 3. The method of claim 1, further comprising: determining whether the interposing hub is busy; and in the event that the interposing hub is not busy, generating the arbitration packet for provision to the interposing hub prior to providing the associated read response to the interposing hub.
  • 4. The method of claim 1, further comprising: determining whether a local data path of the responding hub is idle; in the event that the local data path is idle, generating the arbitration packet for provision to the interposing hub prior to providing the associated read response to the interposing hub.
  • 5. The method of claim 1 wherein generating the arbitration packet comprises generating an arbitration packet including data indicative of enabling a bypass data path in the interposing hub for coupling the arbitration packet and read response through the interposing hub.
  • 6. A method of transmitting a read response on a data path of a hub interposed between a transmitting hub and a receiving hub, the method comprising: receiving a query from the transmitting hub whether the hub is busy and responding to the query by indicating to the transmitting hub that the hub is not busy;receiving at the hub an arbitration packet including data indicative of a data path configuration for an associated read response prior to receiving the associated read response; andconfiguring the data path in accordance with the data of the arbitration packet.
  • 7. The method of claim 6, further comprising: receiving the associated read response at the hub; andcoupling the associated read response to the configured data path for transmitting the same to the receiving hub.
  • 8. The method of claim 6 wherein configuring the data path comprises enabling a bypass data path in the hub to couple the arbitration packet and read response through the hub to the receiving hub.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 12/169,493, filed Jul. 8, 2008, U.S. Pat. No. 8,082,404, which is a continuation of U.S. patent application Ser. No. 11/731,122, filed Mar. 29, 2007, U.S. Pat. No. 7,412,571, which is a continuation of U.S. patent application Ser. No. 10/809,839, filed Mar. 24, 2004, U.S. Pat. No. 7,257,683. These applications and patents are incorporated by reference herein in their entirety and for all purposes.

US Referenced Citations (318)
Number Name Date Kind
3742253 Kronies Jun 1973 A
4045781 Levy et al. Aug 1977 A
4078228 Miyazaki Mar 1978 A
4240143 Besemer et al. Dec 1980 A
4245306 Besemer et al. Jan 1981 A
4253144 Bellamy et al. Feb 1981 A
4253146 Bellamy et al. Feb 1981 A
4608702 Hirzel et al. Aug 1986 A
4707823 Holdren et al. Nov 1987 A
4724520 Athanas et al. Feb 1988 A
4831520 Rubinfeld et al. May 1989 A
4843263 Ando Jun 1989 A
4891808 Williams Jan 1990 A
4930128 Suzuki et al. May 1990 A
4953930 Ramsey et al. Sep 1990 A
4982185 Holmberg et al. Jan 1991 A
5241506 Motegi et al. Aug 1993 A
5243703 Farmwald et al. Sep 1993 A
5251303 Fogg, Jr. et al. Oct 1993 A
5255239 Taborn et al. Oct 1993 A
5269022 Shinjo et al. Dec 1993 A
5299293 Mestdagh et al. Mar 1994 A
5313590 Taylor May 1994 A
5317752 Jewett et al. May 1994 A
5319755 Farmwald et al. Jun 1994 A
5327553 Jewett et al. Jul 1994 A
5355391 Horowitz et al. Oct 1994 A
5432823 Gasbarro et al. Jul 1995 A
5432907 Picazo, Jr. et al. Jul 1995 A
5442770 Barratt Aug 1995 A
5461627 Rypinski Oct 1995 A
5465229 Bechtolscheim et al. Nov 1995 A
5467455 Gay et al. Nov 1995 A
5479370 Furuyama et al. Dec 1995 A
5497476 Oldfield et al. Mar 1996 A
5502621 Schumacher et al. Mar 1996 A
5544319 Acton et al. Aug 1996 A
5566325 Bruce, II et al. Oct 1996 A
5577220 Combs et al. Nov 1996 A
5581767 Katsuki et al. Dec 1996 A
5606717 Farmwald et al. Feb 1997 A
5638334 Farmwald et al. Jun 1997 A
5638534 Mote, Jr. Jun 1997 A
5659798 Blumrich et al. Aug 1997 A
5687325 Chang Nov 1997 A
5706224 Srinivasan et al. Jan 1998 A
5715456 Bennett et al. Feb 1998 A
5729709 Harness Mar 1998 A
5748616 Riley May 1998 A
5799048 Farjad-Rad et al. Aug 1998 A
5818844 Singh et al. Oct 1998 A
5819304 Nilsen et al. Oct 1998 A
5822255 Uchida Oct 1998 A
5832250 Whittaker Nov 1998 A
5875352 Gentry et al. Feb 1999 A
5875454 Craft et al. Feb 1999 A
5900020 Safranek et al. May 1999 A
5928343 Farmwald et al. Jul 1999 A
5966724 Ryan Oct 1999 A
5973935 Schoenfeld et al. Oct 1999 A
5973951 Bechtolsheim et al. Oct 1999 A
5978567 Rebane et al. Nov 1999 A
5987196 Noble Nov 1999 A
6014721 Arimilli et al. Jan 2000 A
6023726 Saksena Feb 2000 A
6029250 Keeth Feb 2000 A
6031241 Silfvast et al. Feb 2000 A
6033951 Chao Mar 2000 A
6038630 Foster et al. Mar 2000 A
6038687 Ho Mar 2000 A
6061263 Boaz et al. May 2000 A
6061296 Ternullo, Jr. et al. May 2000 A
6064706 Driskill et al. May 2000 A
6067262 Irrinki et al. May 2000 A
6067649 Goodwin May 2000 A
6073190 Rooney Jun 2000 A
6076139 Welker et al. Jun 2000 A
6079008 Clery, III Jun 2000 A
6098158 Lay et al. Aug 2000 A
6100735 Lu Aug 2000 A
6105075 Ghaffari Aug 2000 A
6125431 Kobayashi Sep 2000 A
6131149 Lu et al. Oct 2000 A
6134624 Burns et al. Oct 2000 A
6137709 Boaz et al. Oct 2000 A
6144587 Yoshida Nov 2000 A
6167465 Parvin et al. Dec 2000 A
6167486 Lee et al. Dec 2000 A
6175571 Haddock et al. Jan 2001 B1
6185352 Hurley Feb 2001 B1
6186400 Dvorkis et al. Feb 2001 B1
6191663 Hannah Feb 2001 B1
6201724 Ishizaki et al. Mar 2001 B1
6208180 Fisch et al. Mar 2001 B1
6219725 Diehl et al. Apr 2001 B1
6233376 Updegrove May 2001 B1
6243769 Rooney Jun 2001 B1
6243831 Mustafa et al. Jun 2001 B1
6246618 Yamamoto et al. Jun 2001 B1
6247107 Christie Jun 2001 B1
6249802 Richardson et al. Jun 2001 B1
6256325 Park Jul 2001 B1
6256692 Yoda et al. Jul 2001 B1
6272600 Talbot et al. Aug 2001 B1
6272609 Jeddeloh Aug 2001 B1
6278755 Baba et al. Aug 2001 B1
6285349 Smith Sep 2001 B1
6286083 Chin et al. Sep 2001 B1
6289068 Hassoun et al. Sep 2001 B1
6294937 Crafts et al. Sep 2001 B1
6301637 Krull et al. Oct 2001 B1
6324485 Ellis Nov 2001 B1
6327642 Lee et al. Dec 2001 B1
6327650 Bapst et al. Dec 2001 B1
6330205 Shimizu et al. Dec 2001 B2
6347055 Motomura Feb 2002 B1
6349363 Cai et al. Feb 2002 B2
6356573 Jonsson et al. Mar 2002 B1
6367074 Bates et al. Apr 2002 B1
6370068 Rhee Apr 2002 B2
6373777 Suzuki Apr 2002 B1
6381190 Shinkai Apr 2002 B1
6392653 Malandain et al. May 2002 B1
6401213 Jeddeloh Jun 2002 B1
6405280 Ryan Jun 2002 B1
6421744 Morrison et al. Jul 2002 B1
6430696 Keeth Aug 2002 B1
6434639 Haghighi Aug 2002 B1
6434696 Kang Aug 2002 B1
6434736 Schaecher et al. Aug 2002 B1
6438622 Haghighi et al. Aug 2002 B1
6438668 Esfahani et al. Aug 2002 B1
6449308 Knight, Jr. et al. Sep 2002 B1
6453393 Holman et al. Sep 2002 B1
6462978 Shibata et al. Oct 2002 B2
6463059 Movshovich et al. Oct 2002 B1
6467013 Nizar Oct 2002 B1
6470422 Cai et al. Oct 2002 B2
6473828 Matsui Oct 2002 B1
6477592 Chen et al. Nov 2002 B1
6477614 Leddige et al. Nov 2002 B1
6477621 Lee et al. Nov 2002 B1
6479322 Kawata et al. Nov 2002 B2
6487556 Downs et al. Nov 2002 B1
6490188 Nuxoll et al. Dec 2002 B2
6493803 Pham et al. Dec 2002 B1
6496909 Schimmel Dec 2002 B1
6501471 Venkataraman et al. Dec 2002 B1
6505287 Uematsu Jan 2003 B2
6523092 Fanning Feb 2003 B1
6523093 Bogin et al. Feb 2003 B1
6526483 Cho et al. Feb 2003 B1
6539490 Forbes et al. Mar 2003 B1
6552564 Forbes et al. Apr 2003 B1
6564329 Cheung et al. May 2003 B1
6577174 Locker et al. Jun 2003 B2
6587912 Leddige et al. Jul 2003 B2
6590816 Perner Jul 2003 B2
6594713 Fuoco et al. Jul 2003 B1
6594722 Willke, II et al. Jul 2003 B1
6598154 Vaid et al. Jul 2003 B1
6615325 Mailloux et al. Sep 2003 B2
6622186 Moniot et al. Sep 2003 B1
6622188 Goodwin et al. Sep 2003 B1
6622227 Zumkehr et al. Sep 2003 B2
6625687 Halbert et al. Sep 2003 B1
6628294 Sadowsky et al. Sep 2003 B1
6629220 Dyer Sep 2003 B1
6631440 Jenne et al. Oct 2003 B2
6633576 Melaragni et al. Oct 2003 B1
6636110 Ooishi et al. Oct 2003 B1
6636912 Ajanovic et al. Oct 2003 B2
6646929 Moss et al. Nov 2003 B1
6658509 Bonella et al. Dec 2003 B1
6662304 Keeth et al. Dec 2003 B2
6665202 Lindahl et al. Dec 2003 B2
6667895 Jang et al. Dec 2003 B2
6667926 Chen et al. Dec 2003 B1
6670833 Kurd et al. Dec 2003 B2
6681292 Creta et al. Jan 2004 B2
6697926 Johnson et al. Feb 2004 B2
6715018 Farnworth et al. Mar 2004 B2
6718440 Maiyuran et al. Apr 2004 B2
6721195 Brunelle et al. Apr 2004 B2
6721860 Klein Apr 2004 B2
6724685 Braun et al. Apr 2004 B2
6728800 Lee et al. Apr 2004 B1
6735679 Herbst et al. May 2004 B1
6735682 Segelken et al. May 2004 B2
6742098 Halbert et al. May 2004 B1
6745275 Chang Jun 2004 B2
6751703 Chilton Jun 2004 B2
6754812 Abdallah et al. Jun 2004 B1
6756661 Tsuneda et al. Jun 2004 B2
6760833 Dowling Jul 2004 B1
6771538 Shukuri et al. Aug 2004 B2
6775747 Venkatraman Aug 2004 B2
6782435 Garcia et al. Aug 2004 B2
6789173 Tanaka et al. Sep 2004 B1
6792059 Yuan et al. Sep 2004 B2
6792496 Aboulenein et al. Sep 2004 B2
6795899 Dodd et al. Sep 2004 B2
6799246 Wise et al. Sep 2004 B1
6799268 Boggs et al. Sep 2004 B1
6804760 Wiliams Oct 2004 B2
6804764 Laberge et al. Oct 2004 B2
6807630 Lay et al. Oct 2004 B2
6811320 Abbott Nov 2004 B1
6816947 Huffman Nov 2004 B1
6820181 Jeddeloh et al. Nov 2004 B2
6821029 Grung et al. Nov 2004 B1
6823023 Hannah Nov 2004 B1
6845409 Talagala et al. Jan 2005 B1
6877079 Yoo et al. Apr 2005 B2
6889304 Perego et al. May 2005 B2
6901494 Zumkehr et al. May 2005 B2
6904556 Walton et al. Jun 2005 B2
6910109 Holman et al. Jun 2005 B2
6912612 Kapur et al. Jun 2005 B2
6947672 Jiang et al. Sep 2005 B2
6980042 LaBerge Dec 2005 B2
7046060 Minzoni et al. May 2006 B1
7047351 Jeddeloh May 2006 B2
7068085 Gomm et al. Jun 2006 B2
7116143 Deivasigamani et al. Oct 2006 B2
7117316 Jeddeloh Oct 2006 B2
7120743 Meyer et al. Oct 2006 B2
7133991 James Nov 2006 B2
7136958 Jeddeloh Nov 2006 B2
7149874 Jeddeloh Dec 2006 B2
7181584 LaBerge Feb 2007 B2
7187742 Logue et al. Mar 2007 B1
7251714 James Jul 2007 B2
7257683 Jeddeloh et al. Aug 2007 B2
7363419 Cronin et al. Apr 2008 B2
7386649 Jeddeloh Jun 2008 B2
7412571 Jeddeloh et al. Aug 2008 B2
7412574 Jeddeloh Aug 2008 B2
7415567 Jeddeloh Aug 2008 B2
7447240 James Nov 2008 B2
7469316 Dodd Dec 2008 B2
7529273 James May 2009 B2
7768325 Milton Aug 2010 B2
7788451 Larson et al. Aug 2010 B2
8082404 LeBerge Dec 2011 B2
8291173 Larson et al. Oct 2012 B2
20010038611 Darcie et al. Nov 2001 A1
20010039612 Lee Nov 2001 A1
20020016885 Ryan et al. Feb 2002 A1
20020084458 Halbert et al. Jul 2002 A1
20020112119 Halbert et al. Aug 2002 A1
20020116588 Beckert et al. Aug 2002 A1
20020144064 Fanning Oct 2002 A1
20020174284 Garcia et al. Nov 2002 A1
20020196806 Ghodrat et al. Dec 2002 A1
20030005223 Coulson et al. Jan 2003 A1
20030005344 Bhamidipati et al. Jan 2003 A1
20030043158 Wasserman et al. Mar 2003 A1
20030043426 Baker et al. Mar 2003 A1
20030093630 Richard et al. May 2003 A1
20030149809 Jensen et al. Aug 2003 A1
20030156581 Osborne Aug 2003 A1
20030163649 Kapur et al. Aug 2003 A1
20030177320 Sah et al. Sep 2003 A1
20030193927 Hronik Oct 2003 A1
20030200401 Moriwaki et al. Oct 2003 A1
20030217223 Nino, Jr. et al. Nov 2003 A1
20030227798 Pax Dec 2003 A1
20030229762 Maiyuran et al. Dec 2003 A1
20030229770 Jeddeloh Dec 2003 A1
20040015650 Zumkehr et al. Jan 2004 A1
20040022094 Radhakrishnan et al. Feb 2004 A1
20040024948 Winkler et al. Feb 2004 A1
20040034753 Jeddeloh Feb 2004 A1
20040044833 Ryan Mar 2004 A1
20040044857 Jeddeloh et al. Mar 2004 A1
20040047169 Lee et al. Mar 2004 A1
20040107306 Barth et al. Jun 2004 A1
20040126115 Levy et al. Jul 2004 A1
20040128449 Osborne et al. Jul 2004 A1
20040144994 Lee et al. Jul 2004 A1
20040160206 Komaki et al. Aug 2004 A1
20040193821 Ruhovets et al. Sep 2004 A1
20040199739 Jeddeloh Oct 2004 A1
20040225847 Wastlick et al. Nov 2004 A1
20040230718 Polzin et al. Nov 2004 A1
20040236885 Fredriksson et al. Nov 2004 A1
20040251936 Gomm Dec 2004 A1
20050015426 Woodruff et al. Jan 2005 A1
20050044327 Howard et al. Feb 2005 A1
20050071542 Weber et al. Mar 2005 A1
20050086441 Meyer et al. Apr 2005 A1
20050105350 Zimmerman May 2005 A1
20050122153 Lin Jun 2005 A1
20050132159 Jeddeloh Jun 2005 A1
20050149603 DeSota et al. Jul 2005 A1
20050162882 Reeves et al. Jul 2005 A1
20050166006 Talbot et al. Jul 2005 A1
20050169168 Aronson et al. Aug 2005 A1
20050177677 Jeddeloh Aug 2005 A1
20050177695 Larson et al. Aug 2005 A1
20050213611 James Sep 2005 A1
20050268060 Cronin et al. Dec 2005 A1
20060022724 Zerbe et al. Feb 2006 A1
20060066375 Laberge Mar 2006 A1
20060136683 Meyer et al. Jun 2006 A1
20060174070 Jeddeloh Aug 2006 A1
20060218318 James Sep 2006 A1
20060271746 Meyer et al. Nov 2006 A1
20070033317 Jeddeloh Feb 2007 A1
20070180171 Jeddeloh Aug 2007 A1
20070300023 Cronin et al. Dec 2007 A1
20080215792 Jeddeloh Sep 2008 A1
20080294862 Jeddeloh Nov 2008 A1
20090013211 Vogt et al. Jan 2009 A1
20100019822 LaBerge Jan 2010 A1
20100287323 Larson Nov 2010 A1
20130007384 Larson et al. Jan 2013 A1
Foreign Referenced Citations (21)
Number Date Country
0 709 786 May 1996 EP
0849685 Jun 1998 EP
0910021 Apr 1999 EP
2001-161912 Jun 1989 JP
05-342084 Dec 1993 JP
06-104707 Apr 1994 JP
2008-023267 Jan 1996 JP
8-503800 Apr 1996 JP
2001265539 Sep 2001 JP
2002530731 Sep 2002 JP
2002342161 Nov 2002 JP
2007520826 Jul 2007 JP
498215 Aug 2002 TW
502174 Sep 2002 TW
548547 Aug 2003 TW
9319422 Sep 1993 WO
0029921 May 2000 WO
2005076816 Aug 2000 WO
0223353 Mar 2002 WO
0227499 Apr 2002 WO
2004102403 Nov 2004 WO
Non-Patent Literature Citations (15)
Entry
“Flash Erasable Programmable Read-Only Memory”, “Free On-Line Dictionary of Computing”, online May 17, 2004 [http://foldoc.doc.ic.ac.uk/foldoc/foldoc.cgi?flaxh+memory].
“HyperTransport I/O Link Specification”, HyperTransport Technology Consortium, 2001, pp. 46-49.
“HyperTransport I/O Link Specification”. HyperTransport Technology Consortium, Aug. 25, 2003, pp. 1, 28-43 and 80-82.
“Office Action mailed Sep. 7, 2010 for European Application No. 05728609.8”.
INTEL, “Flash Memory PCI Add-In Card for Embedded Systems”, Application Note AP-758, Sep. 1997, pp. i-13.
Intel, “Intel 840 Chipset: 82840 Memory Controller Hub (MCH)”, Datasheet, www.intel.com/design/chipsets/datashts/298020.htm, pp. 1-178.
Micron Technology, Inc., “Synchronous DRAM Module 512MB/1GB (x72, ECC) 168-PIN Registered FBGA SDRAM DIMM”, Micron Technology, Inc., 2002, pp. 1-23.
Rambus, Inc., “Direct Rambus™ Technology Disclosure”, Oct. 1997 pp. 1-16.
Shanley, T et al., “PCI System Architecture”, Third Edition, Mindshare, Inc., Jul. 1996, pp. 24-25.
“Extended Search Report”, Extended Search Report for EP Appl #12155103.7 dated Oct. 26, 2012.
“First Office Action”, First Offie Action for CN Appl #200580016550.1 dated Feb. 15, 2008.
“Office Action”, Office Action for EP Appl #05 728 609.8 dated Sep. 7, 2010.
“Suppl. Search Report”, Suppl. Search Report for EP Appl #05 728 609.8 dated Sep. 14, 2007.
“Written Opinion”, Written Opinion for PCT/US2005/09523, Jul. 12, 2006.
“Micron 1Gb: x4, x8, x16 DDR SDRAM”, Micron Technology, Inc, Jul. 2003.
Related Publications (1)
Number Date Country
20120066461 A1 Mar 2012 US
Continuations (3)
Number Date Country
Parent 12169493 Jul 2008 US
Child 13301653 US
Parent 11731122 Mar 2007 US
Child 12169493 US
Parent 10809839 Mar 2004 US
Child 11731122 US