A multi-processing system includes a plurality of processors that share a single memory. Typically, multi-level caches are used to reduce memory bandwidth demands on the single memory. The multi-level caches may include a first-level private cache in each processor and a second-level cache shared by all of the processors. As the cache is much smaller than the memory in the system, only a portion of the data stored in buffers/blocks in memory is replicated in the cache.
If data stored in a buffer/block requested by a processor is replicated in the cache, there is a cache hit. If the requested data is not replicated in the cache, there is a cache miss and the requested block that stores the data is retrieved from memory and also stored in the cache.
When shared data is cached, the shared value may be replicated in multiple first-level caches. Thus, caching of shared data requires cache coherence. Cache coherence ensures that multiple processors see a consistent view of memory, for example, a read of the shared data by any of the processors returns the most recently written value of the data.
Typically, blocks of memory (cache blocks) are replicated in cache and each cache block has an associated tag that includes a so-called dirty bit. The state of the dirty bit indicates whether the cache block has been modified. In a write back cache, the modified cache block is written back to memory only when the modified cache block is replaced by another cache block in the cache.
In a network services processor, when the modified cache block is replaced in the cache, the modified cache block may not always need to be written back to memory. For example, the cache block can be used to store packet data while it is being processed. After the data has been processed, the processed packet data stored in the cache block is no longer required and the buffer in memory is freed, that is, made available for allocation to store data for another packet. As the processed packet data that is stored in the cache block will not be used when the buffer in memory is re-allocated for storing other packet data, it would be wasteful to write the cache block in the cache back to the buffer in memory. Not performing a write operation to write the cache block back to memory reduces both the time taken for the write operation in the processor and the memory bandwidth to write the data to memory.
Accordingly, a network services processor includes a input/output bridge that avoids unnecessary memory updates when cache blocks storing processed packet data are no longer required, that is, buffers in memory (corresponding to the cache blocks in cache) are freed. Instead of writing the cache block back to memory, only the dirty bit for the selected cache block is cleared, thus avoiding these wasteful write-backs from cache to memory.
A network services processor includes a plurality of processors and a coherent shared memory. The coherent memory includes a cache and a memory and is shared by the plurality of processors. An input/output bridge is coupled to the plurality of processors and the cache. The input/output bridge monitors requests to free a buffer in memory (that is, a buffer that has been allocated for storing packet data) to avoid writing a modified cache block in the cache back to the buffer.
Upon detecting a request to free the block stored in cache memory, the input/output bridge issues a command to clear a dirty bit associated with the cache block. A cache controller may be coupled to the plurality of processors, the cache and the input/output bridge. The cache controller stores the dirty bit associated with the block and clears the dirty bit upon receiving the command from the input/output bridge.
The input/output bridge may also include a don't write back queue which stores commands to be issued to the cache controller. The input/output bridge may include a free queue that stores requests to free blocks to be added to a free pool. The network services processor may also include a plurality of processing units coupled to the input/output bridge. The input/output bridge stores packets to be transferred between the processing units and the coherent shared memory in which packets are stored for processing by the processors.
The network services processor may also include a memory allocator that provides free lists of blocks in shared coherent memory for storing received packets.
The foregoing and other objects, features and advantages of the invention will be apparent from the following more particular description of preferred embodiments of the invention, as illustrated in the accompanying drawings in which like reference characters refer to the same parts throughout the different views. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention.
A description of preferred embodiments of the invention follows.
The network services processor 100 processes Open System Interconnection network L2-L7 layer protocols encapsulated in received packets. As is well-known to those skilled in the art, the Open System Interconnection (OSI) reference model defines seven network protocol layers (L1-7). The physical layer (L1) represents the actual interface, electrical and physical that connects a device to a transmission medium. The data link layer (L2) performs data framing. The network layer (L3) formats the data into packets. The transport layer (L4) handles end to end transport. The session layer (L5) manages communications between devices, for example, whether communication is half-duplex or full-duplex. The presentation layer (L6) manages data formatting and presentation, for example, syntax, control codes, special graphics and character sets. The application layer (L7) permits communication between users, for example, file transfer and electronic mail.
The network services processor performs work (packet processing operations) for upper level network protocols, for example, L4-L7. The packet processing (work) to be performed on a particular packet includes a plurality of packet processing operations (pieces of work). The network services processor allows processing of upper level network protocols in received packets to be performed to forward packets at wire-speed. Wire-speed is the rate of data transfer of the network over which data is transmitted and received. By processing the protocols to forward the packets at wire-speed, the network services processor does not slow down the network data transfer rate.
The network services processor 100 includes a plurality of Ethernet Media Access Control interfaces with standard Reduced Gigabyte Media Independent Interface (RGMII) connections to the off-chip PHYs 104a, 104b.
The network services processor 100 receives packets from the Ethernet ports (Gig E) through the physical interfaces PHY 104a, 104b, performs L7-L2 network protocol processing on the received packets and forwards processed packets through the physical interfaces 104a, 104b to another hop in the network or the final destination or through the PCI bus 106 for further processing by a host processor. The network protocol processing can include processing of network security protocols such as Firewall, Application Firewall, Virtual Private Network (VPN) including IP Security (IPSec) and/or Secure Sockets Layer (SSL), Intrusion detection System (IDS) and Anti-virus (AV).
A DRAM controller in the network services processor 100 controls access to an external Dynamic Random Access Memory (DRAM) 108 that is coupled to the network services processor 100. The DRAM 108 stores data packets received from the PHYs interfaces 104a, 104b or the Peripheral Component Interconnect Extended (PCI-X) interface 106 for processing by the network services processor 100. In one embodiment, the DRAM interface supports 64 or 128 bit Double Data Rate II Synchronous Dynamic Random Access Memory (DDR II SDRAM) operating up to 800 MHz.
A boot bus 110 provides the necessary boot code which is stored in flash memory 112 and is executed by the network services processor 100 when the network services processor 100 is powered-on or reset. Application code can also be loaded into the network services processor 100 over the boot bus 110, from a device 114 implementing the Compact Flash standard, or from another high-volume device, which can be a disk, attached via the PCI bus.
The miscellaneous I/O interface 116 offers auxiliary interfaces such as General Purpose Input/Output (GPIO), Flash, IEEE 802 two-wire Management Interface (MDIO), Universal Asynchronous Receiver-Transmitters (UARTs) and serial interfaces.
The network services processor 100 includes another memory controller for controlling Low latency DRAM 118. The low latency DRAM 118 is used for Internet Services and Security applications allowing fast lookups, including the string-matching that may be required for Intrusion Detection System (IDS) or Anti Virus (AV) applications.
In one embodiment, each processor core 202 is a dual-issue, superscalar processor with instruction cache 206, Level 1 data cache 204, and built-in hardware acceleration (crypto acceleration module) 200 for cryptography algorithms with direct access to low latency memory over the low latency memory bus 230.
The network services processor 100 includes a memory subsystem. The memory subsystem includes level 1 data cache memory 204 in each core 202, instruction cache in each core 202, level 2 cache memory 212, a DRAM controller 216 for access to external DRAM memory 108 (
The memory subsystem is architected for multi-core support and tuned to deliver both high-throughput and low-latency required by memory intensive content networking applications. Level 2 cache memory 212 and external DRAM memory 108 (
Frequently used data values stored in DRAM 108 (
A Free Pool Allocator (FPA) 236 maintains pools of pointers to free memory locations (that is, memory that is not currently used and is available for allocation) in DRAM 108 (
In one embodiment, pointers submitted to the free pools are aligned on a 128 byte boundary and each pointer points to at least 128 bytes of free memory. The free size (number of bytes) of memory can differ in each pool and can also differ within the same pool. In one embodiment, the FPA unit 236 stores up to 2048 pointers. Each pool uses a programmable portion of these 2048 pointers, so higher priority pools can be allocated a larger amount of free memory. If a pool of pointers is too large to fit in the Free Pool Allocator (FPA) 236, the Free Pool Allocator (FPA) 236 builds a tree/list structure in level 2 cache 212 or DRAM using freed memory in the pool of pointers to store additional pointers.
The I/O Bridge (IOB) 232 manages the overall protocol and arbitration and provides coherent I/O partitioning. The IOB 232 includes a bridge 238 and a Fetch and Add Unit (FAU) 240. The bridge 238 includes queues for storing information to be transferred between the I/O bus 262, coherent memory bus 234, and the IO units including the packet input unit 214 and the packet output unit 218. The bridge 238 also includes a Don't Write Back (DWB) engine 260 that monitors requests to free memory in order to avoid unnecessary cache updates to DRAM 108 (
Packet Input/Output processing is performed by an interface unit 210a, 210b, a packet input unit (Packet Input) 214 and a packet output unit (PKO) 218. The input controller and interface units 210a, 210b perform all parsing of received packets and checking of results to offload the cores 202.
The packet input unit 214 allocates and creates a work queue entry for each packet. This work queue entry includes a pointer to one or more buffers (blocks) stored in L2 cache 212 or DRAM 108 (
A packet is received by any one of the interface units 210a, 210b through a SPI-4.2 or RGM II interface. A packet can also be received by the PCI interface 224. The interface unit 210a, 210b handles L2 network protocol pre-processing of the received packet by checking various fields in the L2 network protocol header included in the received packet. After the interface unit 210a, 210b has performed L2 network protocol processing, the packet is forwarded to the packet input unit 214. The packet input unit 214 performs pre-processing of L3 and L4 network protocol headers included in the received packet. The pre-processing includes checksum checks for Transmission Control Protocol (TCP)/User Datagram Protocol (UDP) (L3 network protocols).
The Packet order/work (POW) module (unit) 228 queues and schedules work (packet processing operations) for the processor cores 202. Work is defined to be any task to be performed by a core that is identified by an entry on a work queue. The task can include packet processing operations, for example, packet processing operations for L4-L7 layers to be performed on a received packet identified by a work queue entry on a work queue. The POW module 228 selects (i.e. schedules) work for a core 202 and returns a pointer to the work queue entry that describes the work to the core 202.
After the packet has been processed by the cores 202, a packet output unit (PKO) 218 reads the packet data stored in L2 cache 212 or memory (DRAM 108 (FIG. 1)), performs L4 network protocol post-processing (e.g., generates a TCP/UDP checksum), forwards the packet through the interface unit 210a, 210b and frees the L2 cache 212 or DRAM 108 locations used to store the packet by adding pointers to the locations in a pool in the FPA unit 236.
The network services processor 100 also includes application specific co-processors that offload the cores 202 so that the network services processor achieves high-throughput. The application specific co-processors include a DFA co-processor 244 that performs Deterministic Finite Automata (DFA) and a compression/decompression co-processor 208 that performs compression and decompression.
The Fetch and Add Unit (FAU) 240 is a 2 KB register file supporting read, write, atomic fetch-and-add, and atomic update operations. The PCI interface controller 224 has a DMA engine that allows the processor cores 202 to move data asynchronously between local memory in the network services processor and remote (PCI) memory in both directions.
The CMB 234 includes four busses: ADD 300, STORE 302, COMMIT 304, and FILL 306. The ADD bus 300 transfers address and control information to initiate a CMB transaction. The STORE bus 302 transfers the store data associated with a transaction. The COMMIT bus 304 transfers control information that initiates transaction responses from the L2 cache. The FILL bus 306 transfers fill data (cache blocks) from the L2 cache controller and cache 212 to the L1 data cache 204 and reflection data for transfers from a core 202 to the I/O bus 262. The reflection data includes commands/results that are transferred between the I/O Bridge 232 and cores 202. The CMB 234 is a split-transaction highly pipelined bus. For an embodiment with a cache block size of 128 bytes, a CMB transaction transfers a cache block size at a time.
All of the busses in the CMB 234 are decoupled by queues in the L2 cache controller and cache 212 and the bridge 238. This decoupling allows for variable timing between the different operations required to complete different CMB transactions.
Memory requests to coherent memory space initiated by a core 202 or the IOB 232 are directed to the L2 cache controller 212. The IOB 232 initiates memory requests on behalf of I/O units coupled to the IO bus 262.
A fill transaction initiated by a core 202 replicates contents of a cache block in either L1 instruction cache 206 (
A store transaction puts contents of a cache block stored in L1 instruction cache 206 (
A Don't write back command issued by the IOB 232 results in control information and the address of the cache block placed on the ADD bus 300. The L2 cache controller 212 receives the ADD bus information and services the command by clearing a dirty bit in a tag associated with the cache block, if the cache block is present in the L2 cache. The L2 cache controller and cache 212 will be described later in conjunction with
As already discussed in conjunction with
The packet input unit 214 includes a Packet Input Processing (PIP) unit 302 and an Input Packet Data unit (IPD) 400. The packet input unit 214 uses one of the pools of pointers in the FPA unit 236 to store received packet data in level 2 cache or DRAM.
The I/O busses include an inbound bus (IOBI) 308 and an outbound bus (IOBO) 310, a packet output bus (POB) 312, a PKO-specific bus (PKOB) 316 and an input packet data bus (IPDB) 314. The interface unit 210a, 210b places the 64-bit packet segments from the received packets onto the IOBI bus 308. The IPD 400 in the packet input unit 214 latches each 64-bit packet segment from the IOBI bus for processing. The IPD 400 accumulates the 64 bit packet segments into 128-byte cache blocks. The IPD 400 then forwards the cache block writes on the IPDB bus 314. The I/O Bridge 232 forwards the cache block write onto the Coherent Memory Bus (CMB) 234.
A work queue entry is added to a work queue by the packet input unit 214 for each packet arrival. The work queue entry is the primary descriptor that describes work to be performed by the cores. The Packet Order/Work (POW) unit 228 implements hardware work queuing, hardware work scheduling and tag-based synchronization and ordering to queue and schedule work for the cores.
The L2 cache controller 212 also contains internal buffering and manages simultaneous in-flight transactions. The L2 cache controller 212 maintains copies of tags for L1 data cache 204 in each core 202 and initiates invalidations to the L1 data cache 204 in the cores 202 when other CMB sources update blocks in the L1 data cache.
In one embodiment, the L2 cache is 1 MB, 8-way set associative with a 128 byte cache block. In a set associative cache, a cache block read from memory can be stored in a restricted set of blocks in the cache. A cache block is first mapped to a set of blocks and can be stored in any block in the set. For example, in an 8-way set associative cache, there are 32 blocks in a set of blocks and a 128 byte block in memory can replicated in any block in the set of blocks in the cache. The cache controller includes an address tag for each block that stores the block address. The address tag is stored in the L2 tags 410.
The CMB 234 includes write-invalidate coherence support. The data cache 204 in each core is a write-through cache. The L2 cache is write-back and both the data stored in the L2 cache 612 and the tags stored in L2 tags 410 are protected by a Single Error Correction, Double Error Detection Error Correction Code (SECDED ECC).
The L2 cache controller 212 maintains memory reference coherence and returns the latest copy of a block for every fill request, whether the latest copy of the block is in the cache (L1 data cache 204 or L2 data cache 612), in DRAM 108 (
The L2 cache controller 212 has two memory input queues 602 that receive memory transactions from the ADD bus 300: one for transactions initiated by cores 202 and one for transactions initiated by the IOB 232.
The two queues 602 allow the L2 cache controller 212 to give the IOB memory transactions a higher priority than core transactions. The L2 cache controller 212 processes transactions from the queues 602 in one of two programmable arbitration modes, fixed priority or round-robin allowing IOB transactions required to service real-time packet transfers to be processed at a higher priority.
The L2 cache controller 212 also services CMB reflections, that is, non-memory transactions that are necessary to transfer commands and/or data between the cores and the I0B. The L2 cache controller 212 includes two reflection queues 604, 606, that store the ADD/STORE bus information to be reflected. Two different reflection queues are provided to avoid deadlock: reflection queue 604 stores reflections destined to the cores 202, and reflection queue 606 stores reflections destined to the IOB 232 over the FILL bus and COMMIT bus.
The L2 cache controller 212 can store and process up to 16 simultaneous memory transactions in its in-flight address buffer 610. The L2 cache controller 212 can also manage up to 16 in-flight cache victims, and up to four of these victims may reside in the victim data file 608. On a fill transaction, received data is returned from either the L2 cache or DRAM 108 (
All data movement transactions between the L2 cache controller 212 and the DRAM controller 216 are 128 byte, full-cache blocks. The L2 cache controller 212 buffers DRAM controller fills in one or both of two queues: in a DRAM-to-L2 queue 420 for data destined to be written to L2 cache 612, and in a DRAM-to-CMB queue 422 for data destined for the FILL bus 306. The L2 cache controller 212 buffers stores for the DRAM controller in the victim address/data files 414, 608 until the DRAM controller 216 accepts them.
The cache controller buffers all the COMMIT/FILL bus commands needed from each possible source: the two reflection queues 604, 606, fills from L2/DRAM 420, 422, and invalidates 416.
The IOB 232 includes twelve queues 500a-l to store information to be transferred on different buses. There are six queues 500a-f arbitrating to transfer on the ADD/STORE buses of the Coherent Memory Bus (CMB) 234 and five queues 500g-k arbitrating to transfer on the IOBO bus. Another queue 500l queues packet data to be transferred to the PKO 218 (
As previously discussed, when a buffer in memory is added to a free pool in the FPA unit 236, that buffer may also be replicated in a cache block in cache (L1 data cache 204 in a core 202 or L2 cache 612). Furthermore, these cached blocks may store a more current version of the data than stored in the corresponding block in DRAM 108 (
In a write-back cache, when dirty blocks are replaced in the cache, the dirty cache blocks are written back to DRAM to ensure that the data in the block stored in the DRAM is up-to-date. The memory has just been freed and it will not be used until it is re-allocated for processing another packet, so it would be wasteful to write the cache blocks from the level 2 cache back to the DRAM. It is more efficient to clear the dirty bit for any of these blocks that are replicated in the cache to avoid writing the ‘dirty’ cache blocks to DRAM later.
The core freeing the memory executes a store instruction to add the address to pool of free buffers. The store instruction from the core is reflected through reflection queue 606 on FILL bus 306 of the CMB. The IOB 232 can create Don't write back (DWB) CMB commands as a result of the memory free command.
The DWB command results in a Don't Write Back (DWB) coherent memory bus transaction on the ADD bus 300 that results in clearing the dirty bit in the L2 tags 410, if the cache block is present in the L2 cache. This is an ADD-bus only transaction on the coherent memory bus. This architecture allows the DWB engine 260 to be separated from the free pool unit 236. In one embodiment, the DWB engine 260 resides nearer to the cache controller, so less bandwidth is required to issue the DWB commands on the coherent memory bus 234. The Don't write back operation is used to avoid unnecessary writebacks from the L2 cache to DRAM for free memory locations (that is, memory blocks (buffers) in a free memory pool available for allocation).
When a core 202 or I/O unit coupled to the IO bus 262 adds free memory to a pool in the FPA unit 236, it not only specifies the address of the free memory, but also specifies the number of cache blocks for which the DWB engine 260 can send DWB commands to the L2 cache controller. The core or I/O module need not initiate any DWB commands. Rather, the DWB engine 260 automatically creates the DWB commands when it observes the command to add free memory to a pool in the FPA unit 236.
The DWB engine 260 avoids unnecessary cache memory updates when buffers replicated in cache blocks in cache that store processed packets are freed by intercepting memory free requests destined for the free pool allocator (FPA) 236. The IOB 232 intercepts memory free commands arriving from either the cores (via a reflection onto the COMMIT/FILL busses 304, 306) or from other IO units (via the IOBI bus 308). When the DWB engine 260 observes a memory free operation, it intercepts and queues the memory free operation. The free memory is not made available to the FPA unit 236 while the DWB engine 260 is sending DWB commands for the free memory. Later the DWB engine 260 sends all necessary DWB commands for the free memory. After all of the DWB commands are completed/visible, the memory free operation continues by forwarding the request to the FPA unit 236.
The IOB 232 can buffer a limited number of the memory free commands inside the DWB 260. If buffering is available, the IOB intercepts the memory free request until the IOB 232 has finished issuing the CMB DWB commands through the DWB engine 260 to the L2 cache controller queue 500e for the request, and then forwards the request onto the FPA unit 236 (via the IOBO bus 310). It is optional for the IOB 232 to issue the DWB requests. Thus, if buffering is not available in the DWB 232, the DWB engine 260 does not intercept the memory free request, and instead the memory free request is forwarded directly to the FPA unit 236 and no DWB commands are issued.
The memory free requests include a hint indicating the number of DWB Coherent Memory Bus (CMB) transactions that the IOB 232 can issue. Don't Write Back (DWB) commands are issued on the ADD bus 300 in the Coherent Memory Bus (CMB) 234 for free memory blocks so that DRAM bandwidth is not unnecessarily wasted writing the freed cache blocks back to DRAM. The DWB commands are queued on the DWB-to-L2C queue 500e and result in the L2 cache controller 212 clearing the dirty bits for the selected blocks in the L2 tags 410 in the L2 cache memory controller, thus avoiding these wasteful write-backs to DRAM 108 (
Returning to
No further processing of the address is performed, that is the address is not checked against copies of the L1 tags in the “Duplicate Tags” block 412, and the victim address file 414 as would be the case for other in flight addresses.
Returning to
The DWB engine 260 in the IOB 232 starts issuing DWB commands for cache blocks starting at the beginning of the free memory identified by the pointer 604 and marches forward linearly. As the DWB commands consume bandwidth on the CMB, the DWB count should be selected so that DWB commands are only issued for cache blocks that may have been modified.
While this invention has been particularly shown and described with references to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of the invention encompassed by the appended claims.
This application is a continuation of Ser. No. 11/030,010, filed Jan. 5, 2005, which claims the benefit of U.S. Provisional Application No. 60/609,211, filed on Sep. 10, 2004. The entire teachings of the above applications are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4415970 | Swenson et al. | Nov 1983 | A |
4755930 | Wilson, Jr. et al. | Jul 1988 | A |
4780815 | Shiota | Oct 1988 | A |
5091846 | Sachs et al. | Feb 1992 | A |
5119485 | Ledbetter et al. | Jun 1992 | A |
5155831 | Emma et al. | Oct 1992 | A |
5276852 | Callander et al. | Jan 1994 | A |
5347648 | Stamm et al. | Sep 1994 | A |
5404482 | Stamm et al. | Apr 1995 | A |
5404483 | Stamm et al. | Apr 1995 | A |
5408644 | Schneider et al. | Apr 1995 | A |
5432918 | Stamm | Jul 1995 | A |
5590368 | Heeb et al. | Dec 1996 | A |
5619680 | Berkovich et al. | Apr 1997 | A |
5623627 | Witt | Apr 1997 | A |
5623633 | Zeller et al. | Apr 1997 | A |
5737547 | Zuravleff et al. | Apr 1998 | A |
5737750 | Kumar et al. | Apr 1998 | A |
5742840 | Hansen et al. | Apr 1998 | A |
5754819 | Lynch et al. | May 1998 | A |
5794060 | Hansen et al. | Aug 1998 | A |
5794061 | Hansen et al. | Aug 1998 | A |
5809321 | Hansen et al. | Sep 1998 | A |
5822603 | Hansen et al. | Oct 1998 | A |
5860158 | Pai et al. | Jan 1999 | A |
5890217 | Kabemoto et al. | Mar 1999 | A |
5893141 | Kulkarni | Apr 1999 | A |
5895485 | Loechel et al. | Apr 1999 | A |
5897656 | Vogt et al. | Apr 1999 | A |
5991855 | Jeddeloh et al. | Nov 1999 | A |
6009263 | Golliver et al. | Dec 1999 | A |
6018792 | Jeddeloh et al. | Jan 2000 | A |
6021473 | Davis et al. | Feb 2000 | A |
6026475 | Woodman | Feb 2000 | A |
6065092 | Roy | May 2000 | A |
6070227 | Rokicki | May 2000 | A |
6125421 | Roy | Sep 2000 | A |
6134634 | Marshall, Jr. et al. | Oct 2000 | A |
6188624 | Zheng | Feb 2001 | B1 |
6226715 | Van Der Wolf et al. | May 2001 | B1 |
6279080 | DeRoo | Aug 2001 | B1 |
6408365 | Hosomi | Jun 2002 | B1 |
6438658 | Baliga et al. | Aug 2002 | B1 |
6526481 | Shen et al. | Feb 2003 | B1 |
6546471 | Tarui et al. | Apr 2003 | B1 |
6560680 | Meyer | May 2003 | B2 |
6563818 | Sang et al. | May 2003 | B1 |
6571320 | Hachmann | May 2003 | B1 |
6587920 | Mekhiel | Jul 2003 | B2 |
6598136 | Norrod et al. | Jul 2003 | B1 |
6622214 | Vogt et al. | Sep 2003 | B1 |
6622219 | Tremblay et al. | Sep 2003 | B2 |
6643745 | Palanca et al. | Nov 2003 | B1 |
6647456 | Van Dyke et al. | Nov 2003 | B1 |
6654858 | Asher et al. | Nov 2003 | B1 |
6665768 | Redford | Dec 2003 | B1 |
6718457 | Tremblay et al. | Apr 2004 | B2 |
6725336 | Cherabuddi | Apr 2004 | B2 |
6754810 | Elliott et al. | Jun 2004 | B2 |
6757784 | Lu et al. | Jun 2004 | B2 |
6785677 | Fritchman | Aug 2004 | B1 |
6924810 | Tischler | Aug 2005 | B1 |
7055003 | Cargnoni et al. | May 2006 | B2 |
7093153 | Witek et al. | Aug 2006 | B1 |
7209996 | Kohn et al. | Apr 2007 | B2 |
7558925 | Bouchard et al. | Jul 2009 | B2 |
7594081 | Bouchard et al. | Sep 2009 | B2 |
20010037406 | Philbrick et al. | Nov 2001 | A1 |
20010054137 | Eickemeyer et al. | Dec 2001 | A1 |
20020032827 | Nguyen et al. | Mar 2002 | A1 |
20020099909 | Meyer | Jul 2002 | A1 |
20020112129 | Arimilli et al. | Aug 2002 | A1 |
20030056061 | Sherman | Mar 2003 | A1 |
20030065884 | Lu et al. | Apr 2003 | A1 |
20030067913 | Georgiou et al. | Apr 2003 | A1 |
20030105793 | Guttag et al. | Jun 2003 | A1 |
20030110208 | Wyschogrod et al. | Jun 2003 | A1 |
20030115238 | O'Connor et al. | Jun 2003 | A1 |
20030115403 | Bouchard et al. | Jun 2003 | A1 |
20030172232 | Naffziger | Sep 2003 | A1 |
20030212874 | Alderson | Nov 2003 | A1 |
20040010782 | Moritz | Jan 2004 | A1 |
20040012607 | Witt | Jan 2004 | A1 |
20040059880 | Bennett | Mar 2004 | A1 |
20040073778 | Adiletta et al. | Apr 2004 | A1 |
20040250045 | Dowling | Dec 2004 | A1 |
20050114606 | Matick et al. | May 2005 | A1 |
20050138276 | Navada et al. | Jun 2005 | A1 |
20050138297 | Sodani et al. | Jun 2005 | A1 |
20050166038 | Wang et al. | Jul 2005 | A1 |
20050267996 | O'Connor et al. | Dec 2005 | A1 |
20050273563 | Kim et al. | Dec 2005 | A1 |
20050273605 | Saha et al. | Dec 2005 | A1 |
20060059310 | Asher et al. | Mar 2006 | A1 |
20060059314 | Bouchard et al. | Mar 2006 | A1 |
20060059316 | Asher et al. | Mar 2006 | A1 |
20060143396 | Cabot | Jun 2006 | A1 |
20070038798 | Bouchard et al. | Feb 2007 | A1 |
20100306510 | Olson et al. | Dec 2010 | A1 |
Number | Date | Country |
---|---|---|
2378779 | Feb 2003 | GB |
WO 2006031462 | Mar 2006 | WO |
WO 2006031551 | Mar 2006 | WO |
Entry |
---|
Gharachorloo, Kourosh, et al., “Architecture and Design of AlphaServer GS320.” Ninth International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS-IX) (2000). |
Jouppi, Norman P., “Cache Write Policies and Performance,” WRL Research Report 91/12 (1991). |
Handy, Jim. “The Cache memory Book.” 1998. Academic Press, Inc. Second edition. pp. 85-86. |
Handy, Jim. “The Cache memory Book.” 1998. Academic Press, Inc. Second edition. pp. 126-127. |
“Microsoft Computer Dictionary,” 2002. Microsoft Press, Fifth Edition, p. 466. |
“Double Date Rate SDRAMs operate at 4000MHz”, Oct. 14, 2003. |
Stokes, Jon, “A Look at Centrino's Core: The Pentium M” “Instruction decoding and micro-op fusion,” http://arstechnica.com/articles/paedia/cpu/pentium-m.ars/4, pp. 1-4, Feb. 25, 2004. |
Van Riel, R, “Page Replacement in Linux 2.4 Memory Management”, Collective Inc., pp. 1-10. Retrieved from the internet on Jun. 5, 2007; URL: http://web.archive.org/web/2001821013232/http://surriel.com/lectures.linux24-vm.html. |
Number | Date | Country | |
---|---|---|---|
20140317353 A1 | Oct 2014 | US |
Number | Date | Country | |
---|---|---|---|
60609211 | Sep 2004 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11030010 | Jan 2005 | US |
Child | 14159210 | US |