System and method for flexible storage and networking provisioning in large scalable processor installations

Information

  • Patent Grant
  • 9585281
  • Patent Number
    9,585,281
  • Date Filed
    Friday, December 13, 2013
    11 years ago
  • Date Issued
    Tuesday, February 28, 2017
    7 years ago
Abstract
A system and method for provisioning within a system design to allow the storage and IO resources to scale with compute resources are provided.
Description
FIELD

The disclosure relates generally to provisioning within a system design to allow the storage and networking resources to scale with compute resources.


BACKGROUND

Server systems generally provide a fixed number of options. For example, there are a fixed number of PCI Express IO slots and a fixed number of hard drive bays, which often are delivered empty as they provide future upgradability. The customer is expected to gauge future needs and select a server chassis category that will serve present and future needs. Historically, and particularly with x86-class servers, predicting the future needs has been achievable because product improvements from one generation to another have been incremental.


With the advent of scalable servers, the ability to predict future needs has become less obvious. For example, in the class of servers within a 2U chassis, it is possible to install 120 compute nodes in an incremental fashion. Using this server as a data storage device, the user may require only 4 compute nodes, but may desire 80 storage drives. Using the same server as a pure compute function focused on analytics, the user may require 120 compute nodes and no storage drives. The nature of scalable servers lends itself to much more diverse applications which require diverse system configurations. As the diversity increases over time, the ability to predict the system features that must scale becomes increasingly difficult.


An example of a typical server system is shown in FIG. 1. The traditional server system has fixed areas for 24 hard drives along its front surface and a fixed area for compute subsystem (also called motherboard) and a fixed area for IO expansion (PCI slots). This typical server system does not provide scalability of the various computer components. Thus, it is desirable to create a system and method to scale storage and networking within a server system and it is to this end that this disclosure is directed. The benefit of this scalability is a much more flexible physical system that fits many user applications.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates a traditional server system, depicting fixed areas for 24 hard drives along its front surface and a fixed area for compute subsystem (also called motherboard) and a fixed area for IO expansion (PCI slots).



FIG. 2 illustrates an exemplary system with multiple slots that can house a compute module, a storage module, or an IO module.



FIG. 3 illustrates an exemplary compute module.



FIGS. 4a
1 and 4a2 are a side view and a top view, respectively, of an exemplary storage module which implements industry standard 2.5″ hard drives or SSDs (solid state drives).



FIG. 4b illustrates an exemplary storage module which implements SATA SSD modules.



FIG. 4c illustrates an exemplary storage module which implements mSATA SSD modules.



FIG. 5 illustrates an exemplary IO module.



FIG. 6 illustrates an exemplary hybrid module.



FIG. 7 illustrates a module block (or super module) made up of an integrated collection of modules connected together by way of a private interconnect.



FIG. 8a illustrates an example of how the exemplary system can be populated specifically for high compute applications which require no local storage.



FIG. 8b illustrates an example of how the exemplary system can be populated with a 1:1 ratio of mix of compute and storage. These are useful, for example, for Hadoop applications.



FIG. 8c illustrates another example of how the exemplary system can be populated specifically for storage applications.



FIG. 8d illustrates an example of a straddle slot. For long chassis', a practical limit is reached on system board size. The center columns of slots straddle across system boards.



FIG. 8e illustrates the use of straddle slots in systems with a much larger system board area.





DETAILED DESCRIPTION OF ONE OR MORE EMBODIMENTS

The disclosure is particular applicable to a 2U chassis which is the most widely favored form factor for PC-class servers. The concepts herein apply to any chassis form factor, such as tower and rack chassis' of varying customary sizes and any unconventional form. For example, FIG. 8e shows an unconventional form factor, the sliding door, which relies on rack rails at the top and bottom of a server rack, rather than left and right sides as used by conventional rack chassis'. The sliding door approach expands the usable space for system boards, but at the same time, it creates a new interconnect problem between system boards that should be solved by the flexible provisioning concepts herein.


Computer architecture have various components and those components can be categorized in three categories: compute, storage, and IO wherein the compute category may include computing related or processor components, the storage category are storage type devices and IO are input/output components of the computer architecture. Each category can be further subdivided, and each category can be defined to contain certain element types. For example, compute can be subdivided into an ALU, cache, system memory, and local peripherals. Also for example, the storage category can contain element types of hard drives, solid state storage devices, various industry-standard form factors, or non-standard devices. For this disclosure, the component level (compute, storage, IO) are used with the understanding that each component has dimensions and attributes to which the same concepts may be applied.


The system and method of the disclosure allow the same physical space to be used by any of the computer components: compute devices, storage devices, or IO devices. This provides the greatest flexibility in configuration of systems for different applications. In addition, devices within the computer system that support all three components, such as power supplies and fans, will be assumed to be stationary for simplicity in the examples provided. It is understood that these support devices do not have to be stationary, depending on the goals in differentiation of the system design, meaning that they also can scale as needed.


In this example, a “slot” consists of physical connectors and a defined volume of space above these connectors. In one implementation, two PCI Express x16 connectors are used, along with a volume of 10″ length by 2.7″ height by 1″ width. This volume is selected based on associated component heights, the restrictions of a 2U chassis, and a length driven by the PCB space required to accommodate this implementation. It is understood that other connector types can be used, depending on the signaling frequency and quantity of pins required. It is understood that other volumes can be used, depending on the physical constraints that are acceptable for the application. The connector pin definitions are critical to accommodate the many needs of the computer components, both in power delivery and bandwidth of the electrical interfaces. FIG. 2 depicts the resulting example system 20 that has one or more fixed locations 22 in the system for fans, one or more fixed locations 24 for the power supplies, and one or more slots 26 (30 slots in this example) for processors, storage or IO components of the system in which


An exemplary compute module 30 is shown in FIG. 3. In support of the principle of scaling, the compute module 30 has one or more nodes, such as four nodes 32-38 in this example. Each node consists of a highly integrated SOC (System On Chip) 40, associated DIMM 42 for system memory, nonvolatile memory (NAND) 44 for local storage space, one or more known SATA channels 46 for connectivity to storage components and other necessary small devices which are necessary for general functions of the node (EEPROMs, boot flash memory, sensors, etc). The four nodes 32-38 have local IO connections to each other, which provide intercommunication and redundancy if an external IO connection fails. Each of the nodes runs an independent operating system, although as another example, a cache-coherent compute module is possible which would run one instance of an operating system on each node.


Examples of storage modules 50 that may be used in the system are shown in FIGS. 4a, 4b, and 4c. FIGS. 4a1 and 4a2 illustrate a storage module that leverages the existing industry-standard 2.5″ drive form factor for hard drives (defined to contain spinning mechanical platters which store data) or for solid state drives (defined to have no moving parts and uses integrated circuits for its storage media). In this example, it is possible to use a printed circuit board (PCB) card edge connector for power delivery and/or data delivery using the necessary IO standard, such as SATA or SAS. The IO standard selected is purely a convenience based on support by the implemented devices. Any IO protocol can be routed through this card edge connector as long as the mechanical interface can support the necessary signaling frequency. Alternatively, directly connecting the IO for data delivery to the drive provides further flexibility in system configuration.


In FIG. 4a1, a printed circuit board 52 is shown to which power/data connectors and voltage regulators are integrated for connection to subsequently attached storage devices. The storage modules also have one or more connectors 54, such as SATA power connectors, and power cables to connect power from PCB power rails to the attached storage media (in this case, SATA 2.5″ mechanical spindle hard drives). In this example, these cables are not needed for SATA SSD nor mSATA. The storage module may also have stand-offs 55 that mount the 2.5″ SATA HDD to the blue mounting holes in 4a2. The storage module also has the SATA data cable 56 which do not convey power.


In FIG. 4a2, the storage module has a set of SATA power/data connector 56 that are another method of attaching a hard drive to the PCB. The storage module in FIG. 4s2 may also have one or more mounting holes 57 for the standoffs 55 shown in FIG. 4a1. They also include holes used for standard manufacturing of the PCB assembly.



FIG. 4b depicts a storage module that implements an industry-standard 22-pin SATA connector and interface, along with mechanical support features, to support SATA SSD modules per the JEDEC MO-297 standard. FIG. 4c depicts a storage module that implements an industry-standard x1 PCI connector, along with mechanical support features to support the mSATA modules per the JEDEC MO-300 standard.


The example in FIG. 4c demonstrates an opportunity to expand beyond the industry standard to maximize the benefit of a storage module that can be very close to its associated compute module. The reuse of an x1 PCI connector for the mSATA module left many pins unused, as the JEDEC standard had need for only one SATA channel through this interface. In fact, there is space for 5 additional SATA channels, even when allocating pins for sufficient grounding. This allows up to 6 SATA channels, each with smaller memories, as opposed to one SATA channel with one large memory block, although both scenario's can result in the same total storage space. The advantage of the multiple SATA channels is increased interface bandwidth, created by the possibility of parallel access to memory. Given that the operating system can stripe across multiple physical disks to create a single logical disk, the net change is a boost in SATA interface performance. Thus, mSATA modules with greater than one SATA channel can provide a new solution to IO bottlenecks to disks.


An exemplary IO module 60 for the system is shown in FIG. 5. Unlike a Network Interface Controller (NIC) that would plug into a conventional server and tie into its operating system, this IO module 60 connects to the infrastructural IO of the system at its edge connectors 62 and provides a translation 64 (using an IO translation circuit) from the internal IO protocol to an external IO protocol, such as Ethernet. The IO module 60 operates independent of any particular operating system of any node. The IO module 60 can support one or many external IO ports, and can take on a form factor that is suitable for a particular chassis design. The benefit of modularity allows the quantity of IO modules to be determined by the bandwidth requirement for data traversing from this system to/from others.


An exemplary hybrid module 70 is shown in FIG. 6, demonstrating that a combination of compute 30, storage 50, and IO 60 concepts can be implemented on a single module that are then incorporated into the system.



FIG. 7 illustrates a module block (or super module) 100 made up of an integrated collection of modules 70 connected together by way of a private interconnect 102.


With the compute, storage, and IO module concepts described above, exemplary systems of FIG. 8 are now described. FIGS. 8a, 8b, and 8c depict different system configurations to address the basic categories of compute-intensive applications, Hadoop applications, and storage applications respectively. Of course, many other combinations of modules are possible to form the recipe needed for specific applications. As shown, the module form factor is kept consistent for convenience, but when required, it can change also, as shown by the IO module labeled “Network”. These degrees of flexibility allow creation of a family of modules that can be mixed and matched according to software application needs, with very little volume within the chassis tied to dedicated purposes. For example, FIG. 8a shows a system 20 that has the fans 22 and power supplies 24 and a plurality of compute modules 30 for a compute intensive system. In FIG. 8b, the system 20 has the same form factor and the fans and power supplies, but the slots 26 are filled with a combination of compute modules 20 and storage modules 50 as shown for a system that requires more storage than the system in FIG. 8a. FIG. 8c illustrates a system 20 has the same form factor and the fans and power supplies, but the slots 26 are filled a few compute modules 20 and many more storage modules 50 as shown for a system that requires more storage than computing power than the systems in FIGS. 8a and 8b.



FIG. 8d expands on the system 20 concepts by considering a chassis that is particularly long, such that the system board size is larger than the practical limit allowed by PCB fabrication factories. Typical PCB panel sizes are 18″×24″ or 24″×24″, although panels up to 30″ are also available with limited sources. Given a typical 2U chassis that fits in a 19″ wide rack, the 18″×24″ PCB panel is the preferred size for most server motherboards today. To expand beyond the 24″ limit, board-to-board connectors must be used to interconnect two assemblies. When high speed signaling must pass between the two assemblies, a relatively expensive interconnect solution must be implemented, such as FCI AirMax connectors. The use of these connectors complicates the electrical design by adding signal integrity considerations and complicates the mechanical design due to the volume required for these connectors. Alternatively, the two system boards do not need to be directly connected at all, relying instead on the IO fabric within a Compute module to traverse data between them, called a “straddle slot”. In FIG. 8d, the left system board might be aligned based on controlled mounting points, while the right system board might be designed to “float” on its mounting points such that installed modules can control the alignment of associated edge connectors.



FIG. 8e breaks away from the 2U chassis example with an exemplary vertical system 20 that greatly expands the area possible for system boards. Each section on rails is referred to as a “vertical chassis”. The black dashed lines represent module slots. Note the angled slot orientation enhances air flow due to natural convection, without the consequence of undue heat build-up caused in true vertical chimney rack designs. The straddle slot concept can be employed here to avoid the expense and space requirements of board-to-board high speed connectors. Power and cooling are not shown, as it is self-evident that space in the enclosure can be dedicated to these as needed.


While the foregoing has been with reference to a particular embodiment of the invention, it will be appreciated by those skilled in the art that changes in this embodiment may be made without departing from the principles and spirit of the disclosure, the scope of which is defined by the appended claims.

Claims
  • 1. A method for building a scalable system, the method comprising: providing a chassis having a predetermined physical form factor, wherein the chassis has a plurality of slots into which modules are placed;providing compute modules, storage modules, and input/output (I/O) modules, all of which are housed within the plurality of slots; anddetermining one or more of the compute modules, one or more of the storage modules, and one or more of the I/O modules to form the scalable system based on a desired computing power, a desired storage power, and a desired input/output power for the scalable system, wherein a quantity of the one or more I/O modules is determined based, at least in part, on a bandwidth requirement for data traversing from and to the scalable system.
  • 2. The method of claim 1, further comprising providing one or more support devices within the chassis that support the one or more compute modules, the one or more storage modules, and the one or more I/O modules housed within the chassis.
  • 3. The method of claim 2, wherein the one or more support devices comprise a fan or a power supply.
  • 4. The method of claim 1, wherein the plurality of slots each comprise a set of physical connectors and a volume of space.
  • 5. The method of claim 4, wherein the set of physical connectors comprises one or more peripheral component interconnect express (PCIe) connectors.
  • 6. The method of claim 5, wherein the set of physical connectors comprises two PCI Express x16 connectors, and wherein the volume of space is 10 inches in length by 2.7 inches in height by 1 inch in width.
  • 7. The method of claim 1, wherein the chassis is a 2U chassis or a vertical chassis.
  • 8. The method of claim 1, wherein each compute module further comprises one or more compute nodes.
  • 9. The method of claim 8, wherein each compute node has a system on chip, a set of system memory that is accessible by the system on chip, and a local storage space for the system on chip.
  • 10. The method of claim 1, wherein each storage module comprises a 2.5 inch cased serial advanced technology attachment (SATA) drive, a caseless SATA solid state device (SSD), or a Modular SATA (mSATA) SSD.
  • 11. The method of claim 1, wherein each I/O module comprises a set of connectors and a translation circuit that translates between I/O protocols.
  • 12. The method of claim 1, further comprising providing hybrid modules that are capable of being housed within the plurality of slots, wherein each hybrid module incorporates compute, storage, and I/O functions therein.
  • 13. The method of claim 11, wherein the I/O protocols comprise an internal I/O protocol and an external I/O protocol.
  • 14. The method of claim 1, wherein the scalable system is configured to address Hadoop applications.
  • 15. The method of claim 1, further comprising providing one or more system boards within the chassis.
  • 16. The method of claim 15, wherein the one or more system boards are interconnected by board-to-board connectors.
  • 17. The method of claim 15, wherein the one or more system boards use the one or more compute modules to exchange data.
Parent Case Info

This application is a divisional application and claims the benefit of U.S. patent application Ser. No. 13/284,855 filed on Oct. 28, 2011, the disclosure of which is incorporated herein by reference.

US Referenced Citations (352)
Number Name Date Kind
5451936 Yang et al. Sep 1995 A
5594908 Hyatt Jan 1997 A
5623641 Kadoyashiki Apr 1997 A
5781187 Gephardt et al. Jul 1998 A
5901048 Hu May 1999 A
5908468 Hartmann Jun 1999 A
5968176 Nessett et al. Oct 1999 A
5971804 Gallagher et al. Oct 1999 A
6055618 Thorson Apr 2000 A
6141214 Ahn Oct 2000 A
6181699 Crinion et al. Jan 2001 B1
6192414 Horn Feb 2001 B1
6198741 Yoshizawa et al. Mar 2001 B1
6314487 Hahn et al. Nov 2001 B1
6314501 Gulick et al. Nov 2001 B1
6373841 Goh et al. Apr 2002 B1
6442137 Yu et al. Aug 2002 B1
6446192 Narasimhan et al. Sep 2002 B1
6452809 Jackson et al. Sep 2002 B1
6507586 Satran et al. Jan 2003 B1
6556952 Magro Apr 2003 B1
6574238 Thrysoe Jun 2003 B1
6711691 Howard et al. Mar 2004 B1
6766389 Hayter et al. Jul 2004 B2
6813676 Henry et al. Nov 2004 B1
6816750 Klaas Nov 2004 B1
6842430 Melnik Jan 2005 B1
6857026 Cain Feb 2005 B1
6963926 Robinson Nov 2005 B1
6963948 Gulick Nov 2005 B1
6977939 Joy et al. Dec 2005 B2
6988170 Barroso et al. Jan 2006 B2
6990063 Lenoski et al. Jan 2006 B1
7020695 Kundu et al. Mar 2006 B1
7032119 Fung Apr 2006 B2
7080078 Slaughter et al. Jul 2006 B1
7080283 Songer et al. Jul 2006 B1
7095738 Desanti Aug 2006 B1
7119591 Lin Oct 2006 B1
7143153 Black et al. Nov 2006 B1
7165120 Giles et al. Jan 2007 B1
7170315 Bakker et al. Jan 2007 B2
7180866 Chartre et al. Feb 2007 B1
7203063 Bash et al. Apr 2007 B2
7257655 Burney et al. Aug 2007 B1
7263288 Islam Aug 2007 B1
7274705 Chang et al. Sep 2007 B2
7278582 Siegel et al. Oct 2007 B1
7310319 Awsienko et al. Dec 2007 B2
7325050 O'Connor et al. Jan 2008 B2
7337333 O'Conner et al. Feb 2008 B2
7340777 Szor Mar 2008 B1
7353362 Georgiou et al. Apr 2008 B2
7382154 Ramos et al. Jun 2008 B2
7386888 Liang et al. Jun 2008 B2
7418534 Hayter et al. Aug 2008 B2
7437540 Paolucci et al. Oct 2008 B2
7447147 Nguyen et al. Nov 2008 B2
7447197 Terrell et al. Nov 2008 B2
7466712 Makishima et al. Dec 2008 B2
7467306 Cartes et al. Dec 2008 B2
7467358 Kang et al. Dec 2008 B2
7502884 Shah et al. Mar 2009 B1
7519843 Buterbaugh et al. Apr 2009 B1
7555666 Brundridge et al. Jun 2009 B2
7583661 Chaudhuri Sep 2009 B2
7586841 Vasseur Sep 2009 B2
7596144 Pong Sep 2009 B2
7599360 Edsall et al. Oct 2009 B2
7606225 Xie et al. Oct 2009 B2
7606245 Ma et al. Oct 2009 B2
7616646 Ma et al. Nov 2009 B1
7620057 Aloni et al. Nov 2009 B1
7644215 Wallace et al. Jan 2010 B2
7657677 Huang et al. Feb 2010 B2
7657756 Hall Feb 2010 B2
7660922 Harriman Feb 2010 B2
7664110 Lovett et al. Feb 2010 B1
7673164 Agarwal Mar 2010 B1
7710936 Morales Barroso May 2010 B2
7719834 Miyamoto et al. May 2010 B2
7721125 Fung May 2010 B2
7751433 Dollo et al. Jul 2010 B2
7760720 Pullela et al. Jul 2010 B2
7761687 Blumrich et al. Jul 2010 B2
7783910 Felter et al. Aug 2010 B2
7791894 Bechtolsheim Sep 2010 B2
7792113 Foschiano et al. Sep 2010 B1
7796399 Clayton et al. Sep 2010 B2
7801132 Ofek et al. Sep 2010 B2
7802017 Uemura et al. Sep 2010 B2
7805575 Agarwal et al. Sep 2010 B1
7831839 Hatakeyama Nov 2010 B2
7840703 Arimilli et al. Nov 2010 B2
7865614 Lu et al. Jan 2011 B2
7925795 Tamir et al. Apr 2011 B2
7934005 Fascenda Apr 2011 B2
7970929 Mahalingaiah Jun 2011 B1
7975110 Spaur et al. Jul 2011 B1
7991817 Dehon et al. Aug 2011 B2
7991922 Hayter et al. Aug 2011 B2
7992151 Warrier et al. Aug 2011 B2
8019832 De Sousa et al. Sep 2011 B2
8060760 Shetty et al. Nov 2011 B2
8060775 Sharma et al. Nov 2011 B1
8082400 Chang et al. Dec 2011 B1
8108508 Goh et al. Jan 2012 B1
8122269 Houlihan et al. Feb 2012 B2
8132034 Lambert et al. Mar 2012 B2
8155113 Agarwal Apr 2012 B1
8156362 Branover et al. Apr 2012 B2
8165120 Maruccia et al. Apr 2012 B2
8170040 Konda May 2012 B2
8180996 Fullerton et al. May 2012 B2
8189612 Lemaire et al. May 2012 B2
8194659 Ban Jun 2012 B2
8199636 Rouyer et al. Jun 2012 B1
8205103 Kazama et al. Jun 2012 B2
8379425 Fukuoka et al. Feb 2013 B2
8397092 Karnowski Mar 2013 B2
8407428 Cheriton et al. Mar 2013 B2
8504791 Cheriton et al. Aug 2013 B2
RE44610 Krakirian et al. Nov 2013 E
8599863 Davis Dec 2013 B2
8684802 Gross et al. Apr 2014 B1
8738860 Griffin et al. May 2014 B1
8745275 Ikeya et al. Jun 2014 B2
8745302 Davis et al. Jun 2014 B2
8782321 Harriman et al. Jul 2014 B2
8812400 Faraboschi et al. Aug 2014 B2
8824485 Biswas et al. Sep 2014 B2
8854831 Arnouse Oct 2014 B2
8903964 Breslin et al. Dec 2014 B2
9008079 Davis et al. Apr 2015 B2
9311269 Davis et al. Apr 2016 B2
20010046227 Matsuhira et al. Nov 2001 A1
20020004912 Fung Jan 2002 A1
20020040391 Chaiken et al. Apr 2002 A1
20020083352 Fujimoto et al. Jun 2002 A1
20020097732 Worster et al. Jul 2002 A1
20020107903 Richter et al. Aug 2002 A1
20020124128 Qiu Sep 2002 A1
20020159452 Foster et al. Oct 2002 A1
20020161917 Shapiro et al. Oct 2002 A1
20020172205 Tagore-Brage et al. Nov 2002 A1
20020186656 Vu Dec 2002 A1
20020194412 Bottom Dec 2002 A1
20030007493 Oi et al. Jan 2003 A1
20030033547 Larson et al. Feb 2003 A1
20030041266 Ke et al. Feb 2003 A1
20030076832 Ni Apr 2003 A1
20030093255 Freyensee et al. May 2003 A1
20030093624 Arimilli et al. May 2003 A1
20030110262 Hasan et al. Jun 2003 A1
20030140190 Mahony et al. Jul 2003 A1
20030158940 Leigh Aug 2003 A1
20030159083 Fukuhara et al. Aug 2003 A1
20030172191 Williams Sep 2003 A1
20030188083 Kumar et al. Oct 2003 A1
20030193402 Post et al. Oct 2003 A1
20030202520 Witkowski et al. Oct 2003 A1
20030231624 Alappat et al. Dec 2003 A1
20040013113 Singh et al. Jan 2004 A1
20040017806 Yazdy et al. Jan 2004 A1
20040017808 Forbes et al. Jan 2004 A1
20040030938 Barr et al. Feb 2004 A1
20040068676 Larson et al. Apr 2004 A1
20040111612 Choi et al. Jun 2004 A1
20040141521 George Jul 2004 A1
20040165588 Pandya Aug 2004 A1
20040210693 Zeitler et al. Oct 2004 A1
20040215864 Arimilli et al. Oct 2004 A1
20040215991 McAfee et al. Oct 2004 A1
20040267486 Percer et al. Dec 2004 A1
20050015378 Gammel et al. Jan 2005 A1
20050018604 Dropps et al. Jan 2005 A1
20050018606 Dropps et al. Jan 2005 A1
20050018663 Dropps et al. Jan 2005 A1
20050021606 Davies et al. Jan 2005 A1
20050030954 Dropps et al. Feb 2005 A1
20050033742 Kamvar et al. Feb 2005 A1
20050033890 Lee Feb 2005 A1
20050044195 Westfall Feb 2005 A1
20050077921 Percer et al. Apr 2005 A1
20050105538 Perera et al. May 2005 A1
20050141424 Lim et al. Jun 2005 A1
20050228852 Santos et al. Oct 2005 A1
20050240688 Moerman et al. Oct 2005 A1
20060002311 Iwanaga et al. Jan 2006 A1
20060013218 Shore et al. Jan 2006 A1
20060029053 Roberts et al. Feb 2006 A1
20060090025 Tufford et al. Apr 2006 A1
20060136570 Pandya Jun 2006 A1
20060140211 Huang et al. Jun 2006 A1
20060174342 Zaheer et al. Aug 2006 A1
20060179241 Clark et al. Aug 2006 A1
20060236371 Fish Oct 2006 A1
20060248359 Fung Nov 2006 A1
20060259734 Sheu et al. Nov 2006 A1
20060265609 Fung Nov 2006 A1
20070006001 Isobe et al. Jan 2007 A1
20070076653 Park et al. Apr 2007 A1
20070094486 Moore et al. Apr 2007 A1
20070109968 Hussain et al. May 2007 A1
20070130397 Tsu Jun 2007 A1
20070174390 Silvain et al. Jul 2007 A1
20070180310 Johnson et al. Aug 2007 A1
20070209072 Chen Sep 2007 A1
20070226795 Conti et al. Sep 2007 A1
20070280230 Park Dec 2007 A1
20070286009 Norman Dec 2007 A1
20070288585 Sekiguchi et al. Dec 2007 A1
20080013453 Chiang et al. Jan 2008 A1
20080040463 Brown et al. Feb 2008 A1
20080052437 Loffink et al. Feb 2008 A1
20080059782 Kruse et al. Mar 2008 A1
20080075089 Evans et al. Mar 2008 A1
20080089358 Basso et al. Apr 2008 A1
20080104264 Duerk et al. May 2008 A1
20080140771 Vass et al. Jun 2008 A1
20080140930 Hotchkiss Jun 2008 A1
20080159745 Segal Jul 2008 A1
20080162691 Zhang et al. Jul 2008 A1
20080183882 Flynn et al. Jul 2008 A1
20080186965 Zheng et al. Aug 2008 A1
20080199133 Takizawa et al. Aug 2008 A1
20080212273 Bechtolsheim Sep 2008 A1
20080212276 Bottom et al. Sep 2008 A1
20080217021 Lembcke et al. Sep 2008 A1
20080222434 Shimizu et al. Sep 2008 A1
20080235443 Chow et al. Sep 2008 A1
20080243634 Dworkin et al. Oct 2008 A1
20080250181 Li et al. Oct 2008 A1
20080259555 Bechtolsheim et al. Oct 2008 A1
20080259788 Wang et al. Oct 2008 A1
20080266793 Lee Oct 2008 A1
20080270599 Tamir et al. Oct 2008 A1
20080288660 Balasubramanian et al. Nov 2008 A1
20080288664 Pettey et al. Nov 2008 A1
20080288683 Ramey Nov 2008 A1
20080301794 Lee Dec 2008 A1
20080310848 Yasuda et al. Dec 2008 A1
20080313369 Verdoorn et al. Dec 2008 A1
20080320161 Maruccia et al. Dec 2008 A1
20090021907 Mann et al. Jan 2009 A1
20090044036 Merkin Feb 2009 A1
20090063443 Arimilli et al. Mar 2009 A1
20090064287 Bagepalli et al. Mar 2009 A1
20090080428 Witkowski et al. Mar 2009 A1
20090097200 Sharma et al. Apr 2009 A1
20090113130 He et al. Apr 2009 A1
20090133129 Jeong et al. May 2009 A1
20090135751 Hodges et al. May 2009 A1
20090135835 Gallatin et al. May 2009 A1
20090158070 Gruendler Jun 2009 A1
20090172423 Song et al. Jul 2009 A1
20090198958 Arimilli et al. Aug 2009 A1
20090204834 Hendin et al. Aug 2009 A1
20090204837 Raval et al. Aug 2009 A1
20090219827 Chen et al. Sep 2009 A1
20090222884 Shaji et al. Sep 2009 A1
20090225751 Koenck et al. Sep 2009 A1
20090235104 Fung Sep 2009 A1
20090248943 Jiang et al. Oct 2009 A1
20090259863 Williams et al. Oct 2009 A1
20090259864 Li et al. Oct 2009 A1
20090265045 Coxe, III Oct 2009 A1
20090271656 Yokota et al. Oct 2009 A1
20090276666 Haley et al. Nov 2009 A1
20090279518 Falk et al. Nov 2009 A1
20090282274 Langgood et al. Nov 2009 A1
20090282419 Mejdrich et al. Nov 2009 A1
20090313390 Ahuja et al. Dec 2009 A1
20100005331 Somasundaram et al. Jan 2010 A1
20100008038 Coglitore Jan 2010 A1
20100008365 Porat Jan 2010 A1
20100026408 Shau Feb 2010 A1
20100040053 Gottumukkula et al. Feb 2010 A1
20100049822 Davies et al. Feb 2010 A1
20100051391 Jahkonen Mar 2010 A1
20100106987 Lambert et al. Apr 2010 A1
20100118880 Kunz et al. May 2010 A1
20100125742 Ohtani May 2010 A1
20100125915 Hall et al. May 2010 A1
20100138481 Behrens Jun 2010 A1
20100158005 Mukhopadhyay et al. Jun 2010 A1
20100161909 Nation et al. Jun 2010 A1
20100165983 Aybay et al. Jul 2010 A1
20100169479 Jeong et al. Jul 2010 A1
20100198972 Umbehocker Aug 2010 A1
20100218194 Dallman et al. Aug 2010 A1
20100220732 Hussain et al. Sep 2010 A1
20100250914 Abdul et al. Sep 2010 A1
20100265650 Chen et al. Oct 2010 A1
20100281246 Bristow et al. Nov 2010 A1
20100299548 Chadirchi et al. Nov 2010 A1
20100308897 Evoy et al. Dec 2010 A1
20100312910 Lin et al. Dec 2010 A1
20100312969 Yamazaki et al. Dec 2010 A1
20100318812 Auradkar et al. Dec 2010 A1
20110023104 Franklin Jan 2011 A1
20110026397 Saltsidis et al. Feb 2011 A1
20110029652 Chhuor et al. Feb 2011 A1
20110058573 Balakavi et al. Mar 2011 A1
20110075369 Sun et al. Mar 2011 A1
20110090633 Rabinovitz Apr 2011 A1
20110103391 Davis et al. May 2011 A1
20110113115 Chang et al. May 2011 A1
20110119344 Eustis May 2011 A1
20110123014 Smith May 2011 A1
20110138046 Bonnier et al. Jun 2011 A1
20110185370 Tamir et al. Jul 2011 A1
20110191514 Wu et al. Aug 2011 A1
20110191610 Agarwal et al. Aug 2011 A1
20110197012 Liao et al. Aug 2011 A1
20110210975 Wong et al. Sep 2011 A1
20110239014 Karnowski Sep 2011 A1
20110271159 Ahn et al. Nov 2011 A1
20110273840 Chen Nov 2011 A1
20110295991 Aida Dec 2011 A1
20110296141 Daffron Dec 2011 A1
20110320690 Petersen et al. Dec 2011 A1
20120011500 Faraboschi et al. Jan 2012 A1
20120020207 Corti et al. Jan 2012 A1
20120050981 Xu et al. Mar 2012 A1
20120054469 Ikeya et al. Mar 2012 A1
20120054511 Brinks et al. Mar 2012 A1
20120081850 Regimbal et al. Apr 2012 A1
20120096211 Davis et al. Apr 2012 A1
20120099265 Reber Apr 2012 A1
20120131201 Matthews et al. May 2012 A1
20120155168 Kim et al. Jun 2012 A1
20120198252 Kirschtein et al. Aug 2012 A1
20120207165 Davis Aug 2012 A1
20120297042 Davis et al. Nov 2012 A1
20130010639 Armstrong et al. Jan 2013 A1
20130024645 Cheriton et al. Jan 2013 A1
20130031331 Cheriton et al. Jan 2013 A1
20130058250 Casado et al. Mar 2013 A1
20130094499 Davis et al. Apr 2013 A1
20130097448 Davis et al. Apr 2013 A1
20130111107 Chang et al. May 2013 A1
20130148667 Hama et al. Jun 2013 A1
20130163605 Chandra et al. Jun 2013 A1
20130290643 Lim et al. Oct 2013 A1
20130290650 Chang et al. Oct 2013 A1
20130318269 Dalal et al. Nov 2013 A1
20140122833 Davis et al. May 2014 A1
20140359044 Davis et al. Dec 2014 A1
20140365596 Kanevsky et al. Dec 2014 A1
20150039840 Chandra et al. Feb 2015 A1
20150103826 Davis Apr 2015 A1
Foreign Referenced Citations (9)
Number Date Country
2005-223753 Aug 2005 JP
2005-536960 Dec 2005 JP
M377621 Apr 2010 TW
201017430 May 2010 TW
WO-2004021641 Mar 2004 WO
WO-2005013143 Feb 2005 WO
WO-2008000193 Jan 2008 WO
WO-2011044271 Apr 2011 WO
WO-2012037494 Mar 2012 WO
Non-Patent Literature Citations (103)
Entry
PCT Search Report and Written Opinion mailed on Mar. 1, 2013 corresponding to the related PCT Patent Application No. US12/61747.
Elghany et al., “High Throughput High Performance NoC Switch,” NORCHIP 2008, Nov. 2008, pp. 237-240.
Grecu et al., “A Scalable Communication-Centric SoC Interconnect Architecture” Proceedings 5th International Symposium on Quality Electronic Design, 2005, pp. 343, 348 (full article included).
Hossain et al., “Extended Butterfly Fat Tree Interconnection (EFTI) Architecture for Network on Chip,” 2005 IEEE Pacific Rim Conference on Communicatinos, Computers and Signal Processing, Aug. 2005, pp. 613-616.
Non-Final Office Action on U.S. Appl. No. 14/106,698, mailed Feb. 12, 2015.
Non-Final Office Action on U.S. Appl. No. 14/334,931, mailed Jan. 5, 2015.
Pande et al., “Design of a Switch for Network on Chip Applications,” May 25-28, 2003 Proceedings of the 2003 International Symposium on Circuits and Systems, vol. 5, pp. V217-V220.
Final Office Action on U.S. Appl. No. 13/692,741, mailed Mar. 11, 2015.
Notice of Allowance on U.S. Appl. No. 13/475,713, mailed Feb. 5, 2015.
Notice of Allowance on U.S. Appl. No. 13/475,722, mailed Feb. 27, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,498, mailed Feb. 23, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,505, mailed Mar. 6, 2015.
Notice of Allowance on U.S. Appl. No. 13/624,731, mailed Mar. 5, 2015.
Reexamination Report on Japanese Application 2012-536877, mailed Jan. 22, 2015 (English Translation not available).
Search Report on EP Application 10827330.1, mailed Feb. 12, 2015.
International Preliminary Report on Patentability for PCT/US2012/061747, mailed Apr. 29, 2014.
Non-Final Office Action on U.S. Appl. No. 13/284,855, mailed Dec. 19, 2013.
Notice of Allowance on U.S. Appl. No. 13/284,855, mailed Jul. 14, 2014.
Advanced Switching Technology Tech Brief, published 2005, 2 pages.
Chapter 1 Overview of the Origin Family Architecture from Origin and Onyx2 Theory of Operations Manual, published 1997, 18 pages.
Cisco MDS 9000 Family Multiprotocol Services Module, published 2006, 13 pages.
Comparing the I2C BUS to the SMBUS, Maxim Integrated, Dec. 1, 2000, p. 1.
Deering, “IP Multicast Extensions for 4.3BSD UNIX and related Systems,” Jun. 1999, 5 pages.
Extended European Search Report for EP 10827330.1, mailed Jun. 5, 2013.
Final Office Action on U.S. Appl. No. 12/889,721, mailed Apr. 17, 2014.
Final Office Action on U.S. Appl. No. 12/794,996, mailed Jun. 19, 2013.
Final Office Action on U.S. Appl. No. 13/475,713, mailed Oct. 17, 2014.
Final Office Action on U.S. Appl. No. 13/475,722, mailed Oct. 20, 2014.
Final Office Action on U.S. Appl. No. 13/527,498, mailed Nov. 17, 2014.
Final Office Action on U.S. Appl. No. 13/624,725, mailed Nov. 13, 2013.
Final Office Action on U.S. Appl. No. 13/624,731, mailed Jul. 25, 2014.
Final Office Action on U.S. Appl. No. 13/705,340, mailed Aug. 2, 2013.
Final Office Action on U.S. Appl. No. 13/705,414, mailed Aug. 9, 2013.
Final Office Action on U.S. Appl. No. 13/624,731, mailed Nov. 12, 2013.
fpga4fun.com,“What is JTAG?”, 2 pages, Jan. 31, 2010.
From AT to BTX: Motherboard Form Factor, Webopedia, Apr. 29, 2005, p. 1.
HP Virtual Connect Traffic Flow—Technology brief, Jan. 2012, 22 pages.
International Preliminary Report on Patentability for PCT/US2009/044200, mailed Nov. 17, 2010.
International Preliminary Report on Patentability for PCT/US2012/038986 issued on Nov. 26, 2013.
International Preliminary Report on Patentability issued on PCT/US12/62608, issued May 6, 2014.
International Search Report and Written Opinion for PCT/US12/38987, mailed Aug. 16, 2012.
International Search Report and Written Opinion for PCT/US12/62608, mailed Jan. 18, 2013.
International Search Report and Written Opinion for PCT/US2010/053227, mailed May 10, 2012.
International Search Report and Written Opinion on PCT/US09/44200, mailed Jul. 1, 2009.
International Search Report and Written Opinion on PCT/US2012/038986, mailed Mar. 14, 2013.
Jansen et al., “SATA-IO to Develop Specification for Mini Interface Connector” Press Release Sep. 21, 2009, Serial ATA3 pages.
Nawathe et al., “Implementation of an 8-Core, 64-Thread, Power Efficient SPARC Server on a Chip”, IEEE Journal of Solid-State Circuits, vol. 43, No. 1, Jan. 2008, pp. 6-20.
Non-Final Action on U.S. Appl. No. 13/728,362, mailed Feb. 21, 2014.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Jul. 2, 2013.
Non-Final Office Action on U.S. Appl. No. 13/475,722, mailed Jan. 17, 2014.
Non-Final Office Action on U.S. Appl. No. 12/794,996, mailed Sep. 17, 2012.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Oct. 11, 2012.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Sep. 29, 2014.
Non-Final Office Action on U.S. Appl. No. 13/234,054, mailed Oct. 23, 2014.
Non-Final Office Action on U.S. Appl. No. 13/453,086, mailed Mar. 12, 2013.
Non-Final Office Action on U.S. Appl. No. 13/475,713, mailed Apr. 1, 2014.
Non-Final Office Action on U.S. Appl. No. 13/527,505, mailed May 8, 2014.
Non-Final Office Action on U.S. Appl. No. 13/527,498, Mailed May 8, 2014.
Non-Final Office Action on U.S. Appl. No. 13/624,725, mailed Jan. 10, 2013.
Non-final office action on U.S. Appl. No. 13/624,731 mailed Jan. 29, 2013.
Non-Final Office Action on U.S. Appl. No. 13/662,759, mailed Nov. 6, 2014.
Non-Final Office Action on U.S. Appl. No. 13/692,741, mailed Sep. 4, 2014.
Non-Final Office Action on U.S. Appl. No. 13/705,286, mailed May 13, 2013.
Non-Final Office Action on U.S. Appl. No. 13/705,340, mailed Mar. 12, 2014.
Non-Final Office Action on U.S. Appl. No. 13/705,340, mailed Mar. 29, 2013.
Non-Final Office Action on U.S. Appl. No. 13/705,414, mailed Apr. 9, 2013.
Non-Final Office Action on U.S. Appl. No. 13/705,428, mailed Jul. 10, 2013.
Notice of Allowance on U.S. Appl. No. 13/453,086, mailed Jul. 18, 2013.
Notice of Allowance on U.S. Appl. No. 13/705,414, mailed Nov. 4, 2013.
Venaas, “IPv4 Multicast Address Space Registry,” 2013, http://www.iana.org/assignments/multicast-addresses/multicast-addresses.xhtml.
Final Office Action on U.S. Appl. No. 13/527,505, mailed Dec. 5, 2014.
Notice of Allowance on U.S. Appl. No. 13/705,340, mailed Dec. 3, 2014.
Hossain et al., “Extended Butterfly Fat Tree Interconnection (EFTI) Architecture for Network on Chip,” 2005 IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Aug. 2005, pp. 613-616.
Final Office Action on U.S. Appl. No. 12/889,721, mailed May 22, 2015.
Final Office Action on U.S. Appl. No. 13/234,054, mailed Apr. 16, 2015.
Final Office Action on U.S. Appl. No. 14/334,931, mailed Jul. 9, 2015.
Non-Final Office Action on U.S. Appl. No. 13/234,054, mailed Aug. 6, 2015.
Non-Final Office Action on U.S. Appl. No. 13/624,725, mailed Apr. 23, 2015.
Non-Final Office Action on U.S. Appl. No. 13/692,741, mailed Jul. 1, 2015.
Non-Final Office Action on U.S. Appl. No. 13/728,308, mailed May 14, 2015.
Non-Final Office Action on U.S. Appl. No. 13/728,428, mailed Jun. 12, 2015.
Non-Final Office Action on U.S. Appl. No. 14/052,723, mailed May 1, 2015.
Office Action on Taiwan Application 101139729, mailed May 25, 2015 (English translation not available).
Final Office Action on U.S. Appl. No. 14/106,698, mailed Aug. 19, 2015.
Notice of Allowance U.S. Appl. No. 13/728,308, mailed Oct. 7, 2015.
Das et al., “Unifying Packet and Circuit Switched Networks,” IEEE Globecom Workshops 2009, Nov. 30 2009, pp. 1-6.
Final Office Action on U.S. Appl. No. 13/624,725 mailed Mar. 10, 2016.
Final Office Action on U.S. Appl. No. 13/728,428 mailed May 6, 2016.
Non-Final Office Action on U.S. Appl. No. 14/725,543 mailed Apr. 7, 2016.
Notice of Allowance on U.S. Appl. No. 14/334,931 mailed May 20, 2016.
Notice of Allowance on U.S. Appl. No. 13/624,725, mailed Mar. 30, 2016.
Notice of Allowance on U.S. Appl. No. 13/662,759 mailed May 10, 2016.
Final Office Action on U.S. Appl. No. 13/234,054, mailed Jan. 26, 2016.
Final Office Action on U.S. Appl. No. 13/662,759, mailed Feb. 22, 2016.
Final Office Action on U.S. Appl. No. 14/052,723, mailed Dec. 3, 2015.
Non-Final Office Action on U.S. Appl. No. 14/334,178 mailed Dec. 18, 2015.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Feb. 24, 2016.
Non-Final Office Action on U.S. Appl. No. 14/334,931 Mailed Dec. 11, 2015.
Notice of Allowance on U.S. Appl. No. 13/692,741 mailed Dec. 4, 2015.
Notice of Allowance on U.S. Appl. No. 13/728,428 mailed Jul. 18, 2016.
Notice of Allowance on U.S. Appl. No. 14/334,178 mailed Jun. 8, 2016.
Non-Final Office Action on U.S. Appl. No. 13/234,054 mailed Oct. 20, 2016.
Non-Final Office Action on U.S. Appl. No. 14/753,948 mailed Nov. 4, 2016.
Related Publications (1)
Number Date Country
20140101932 A1 Apr 2014 US
Divisions (1)
Number Date Country
Parent 13284855 Oct 2011 US
Child 14106697 US