System and method for data center security enhancements leveraging server SOCs or server fabrics

Information

  • Patent Grant
  • 9749326
  • Patent Number
    9,749,326
  • Date Filed
    Wednesday, November 23, 2016
    8 years ago
  • Date Issued
    Tuesday, August 29, 2017
    7 years ago
Abstract
A data center security system and method are provided that leverage server systems on a chip (SOCs) and/or server fabrics. In more detail, server interconnect fabrics may be leveraged and extended to dramatically improve security within a data center.
Description
FIELD

The disclosure relates generally to security aspects for data centers and in particular to data center security enhancements leveraging server systems on a chip (SOCs) or server switch fabrics.


BACKGROUND


FIGS. 1A and 1B show a classic data center network aggregation as is currently well known. FIG. 1A shows a diagrammatical view of a typical network data center architecture 100 wherein top level switches 101a-n are at the tops of racks 102a-n filled with blade servers 107a-n interspersed with local routers 103a-f. Additional storage routers and core switches. 105a-b and additional rack units 108a-n contain additional servers 104e-k and routers 106a-g FIG. 1b shows an exemplary physical view 110 of a system with peripheral servers 111a-bn arranged around edge router systems 112a-h, which are placed around centrally located core switching systems 113. Typically such an aggregation 110 has 1-Gb Ethernet from the rack servers to their top of rack switches, and often 10 Gb Ethernet ports to the edge and core routers. These typical data centers do not have good security.


The idea of network security is well known. The terms used in field of network security may include deep packet inspection (DPI) and intrusion prevention systems (IPS) which are also known as Intrusion Detection and Prevention Systems (IDPS) and are network security appliances that monitor network and/or system activities for malicious activity. The main functions of intrusion prevention systems are to identify malicious activity, log information about said activity, attempt to block/stop activity, and report activity. The network security may also utilize an intrusion detection system (IDS), which is a device or software application that monitors network and/or system activities for malicious activities or policy violations and produces reports to a Management Station.



FIG. 2 shows a typical implementation of an IDS and IPS within a corporate network. In the typical implementation, the IDS is focused on detection, monitoring, and reporting of potential intrusions. As such, the IDS is implemented out-of-line of the core network flow and is not invasive (located outside of the firewall and attached to a DMZ switch as shown in FIG. 2). The IPS adds the capability to prevent and block potential intrusion or undesired network flows and the IPS is implemented in-line of the core network flow.


Thus, it is desirable to provide a data center security system and method that leverage server systems on a chip (SOCs) and/or server fabrics, and it is to this end that the disclosure is directed.





BRIEF DESCRIPTION OF THE DRAWINGS


FIGS. 1A and 1B illustrate a typical data center system;



FIG. 2 shows a typical implementation of an IDS and IPS within a corporate network;



FIG. 3 illustrates a high-level topology of a network aggregating system that may be leveraged for increased security in a data center;



FIG. 4 illustrates a block diagram of an exemplary switch of the network aggregation system that may be leveraged for increased security in a data center;



FIG. 5 illustrates a network aggregation system with a network switch and enhanced security;



FIG. 6 illustrates a four-node server fabric with a network switch and enhanced security; and



FIG. 7 illustrates a small three-node server fabric with a network switch and enhanced security.





DETAILED DESCRIPTION OF ONE OR MORE EMBODIMENTS

The disclosure is particularly applicable to a Calxeda™ server system on a chip and Calxeda™ switch fabrics as illustrated and described below with the security aspects and it is in this context that the disclosure will be described. However, the principles described below can be applied to other server-on-a-chip systems.


A server-on-a-chip (SOC) with packet switch functionality is focused on network aggregation. It contains a layer 2 packet switch, with routing based on source/destination MAC addresses. It further supports virtual local area network (VLAN), with configurable VLAN filtering on domain incoming packets to minimize unnecessary traffic in a domain. The embedded MACs within the SOC do have complete VLAN support providing VLAN capability to the overall SOC without the embedded switch explicitly having VLAN support.



FIG. 3 shows a high-level topology 800 of the network system that illustrates XAUI (a well-known interface standard) connected SoC nodes connected by the switching fabric. Two 10 Gb Ethernet ports Eth0 801a and Eth1 801b come from the top of the tree. Ovals 802a-n are Calxeda™ nodes that comprise at least one computational processors and an embedded switch. Each node may have five XAUI links connected to the internal switch. The switching layers use all five XAUI links for switching. Level 0 leaf nodes 802d, e (i.e., N0n nodes, or Nxy, where x=level and y=item number) only use one XAUI link to attach to the interconnect, leaving four high-speed ports that can be used as XAUI, 10 Gb Ethernet, PCIe, SATA, etc., for attachment to I/O. The vast majority of trees and fat trees have active nodes only as leaf nodes, and the other nodes are pure switching nodes. This approach makes routing much more straightforward. Topology 800 has the flexibility to permit every node to be a combination computational and switch node, or just a switch node. Most tree-type implementations have I/O on the leaf nodes, but topology 800 let the I/O be on any node. In general, placing the Ethernet at the top of the tree (the Ethernet ports) minimizes the average number of hops to the Ethernet.


The system and method also supports a routing using a tree-like or graph topology that supports multiple links per node, where each link is designated as an Up, Down, or Lateral link, or both, within the topology. In addition, each node in the system may be a combination computational/switch node, or just a switch node, and input/outpout (I/O) can reside on any node as described below in more detail. The system may also provide a system with a segmented Ethernet Media Access Control (MAC) architecture which may have a method of re-purposing MAC IP addresses for inside MACs and outside MACs, and leveraging what would normally be the physical signaling for the MAC to feed into the switch. The system may also provide a method of non-spoofing communication, as well as a method of fault-resilient broadcasting, which may have a method of unicast misrouting for fault resilience.


A data center with the Calxeda™ server system on a chip may be implemented using the set of fabric connected nodes with Ethernet uplinks as shown in FIG. 3. Each node may be one or more Calxeda server boxes each of which has at least one Calxeda™ server system on a chip.


The system may also provide a rigorous security between the management processor cores, such that management processors can “trust” one another. In the example node 900 shown in FIG. 4 (which is described below in more detail), there is a management processor core within each SoC (block 906, FIG. 4). The software running on the management processor is trusted because a) the vendor (in this case Calxeda™) has developed and verified the code, b) non-vendor code is not allowed to run on the processor. Maintaining a Trust relationship between the management processors allow them to communicate commands (e.g. reboot another node) or request sensitive information from another node without worrying that a user could spoof the request and gain access to information or control of the system.


Typically the management processor, block 906, is running an embedded OS, while the multiple processor cores represented by block 905 are more typically running a standard operating system, such as Linux. The management processor would typically use one of the Ethernet MACs, in this case block 907, while the main processors, block 905, would utilize the remaining Ethernet MACs, in this case blocks 902 and 903.


Each routing header unit 901, that may be implemented as a processing unit or processor, prepends routing headers to layer 2 Ethernet frames to form a routing frame going into the fabric switch, and removes the routing headers as they leave the switch and enter standard Ethernet MACs. The routing frame is composed of the routing frame header plus the core part of the Ethernet frame, and is structured as shown in Table 1, below:









TABLE 1







Routing Header Prepended to Layer 2 Frame








Routing Frame



Header
Ethernet Frame Packet















RF Header
MAC
MAC
Ethertype/
Payload
CRC32



destination
Source
Length
(data and






padding)









The routing frame header (RF Header) typically consists of the fields shown in Table 2, below:









TABLE 2







Routing Header Fields










Width



Field
(Bits)
Notes












Domain ID
5
Domain ID associated with this packet. 0 indi-




cates that no domain has been specified.


Mgmt
1
Specifies that the packet is allowed on the


Domain

private management domain.


Source Node
12
Source node ID


Source Port
2
0 = MAC0, 1 = MAC1, 2 = MAC_management




processor, 3 = MAC_OUT


Dest Node
12
Destination node ID


Dest Port
2
0 = MAC0, 1 = MAC1, 2 = MAC_management




processor, 3 = MAC_OUT


RF Type
2
Routing Frame Type (0 = Unicast, 1 =




Multicast, 2 = Neighbor




Multicast, 3 = Link Directed)


TTL
6
Time to Live—# of hops that this frame has




existed. Switch will drop packet if the TTL




threshold is exceeded (and notify management




processor of exception).


Broadcast
5
Broadcast ID for this source node for this


ID

broadcast packet.


Checksum

Checksum of the frame header fields.


Total
46
+checksum









The Routing Header processor 901 contains a MAC Lookup CAM (Content Addressable Memory) (MCAM), macAddrLookup, that maps from 6 byte MAC addresses to 12-bit Node IDs, as shown in Table 3, below.









TABLE 3







MAC Address CAM (MCAM)












MAC Lookup CAM Input

MAC Lookup CAM Output













Node Local
MAC Address
Node ID
Port ID







1 bit
6 bytes
12 bits
2 bits










The approach to security domain management in the system and method disclosed here is as follows: Support multiple domain IDs within the fabric. Allow each of the MACs within a node (management processor, MAC0, MAC1, Gateway) to be assigned to a domain ID individually (and tagged with domain 0 if not set). Allow each of the MACs within a node to have a bit indicating access to the management domain. The domain IDs associated with a MAC could only be assigned by the management processor, and could not be altered by the A9. For frames generated by MACs (both inside and outside), the routing frame processor would tag the routing frame with the domain ID and management domain state associated with that MAC. Domains would provide the effect of tunnels or VLANs, in that they keep packets (both unicast and multicast) within that domain, allowing MACs outside that domain to be able to neither sniff or spoof those packets. Additionally, this approach would employ a five-bit domain ID. It would add options to control domain processing, such as, for example, a switch with a boolean per MAC that defines whether packets are delivered with non-defined (i.e., zero) domain ID, or a switch that has a boolean per MAC that defines whether packets are delivered with defined (non-zero) but non-matching domain IDs. A further option in the switch could turn off node encoded MAC addresses per MAC (eliminating another style of potential attack vector). Each of these options described in this paragraph are options that are implemented in the fabric switch, controlled by bits in the control status registers (CSRs) of the fabric switch. Software initializes the CSRs to the desired set of options.


To keep management processor to management processor communication secure, the management domain bit on all management processor MACs could be marked. Generally, the management processor should route on domain 1 (by convention). Such a technique allows all the management processor's to tunnel packets on the management domain so that they cannot be inspected or spoofed by any other devices (inside or outside the fabric), on other VLANs or domains. Further, to provide a secure management LAN, a gateway MAC that has the management domain bit set could be assigned, keeping management packets private to the management processor domain. Additionally, the switch fabric could support “multi-tenant” within itself, by associating each gateway MAC with a separate domain. For example, each gateway MAC could connect to an individual port on an outside router, allowing that port to be optionally associated with a VLAN. As the packets come into the gateway, they are tagged with the domain ID, keeping that traffic private to the MACs associated with that domain across the fabric.


Unicast routing is responsible for routing non-multicast (i.e. unicast) packets to the next node. This is done by utilizing a software computed unicastRoute[ ] next node routing table that provides a vector of available links to get to the destination node.


Server Interconnect Fabric Security


The above server fabric and switch fabric can benefit by enhanced security and a number of techniques to leverage and extend upon server interconnect fabrics that have some or all of the characteristics described above to dramatically improve security within a data center are described. The different embodiments implement “packet processing” which may include a wide range of packet processing including, but not limited to: IDS functionality, IPS functionality, sFlow monitoring (wherein sFlow is a specification for monitoring computer networks set forth in an sFlow specification that is RFC 3176) Packet routing or bridging between networks, Deep packet inspection, Packet logging, Transparent VPN encapsulation, Packet encryption/decryption and/or Packet compression/decompression.


Multi-Tenant Fabric Use Case


In a first embodiment, the server fabric domains are used to enhance security in fabric multi-tenant use case. In particular, there are data centers that host applications and data for multiple clients and networked servers within a single rack may host multiple clients. In the case of servers and nodes connected via interconnect fabrics, one example of which is described above, multiple clients may exist on separate nodes (such as the nodes shown in FIGS. 3 and 4 above) within a single fabric which is a multi-tenant fabric use case.


There are a couple of network security goals in this multi-tenant fabric use case:

    • Client A should have no possible way to inspect data from Client B, including Client B's network traffic.
    • Client A should have no possible way to spoof data to Client B's network. This case specifically covers the case where network packets cannot be hand crafted to look like they came from a Client B node, and routed to a Client B node.


To illustrate this embodiment, FIG. 5 illustrates a network aggregation system with a network switch and enhanced security. As shown in FIG. 5, a number of server nodes, 802a-n, are connected by a server interconnect fabric, there are two gateway nodes, N30 and N31, that serve as Ethernet gateways to the outside Ethernet network and there are two gateway Ethernet ports, 801a and 801b that are connected to a network switch 804, typically a top of rack switch, connecting to two ports on the switch, Port A and Port B.


When Client A's network traffic comes from Port A on the switch and Client B′s network traffic comes from Port B on the switch, a common way for a network engineer to manage this multi client use would be to have a VLAN assigned to Client A and a different VLAN assigned to Client B. To guarantee isolation of Client A's traffic from Client B′s traffic on the fabric, the following techniques (alone or in combination) can be used:

    • Map Client A's VLAN to Port A and Map Client B's VLAN to Port B.
    • Assign Fabric Domain A to Gateway Eth0 801a and assign Fabric Domain B to Gateway Eth1 801b.


Initialize every node in the fabric such that the node's MACs will only accept packets from that particular client's fabric domain. As an example, all the nodes in the cluster assigned to Client A will have the MAC fabric ports within that node to be assigned to only accept Domain A packets, and drop other domain packets.


Using this technique, there will be no packet visibility between the clients, and no packets (unicast or multicast) can be transferred directly between them on the fabric, which improves the security of the system by leveraging the server fabric.


Securing Inter-Management Processor Traffic within the Fabric


In a second embodiment, the inter-management processor (906 in FIG. 4) traffic with the fabric is secured. In particular, the management processors within a server fabric (at each node as shown in FIG. 4) need a secure way to communicate between themselves with no possibility of sniffing or spoofing by the application processors within the fabric. The following techniques (alone or in combination) can be used to secure inter-management traffic:


Either set the management domain bit within the Routing Header (see above) Processor for the management processor and/or assign that MAC the Fabric Domain of 0.


Configure the fabric such that the Ethernet MAC for the management processor only accepts routing headers marked with the management domain bit, or having Fabric Domain of 0.


Configure the fabric such that the Ethernet MACs for the application processors do not have the management domain bit set, and have a non-zero Fabric Domain.


Creating Secure Private Management LAN


In a third embodiment, the fabric may be used to create a secure private management local area network (LAN.) Traditional rack-oriented servers may have an embedded BMC (baseboard management controller) and the BMC will have two paths for network connectivity including a shared management LAN with BMC traffic being routed out the main network port of the server and a Private management LAN with BMC traffic being routed out a private network port of the server.


To illustrate this embodiment, FIG. 6 illustrates a four-node server fabric with a network switch and enhanced security with the goal of creating a private management LAN for the server fabric. The following technique (alone or in combination) can be used to secure the management traffic out of Eth1 801b:


Set the management domain bit within the Routing Header Processor for the management processor and assign that MAC the Fabric Domain of 0.


Configure the fabric such that the Ethernet MAC for the management processor only accepts routing headers marked with the management domain bit, or having Fabric Domain of 0.


Configure the fabric such that the Routing Header Processor for the outgoing MAC, block 910D of FIG. 4, of N31, Eth1 is configured to tag and only accept Fabric Domain of 0.


Configure the fabric such that the Ethernet MACs for the application processors do not have the management domain bit set, and have a non-zero Fabric Domain


In this way, the management processors can securely communicate using the Management Domain, and management traffic will be secured on Eth1.


Using Constrained Routing Tables to Enhance Security in Multi-Tenant Fabrics


In a fourth embodiment, constrained routing tables are used to enhance security in multi-tenant fabric. To illustrate this embodiment, FIG. 6 illustrates a four-node server fabric (nodes 0, 1, 2 and 3 in FIG. 6) with a network switch and enhanced security. The link numbers are depicted in the figure, as an example, packets leaving Node 0 to Node 1 would leave on link 2 (L2 as shown in FIG. 7). A typical unicast routing table for this fabric for Node 1 would look like the following:









TABLE 4







Node 0 Full Fabric Routing Table










Destination Node
Outgoing Link







0




1
L2



2
L0



3
L1










In the case in a multi-tenant fabric where Nodes 0 and 1 are being used by Customer A and Nodes 2 and 3 are being used by Customer B, routing can actually be denied from one customer to another by not having the routes such as in the below constrained routing table.









TABLE 5







Node 0 Constrained Routing Table










Destination Node
Outgoing Link







0




1
L2



2




3











Packet Processing Using OS Routing on Gateway Node


In a fifth embodiment, the fabric can perform packet processing using operating system (OS) routing on a gateway node. This embodiment is illustrated in FIG. 7 that shows a small three-node server fabric. The following technique can be used to create an IPS using the gateway node, node 0:


Assign the incoming Ethernet gateway traffic to the Eth0 MAC (block 902 of FIG. 4) and it can be designated as the Outside MAC.


Assign the fabric-side Ethernet traffic to the Eth1 MAC (block 903 of FIG. 4) and it can be designates as the Inside MAC.


Use Linux (or other OS equivalent) routing features to route traffic between the Inside MAC and the Outside MAC.


Linux (or other OS equivalent) IPS (e.g. Snort) or IDS software can then be run on the application processors (block 905 of FIG. 4) to inspect or block traffic between the fabric and the outside Ethernet.


Packet Processing on Arbitrary Nodes Using Non-Symmetric MCAMs


The sixth embodiment is directed to packet processing on arbitrary nodes using non-symmetric MCAMs. This embodiment is illustrated in FIG. 7 that shows the small three-node server fabric. The following technique can be used to create an IPS using an arbitrary node as the IPS (in this example node 2):


Initialize the MCAM on Node 0, the gateway node, such that all fabric MAC addresses map to Node 2









TABLE 6







Asymmetric MCAM for Node 0 for Node 2 IPS/IDS










MAC Address
Node







Node 0 MAC
2



Node 1 MAC
2



Node 2 MAC
2










Initialize the MCAM on Node 2 to map the MAC addresses to the correct nodes.









TABLE 7







Asymmetric MCAM for Node 2 for Node 2 IPS/IDS










MAC Address
Node







Node 0 MAC
0



Node 1 MAC
1



Node 2 MAC
2










Packets coming into the gateway node hit the fabric switch on node 0.


The destination MAC address on the packet gets translated by the Node 0 MCAM to a destination node, in this case Node 2 (for all fabric MAC addresses).


Packet gets routed to Node 2 and delivered to the application processor MAC on Node 2.


IPS/IDS software runs on node 2, then assuming the packet is not blocked forwards the packet back into the fabric for delivery.


The destination MAC address on the packet gets translated by the Node 2 MCAM to a destination node, in this case the correct destination node within the fabric, and gets delivered to the targeted destination node.


Packet Processing Using Local Management Processor


The seventh embodiment relates to packet processing using local management processor(s), which can be illustrated by the small three-node server fabric depicted in FIG. 7. The following technique can be used to create an IDS or other packet inspection and logging using the local management processor on each node:

    • Configure the fabric Promiscuous Vector to replicate packets to the management processor MAC (block 906 of FIG. 4). The Promiscuous Vector defines a list of ports to which the incoming packet should be replicated. This allows the management processor to declare to the switch that it should get a copy of the incoming packets, without knowledge or intervention of the source or destination of the packet transfer.


Packets entering or leaving Eth0 and Eth1 MACs (blocks 902 and 903 of FIG. 4) will be replicated to management processor MAC, block 906.


The management processor can then run IDS or other packet inspection or logging software not only unobtrusively to the OS and applications on the application processor, but without the OS or applications processor being aware of the management processor packet processing.


Security Enhancement of Having Non-Whitelisted Destination Macs Dropped at the Ingress Node


The eighth embodiment is directed to a security enhancement of having Non-whitelisted destination MACs dropped at the ingress node which can be illustrated using the switch fabric in FIG. 7 to be able to enforce white-listing of destination MAC addresses (meaning that the network manager will have a list of known MAC addresses within the fabric or within the broadcast domain, and packets ingressing into the fabric that are not on the destination MAC whitelist will be immediately dropped.) The following technique can be used to create MAC address whitelists:


The Network administrator collects all the MAC addresses within the broadcast domain, both inside and outside the fabric.


All the MCAMs are initialized with the complete list of (MAC Address, Node ID, Port ID) mappings.


For those MAC addresses that are outside the fabric, the MCAM mapping is to (Gateway Node ID, Outlink Port).


The primary and secondary gateway node ID mappings in each switch are disabled.


This results in packets entering the fabric that don't match the MAC address whitelist to be routed to the gateway node, but by invalidating the gateway node entries, the packets are dropped.


This results in packets being dropped immediately at the ingress point that don't match the MAC address whitelist.


Additional Security Aspects


The security may also include secure fabric local Network Attached Storage (NAS) through private internal domains. There are domains do not have to go all the way to an uplink. Thus, the system can establish a domain between one server node and a node acting as a NAS server.


The system may also provide port scan and port sweep monitoring. A port sweep is the act of systematically scanning the ports on one or more computers by security attackers to find weakened access points to break into computer systems. A port scan is a series of messages sent by someone attempting to break into a computer to learn which computer network services, each associated with a port number, the computer provides. The port scan and port sweep are generally hard to detect at the IPS/IPD level because that are a large number of data flows to watch (and with port sweep many systems) and tracking of the accesses over time. Since the switch system described above has all traffic going into the cluster, the system can monitor for port scan/port sweep better than external appliances.


The system also may allow for the monitoring for a typical network traffic to/from a node. Since the system can monitor all rates over time, the system can monitor traffic to/from a node and isolate it, or flag it, if it exceeds (customer settable) tolerances.


The system may also provide isolation of traffic. In particular, in addition to operating system (OS) routing to separate multi-tenant traffic, the system can also provide physical isolation by cutting links.


The system may also permit customers to configure the topology of the switch. The configuration of the switch may prevent the sharing of links (avoiding a DOS at a link), or sharing of boards (to avoid fault sharing.)


The system may also use IP reputation processing for security. In particular, the blocking or allowing of access based on source address may be incorporated into any place in the switch that packet processing occurs. Using IP reputation processing, the system can support multiple equivalent servers with one server receiving traffic from trusted systems, one receiving traffic from less trusted systems, and one receiving from untrusted system. This could allow for faster/streamlined processing of trusted traffic, and more security checking of less trusted traffic.


The switch security (and the management processor in particular) may provide encryption services in which the keys never leave the trusted zone.


The switch system may also perform real mapping of external virtual local area networks (VLANs) to domains by having the uplink nodes being in their own domains. To provide the real mapping, the switch uses their downlinks as MACLinks (even though they go to our nodes) and uses routing through the downlinks to pick the desired domain (based on VLAN). For example, if the user wants to map a VLAN 101 packet to Domain 10, the uplink node would have the four other links configured as MACLinks, one of those links would go to another node (whose link is also configured as a MACLink with a Domain of 10, so any packet sent down that link goes into the fabric as Domain 10.)


The use of the Outside Ethernet MAC (904 in FIG. 4) gives the system the ability to filter on VLANs, Source MAC, Destination MAC, etc. with perfect filters and wildcards or hashes within the Outlink. Thus, packets can be dropped before they enter the fabric and that is another way of implementing security enhancement of having Non-Whitelisted Destination MACs dropped at the Ingress Node or dropping of packets based on a source address.


While the foregoing has been with reference to a particular embodiment of the invention, it will be appreciated by those skilled in the art that changes in this embodiment may be made without departing from the principles and spirit of the disclosure, the scope of which is defined by the appended claims.

Claims
  • 1. A server system-on-a-chip device comprising: a management processor configured to transmit one or more frames;a routing header unit connected to the management processor, wherein the routing header unit is configured to set a management domain bit in a routing header of the one or more frames, and wherein the management domain bit comprises an identifier that indicates a particular network domain for which the routing header unit belongs and indicates access to a management domain; anda fabric switch connected to the routing header unit and to a plurality of external ports, wherein the fabric switch is configured to perform packet processing based, at least in part, on the management domain bit.
  • 2. The device of claim 1, further comprising a media access control (MAC) unit connected to the fabric switch, wherein the MAC unit is configured to only accept frames having the management domain bit.
  • 3. The device of claim 2, wherein the management domain bit further indicates a fabric domain to which the management processor belongs, and wherein the MAC unit is assigned the fabric domain.
  • 4. The device of claim 3, wherein the fabric domain is configured to provide a trusted communication channel between multiple management processors.
  • 5. The device of claim 3, wherein the fabric domain is set to zero.
  • 6. The device of claim 5, further comprising a plurality of application processors and a plurality of MAC units associated with respective application processors, wherein the plurality of MAC units associated with the respective application processors do not have management domain bits set, and wherein the plurality of MAC units associated with the respective application processors have a non-zero fabric domain.
  • 7. The device of claim 5, further comprising another routing header unit associated with an outgoing MAC unit, wherein the other routing header unit is configured to tag and only accept frames having management domain bits that indicate the fabric domain of zero.
  • 8. The device of claim 1, further comprising a plurality of application processors and a plurality of MAC units associated with respective application processors, wherein the plurality of MAC units do not have management domain bits set.
  • 9. The device of claim 1, wherein the routing header unit is configured to prepend routing heads to packets generated by an associated MAC unit, and wherein the routing heads include the management domain identifier associated with the associated MAC unit.
  • 10. The device of claim 1, wherein the fabric switch is further configured to stop a packet that has been assigned a default domain identifier.
  • 11. A method comprising: interconnecting a plurality of nodes with a plurality of links to form a server fabric, wherein each of the plurality of nodes includes:
  • 12. The method of claim 11, wherein said routing the data packets comprises routing the data packets to another MAC unit, and wherein the other MAC unit is configured to only accept data packets having a header with the management domain bit.
  • 13. The method of claim 12, wherein the management domain bit further indicates a fabric domain to which the management processor belongs, and wherein the MAC unit is assigned the fabric domain.
  • 14. The method of claim 13, wherein the fabric domain is set to zero.
  • 15. The method of claim 14, further comprising tagging and only accepting, by another routing header unit, frames having management domain bits that indicate the fabric domain of zero.
  • 16. The method of claim 11, further comprising prepending, by the routing header unit, routing heads to the data packets, wherein the routing heads include the management domain identifier associated with the MAC unit.
  • 17. The method of claim 11, further comprising stopping, by the fabric switch, a packet that has been assigned a default domain identifier.
  • 18. A system comprising: a plurality of nodes, wherein each node in the plurality of nodes includes:
  • 19. The system of claim 18, wherein each node further comprises a media access control (MAC) unit connected to the fabric switch, and wherein the MAC unit is configured to only accept frames having the management domain bit.
  • 20. The system of claim 19, wherein the management domain bit further indicates a fabric domain to which the management processor belongs, and wherein the MAC unit is assigned the fabric domain.
CROSS-REFERENCE TO RELATED PATENT APPLICATIONS

This application is a Continuation of U.S. application Ser. No. 14/725,543, filed May 29, 2015, which is a Continuation of U.S. application Ser. No. 13/475,713, filed May 18, 2012, which is a Continuation-In-Part of U.S. application Ser. No. 12/794,996, filed Jun. 7, 2010, which claims priority to U.S. Provisional Application 61/256723, filed Oct. 30, 2009. U.S. application Ser. No. 13/475,713 also claims priority to U.S. Provisional Application 61/489,569, filed May 24, 2011. Each of the above-referenced application is incorporated herein by reference in their respective entireties.

US Referenced Citations (360)
Number Name Date Kind
5451936 Yang et al. Sep 1995 A
5594908 Hyatt Jan 1997 A
5623641 Kadoyashiki Apr 1997 A
5781187 Gephardt et al. Jul 1998 A
5901048 Hu May 1999 A
5908468 Hartmann Jun 1999 A
5968176 Nessett et al. Oct 1999 A
5971804 Gallagher et al. Oct 1999 A
6055618 Thorson Apr 2000 A
6141214 Ahn Oct 2000 A
6181699 Crinion et al. Jan 2001 B1
6192414 Horn Feb 2001 B1
6198741 Yoshizawa et al. Mar 2001 B1
6252878 Locklear, Jr. et al. Jun 2001 B1
6314487 Hahn et al. Nov 2001 B1
6314501 Gulick et al. Nov 2001 B1
6373841 Goh et al. Apr 2002 B1
6442137 Yu et al. Aug 2002 B1
6446192 Narasimhan et al. Sep 2002 B1
6452809 Jackson et al. Sep 2002 B1
6507586 Satran et al. Jan 2003 B1
6556952 Magro Apr 2003 B1
6574238 Thrysoe Jun 2003 B1
6711691 Howard et al. Mar 2004 B1
6766389 Hayter et al. Jul 2004 B2
6813676 Henry et al. Nov 2004 B1
6816750 Klaas Nov 2004 B1
6842430 Melnik Jan 2005 B1
6857026 Cain Feb 2005 B1
6963926 Robinson Nov 2005 B1
6963948 Gulick Nov 2005 B1
6977939 Joy et al. Dec 2005 B2
6988170 Barroso et al. Jan 2006 B2
6990063 Lenoski et al. Jan 2006 B1
7020695 Kundu et al. Mar 2006 B1
7032119 Fung Apr 2006 B2
7080078 Slaughter et al. Jul 2006 B1
7080283 Songer et al. Jul 2006 B1
7095738 Desanti Aug 2006 B1
7119591 Lin Oct 2006 B1
7143153 Black et al. Nov 2006 B1
7165120 Giles et al. Jan 2007 B1
7170315 Bakker et al. Jan 2007 B2
7180866 Chartre et al. Feb 2007 B1
7203063 Bash et al. Apr 2007 B2
7257655 Burney et al. Aug 2007 B1
7263288 Islam Aug 2007 B1
7274705 Chang et al. Sep 2007 B2
7278582 Siegel et al. Oct 2007 B1
7310319 Awsienko Dec 2007 B2
7325050 O'Connor et al. Jan 2008 B2
7337333 O'Conner et al. Feb 2008 B2
7340777 Szor Mar 2008 B1
7353362 Georgiou et al. Apr 2008 B2
7382154 Ramos et al. Jun 2008 B2
7386888 Liang et al. Jun 2008 B2
7418534 Hayter et al. Aug 2008 B2
7437540 Paolucci et al. Oct 2008 B2
7447147 Nguyen et al. Nov 2008 B2
7447197 Terrell et al. Nov 2008 B2
7466712 Makishima et al. Dec 2008 B2
7467306 Cartes et al. Dec 2008 B2
7467358 Kang et al. Dec 2008 B2
7502884 Shah et al. Mar 2009 B1
7519843 Buterbaugh et al. Apr 2009 B1
7555666 Brundridge et al. Jun 2009 B2
7583661 Chaudhuri Sep 2009 B2
7586841 Vasseur Sep 2009 B2
7596144 Pong Sep 2009 B2
7599360 Edsall et al. Oct 2009 B2
7606225 Xie et al. Oct 2009 B2
7606245 Ma et al. Oct 2009 B2
7616646 Ma et al. Nov 2009 B1
7620057 Aloni et al. Nov 2009 B1
7644215 Wallace et al. Jan 2010 B2
7657677 Huang et al. Feb 2010 B2
7657756 Hall Feb 2010 B2
7660922 Harriman Feb 2010 B2
7664110 Lovett et al. Feb 2010 B1
7673164 Agarwal Mar 2010 B1
7710936 Morales Barroso May 2010 B2
7719834 Miyamoto et al. May 2010 B2
7721125 Fung May 2010 B2
7751433 Dollo et al. Jul 2010 B2
7760720 Pullela et al. Jul 2010 B2
7761687 Blumrich et al. Jul 2010 B2
7783910 Felter et al. Aug 2010 B2
7791894 Bechtolsheim Sep 2010 B2
7792113 Foschiano et al. Sep 2010 B1
7796399 Clayton et al. Sep 2010 B2
7801132 Ofek et al. Sep 2010 B2
7802017 Uemura et al. Sep 2010 B2
7805575 Agarwal et al. Sep 2010 B1
7831839 Hatakeyama Nov 2010 B2
7840703 Arimilli et al. Nov 2010 B2
7865614 Lu et al. Jan 2011 B2
7925795 Tamir et al. Apr 2011 B2
7934005 Fascenda Apr 2011 B2
7970929 Mahalingaiah Jun 2011 B1
7975110 Spaur et al. Jul 2011 B1
7991817 Dehon et al. Aug 2011 B2
7991922 Hayter et al. Aug 2011 B2
7992151 Warrier et al. Aug 2011 B2
8019832 De Sousa et al. Sep 2011 B2
8060760 Shetty et al. Nov 2011 B2
8060775 Sharma et al. Nov 2011 B1
8082400 Chang et al. Dec 2011 B1
8108508 Goh et al. Jan 2012 B1
8122269 Houlihan et al. Feb 2012 B2
8132034 Lambert et al. Mar 2012 B2
8155113 Agarwal Apr 2012 B1
8156362 Branover et al. Apr 2012 B2
8165120 Maruccia et al. Apr 2012 B2
8170040 Konda May 2012 B2
8180996 Fullerton et al. May 2012 B2
8189612 Lemaire et al. May 2012 B2
8194659 Ban Jun 2012 B2
8199636 Rouyer et al. Jun 2012 B1
8205103 Kazama et al. Jun 2012 B2
8379425 Fukuoka et al. Feb 2013 B2
8397092 Karnowski Mar 2013 B2
8407428 Cheriton et al. Mar 2013 B2
8504791 Cheriton et al. Aug 2013 B2
RE44610 Krakirian et al. Nov 2013 E
8599863 Davis Dec 2013 B2
8684802 Gross et al. Apr 2014 B1
8738860 Griffin et al. May 2014 B1
8745275 Ikeya et al. Jun 2014 B2
8745302 Davis et al. Jun 2014 B2
8782321 Harriman et al. Jul 2014 B2
8812400 Faraboschi et al. Aug 2014 B2
8824485 Biswas et al. Sep 2014 B2
8854831 Arnouse Oct 2014 B2
8903964 Breslin Dec 2014 B2
9008079 Davis et al. Apr 2015 B2
9311269 Davis et al. Apr 2016 B2
9465771 Davis et al. Oct 2016 B2
20010046227 Matsuhira et al. Nov 2001 A1
20020004912 Fung Jan 2002 A1
20020040391 Chaiken et al. Apr 2002 A1
20020083352 Fujimoto et al. Jun 2002 A1
20020097732 Worster et al. Jul 2002 A1
20020107903 Richter et al. Aug 2002 A1
20020124128 Qiu Sep 2002 A1
20020159452 Foster Oct 2002 A1
20020161917 Shapiro et al. Oct 2002 A1
20020172205 Tagore-Brage et al. Nov 2002 A1
20020186656 Vu Dec 2002 A1
20020194412 Bottom Dec 2002 A1
20020196611 Ho et al. Dec 2002 A1
20030007493 Oi et al. Jan 2003 A1
20030033547 Larson et al. Feb 2003 A1
20030041266 Ke et al. Feb 2003 A1
20030076832 Ni Apr 2003 A1
20030093255 Freyensee et al. May 2003 A1
20030093624 Arimilli et al. May 2003 A1
20030110262 Hasan et al. Jun 2003 A1
20030140190 Mahony et al. Jul 2003 A1
20030158940 Leigh Aug 2003 A1
20030159083 Fukuhara et al. Aug 2003 A1
20030172191 Williams Sep 2003 A1
20030188083 Kumar et al. Oct 2003 A1
20030193402 Post et al. Oct 2003 A1
20030202520 Witkowski et al. Oct 2003 A1
20030231624 Alappat et al. Dec 2003 A1
20040013113 Singh et al. Jan 2004 A1
20040017806 Yazdy et al. Jan 2004 A1
20040017808 Forbes et al. Jan 2004 A1
20040030938 Barr et al. Feb 2004 A1
20040068676 Larson et al. Apr 2004 A1
20040111612 Choi et al. Jun 2004 A1
20040141521 George Jul 2004 A1
20040165588 Pandya Aug 2004 A1
20040210693 Zeitler et al. Oct 2004 A1
20040215864 Arimilli et al. Oct 2004 A1
20040215991 McAfee et al. Oct 2004 A1
20040267486 Percer et al. Dec 2004 A1
20050015378 Gammel et al. Jan 2005 A1
20050018604 Dropps et al. Jan 2005 A1
20050018606 Dropps et al. Jan 2005 A1
20050018663 Dropps et al. Jan 2005 A1
20050021606 Davies et al. Jan 2005 A1
20050021728 Sugimoto Jan 2005 A1
20050030954 Dropps et al. Feb 2005 A1
20050033742 Kamvar et al. Feb 2005 A1
20050033890 Lee Feb 2005 A1
20050044195 Westfall Feb 2005 A1
20050077921 Percer et al. Apr 2005 A1
20050105538 Perera et al. May 2005 A1
20050141424 Lim et al. Jun 2005 A1
20050228852 Santos et al. Oct 2005 A1
20050240688 Moerman et al. Oct 2005 A1
20060002311 Iwanaga et al. Jan 2006 A1
20060013218 Shore et al. Jan 2006 A1
20060029053 Roberts et al. Feb 2006 A1
20060090025 Tufford et al. Apr 2006 A1
20060136570 Pandya Jun 2006 A1
20060140211 Huang et al. Jun 2006 A1
20060174342 Zaheer et al. Aug 2006 A1
20060179241 Clark et al. Aug 2006 A1
20060236371 Fish Oct 2006 A1
20060248359 Fung Nov 2006 A1
20060259734 Sheu et al. Nov 2006 A1
20060265609 Fung Nov 2006 A1
20070006001 Isobe et al. Jan 2007 A1
20070076653 Park et al. Apr 2007 A1
20070081315 Mondor et al. Apr 2007 A1
20070094486 Moore et al. Apr 2007 A1
20070109968 Hussain et al. May 2007 A1
20070130397 Tsu Jun 2007 A1
20070174390 Silvain et al. Jul 2007 A1
20070180310 Johnson et al. Aug 2007 A1
20070209072 Chen Sep 2007 A1
20070226795 Conti et al. Sep 2007 A1
20070280230 Park Dec 2007 A1
20070286009 Norman Dec 2007 A1
20070288585 Sekiguchi et al. Dec 2007 A1
20080013453 Chiang et al. Jan 2008 A1
20080040463 Brown et al. Feb 2008 A1
20080052437 Loffink et al. Feb 2008 A1
20080059782 Kruse et al. Mar 2008 A1
20080075089 Evans et al. Mar 2008 A1
20080089358 Basso et al. Apr 2008 A1
20080104264 Duerk et al. May 2008 A1
20080140771 Vass et al. Jun 2008 A1
20080140930 Hotchkiss Jun 2008 A1
20080159745 Segal Jul 2008 A1
20080162691 Zhang et al. Jul 2008 A1
20080183882 Flynn et al. Jul 2008 A1
20080186965 Zheng et al. Aug 2008 A1
20080199133 Takizawa et al. Aug 2008 A1
20080212273 Bechtolsheim Sep 2008 A1
20080212276 Bottom et al. Sep 2008 A1
20080217021 Lembcke et al. Sep 2008 A1
20080222434 Shimizu et al. Sep 2008 A1
20080235443 Chow et al. Sep 2008 A1
20080239649 Bradicich et al. Oct 2008 A1
20080243634 Dworkin et al. Oct 2008 A1
20080250181 Li et al. Oct 2008 A1
20080259555 Bechtolsheim et al. Oct 2008 A1
20080259788 Wang et al. Oct 2008 A1
20080266793 Lee Oct 2008 A1
20080270599 Tamir et al. Oct 2008 A1
20080288660 Balasubramanian et al. Nov 2008 A1
20080288664 Pettey et al. Nov 2008 A1
20080288683 Ramey Nov 2008 A1
20080301794 Lee Dec 2008 A1
20080310848 Yasuda et al. Dec 2008 A1
20080313369 Verdoorn et al. Dec 2008 A1
20080320161 Maruccia et al. Dec 2008 A1
20090021907 Mann et al. Jan 2009 A1
20090044036 Merkin Feb 2009 A1
20090063443 Arimilli et al. Mar 2009 A1
20090064287 Bagepalli et al. Mar 2009 A1
20090080428 Witkowski Mar 2009 A1
20090097200 Sharma et al. Apr 2009 A1
20090113130 He et al. Apr 2009 A1
20090133129 Jeong et al. May 2009 A1
20090135751 Hodges et al. May 2009 A1
20090135835 Gallatin et al. May 2009 A1
20090158070 Gruendler Jun 2009 A1
20090172423 Song et al. Jul 2009 A1
20090198958 Arimilli et al. Aug 2009 A1
20090204834 Hendin et al. Aug 2009 A1
20090204837 Raval et al. Aug 2009 A1
20090216920 Lauterbach et al. Aug 2009 A1
20090219827 Chen et al. Sep 2009 A1
20090222884 Shaji Sep 2009 A1
20090225751 Koenck et al. Sep 2009 A1
20090235104 Fung Sep 2009 A1
20090248943 Jiang et al. Oct 2009 A1
20090251867 Sharma et al. Oct 2009 A1
20090259863 Williams et al. Oct 2009 A1
20090259864 Li et al. Oct 2009 A1
20090265045 Coxe, III Oct 2009 A1
20090271656 Yokota et al. Oct 2009 A1
20090276666 Haley et al. Nov 2009 A1
20090279518 Falk et al. Nov 2009 A1
20090282274 Langgood et al. Nov 2009 A1
20090282419 Mejdrich et al. Nov 2009 A1
20090313390 Ahuja et al. Dec 2009 A1
20100005331 Somasundaram et al. Jan 2010 A1
20100008038 Coglitore Jan 2010 A1
20100008365 Porat Jan 2010 A1
20100026408 Shau Feb 2010 A1
20100040053 Gottumukkula et al. Feb 2010 A1
20100049822 Davies et al. Feb 2010 A1
20100051391 Jahkonen Mar 2010 A1
20100106987 Lambert et al. Apr 2010 A1
20100118880 Kunz et al. May 2010 A1
20100125742 Ohtani May 2010 A1
20100125915 Hall et al. May 2010 A1
20100138481 Behrens Jun 2010 A1
20100158005 Mukhopadhyay et al. Jun 2010 A1
20100161909 Nation et al. Jun 2010 A1
20100165983 Aybay et al. Jul 2010 A1
20100169479 Jeong et al. Jul 2010 A1
20100198972 Umbehocker Aug 2010 A1
20100218194 Dallman et al. Aug 2010 A1
20100220732 Hussain et al. Sep 2010 A1
20100250914 Abdul et al. Sep 2010 A1
20100265650 Chen et al. Oct 2010 A1
20100281246 Bristow et al. Nov 2010 A1
20100299548 Chadirchi et al. Nov 2010 A1
20100308897 Evoy et al. Dec 2010 A1
20100312910 Lin et al. Dec 2010 A1
20100312969 Yamazaki et al. Dec 2010 A1
20100318812 Auradkar et al. Dec 2010 A1
20110023104 Franklin Jan 2011 A1
20110026397 Saltsidis et al. Feb 2011 A1
20110029652 Chhuor et al. Feb 2011 A1
20110058573 Balakavi et al. Mar 2011 A1
20110075369 Sun et al. Mar 2011 A1
20110090633 Rabinovitz Apr 2011 A1
20110103391 Davis et al. May 2011 A1
20110113115 Chang et al. May 2011 A1
20110119344 Eustis May 2011 A1
20110123014 Smith May 2011 A1
20110138046 Bonnier et al. Jun 2011 A1
20110185370 Tamir et al. Jul 2011 A1
20110191514 Wu et al. Aug 2011 A1
20110191610 Agarwal et al. Aug 2011 A1
20110197012 Liao et al. Aug 2011 A1
20110210975 Wong et al. Sep 2011 A1
20110239014 Karnowski Sep 2011 A1
20110271159 Ahn et al. Nov 2011 A1
20110273840 Chen Nov 2011 A1
20110295991 Aida Dec 2011 A1
20110296141 Daffron Dec 2011 A1
20110320690 Petersen et al. Dec 2011 A1
20120011500 Faraboschi et al. Jan 2012 A1
20120020207 Corti et al. Jan 2012 A1
20120050981 Xu et al. Mar 2012 A1
20120054469 Ikeya et al. Mar 2012 A1
20120054511 Brinks et al. Mar 2012 A1
20120081850 Regimbal et al. Apr 2012 A1
20120096211 Davis et al. Apr 2012 A1
20120099265 Reber Apr 2012 A1
20120131201 Matthews et al. May 2012 A1
20120155168 Kim et al. Jun 2012 A1
20120198252 Kirschtein et al. Aug 2012 A1
20120207165 Davis Aug 2012 A1
20120297042 Davis et al. Nov 2012 A1
20130010639 Armstrong et al. Jan 2013 A1
20130024645 Cheriton et al. Jan 2013 A1
20130031331 Cheriton et al. Jan 2013 A1
20130058250 Casado et al. Mar 2013 A1
20130094499 Davis et al. Apr 2013 A1
20130097448 Davis et al. Apr 2013 A1
20130111107 Chang et al. May 2013 A1
20130148667 Hama et al. Jun 2013 A1
20130163605 Chandra et al. Jun 2013 A1
20130290643 Lim et al. Oct 2013 A1
20130290650 Chang et al. Oct 2013 A1
20130318269 Dalal et al. Nov 2013 A1
20140122833 Davis et al. May 2014 A1
20140359044 Davis et al. Dec 2014 A1
20140365596 Kanevsky et al. Dec 2014 A1
20150039840 Chandra et al. Feb 2015 A1
20150103826 Davis Apr 2015 A1
Foreign Referenced Citations (9)
Number Date Country
2005-223753 Aug 2005 JP
2005-536960 Dec 2005 JP
M377621 Apr 2010 TW
201017430 May 2010 TW
WO-2004021641 Mar 2004 WO
WO-2005013143 Feb 2005 WO
WO-2008000193 Jan 2008 WO
WO-2011044271 Apr 2011 WO
WO-2012037494 Mar 2012 WO
Non-Patent Literature Citations (120)
Entry
Advanced Switching Technology Tech Brief, published 2005, 2 pages.
Chapter 1 Overview of the Origin Family Architecture from Origin and Onyx2 Theory of Operations Manual, published 1997, 18 pages.
Cisco MDS 9000 Family Multiprotocol Services Module, published 2006, 13 pages.
Comparing the I2C BUS to the SMBUS, Maxim Integrated, Dec. 1, 2000, p. 1.
Das et al., “Unifying Packet and Circuit Switched Networks,” EEE Globecom Workshops 2009, Nov. 30, 2009, pp. 1-6.
Deering, “IP Multicast Extensions for 4.3BSD UNIX and related Systems,” Jun. 1999, 5 pages.
Elghany et al., “High Throughput High Performance NoC Switch,” NORCHIP 2008, Nov. 2008, pp. 237-240.
Extended European Search Report for EP 10827330.1, mailed Jun. 5, 2013.
Final Office Action on U.S. Appl. No. 12/889,721, mailed Apr. 17, 2014.
Final Office Action on U.S. Appl. No. 13/234,054, mailed Jan. 26, 2016.
Final Office Action on U.S. Appl. No. 13/692,741, mailed Mar. 11, 2015.
Final Office Action on U.S. Appl. No. 12/794,996, mailed Jun. 19, 2013.
Final Office Action on U.S. Appl. No. 12/889,721 mailed Aug. 2, 2016.
Final Office Action on U.S. Appl. No. 12/889,721, mailed May 22, 2015.
Final Office Action on U.S. Appl. No. 13/234,054, mailed Apr. 16, 2015.
Final Office Action on U.S. Appl. No. 13/475,713, mailed Oct. 17, 2014.
Final Office Action on U.S. Appl. No. 13/475,722, mailed Oct. 20, 2014.
Final Office Action on U.S. Appl. No. 13/527,498, mailed Nov. 17, 2014.
Final Office Action on U.S. Appl. No. 13/527,505, mailed Dec. 5, 2014.
Final Office Action on U.S. Appl. No. 13/624,725 mailed Mar. 10, 2016.
Final Office Action on U.S. Appl. No. 13/624,725, mailed Nov. 13, 2013.
Final Office Action on U.S. Appl. No. 13/624,731, mailed Jul. 25, 2014.
Final Office Action on U.S. Appl. No. 13/662,759, mailed Feb. 22, 2016.
Final Office Action on U.S. Appl. No. 13/705,340, mailed Aug. 2, 2013.
Final Office Action on U.S. Appl. No. 13/705,414, mailed Aug. 9, 2013.
Final Office Action on U.S. Appl. No. 13/728,428 mailed May 6, 2016.
Final Office Action on U.S. Appl. No. 14/052,723, mailed Dec. 3, 2015.
Final Office Action on U.S. Appl. No. 14/106,697 mailed Feb. 2, 2016.
Final Office Action on U.S. Appl. No. 14/106,698, mailed Aug. 19, 2015.
Final Office Action on U.S. Appl. No. 14/334,178, mailed Nov. 4, 2015.
Final Office Action on U.S. Appl. No. 14/334,931, mailed Jul. 9, 2015.
Final Office Action on U.S. Appl. No. 13/624,731, mailed Nov. 12, 2013.
fpga4fun.com,“What is JTAG?”, 2 pages, Jan. 31, 2010.
From AT to BTX: Motherboard Form Factor, Webopedia, Apr. 29, 2005, p. 1.
Grecu et al., “A Scalable Communication-Centric SoC Interconnect Architecture” Proceedings 5th International Symposium on Quality Electronic Design, 2005, pp. 343, 348 (full article included).
Hossain et al., “Extended Butterfly Fat Tree Interconnection (EFTI) Architecture for Network on CHIP,” 2005 IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Aug. 2005, pp. 613-616.
HP ProLiant 5L6500 Scalable System, Family data sheet, HP Technical sheet, Sep. 2010 4 pages.
HP Virtual Connect Traffic Flow—Technology brief, Jan. 2012, 22 pages.
International Preliminary Report on Patentability for PCT/US2009/044200, mailed Nov. 17, 2010.
International Preliminary Report on Patentability for PCT/US2012/038986 issued on Nov. 26, 2013.
International Preliminary Report on Patentability for PCT/US2012/061747, mailed Apr. 29, 2014.
International Preliminary Report on Patentability issued on PCT/US12/62608, issued May 6, 2014.
International Search Report and Written Opinion for PCT/US12/38987, mailed Aug. 16, 2012.
International Search Report and Written Opinion for PCT/US12/61747, mailed Mar. 1, 2013.
International Search Report and Written Opinion for PCT/US12/62608, mailed Jan. 18, 2013.
International Search Report and Written Opinion for PCT/US2010/053227, mailed May 10, 2012.
International Search Report and Written Opinion for PCT/US2011/051996, mailed Jan. 19, 2012.
International Search Report and Written Opinion on PCT/US09/44200, mailed Jul. 1, 2009.
International Search Report and Written Opinion on PCT/US2012/038986, mailed Mar. 14, 2013.
Jansen et al., “SATA-IO to Develop Specification for Mini Interface Connector” Press Release Sep. 21, 2009, Serial ATA3 pages.
Nawathe et al., “Implementation of an 8-Core, 64-Thread, Power Efficient SPARC Server on a Chip”, IEEE Journal of Solid-State Circuits, vol. 43, No. 1, Jan. 2008, pp. 6-20.
Non-Final Action on U.S. Appl. No. 13/728,362, mailed Feb. 21, 2014.
Non-Final Office Action on U.S. Appl. No. 15/281,462 mailed Feb. 10, 2017.
Non-Final Office Action on U.S. Appl. No. 14/334,178 mailed Dec. 18, 2015.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Feb. 24, 2016.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Jul. 2, 2013.
Non-Final Office Action on U.S. Appl. No. 13/475,722, mailed Jan. 17, 2014.
Non-Final Office Action on U.S. Appl. No. 12/794,996, mailed Sep. 17, 2012.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Oct. 11, 2012.
Non-Final Office Action on U.S. Appl. No. 12/889,721, mailed Sep. 29, 2014.
Non-Final Office Action on U.S. Appl. No. 13/234,054 mailed Oct. 20, 2016.
Non-Final Office Action on U.S. Appl. No. 13/234,054, mailed Oct. 23, 2014.
Non-Final Office Action on U.S. Appl. No. 13/234,054, mailed Aug. 6, 2015.
Non-Final Office Action on U.S. Appl. No. 13/284,855, mailed Dec. 19, 2013.
Non-Final Office Action on U.S. Appl. No. 13/453,086, mailed Mar. 12, 2013.
Non-Final Office Action on U.S. Appl. No. 13/475,713, mailed Apr. 1, 2014.
Non-Final Office Action on U.S. Appl. No. 13/527,505, mailed May 8, 2014.
Non-Final Office Action on U.S. Appl. No. 13/527,498, Mailed May 8, 2014.
Non-Final Office Action on U.S. Appl. No. 13/624,725, mailed Jan. 10, 2013.
Non-Final Office Action on U.S. Appl. No. 13/624,725, mailed Apr. 23, 2015.
Non-final office action on U.S. Appl. No. 13/624,731 mailed Jan. 29, 2013.
Non-Final Office Action on U.S. Appl. No. 13/662,759, mailed Nov. 6, 2014.
Non-Final Office Action on U.S. Appl. No. 13/692,741, mailed Sep. 4, 2014.
Non-Final Office Action on U.S. Appl. No. 13/692,741, mailed Jul. 1, 2015.
Non-Final Office Action on U.S. Appl. No. 13/705,286, mailed May 13, 2013.
Non-Final Office Action on U.S. Appl. No. 13/705,340, mailed Mar. 12, 2014.
Non-Final Office Action on U.S. Appl. No. 13/705,340, mailed Mar. 29, 2013.
Non-Final Office Action on U.S. Appl. No. 13/705,414, mailed Apr. 9, 2013.
Non-Final Office Action on U.S. Appl. No. 13/728,308, mailed May 14, 2015.
Non-Final Office Action on U.S. Appl. No. 13/728,428, mailed Jun. 12, 2015.
Non-Final Office Action on U.S. Appl. No. 14/052,723, mailed May 1, 2015.
Non-Final Office Action on U.S. Appl. No. 14/106,697, mailed Aug. 17, 2015.
Non-Final Office Action on U.S. Appl. No. 14/106,698, mailed Feb. 12, 2015.
Non-Final Office Action on U.S. Appl. No. 14/334,931, mailed Jan. 5, 2015.
Non-Final Office Action on U.S. Appl. No. 14/725,543 mailed Apr. 7, 2016.
Non-Final Office Action on U.S. Appl. No. 14/753,948 mailed Nov. 4, 2016.
Non-Final Office Action on U.S. Appl. No. 14/809,723 mailed Dec. 30, 2016.
Non-Final Office Action on U.S. Appl. No. 13/705,428, mailed Jul. 10, 2013.
Non-Final Office Action on U.S. Appl. No. 14/334,931 Mailed Dec. 11, 2015.
Notice of Allowance issued on U.S. Appl. No. 14/052,723, mailed Feb. 8, 2017.
Notice of Allowance on U.S. Appl. No. 14/334,931 mailed May 20, 2016.
Notice of Allowance on U.S. Appl. No. 13/453,086, mailed Jul. 18, 2013.
Notice of Allowance on U.S. Appl. No. 13/475,713, mailed Feb. 5, 2015.
Notice of Allowance on U.S. Appl. No. 13/475,722, mailed Feb. 27, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,498, mailed Feb. 23, 2015.
Notice of Allowance on U.S. Appl. No. 13/527,505, mailed Mar. 6, 2015.
Notice of Allowance on U.S. Appl. No. 13/624,725, mailed Mar. 30, 2016.
Notice of Allowance on U.S. Appl. No. 13/624,731, mailed Mar. 5, 2015.
Notice of Allowance on U.S. Appl. No. 13/705,340, mailed Dec. 3, 2014.
Notice of Allowance on U.S. Appl. No. 13/705,386, mailed Jan. 24, 2014.
Notice of Allowance on U.S. Appl. No. 13/705,414, mailed Nov. 4, 2013.
Notice of Allowance on U.S. Appl. No. 13/728,428 mailed Jul. 18, 2016.
Notice of Allowance on U.S. Appl. No. 14/106,697 mailed Oct. 24, 2016.
Notice of Allowance on U.S. Appl. No. 14/725,543 mailed Jul. 21, 2016.
Notice of Allowance on U.S. Appl. No. 13/284,855, mailed Jul. 14, 2014.
Notice of Allowance on U.S. Appl. No. 13/662,759 mailed May 10, 2016.
Notice of Allowance on U.S. Appl. No. 13/692,741 mailed Dec. 4, 2015.
Notice of Allowance on U.S. Appl. No. 14/334,178 mailed Jun. 8, 2016.
Notice of Allowance U.S. Appl. No. 13/728,308, mailed Oct. 7, 2015.
Office Action on Taiwan Application 100133390, mailed Aug. 25, 2015 (English translation not available).
Office Action on Taiwan Application 101139729, mailed May 25, 2015 (English translation not available).
Pande et al., “Design of a Switch for Network on Chip Applications,” May 25-28, 2003 Proceedings of the 2003 International Symposium on Circuits and Systems, vol. 5, pp. V217-V220.
Reexamination Report on Japanese Application 2012-536877, mailed Jan. 22, 2015 (English Translation not available).
Search Report on EP Application 10827330.1, mailed Feb. 12, 2015.
Venaas, “IPv4 Multicast Address Space Registry,” 2013, http://www.iana.org/assignments/multicast-addresses/multicast-addresses.xhtml.
Final Office Action on U.S. Appl. No. 13/234,054 dated May 31, 2017.
Final Office Action on U.S. Appl. No. 15/281,462 dated Jun. 13, 2017.
Non-Final Office Action on U.S. Appl. No. 15/254,111 dated Jun. 20, 2017.
Non-Final Office Action on U.S. Appl. No. 15/270,418 dated Apr. 21, 2017.
Notice of Allowance on U.S. Appl. No. 14/753,948 dated Jun. 14, 2017.
Related Publications (1)
Number Date Country
20170078296 A1 Mar 2017 US
Provisional Applications (2)
Number Date Country
61489569 May 2011 US
61256723 Oct 2009 US
Continuations (2)
Number Date Country
Parent 14725543 May 2015 US
Child 15360668 US
Parent 13475713 May 2012 US
Child 14725543 US
Continuation in Parts (1)
Number Date Country
Parent 12794996 Jun 2010 US
Child 13475713 US