This invention relates to cluster interconnect architecture for high-speed and low latency information and data transfer between the systems in the configuration.
The need for high speed and low latency cluster interconnect scheme for data and information transport between systems have been recognized as one needing attention in recent times. The growth of interconnected and distributed processing schemes have made it essential that high speed interconnect schemes be defined and established to provide the speed up the processing and data sharing between these systems.
There are interconnect schemes that allow data transfer at high speeds, the most common and fast one existing today is the Ethernet connection allowing transport speeds from 10 MB to as high as 10 GB/sec. TCP/IP protocols used with Ethernet have high over head with inherent latency that make it unsuitable for some distributed applications. Effort is under way in different areas of data transport to reduce the latency of the interconnect as this is a limitation on growth of the distributed computing power.
What is Proposed
PCI Express (PCIE) is an emerging I/O interconnect standard for use inside computers, or embedded systems that allow serial high speed data transfer to and from peripheral devices. The typical PCIE provides 2.5 GB transfer rate per link (this may change as the standard and data rates change). Since the PCIE standard is starting become firm and used within the systems, what is disclosed is the use of PCIE standard based peripheral to PCIE standard based peripheral connected directly using data links, as an interconnect between individual stand alone systems, typically through an interconnect module or a network switch. This interconnect scheme by using only PCIE based protocols for data transfer over direct physical connection links between the PCIE based peripheral devices, (see
Some Advantages of the Proposed Connection Scheme:
PCI Express is a Bus or I/O interconnect standard for use inside the computer or embedded system enabling faster data transfers to and from peripheral devices. The standard is still evolving but has achieved a degree of stability such that other applications can be implemented using PCIE as basis. A PCIE based interconnect scheme to enable switching and inter-connection between external systems, such that the scalability can be applied to enable data transport between connected systems to form a cluster of systems is proposed. These connected systems can be any computing or embedded system. The scalability of the interconnect will allow the cluster to grow the bandwidth between the systems as they become necessary without changing to a different connection architecture.
When data has to be transferred between say system 1 and system 5, in the simple case, the control is used to establish an internal link between PCIE based peripheral modules 1b and 5b inside the switch. The hand shake is established between outbound PCIE based peripheral module (PCIE Module) 1a and inbound PCIE module 1b and outbound PCIE module 5a and inbound PCIE module 5b. This provides a through connection between the PCIE modules 1a to 5b through the switch allowing data transfer. Data can then be transferred at speed between the modules and hence between systems. In more complex cases data can also be transferred and qued in storage implemented in the switch and then when links are free transferred out to the right systems at speed.
Multiple systems can be interconnected at one time to form a multi-system that allow data and information transfer and sharing through the switch. It is also possible to connect smaller clusters together to take advantage of the growth in system volume by using an available connection scheme that interconnects the switches that form a node of the cluster.
If need for higher bandwidth and low latency data transfers between systems increase, the connections can grow by increasing the number of links connecting the PCIE modules between the systems in the cluster and the switch without completely changing the architecture of the interconnect. This scalability is of great importance in retaining flexibility for growth and scaling of the cluster.
It should be understood that the system may consist of peripheral devices, storage devices and processors and any other communication devices. The interconnect is agnostic to the type of device as long as they have a PCIE module at the port to enable the connection to the switch. This feature will reduce the cost of expanding the system by changing the switch interconnect density alone for growth of the multi-system.
PCIE is currently being standardized and that will enable the use of the existing PCIE modules to be used from different vendors to reduce the over all cost of the system. In addition using a standardized module in the system as well as the switch will allow the cost of software development to be reduced and in the long run use available software to configure and run the systems.
As the expansion of the cluster in terms of number of systems, connected, bandwidth usage and control will all be cost effective, it is expected the over all system cost can be reduced and over all performance improved by standardized PCIE module use with standardized software control.
Typical connect operation may be explained with reference to two of the systems, example system (1) and system (5). System (1) has a PCIE module (1a) at the interconnect port and that is connected by the connection link or data-link or link (1L) to a PCIE module (1b) at the IO port of the switch (9). System (5) is similarly connected to the switch trough the PCIE module (5a) at its interconnect port to the PCIE module (5b) at the switch (9) IO port by link (5L). Each PCIE module operates for transfer of data to and from it by standard PCI Express protocols, provided by the configuration software loaded into the PCIE modules and switch. The switch operates by the software control and configuration loaded in through the software configuration input.
The following are some of the advantages of the disclosed interconnect scheme
In fact the disclosed interconnect scheme provides advantages for low latency multi-system cluster growth that are not available from any other source.
Number | Name | Date | Kind |
---|---|---|---|
5805597 | Edem | Sep 1998 | A |
5859844 | Kanno et al. | Jan 1999 | A |
5920545 | Rasanen et al. | Jul 1999 | A |
6108739 | James et al. | Aug 2000 | A |
6260092 | Story et al. | Jul 2001 | B1 |
6362908 | Kimbrough et al. | Mar 2002 | B1 |
6615306 | Ajanovic | Sep 2003 | B1 |
7000037 | Rabinovitz et al. | Feb 2006 | B2 |
7058738 | Stufflebeam, Jr. | Jun 2006 | B2 |
7145866 | Ting et al. | Dec 2006 | B1 |
7151774 | Williams et al. | Dec 2006 | B1 |
7457322 | Flood et al. | Nov 2008 | B1 |
20030099247 | Toutant et al. | May 2003 | A1 |
20030123461 | Riley | Jul 2003 | A1 |
20030158940 | Leigh | Aug 2003 | A1 |
20030188079 | Singhal et al. | Oct 2003 | A1 |
20040083323 | Rabinovitz et al. | Apr 2004 | A1 |
20040083325 | Rabinovitz et al. | Apr 2004 | A1 |
20040170196 | Susnow | Sep 2004 | A1 |
20050147119 | Tofano | Jul 2005 | A1 |
20060015537 | Marks | Jan 2006 | A1 |
20060050693 | Bury et al. | Mar 2006 | A1 |
20060088046 | Wong et al. | Apr 2006 | A1 |
20060098659 | Chinner | May 2006 | A1 |
20060101185 | Kapoor et al. | May 2006 | A1 |
20060114918 | Ikeda et al. | Jun 2006 | A1 |
20060143311 | Madukkarumukumana et al. | Jun 2006 | A1 |
20060206655 | Chappell et al. | Sep 2006 | A1 |
20060251096 | Metsker | Nov 2006 | A1 |
Number | Date | Country | |
---|---|---|---|
20070098001 A1 | May 2007 | US |