Claims
- 1. A massively parallel computing structure comprising:
a plurality of processing nodes interconnected by multiple independent networks, each node including one or more processing elements for performing computation or communication activity as required when performing parallel algorithm operations; and, said multiple independent networks comprising networks for enabling point-to-point, global tree communications and global barrier and notification operations among said nodes or independent partitioned subsets thereof, wherein combinations of said multiple independent networks interconnecting said nodes are collaboratively or independently utilized according to bandwidth and latency requirements of an algorithm for optimizing algorithm processing performance.
- 2. The massively parallel computing structure as claimed in claim 1, wherein a first of said multiple independent networks includes an n-dimensional torus network including communication links interconnecting said nodes in a manner optimized for providing high-speed, low latency point-to-point and multicast packet communications among said nodes or independent partitioned subsets thereof.
- 3. The massively parallel computing structure as claimed in claim 2, wherein said multiple independent networks are collaboratively utilized to enable simultaneous computing and message communication activities among individual nodes and partitioned subsets of nodes according to bandwidth and latency requirements of an algorithm being performed.
- 4. The massively parallel computing structure as claimed in claim 2, wherein said multiple independent networks are collaboratively utilized to enable simultaneous independent parallel processing among one or more partitioned subsets of said plurality of nodes according to needs of a parallel algorithm.
- 5. The massively parallel computing structure as claimed in claim 3, wherein said multiple independent networks are collaboratively utilized to enable dynamic switching between computing and message communication activities among individual nodes according to needs of an algorithm.
- 6. The massively parallel computing structure as claimed in claim 2, wherein said first independent network includes means for enabling virtual cut-through (VCT) routing of packets along interconnected links from a source node to a destination node to optimize throughput and latency, said VCT means providing individual buffered virtual channels for facilitating packet routing along network links.
- 7. The massively parallel computing structure as claimed in claim 6, wherein said means for enabling virtual cut-through of message packets utilizes adaptive-routing algorithm for avoiding network contention.
- 8. The massively parallel computing structure as claimed in claim 2, wherein said first independent network includes means for enabling deterministic shortest-path routing for parallel calculations.
- 9. The massively parallel computing structure as claimed in claim 2, wherein said first independent network includes means for automatic multi-casting of packets whereby packets are deposited to multiple destinations according to a node or packet class.
- 10. The massively parallel computing structure as claimed in claim 2, wherein said first independent network includes embedded virtual networks for enabling adaptive and deadlock free deterministic minimal-path routing of packets.
- 11. The massively parallel computing structure as claimed in claim 10, wherein each said plurality of nodes includes routing devices, said first network implementing token-based flow-control means for controlling routing of packets between routers.
- 12. The massively parallel computing structure as claimed in claim 2, wherein a second of said multiple independent networks includes a scalable global tree network comprising nodal interconnections that facilitate simultaneous global operations among nodes or sub-sets of nodes of said network.
- 13. The massively parallel computing structure as claimed in claim 12, wherein said global operations include global broadcast operations initiated at any node of said tree for downstream broadcast from a root node to leaf nodes of said tree network or sub-tree network thereof, and global reduction operations upstream from nodes toward said root node in each tree or sub-tree network.
- 14. The massively parallel computing structure as claimed in claim 12, wherein said root node of a plurality of tree or sub-tree networks couples with an I/O node for performing high-speed I/O operations for that tree network independent of processing performed in other networks.
- 15. The massively parallel computing structure as claimed in claim 14, further including programmable means enabling point-to-point and sub-tree messaging among nodes of each said global tree network, each node having a unique address associated therewith to enable a host system to directly communicate to every node.
- 16. The massively parallel computing structure as claimed in claim 15, wherein said unique address associated includes an encoded geometric location of the node in the computing structure.
- 17. The massively parallel computing structure as claimed in claim 12, wherein a ratio of a service node to sub-set of nodes is configurable to enable optimized packaging and utilization of said computing structure.
- 18. The massively parallel computing structure as claimed in claim 2, wherein a third of said multiple independent networks includes a global signal network for enabling asynchronous global operations.
- 19. The massively parallel computing structure as claimed in claim 18, wherein said asynchronous global operations include implementation of global logical AND functionality for enabling global barrier operations.
- 20. The massively parallel computing structure as claimed in claim 2, wherein said asynchronous global signal operations include implementation of global logical OR functionality for enabling global notification operations.
- 21. The massively parallel computing structure as claimed in claim 14, wherein a fourth of said multiple independent networks includes an external high-speed network connecting each I/O node to an external host system.
- 22. The massively parallel computing structure as claimed in claim 21, wherein said external high-speed network is a Gigabit Ethernet.
- 23. The massively parallel computing structure as claimed in claim 2, wherein a fifth of said multiple independent networks includes an independent network for providing low-level debug, diagnostic and configuration capabilities for all nodes or sub-sets of nodes in said computing structure.
- 24. The massively parallel computing structure as claimed in claim 23, wherein said low-level debug and inspection of internal processing elements of a node may be conducted transparent from any software executing on that node via said fifth network.
- 25. The massively parallel computing structure as claimed in claim 23, wherein said fifth network comprises an IEEE 1149 (JTAG) network.
- 26. The massively parallel computing structure as claimed in claim 15, wherein a sixth of said multiple independent networks includes an independent control network for providing diagnostic and control functionality to individual nodes.
- 27. The massively parallel computing structure as claimed in claim 1, wherein each node includes two or more processing elements each capable of individually or simultaneously working on any combination of computation or communication activity as required when performing particular classes of parallel algorithms.
- 28. The massively parallel computing structure as claimed in claim 27, further including means for enabling rapid shifting of computation or communication activities between each of said processing elements.
- 29. The massively parallel computing structure as claimed in claim 28, wherein each processing element includes a central processing unit (CPU) and one or more floating point processing units, said node further comprising a local embedded multi-level cache memory and a programmable prefetch engine incorporated into a lower level cache for prefetching data for a higher level cache.
- 30. The massively parallel computing structure as claimed in claim 1, wherein each node comprises a system-on-chip Application Specific Integrated Circuit (ASIC) enabling high packaging density and decreasing power utilization and cooling requirements.
- 31. The massively parallel computing structure as claimed in claim 1, wherein said computing structure comprises a predetermined plurality of ASIC nodes packaged on a circuit card, a plurality of circuit cards being configured on an indivisible midplane unit packaged within said computing structure.
- 32. The massively parallel computing structure as claimed in claim 31, wherein a circuit card is organized to comprise nodes logically connected as a sub-cube, or a rectangle.
- 33. The massively parallel computing structure as claimed in claim 31, further including means for partitioning sub-sets of nodes according to various logical network configurations for enabling independent processing among said nodes according to bandwidth and latency requirements of a parallel algorithm being processed.
- 34. The massively parallel computing structure as claimed in claim 33, said partitioning means includes link devices for redriving signals over conductors interconnecting different mid-planes and, redirecting signals between different ports for enabling the supercomputing system to be partitioned into multiple, logically separate systems.
- 35. The massively parallel computing structure as claimed in claim 34, further including means for programming said link devices for mapping communication and computing activities around any midplanes determined as being faulty for servicing thereof without interfering with the remaining system operations.
- 36. The massively parallel computing structure as claimed in claim 34, wherein one of said multiple independent networks includes an independent control network for controlling said link chips to program said partitioning.
- 37. The massively parallel computing structure as claimed in claim 31, further comprising a clock distribution system for providing clock signals to every circuit card of a midplane unit at minimum jitter.
- 38. The massively parallel computing structure as claimed in claim 37, wherein said clock distribution system utilizes tunable redrive signals for enabling in phase clock distribution to all nodes of said computing structure and networked partitions thereof.
- 39. The massively parallel computing structure as claimed in claim 1, further including high-speed, bi-directional serial links interconnecting said processing nodes for carrying signals in both directions at the same time.
- 40. The massively parallel computing structure as claimed in claim 30, wherein each node ASIC further comprises a shared resource in a memory accessible by said processing units configured for lock exchanges to prevent bottlenecks in said processing units.
- 41. The massively parallel computing structure as claimed in claim 6, wherein each packet communicated includes a header including one or more fields for carrying information, one said field including error correction capability for improved bit-serial network communications.
- 42. The massively parallel computing structure as claimed in claim 41, wherein one said field of said packet header includes a defined number of bits representing possible output directions for routing packets at a node in said network, said bit being set to indicate a packet needs to progress in a corresponding direction to reach a node destination for reducing network contention.
- 43. The massively parallel computing structure as claimed in claim 39, further implementing means for capturing data sent over said links that permits optimal sampling and capture of an asynchronous data stream without sending a clock signal with the data stream
- 44. A scalable, massively parallel computing structure comprising:
a plurality of processing nodes interconnected by independent networks, each node including one or more processing elements for performing computation or communication activity as required when performing parallel algorithm operations; and, a first independent network comprising an n-dimensional torus network including communication links interconnecting said nodes in a manner optimized for providing high-speed, low latency point-to-point and multicast packet communications among said nodes or sub-sets of nodes of said network; a second of said multiple independent networks includes a scalable global tree network comprising nodal interconnections that facilitate simultaneous global operations among nodes or sub-sets of nodes of said network; and, partitioning means for dynamically configuring one or more combinations of independent processing networks according to needs of one or more algorithms, each independent network including a configurable sub-set of processing nodes interconnected by divisible portions of said first and second networks, wherein each of said configured independent processing networks is utilized to enable simultaneous collaborative processing for optimizing algorithm processing performance.
- 45. The scalable, massively parallel computing structure as claimed in claim 44, wherein a third of said multiple independent networks includes a scalable global signal network comprising nodal interconnections for enabling asynchronous global operations among nodes or sub-sets of nodes of said network.
- 46. The scalable, massively parallel computing structure as claimed in claim 44, wherein each node comprises a system-on-chip Application Specific Integrated Circuit (ASIC) comprising two processing elements each capable of individually or simultaneously working on any combination of computation or communication activity, or both, as required when performing particular classes of algorithms.
- 47. The scalable, massively parallel computing structure as claimed in claim 44, further including means for enabling switching of processing among one or more configured independent processing networks when performing particular classes of algorithms.
- 48. In a massively parallel computing structure comprising a plurality of processing nodes interconnected by multiple independent networks, each processing node comprising:
a system-on-chip Application Specific Integrated Circuit (ASIC) comprising two or more processing elements each capable of performing computation or message passing operations; means enabling rapid coordination of processing and message passing activity at each said processing element, wherein one or both of the processing elements performs calculations needed by the algorithm, while the other or both processing element performs message passing activities for communicating with other nodes of said network, as required when performing particular classes of algorithms.
- 49. A scalable, massively parallel computing system comprising:
a plurality of processing nodes interconnected by links to form a torus network, each processing node being connected by a plurality of links including links to all adjacent processing nodes; communication links for interconnecting said processing nodes to form a global combining tree network, and a similar combining tree for communicating global signals including interrupt signals; link means for receiving signals from said torus and global tree networks, and said global interrupt signals, for redirecting said signals between different ports of the link means to enable the computing system to be partitioned into multiple, logically separate computing systems.
- 50. The massively parallel computing system as claimed in claim 49, wherein the link means provides a function of redriving signals over cables between midplane devices that include a plurality of processing nodes, to improve the high speed shape and amplitude of the signals.
- 51. The massively parallel computing system as claimed in claim 49, wherein the link means performs a first type of signal redirection for removing one midplane from one logical direction along a defined axis of the computing system, and a second type of redirection that permits dividing the computing system into two halves or four quarters.
- 52. A massively parallel computing system comprising:
a plurality of processing nodes interconnected by independent networks, each processing node comprising a system-on-chip Application Specific Integrated Circuit (ASIC) comprising two or more processing elements each capable of performing computation or message passing operations; a first independent network comprising an n-dimensional torus network including communication links interconnecting said nodes in a manner optimized for providing high-speed, low latency point-to-point and multicast packet communications among said nodes or sub-sets of nodes of said network; a second of said multiple independent networks includes a scalable global tree network comprising nodal interconnections that facilitate simultaneous global operations among nodes or sub-sets of nodes of said network; and, partitioning means for dynamically configuring one or more combinations of independent processing networks according to needs of one or more algorithms, each independent network including a configured sub-set of processing nodes interconnected by divisible portions of said first and second networks, and, means enabling rapid coordination of processing and message passing activity at each said processing element in each independent processing network, wherein one, or both, of the processing elements performs calculations needed by the algorithm, while the other, or both, of the processing elements performs message passing activities for communicating with other nodes of said network, as required when performing particular classes of algorithms wherein each of said configured independent processing networks and node processing elements thereof are dynamically utilized to enable collaborative processing for optimizing algorithm processing performance.
CROSS-REFERENCE TO RELATED APPLICATION
[0001] The present invention claims the benefit of commonly-owned, co-pending U.S. Provisional Patent Application Serial No. 60/271,124 filed Feb. 24, 2001 entitled MASSIVELY PARALLEL SUPERCOMPUTER, the whole contents and disclosure of which is expressly incorporated by reference herein as if fully set forth herein. This patent application is additionally related to the following commonly-owned, co-pending United States patent applications filed on even date herewith, the entire contents and disclosure of each of which is expressly incorporated by reference herein as if fully set forth herein. U.S. patent application Ser. No. ______ (YOR920020027US1, YOR920020044US1 (15270)), for “Class Networking Routing”; U.S. patent application Ser. No. ______ (YOR920020028US1 (15271)), for “A Global Tree Network for Computing Structures”; U.S. patent application Ser. No. ______ (YOR920020029US1 (15272)), for ‘Global Interrupt and Barrier Networks”; U.S. patent application Ser. No. ______ (YOR920620030US1 (15273)), for ‘Optimized Scalable Network Switch”; U.S. patent application Ser. No. ______ (YOR920020031US1, YOR920020032US1 (15258)), for “Arithmetic Functions in Torus and Tree Networks’; U.S. patent application Ser. No. ______ (YOR920020033US1, YOR920020034US1 (15259)), for ‘Data Capture Technique for High Speed Signaling”; U.S. patent application Ser. No. ______ (YOR920020035US1 (15260)), for ‘Managing Coherence Via Put/Get Windows’; U.S. patent application Ser. No. ______ (YOR920020036US1, YOR920020037US1 (15261)), for “Low Latency Memory Access And Synchronization”; U.S. patent application Ser. No. ______ (YOR920020038US1 (15276), for ‘Twin-Tailed Fail-Over for Fileservers Maintaining Full Performance in the Presence of Failure”; U.S. patent application Ser. No. ______ (YOR920020039US1 (15277)), for “Fault Isolation Through No-Overhead Link Level
[0002] Checksums’; U.S. patent application Ser. No. ______ (YOR920020040US1 (15278)), for “Ethernet Addressing Via Physical Location for Massively Parallel Systems”; U.S. patent application Ser. No. ______ (YOR920020041US1 (15274)), for “Fault Tolerance in a Supercomputer Through Dynamic Repartitioning”; U.S. patent application Ser. No. ______ (YOR920020042US1 (15279)), for “Checkpointing Filesystem”; U.S. patent application Ser. No. ______ (YOR920020043US1 (15262)), for “Efficient Implementation of Multidimensional Fast Fourier Transform on a Distributed-Memory Parallel Multi-Node Computer”; U.S. patent application Ser. No. ______ (YOR9-20010211US2 (15275)), for “A Novel Massively Parallel Supercomputer”; and U.S. patent application Ser. No. ______ (YOR920020045US1 (152.63)), for “Smart Fan Modules and System”.
PCT Information
Filing Document |
Filing Date |
Country |
Kind |
PCT/US02/05571 |
2/25/2002 |
WO |
|