The present invention will become more fully understood from the detailed description given hereinbelow illustration only, and thus are not limitative of the present invention, and wherein:
Reference will now be made in detail to the present preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the description refers to the same or the like parts.
Please refer to
Each of the computer nodes CN 320, 330, 340, 350 mainly includes two processors CPU, chipset(s) and a BMC (Baseboard Management Controller). A typical implementation for the computer nodes CN 320, 330, 340, 350 is pretty much similar as 1-U type standalone server hardware. The chipset such as South Bridge or other integrated bridge chips connects with the BMC on each of the computer nodes CN 320, 330, 340, 350; and the BMC connects with a corresponding first network interface 321/331/341/351 to provides connections with a management network switch 360 and form private network connections. The management network switch 360 is used mainly for system management as well as clustering management of the clustering system. The BMC collects system management information on each of the computer nodes CN 320, 330, 340, 350 respectively, including operating parameters such as system events, temperature, cooling fan speeds, power mode, operating system (OS) status, etc. and sends alerts to a remote management host. The BMC also executes commands sent from the remote management host to manage the operation of the computer nodes CN 320, 330, 340, 350 respectively. The communication paths between the external management host and each of the BMCs on the compute node CN320/330/340/350 will be further disclosed in the following.
To avoid making fundamental system changes caused by service processor or chassis level central management module in the prior art, only common hardware are added and modified on the preset compute node CN 310. Instead of a first network interface 311, the preset compute node CN 310 includes an extra network interface 312 connecting with the management network switch 360 and the chipset, thereby allows the preset compute node CN 310 to join the private network connections with the management network switch 360 and the other compute nodes CN320, 330, 340, 350. On the other hand, the BMC on the preset compute node CN 310 is used to connect with an external management host through the first network interface 311. Meanwhile, the chipset on the preset compute node CN 310 also connects with the first network interface 311 through a system I/O bus 313 such as PCI or PCI-Express. In other words, on the preset compute nodes CN 310 the chipset connects “directly” with the first network interface 311 through the system I/O bus 313, as well as connects “indirectly” with the first network interface 311 through the BMC. Such design will allow the preset compute node CN 310 to provide the same function as the service controller or the central management module in the prior art.
First of all, through the first network interface 311 the external management host will be able to access the BMC on the preset compute node CN 310. This BMC collects the system information directly from some sensors configured on the preset compute node CN 310 and collects indirectly from the chipset (through the KCS interface 314) and a hardware monitor controller (not shown). Then, the system information will be sent to the external management host through the BMC, the sideband SMBus 315, the network interface controller NIC0 and the port interface; namely through the BMC and the first network interface 311. Oppositely, the external management host may send direct commands to control the preset compute node CN 310 through the BMC and manage the preset compute node CN 310.
In actual implementation, the preset compute node CN 310 needs to power on first since it has the feature of chassis level management function. To turn it ON, user needs to either use remote power-on scheme (as defined for IPMI based interface), or simply push a physical power button. Once the preset compute node CN 310 is boot-up, an application program called “System Management Software” will be invoked automatically. The system management software may turn on the rest of compute nodes CN 320, 330, 340, 350 in
The System Management Software operating on the preset compute node CN 310 can request the BMCs configured on the rest of other compute nodes CN 320, 330, 340, 350 in
To access individual compute node CN 310/320/330/340/350 from the external management host, an OS (Operating System) operating on the preset compute node CN 310 needs to provide Network Address Translation service. That is, by means of the Network Address Translation service, the external management host will be able to identify the network interface controller NIC0 or NIC1 through which the data is originally sending. Eventually, the external management host can reach the preset compute node CN310 through the first network interface 311 and the BMC, as well as the other compute nodes CN320, 330, 340, 350 through the extra network interface 312, the chipset, the system I/O bus 313 and the first network interface 311.
Please refer to
The preset compute node CN 310 now includes three network interface controllers NIC0, NIC1, NIC2 to fulfill all the network functions in the clustering system. Other computer nodes CN320, 330, 340, 350 will have only two network interface controllers. In addition, other network interfaces such as Ethernet, InfiniBand, 10 Giga-bit Ethernet and etc. may also be configured on each of the compute nodes.
In
In short, the present invention provides chassis-level central management function without any special hardware such as a dedicated service processor module. The present invention uses only common hardware to approach this feature, which is a very cost effective implementation for a small clustering system. Besides, all internal network topology will be completely encapsulated and users do not have to touch internal network structure. For user's viewpoint, this type of implementation is just like a single computer system. Then, users are released from very complicated network setup to make a clustering system. Moreover, the present invention utilizes only common and standard interface for the system management, such as IPMI. Therefore, the development for providing the service processor application is easy. Most of the basic functions are defined in the standard and the actual development is an application level running on a regular OS on the preset compute node, which actually plays the role like a head node. This would be much easier than developing a dedicated service processor using an independent OS, low-level device driver and management application.
The invention being thus described, it will be obvious that the same may be varied in many ways. Such variations are not to be regarded as a departure from the spirit and scope of the invention, and all such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the following claims.
This application is a non-provisional application of the U.S. provisional application Ser. No. 60/822,540 to Tomonori Hirai, entitled “System Management for a Small Clustering System” filed on Aug. 16, 2006.
| Number | Date | Country | |
|---|---|---|---|
| 60822540 | Aug 2006 | US |