The present invention relates generally to shared file systems for use in computing environments, and more particularly to techniques for providing a highly available shared file system.
In today's computing environments, many computers can have access to many areas of storage on one or several storage subsystems. The areas of storage on disks, CD-ROMs and the like are usually arranged into file systems consisting of organized directories and files for ease of access by the user of a given computer. In many applications, it is useful to be able to share such file systems between multiple computers.
Software is available to enable such file sharing for different types of computing systems. For example, the Network File System (“NFS”), developed by Sun Microsystems Inc., allows sharing of file systems in such computing environments. NFS allows a server, known as the NFS server, to “export” a file system such that it is available for use by other computers, known as NFS clients. The NFS clients can then “mount” the file system, allowing them to access the file system through the NFS server that exported the file system.
Many computing environments implement critical software, such as database applications, that must be highly available. That is, little down time can be tolerated for these applications. Such critical software often requires a file sharing system. NFS, though simple to implement, presents a single point of failure for a given file system because all the NFS clients access the file system through the NFS server. NFS is therefore not an ideal file sharing system in its present form for use in a highly available environment.
Highly available file systems, known as “clustered file systems” (“CFS”) have been developed in order to provide a more robust file system sharing solution. A typical CFS allows storage devices such as disks, CD-ROMs and the like to be attached to all nodes in the cluster and to be accessed directly by every node. The storage devices appear to be controlled and managed by a single entity, which continues to operate as nodes in the cluster fail and recover. Since all nodes have direct access to all storage, there is no single point of failure. Though a CFS provides a highly available shared file system, a CFS is very complicated, expensive, and difficult to implement.
There is a need for a shared file system that is highly available but relatively simple to implement and that avoids the other inadequacies and disadvantages associated with implementing a full scale CFS.
In accordance with the invention, there is provided a method for sharing files amongst a plurality of servers. A floating network address is provided that can be associated with any one of the servers. The floating network address is associated with a first server of the plurality of servers. An area of data on a storage device is made available by the first server to be accessed by the other servers. The other servers then refer to the floating network address to access the area of data through the first server. In accordance with a preferred embodiment, all servers, including the first server, refer to the floating network address to access the area of data through the first server.
Further in accordance with the invention, the floating network address is transferred from the first server to a second server. The area of data is then made available by the second server to be accessed by the other servers. All the servers continue to access the area of data by referring to the floating network address.
According to an embodiment of the invention, the first and second servers are NFS servers and the other servers are NFS clients. The first server makes the area of data accessible to the other servers by issuing an NFS export command to export the area of data. The other servers, and preferably all the servers, issue an NFS mount command to mount the area of data at the floating network address. When the floating network address is transferred from the first server to the second server, the second server then issues an NFS export command to export the area of data. All the servers continue to refer to the floating network address to access the area of data. The failover of the NFS server is thus transparent to the NFS clients.
Similar apparatus and program products are provided in accordance with the invention. In the preferred embodiments, all servers access a shared file system in the same manner, through an NFS mount at the floating network address. At any given time, one of the servers exports a given file system at the floating network address. Thus, all servers always have access to the file system, no matter which server exported it. The invention is advantageous in that a highly available shared file system is provided that is relatively uncomplicated and inexpensive to implement.
In
The file sharing utility preferably employed by the servers is the Network File Sharing system, or NFS. As is known in the art, NFS allows a server, for example server 12, to attach an area on the disk array 22, and issue an NFS “export” command in order to cause the file system to be shared by the other servers, either globally or selectively. The other servers, if selected as part of the export, can issue an NFS “mount” command to gain access to the shared file system. After mounting the file system, a given server 14, 16, or 18 can now access the file system via the server 12. The server 12 that issued the export command is referred to as an NFS server, while the other servers 14, 16, and 18 that mounted the shared file system are referred to as NFS clients. The servers 14, 16, and 18 may also be in communication with the storage system 20 or with another storage system or device (not shown) and can thus also operate as an NFS server. The server 12 is also capable of operating as an NFS client, for accessing files exported by another server. Furthermore, the NFS server 12 can export the shared file system to itself. The server 12 then, operating as an NFS client, accesses the filesystem in a “loopback” fashion as if the file system was remote, even though it is actually local. Generally, a given server operates as an NFS server when exporting file systems located on storage systems to which it is attached, and operates as an NFS client when mounting and accessing filesystems located on storage systems to which the other servers are attached, or when mounting and accessing filesystems that it has exported to itself.
NFS is a connectionless, stateless protocol, employing UDP datagrams for protocol information transfer. As such, the protocol will not fail in the event that the NFS server becomes unavailable. Rather, an NFS client will simply wait for a response to its outstanding request—or continue to re-issue it—until the NFS server recovers. However, some applications that utilize the NFS client are not tolerant of extended wait times and may fail ungracefully in the event that an NFS server becomes unavailable. It is therefore advantageous to implement a highly available NFS system in accordance with the principles of the invention.
In accordance with the invention, the servers 12, 14, 16, and 18 are members of a cluster. Each server implements certain functionality common to available cluster package software. Cluster software is available to enable clustered services to “fail over” between servers. If for example server 12 is hosting a particular service, and server 12 becomes unavailable, for example due to failure or in response to a command to relinquish the service, one of the other servers such as server 14 will assume the hosting of the service in a manner transparent to the other servers, thus preventing interruption of the service. Known cluster software such as the Linux Global File System (GFS), available from Sistina Inc., or the Linux Virtual Server (LVS) software, available from Red Hat Inc., provide resources for server failure detection and for the transfer of services between servers. In particular, such cluster software provides a resource for associating a service with a “floating IP address”, a unique IP address that moves between servers as servers fail or otherwise relinquish a service. Server 12 may for instance host a clustered service associated with the floating IP address. If server 12 fails, or relinquishes the service for other reasons, one of the other servers such as server 14 assumes ownership of the floating IP address. The clustered service fails over with the floating IP address to be hosted by the server 14. The failover is transparent to the other servers in the cluster, all of which continue to utilize the service associated with the floating IP address. Most available cluster software packages allow software scripts to be written to associate a given service with the failure detection and floating IP address resources to enable transparent failover of the service.
Referring now to
This can be done, for example, by providing a software NFS script for the cluster software package employed. Referring to
Referring back to
When the server 12 needs to set up the file system for sharing, it first attaches the areas on the disk array 22 where the file system is located, and locally mounts the file system at a private location shown as “/local/xyz”. The server 10 then exports the file system for use by all servers. The “export/local/xyz” function depicted in
The file system has thus been exported to all the servers 12, 14, 16, and 18, with options controlling for example the type of access each server will have to the file system “/local/xyz”, for example read-only, read-write, etc.
Meanwhile, the file system is mounted by all servers. Referring to
This command mounts the filesystem xyz owned by the server with address IP_Float to a directory /mnt/xyz which is now available to the servers 12, 14, 16, and 18. (This mount functionality is depicted in
Note that, as previously generally described, the server 12 also exported the file system “/local/xyz” to itself. The server 12 executes the mount command as well. Thus, the server 12 accesses the file system “/xyz” via the floating address IP_Float rather than its own local address IP_A. As such, the accesses by the server 12 to the file system “/xyz” are looped back via the floating IP address IP_Float to its local mount. All servers 12, 14, 16, and 18 thus access the file system “/xyz” at the same remote address. Thus, a clustered application running on several or all of the servers accesses the file system “/xyz” via NFS in a consistent manner, allowing functions such as file locking to be properly serviced.
Referring now to
As shown in
Note that, in accordance with the principles of the invention, it is not necessary to provide a dedicated backup server in order to provide a highly available NFS file system. Also, in accordance with the principles of the invention, NFS clients on any of the servers 12, 14, 16, and 18 within the cluster 10 can mount the file system IP_Float:/local/xyz and will be able to continually access it in the event of a failure. Furthermore, multiple file systems may be owned by different servers at the same time. A floating IP address can be associated with each server that owns one or more file systems. All the file systems on a given server can then fail over to another server when the server owning the file systems fails.
The present invention is not to be limited in scope by the specific embodiments described herein. Indeed, various modifications of the present invention, in addition to those described herein, will be apparent to those of ordinary skill in the art from the foregoing description and accompanying drawings. Thus, such modifications are intended to fall within the scope of the following appended claims. Further, although the present invention has been described herein in the context of a particular implementation in a particular environment for a particular purpose, those of ordinary skill in the art will recognize that its usefulness is not limited thereto and that the present invention can be beneficially implemented in any number of environments for any number of purposes. For example, though the preferred embodiment of the invention employs NFS as the shared file system, the principles of the invention could be employed with other file systems, for example Samba, which is a Microsoft Windows file system, or the Andrew file system, which is used by IBM computers, or Coda, another known shared file system. Accordingly, the claims set forth below should be construed in view of the full breadth and spirit of the present invention as disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
5941956 | Shirakihara et al. | Aug 1999 | A |
6052725 | McCann et al. | Apr 2000 | A |
6330605 | Christensen et al. | Dec 2001 | B1 |
6356863 | Sayle | Mar 2002 | B1 |
6360331 | Vert et al. | Mar 2002 | B1 |
6393485 | Chao et al. | May 2002 | B1 |
6681327 | Jardin | Jan 2004 | B1 |
6711559 | Kogota et al. | Mar 2004 | B1 |
6725047 | Farley et al. | Apr 2004 | B1 |
6816891 | Vahalia et al. | Nov 2004 | B1 |
7020669 | McCann et al. | Mar 2006 | B1 |
20020078174 | Sim et al. | Jun 2002 | A1 |
20020120763 | Miloushev et al. | Aug 2002 | A1 |
20020161848 | Willman et al. | Oct 2002 | A1 |
20030078965 | Cocotis et al. | Apr 2003 | A1 |
20030079030 | Cocotis et al. | Apr 2003 | A1 |
20030088650 | Fassold et al. | May 2003 | A1 |
20030097454 | Yamakawa et al. | May 2003 | A1 |
20030158836 | Venkatesh et al. | Aug 2003 | A1 |
20030177107 | Brown et al. | Sep 2003 | A1 |
20040107288 | Menninger et al. | Jun 2004 | A1 |
20040236745 | Keohane et al. | Nov 2004 | A1 |
20050210074 | Nakatani et al. | Sep 2005 | A1 |
20050267920 | Helliker et al. | Dec 2005 | A1 |
Number | Date | Country |
---|---|---|
1363441 | Nov 2003 | EP |
1450538 | Aug 2004 | EP |
2418326 | Mar 2006 | GB |
WO 0022544 | Apr 2000 | WO |
WO20040038666 | May 2004 | WO |
WO2004066588 | Aug 2004 | WO |
WO 2005109223 | Nov 2005 | WO |