The present invention is related to computer networking devices. More specifically, the present invention is related to address translation systems for mapping local Internet Protocol “IP” addresses used by hosts on a private network to globally unique IP addresses for communication with hosts on the Internet.
Private networks are commonly connected to the Internet through one or more routers so that hosts (PCs or other arbitrary network entities) on the private network can communicate with nodes on the Internet. Typically, the host will send packets to locations both within its private network and on the Internet. To receive packets from the Internet, a private network or a host on that network must have a globally unique 32-bit IP address. Each such IP address has a four octet format. Typically, humans communicate IP addresses in a dotted decimal format, with each octet written as a decimal integer separated from other octets by decimal points.
Global IP addresses are issued to enterprises by a central authority known as the Internet Assigned Number Authority (“IANA”). The IANA issues such addresses in one of three commonly used classes. Class A IP addresses employ their first octet as a “netid” and their remaining three octets as a “hostid.” The netid identifies the enterprise network and the hostid identifies a particular host on that network. As three octets are available for specifying a host, an enterprise having class A addresses has 224 (nearly 17 million) addresses at its disposal for use with possible hosts. Thus, even the largest companies vastly under use available class A addresses. Not surprisingly, Class A addresses are issued to only very large entities such as IBM and ATT. Class B addresses employ their first two octets to identify a network (“netid”) and their second two octets to identify a host (“hostid”). Thus, an enterprise having class B addresses can use those addresses on approximately 64,000 hosts. Finally, class C addresses employ their first three octets as a netid and their last octet as a hostid. Only 254 host addresses are available to enterprises having a single class C netid.
Unfortunately, there has been such a proliferation of hosts on the Internet, coupled with so many class A and B licenses issued to large entities (who have locked up much address space), that it is now nearly impossible to obtain a class B address. Many organizations now requiring Internet access have far more than 254 hosts—for which unique IP addresses are available with a single class C network address. It is more common for a mid to large size enterprise to have 1000 to 10,000 hosts. Such companies simply can not obtain enough IP addresses for each of their hosts.
To address this problem, Network Address Translation (“NAT”) has been proposed. See K. Egevang and P. Francis, “The IP Network Address Translator (NAT),” Request For Comments “RFC” 1631, Cray Communications, NTT, May 1994 which is available at http://andrew2.andrew.cmuedu/rfc/rfc1631.html and is incorporated herein by reference for all purposes. NAT is based on the concept of address reuse by private networks, and operates by mapping the reusable IP addresses of the leaf domain to the globally unique ones required for communication with hosts on the Internet. In implementation, a local host wishing to access the Internet receives a temporary IP address from a pool of such addresses available to the enterprise (e.g., class C 254 addresses). While the host is sending and receiving packets on the Internet, it has a global IP address which is unavailable to any other host. After the host disconnects from the Internet, the enterprise takes back its global IP address and makes it available to other hosts wishing to access outside networks.
To implement NAT, a translation system must be provided between the enterprise private network and the Internet. Unfortunately, as the translation system is positioned between the enterprise and the Internet, communication performance can be degraded if NAT is not performed efficiently.
A major source of overhead of the NAT translation process is the allocation and management of client NAT addresses and their associated address pools each time a new address is required for each new connection that is initiated by the client device. Prior art solutions to this problem have either required substantial hardware resources or have otherwise created a bottleneck between the enterprise private network and the Internet. There is, therefore, a need in the art for providing a NAT that consumes fewer resources and yet provides faster performance.
In accordance with the present invention, the disadvantages and problems associated with prior art client NAT systems have been substantially reduced or eliminated. In particular, performance of the client NAT system is greatly enhanced.
In accordance with one embodiment of the present invention, a faster client NAT system is provided. Specifically, an embodiment of the present invention reduces computational overhead by initializing (internal) connection control blocks with the client NAT addresses when a storage pool of connection control blocks are allocated. This avoids the costly overhead as connections are created and destroyed. Unlike the prior art solutions, once the connection control blocks of the present invention are allocated, they remain in effect until the storage subpool is deleted.
Alternate embodiments of the present invention provide additional methods for using intervals to allocate and reclaim client NAT addresses as subpools are allocated and freed. These methods enable as large a pool of client NAT addresses as possible during the operation of the present invention.
These and other features and advantages of the present invention will be presented in more detail in the following specification of the invention and the figures.
For a more complete understanding of the present invention and its advantages, reference is now made to the following description, taken in conjunction with the accompanying drawings, in which:
1. Definitions
The following are terms that are used in the instant specification. Their definitions are provided to assist in the understanding of the embodiments that are described herein.
A “packet” is a collection of data and control information including source and destination node addresses and source and destination ports.
A “router” is a piece of hardware or software which operates at the network layer to direct packets between various nodes of one or more networks. The network layer generally allows pairs of entities in a network to communicate with each other by finding a path through a series of connected nodes.
A “client” is a PC or other arbitrary network entity residing on a network and capable of communicating with entities outside of its own network through a router or bridge.
A “server” is a node to which client requests are sent to execute a service.
A “server load balancer” is network device that receives connection requests and distributes them to servers in a server farm.
A “server farm” is a collection of servers.
A “virtual server” is a server farm identified by a single IP address called a virtual IP address.
A “virtual IP address” is an address defined at a server load balancer that is used to intercept flows that are to be directed to one of the servers in a server farm.
A “flow” is a five-tuple consisting of: source node IP address, source port number, destination node IP address, destination port number, and protocol number.
A “connection” is a pair of flows that define the traffic pattern of data sent between a given client and server. From the perspective of the client, the source and destination addresses and the source and destination ports of the response flow are reversed from those of the request flow.
A “connection control block” is a control block allocated by a server load balancer to manage a connection.
A “client NAT address” is the combination of an IP address and a port number. The IP address and port number of a client request are translated to a client NAT address by the server load balancer.
A “client NAT address range” is a sequenced set of client NAT addresses from which a client NAT address is drawn for the client NAT function.
2. Overview
Once the volume of Internet traffic grew to substantial levels, single servers proved to be a single point of failure. As a remedy, servers are now clustered to form a server farm. Server load balancers provide a means for evenly distributing work among servers in the server farm with scalability and redundancy.
The method and apparatus of the present invention concerns the allocation and deallocation of client NAT address ranges. The apparatus and method of the present invention can be implemented in hardware or software, or any combination there between. At the beginning of the process of defining the client NAT address range, at least one subpool is allocated. During the subpool allocation process, a special sub-process is called to assign client NAT addresses and register the allocated subpool. Upon termination (deletion) of the subpool, a termination routine is called to return addresses to the client NAT address range.
Load balancing devices utilize a database of active connections having a pair of flows and client NAT address information. When a packet is received, a check is made in the connection database for an entry corresponding to the packet's flow information. If no match is found, a connection object is allocated from the connection control block pool. Next, the client NAT address is either calculated or retrieved from a pool of preallocated client NAT addresses. The latter option is faster in practice, but results in poorer utilization of memory resources because a separate memory pool of client NAT addresses is required. After the client NAT address is allocated, the connection object is added to the active connection database.
An embodiment of the present invention utilizes a connection control block pool comprised of connection objects that already contain client NAT address information. Unlike the prior art, the present invention reduces computational overhead by initializing the connection control blocks with the client NAT addresses when a storage pool of connection control blocks is allocated. This avoids the costly overhead as connections are created and destroyed. Moreover, unlike the prior art solutions, once the connection control blocks of the present invention are allocated, they remain in effect until the storage subpool is deleted.
With respect to the connection database, there are five specific pieces of information in each flow: the destination IP address; the source IP address; the destination port number; the source port number; and the protocol (such as TCP or UDP, although other protocols may be identified). Using the client NAT of the present invention, an example inbound flow and an example outbound flow are compared in Table 1.
It should be noted that the “Client IP” and “Client Port” rows in Table 1 in the “From Server” column are values that are pre-populated.
Typically, a client NAT address range is used for connections to all of the servers in a given server farm. The first client NAT address range is allocated when the server farm is brought into service as illustrated in
3. Example Embodiments
The interval list address (ILA) 318 is a pointer to an interval list that is used to keep track of free and allocated client NAT addresses and the associated subpools that are allocated. The ILA 318 consists of a sequence of interval list elements (ILE's) 718 as shown in
As illustrated in
The general configuration of the memory pool 400 is shown in
The fixed block storage 407 of
The process begins in step 504, where the user invokes the procedure to configure a client NAT address range. Specifically, a client NAT address range pointer (CNP 302 of
In step 510, another procedure is used to invoke the memory management function 505 in order to allocate the memory pool. Initially, a single subpool is allocated. In step 512, the client NAT address range parameters are passed on to the memory manager 505 as well as the illustratively labeled initialization procedure “INIT_PROC” of step 516, the illustratively labeled delete procedure “DEL_PROC” of step 612 (of
After the memory pool has been created, execution continues to step 514 as illustrated in
In step 518, the interval list elements (ILE's) 718 (of
Within the memory manager 505, a condition may arise when the server load balancer has allocated all the connection blocks contained in the subpool. The dashed arrow 527 illustrates that when the number of free blocks reaches a minimum threshold, the “grow pool” procedure of steps 514 and 528 are invoked to create another subpool. In step 524, the calling server load balancer requests allocation of a connection block. In step 526, the memory manager 505 selects a block from the free list and returns its address (CONN_BLOCK) to the calling server load balancer. Recall that the client NAT address range is pre-assigned when the memory pool is allocated, the server load balancer never has to incur the overhead of allocating a client NAT address on a per connection basis after the subpool is created. This feature of this embodiment of the present invention precludes a considerable amount of just-in-time (runtime) processing that was necessary in prior art systems because the connection block of the present invention can be used repeatedly for many connections.
All of the blocks in the subpool are freed in step 610 of
The present invention, therefore, is well adapted to carry out the objects and attain both the ends and the advantages mentioned, as well as other benefits inherent therein. While the present invention has been depicted, described, and is defined by reference to particular embodiments of the invention, such references do not imply a limitation on the invention, and no such limitation is to be inferred. The invention is capable of considerable modification, alternation, alteration, and equivalents in form and/or function, as will occur to those of ordinary skill in the pertinent arts. The depicted and described preferred embodiments of the invention are exemplary only, and are not exhaustive of the scope of the invention. Consequently, the invention is intended to be limited only by the spirit and scope of the appended claims, giving full cognizance to equivalents in all respects.
This application claims priority to U.S. application Ser. No. 09/751,317, filed Dec. 29, 2000, now abandoned, entitled “Method Allocation Scheme for Server Load Balancers Services in a High Throughput Environment”. This application also relates to and was filed concurrently with U.S. Patent Application Ser. No. 09/780,755 filed Feb. 8, 2001, entitled “Preallocation of Client Network Address Translation Addresses for Client-Server Networks”.
Number | Name | Date | Kind |
---|---|---|---|
5559980 | Connors et al. | Sep 1996 | A |
5636216 | Fox et al. | Jun 1997 | A |
5774660 | Brendel et al. | Jun 1998 | A |
5793763 | Mayes et al. | Aug 1998 | A |
5860082 | Smith et al. | Jan 1999 | A |
5862452 | Cudak et al. | Jan 1999 | A |
5899994 | Mohamed et al. | May 1999 | A |
6006264 | Colby et al. | Dec 1999 | A |
6006272 | Aravamudan et al. | Dec 1999 | A |
6038233 | Hamamoto et al. | Mar 2000 | A |
6055236 | Nessett et al. | Apr 2000 | A |
6058431 | Srisuresh et al. | May 2000 | A |
6125430 | Noel et al. | Sep 2000 | A |
6154839 | Arrow et al. | Nov 2000 | A |
6178455 | Schutte et al. | Jan 2001 | B1 |
6212613 | Belair | Apr 2001 | B1 |
6243720 | Munter et al. | Jun 2001 | B1 |
6266707 | Boden et al. | Jul 2001 | B1 |
6286089 | Greiner et al. | Sep 2001 | B1 |
6353614 | Borella et al. | Mar 2002 | B1 |
6442662 | Komatsu | Aug 2002 | B1 |
6457061 | Bal et al. | Sep 2002 | B1 |
6470436 | Croft et al. | Oct 2002 | B1 |
6477612 | Wang | Nov 2002 | B1 |
6510496 | Tarui et al. | Jan 2003 | B1 |
6567405 | Borella et al. | May 2003 | B1 |
6574737 | Kingsford et al. | Jun 2003 | B1 |
6615357 | Boden et al. | Sep 2003 | B1 |
6618799 | Hagersten | Sep 2003 | B1 |
6650641 | Albert et al. | Nov 2003 | B1 |
6725321 | Sinclair et al. | Apr 2004 | B1 |
6765920 | Tari et al. | Jul 2004 | B1 |
6813645 | Meyer | Nov 2004 | B1 |
6904039 | Hung et al. | Jun 2005 | B1 |
Number | Date | Country | |
---|---|---|---|
Parent | 09751317 | Dec 2000 | US |
Child | 09781522 | US |