System and method to balance servers based on server load status

Information

  • Patent Grant
  • 10447775
  • Patent Number
    10,447,775
  • Date Filed
    Tuesday, March 27, 2018
    6 years ago
  • Date Issued
    Tuesday, October 15, 2019
    5 years ago
Abstract
Provided are methods and systems for balancing servers based on a server load status. A method for balancing servers based on a server load status may commence with receiving, from a server of a plurality of servers, a service response to a service request. The service response may include a computing load of the server. The method may continue with receiving a next service request from a host. The method may further include determining, based on the computing load of the server, whether the server is available to process the next service request. The method may include selectively sending the next service request to the server based on the determination that the server is available to process the next service request.
Description
BACKGROUND OF THE INVENTION
Field

This invention relates generally to data communications, and more specifically, to a method and system to service load balancers.


Background

Web services and cloud computing are deployed in an unprecedented pace. New servers are unloaded and installed at datacenters every day. Demands of web services and corporate computing come from all directions. Consumer oriented services include iPhone™ apps, mobile applications such as location based services, turn-by-turn navigation services, e-book services such as Kindle™, video applications such as YouTube™ or Hulu™, music applications such as Pandora™ or iTunes™, Internet television services such as Netflix™, and many other fast growing consumer Web services. On the corporate front, cloud computing based services such as Google™ docs, Microsoft™ Office Live and Sharepoint™ software, Salesforce.com™'s on-line software services, tele-presence and web conferencing services, and many other corporate cloud computing services.


As a result, more and more servers are deployed to accommodate the increasing computing needs. Traditionally these servers are managed by server load balancers (SLB). SLB are typically network appliances, such as A10 Network's AX-Series traffic managers. SLB manage the load balancing of servers based on incoming service requests. Common methods to balance load among servers is to distribute the service requests based on the applications (HTTP, FTP, etc.), service addresses such as URL, priorities based on network interfaces or host IP addresses. SLB may distribute service requests additionally in a round robin fashion to the servers, assuming and ensuring the servers would be evenly loaded. However, different service requests have different service computing consequences. A server may be fully loaded with only a handful of service requests while another server remains mostly idle even with plenty of service requests. SLB may inappropriately send another request to a busy server, incorrectly considering the busy server being readily available, instead of sending the request to an idle server.


It would be beneficial if the SLB are aware of the computing load situation of a server so that SLB can better select a server to process a service request.


Therefore, there is a need for a system and method for a server load balancer to select a server based on the server load status.


BRIEF SUMMARY OF THE INVENTION

Provided are computer-implemented methods and systems for balancing servers based on a server load status. According to one example embodiment, a system for balancing servers based on a server load status may include a service gateway and a plurality of servers configured to process service requests. The service gateway may be configured to receive, from a server of the plurality of servers, a service response to a service request. The service response may include a computing load of the server. The service gateway may be configured to receive a next service request from a host. The service gateway may be configured to determine whether the server is available to process the next service request. The determination may be made based on the computing load of the server. Based on the determination that the server is available to process the next service request, the service gateway may selectively send the next service request to the server.


According to one example embodiment, a method for balancing servers based on a server load status may commence with receiving, from a server of a plurality of servers, a service response to a service request. The service response may include a computing load of the server. The method may continue with receiving a next service request from a host. The method may further include determining, based on the computing load of the server, whether the server is available to process the next service request. The method may continue with selectively sending the next service request to the server based on the determination that the server is available to process the next service request.


System and computer program products corresponding to the above-summarized methods are also described and claimed herein.





BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE FIGURES


FIG. 1 illustrates an embodiment of a service gateway and an embodiment of the server pool according to the present invention.



FIG. 2 is a block diagram illustrating an embodiment of the processing of service requests by the service gateway.



FIG. 3 illustrates possible values for server status.



FIG. 4 is a block diagram illustrating an embodiment of the processing of service requests by the service gateway after receiving server status.



FIG. 5 illustrates the service gateway processing the service request according to service priorities.



FIG. 6 illustrates an embodiment of processing service requests by the service gateway 110.



FIG. 7 is a flowchart illustrating an embodiment of the method for processing service requests by the service gateway.





DETAILED DESCRIPTION OF THE INVENTION

The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments. Thus, the present invention is not intended to be limited to the embodiment shown but is to be accorded the widest scope consistent with the principles and features described herein.


The invention can take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment containing both hardware and software elements. In a preferred embodiment, the invention is implemented in software, which includes but is not limited to firmware, resident software, microcode, etc.


Furthermore, the invention can take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.


The medium can be an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium. Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette, a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk and an optical disk. Current examples of optical disks include compact disk-read only memory (CD-ROM), compact disk-read/write (CD-R/W) and DVD.


A data processing system suitable for storing and/or executing program code will include at least one processor coupled directly or indirectly to memory elements through a system bus. The memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution.


Input/output or I/O devices (including but not limited to keyboards, displays, point devices, etc.) can be coupled to the system either directly or through intervening I/O controllers.


Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.


The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified local function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.


The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.



FIG. 1 illustrates an embodiment of the service gateway 110 and an embodiment of the server pool 200 according to the present invention. The service gateway 110 receives a service request 301 from a host 100. Service request 301 is delivered over a data network 153. In one embodiment, service request 301 is a Web service request such as an HTTP (Hypertext Transport Protocol) request, a secure HTTP request, an FTP (File Transfer Protocol) request, a file transfer request, an SIP (Session Initiation Protocol) session request, a request based on Web technology, a video or audio streaming request, a Web conferencing session request, or any request over the Internet or corporate network.


Host 100 is a computing device with network access capabilities. The host 100 is operationally coupled to a processor 103 and a computer readable medium 104. The computer readable medium 104 stores computer readable program code for implementing the various embodiments of the present invention as described herein. In one embodiment, host 100 is a workstation, a desktop personal computer or a laptop personal computer. In one embodiment, host 100 is a Personal Data Assistant (PDA), a smartphone, or a cellular phone. In one embodiment, host 100 is a set-top box, an Internet media viewer, an Internet media player, a smart sensor, a smart medical device, a net-top box, a networked television set, a networked DVR, a networked Blu-ray player, or a media center.


In one embodiment, data network 153 is an Internet Protocol (IP) network. In one embodiment, data network 153 is a corporate data network or a regional corporate data network. In one embodiment, data network 153 is an Internet service provider network. In one embodiment, data network 153 is a residential data network. In one embodiment, data network 153 includes a wired network such as Ethernet. In one embodiment, data network 153 includes a wireless network such as a WiFi network, or cellular network.


The service gateway 110 is operationally coupled to a processor 113 and a computer readable medium 114. The computer readable medium 114 stores computer readable program code, which when executed by the processor 113, implements the various embodiments of the present invention as described herein. In some embodiments, service gateway 110 is implemented as a server load balancer, an application delivery controller, a service delivery platform, a traffic manager, a security gateway, a component of a firewall system, a component of a virtual private network (VPN), a load balancer for video servers, or a gateway to distribute load to a plurality of servers.


Server pool 200 comprises a plurality of servers, for example server 210. Server 210 is operationally coupled to a processor 213 and a computer readable medium 214. The computer readable medium 214 stores computer readable program code, which when executed by the processor 213, implements the various embodiments of the present invention as described herein. In some embodiments, the computer readable program code implements server 210 as a Web server, a file server, a video server, a database server, an application server, a voice system, a conferencing server, a media gateway, an SIP server, a remote access server, a VPN server, or a media center.


In one embodiment, server pool 200 further includes server 220 and server 230. In an embodiment, server pool 200 is located in a datacenter, a server room, or an office. In an embodiment, the plurality of servers in server pool 200 may be located geographically over several locations or several datacenters. Service gateway 110 connects to server pool 200 via data network 155. In one embodiment, data network 155 is the same as data network 153. In one embodiment, data network 155 is different from data network 153. In one embodiment, host 100 does not have direct access to data network 155. In one embodiment, host 100 has direct access to data network 155.



FIG. 2 is a block diagram illustrating an embodiment of the processing of service requests by the service gateway 110. Based on information in service request 301, service gateway 110 selects server 210 to process service request 301. In one embodiment, service request 301 includes a Universal Resource Location (URL) 311.


Service gateway 110 selects server 210 based on service request URL 311. Server pool 200 is configured to process service request 301 efficiently, by using service request URL 311. The servers in the server pool 200 are configured as primary servers for particular URL's, and as secondary servers for other URLs. In an embodiment, server 210 is configured as a primary server for URL 311, whereas server 220 is configured as a secondary server for URL 311. In this embodiment, service gateway 110 preferably selects server 210 to process service request 301 as server 210 is configured as the primary server for URL 311. Service gateway 110 may select server 220 under certain circumstances to process service request 301 as server 220 is configured as the secondary server for URL 311. In one embodiment, there is a second service request URL 312 corresponding to a second service request 302. Server 220 may be configured to process second service request 302 with request URL 312 as a primary server.



FIG. 7 is a flowchart illustrating an embodiment of the method for processing service requests by the service gateway 110. Referring to both FIGS. 2 and 7, service gateway 110 establishes service session 160 with server 210 and sends service request 301 to server 210 (700). Upon processing service request 301, server 210 sends a service response 321 to service gateway 110 (701). Server 210 includes in the service response 321 a server status 218. Server status 218 indicates the availability or the computing load status of server 210. In one embodiment, server status 218 reflects a load representing CPU utilization, memory utilization, network utilization, storage utilization or a combination of one or more of the utilizations. In general, server status 218 summarizes how busy server 210 is.


Service gateway 110 obtains the server status 218 from the service response 321 (702) and relays the service response 321 to host 100 (703). In one embodiment, service gateway 110 modifies the service response 321 by removing server status 218 from service response 321. The service gateway 110 then sends the modified service response to host 100.


Service response 321 includes a result from the servicing of the service request 301. The service response 321 further includes the server status 218 associated with a tag 324. Service gateway 110 identifies the tag 324 from service response 321 and extracts server status 218 associated with tag 324. In one embodiment, service request 301 is an HTTP request, and service response 321 is an HTTP response. In this embodiment, tag 324 is in the HTTP header of the HTTP response. In one embodiment, service response 321 includes an HTML document. In this embodiment, tag 324 is an HTML tag. In another embodiment, service response 321 includes an XML document, and tag 324 can be an XML tag. In one embodiment, service response 321 is an SIP response packet, and tag 324 is an SIP tag. In one embodiment, service response 321 is an FTP response, and tag 324 is a special FTP reply code.



FIG. 3 illustrates possible values for server status 218. Other values for the server status 218 may be configured according to the needs of the system. In one embodiment, a value of 0 for server status 218 indicates that server 210 is not busy. Server 210 can handle new requests without any delay. For example, if service gateway 110 receives service request 301 with a request URL 311, service gateway 110 will select server 210 to process service request 301.


A value of 1 for server status 218 indicates that server 210 is busy. While server 210 can continue to serve as a primary server for URL 311, server 210 cannot serve as a secondary server. For example, server 210 is configured as a secondary server for URL 312. If service gateway 110 receives service request 302 with a request URL 312, service gateway 110 does not select server 210 to process service request 302.


A value of 2 for server status 218 indicates that server 210 is very busy. In addition to indicating that server 210 cannot serve as a secondary server; the server status 218 of 2 also indicates that service gateway 110 should apply a restriction prior to selecting server 210 to process a new service request as a primary server. For example, if service gateway 110 receives service request 301 with a request URL 311, service gateway 110 applies restriction prior to selecting server 210 to process service request 301. The restriction will be explained in further details with reference to FIG. 4.



FIG. 4 is a block diagram illustrating an embodiment of the processing of service requests by the service gateway 110 after receiving server status 218. In one embodiment, service gateway 110 includes a service request buffer 331. Service request buffer 331 is configured to store service requests with request URL 311. In an embodiment, server 210 is configured as a primary server for URL 311 and as a secondary server for request URL 313. Service gateway 110 further includes service request buffer 332, which is configured to store service requests with request URL 313. In an example, service request buffer 332 includes service request 411 with request URL 313.


Service gateway 110 received server status 218 from server 210 in a service response to a previous service request according to FIG. 2 above. Referring to FIGS. 4 and 7, the service gateway 110 receives a next service request 410 from the host 100 (704). The service gateway 110 stores the service request 410 with request URL 311 in service request buffer 331 and processes service request 410 in service request buffer 331 according the value of server status 218.


In one embodiment, server status 218 has a value of 0, indicating server 210 is “not busy”. Service gateway 110 examines (first) service request buffer 331 and finds service request 410 with (first) request URL 311 for which server 210 is configured as the primary server (705). Service gateway 110 selects server 210 and sends service request 410 to server 210 (707). In one embodiment, service request buffer 331 is empty when service gateway 110 receives the next service request with request URL 311. The service gateway 110 sends this service request to server 210 without placing it in the service request buffer 331.


In this embodiment with server status 218 of value 0, service gateway 110 examines (second) service request buffer 332 and finds service request 411 with (second) request URL 313 for which server 210 is configured as the secondary server (706). As server status 218 indicates server 210 is available to serve as a secondary server, service gateway 110 may select server 210 to process service request 411 (711).


In one embodiment, request buffer 332 is empty when service gateway 110 receives the next server request which includes request URL 313. Service gateway 110 may select server 210 to process this service request without placing it in the service request buffer 332.


In one embodiment, server status 218 has a value or 1, indicating server 210 is busy and is not available to serve as a secondary server for URL 313 but is available to serve as a primary server. Service gateway 110 examines service request buffer 331. In one embodiment, service gateway 110 finds service request 410 in service request buffer 331 (705). Service gateway 110 sends service request 410 to server 210 (708). In one embodiment, service request buffer 331 is empty when service gateway 110 receives the next service request which includes request URL 311. Service gateway 110 sends this service request to server 210 without placing it in service request buffer 331.


In this embodiment with server status 218 of value 1, service gateway 110 examines service request buffer 332 and finds service request 411 in service request buffer 332 (706). Service gateway 110 does not select server 210 to process service request 411 (712). Instead, the service gateway 110 may select another server in the server pool 200 to process the service request 411 (713). In one scenario, the service request buffer 332 is empty when the service gateway 110 receives the next service request which includes request URL 313. Service gateway 110 does not select server 210 to service request (712).


In one embodiment server status 218 is of value 2, indicating server 210 is “very busy”. In this embodiment, server 210 is not capable of serving as a secondary server to URL 313 but may serve as a primary server with restrictions.


In this embodiment with server status 218 of value 2, service gateway 110 examines service request buffer 331 and finds service request 410 (705). Service gateway 110 does not automatically send request 410 to server 210. In one embodiment, service gateway 110 includes a timer 117. Service gateway 110 waits for the timer 117 to expire (709) before selecting server 210 and sending request 410 to server 210 (710). In one embodiment, service gateway 110 configures timer 117 after receiving service response 321 with server status 218 of value 2. Timer 117 may be configured for duration of, for example 1 minute, 30 seconds, 1 second, 400 milliseconds, 5 milliseconds, 300 microseconds, or any other duration such that service to service request 410 is not severely affected. In one embodiment, timer 117 duration is based on the session protocol time out duration for service request 410.


In one embodiment with server status 218 of value 2, service request buffer 331 is empty when service gateway 110 receives a next service request with request URL 311. Service gateway 110 stores this service request in service request buffer 331. At a later time when timer 117 expires, service gateway 110 examines service request buffer 331 and finds this service request (705). When the timer 117 expires (709), service gateway 110 selects the server 210 and sends this service request to server 210 (710). In one embodiment, service gateway 110 sets up timer 117 again after processing service request buffer 331. In one embodiment, service gateway 110 processes a plurality of requests in service request buffer 331 before setting timer 117 again. In one embodiment, service gateway 110 cancels timer 117 when service gateway 110 receives server status 218 of value 0 or 1.


In this embodiment with server status 218 of value 2, service gateway 110 examines service request buffer 332 and finds service request 411 in service request buffer 332 (706). Service gateway 110 does not select server 210 to process service request 411 (714). Instead, the service gateway 110 may select another server in the server pool 200 to process the service request 411 (713). In one embodiment, the service request buffer 332 is empty when the service gateway 110 receives the next service request which includes request URL 313. Service gateway 110 does not select server 210 to service the service request.



FIG. 5 illustrates the service gateway 110 processing the service request according to service priorities. Service priorities may be configured based on a variety of parameters, such as the request URL of the service request, the host from which the service request is sent, and the data network where service gateway 110 receives the service request. In one embodiment, the service priority association 550 between a priority and the parameter value is stored in service gateway 110. For example, priority 401 is associated with URL 311, and the association 550 between priority 401 and URL 311 is stored in service gateway 110. In another example, priority 401 is associated with host 100, such as the IP address of host 100 or the user identity of host 100, and the association 550 between the priority 401 and the host 100 is stored in service gateway 110. In another example, priority 401 is associated with the network interface from which service request 410 is received, and the association 550 between the priority 401 and the network interface is stored in service gateway 110. In one embodiment, service gateway 110 includes a datastore (not shown) storing the association 550. Service gateway 110 determines priority 401 for the service request 410 by matching the appropriate parameter of the service request 410 to the stored association 550.


In one embodiment, service gateway 110 includes service request buffer 333 configured to store service requests with request URL 311, and service request buffer 331 also configured to store service requests for URL 311. Service request buffer 333 is configured to store service requests with priority 403, which is higher than priority 401. Service request buffer 331 is configured to store service requests with priority 401. In one embodiment, the server status 218 has a value of 2, and service gateway 110 stores service request 410 with request URL 311 and priority 401 in service request buffer 331. Service gateway 110 further receives a service request 412 with request URL 311 and service priority 403. Service gateway 110 stores service request 412 in service request buffer 333.


When the timer 117 expires, service gateway 110 examines service request buffer 333 of higher priority 403 before examining service request buffer 331 of lower priority 401. In this example, service gateway 110 processes service request 410 before processing service request 412.


In one embodiment, service gateway 110 receives a new server status 219 for server 210 of value 0 or 1, and service gateway 110 continues to examine service request buffer 333 of higher priority 403 before examining service request buffer 331 of lower priority 401. In this example, service gateway 110 processes service request 412 before processing service request 410.


In one embodiment, service gateway 110 processes all service requests from service request buffer 333 before processing service request buffer 331, regardless of the value of the server status 218. In one embodiment, service gateway 110 includes a serving ratio 119 where service gateway 110 processes service requests in service request buffer 333 and service request buffer 331 according to ratio 119, where ratio 119 favors the higher priority service request buffer 333 to the lower priority service request buffer 331 in order to avoid starving the lower priority service requests in service request buffer 331.


In a preferred embodiment, server 210 is configured to serve as primary server for a plurality of URLs. FIG. 6 illustrates an embodiment of processing service requests by the service gateway 110. In one embodiment, a hashing function 502 is applied to URL 311 and URL 313. When hashing function 502 is applied to URL 311, the result is a hashed value 511. Similarly, when hashing function 502 is applied to URL 313, the result is a hashed value 513. In one embodiment, Server 210 is configured to serve as primary server for hashed value 520. In one embodiment, service gateway 110 stores hashing function 502 and hashed value 520. Examples for hashing function 502 include MD5 (Message-Digest algorithm 5) hashing function, a Jenkins hashing function, a hashing function applicable for a sequence of characters, or a hashing function for table lookup.


When service gateway 110 receives service request 410 with request URL 311. Service gateway 110 applies hashing function 502 to request URL 311 to yield hashed value 511. Service gateway 110 compares hashed value 511 to hashed value 520. If hashed value 511 matches hashed value 520, service gateway 110 selects server 210 to process service request 410 in the manner described above. In one embodiment, service gateway 110 receives service request 411 with request URL 313. Service gateway 110 applies hashing function 502 to request URL 313 to yield hashed value 513. Service gateway 110 compares hashed value 513 to hashed value 520. If hashed value 513 matches hashed value 520, service gateway 110 selects server 210 to process service request 411. In one embodiment, hashed value 513 does not match hashed value 520, and service gateway 110 does not automatically select server 210.


In one embodiment, server 210 is configured as a secondary server to hashed value 513. Service gateway 110 may select server 210. In one embodiment, server 210 is not configured to serve hashed value 513, and service gateway 110 does not select server 210 to process service request 411.


In this embodiment, service gateway 110 applies hashing function 502 to a service request when service gateway 110 processes the service request. Once the service gateway 110 determines that the server 210 is configured either as the primary or the secondary server to the hashed value 513, the service gateway 110 processes the service request 411 using the server status 218 as described above. One of ordinary skill in the art will understand how to combine the application of the hashing function to the process described above without departing from the spirit and scope of the present invention.


Although the present invention has been described in accordance with the embodiments shown, one of ordinary skill in the art will readily recognize that there could be variations to the embodiments and those variations would be within the spirit and scope of the present invention. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Claims
  • 1. A system for balancing servers based on a server load status, the system comprising: a plurality of servers configured to process service requests; anda service gateway comprising a processor and a computer readable storage medium having a computer readable program code embodied therewith, wherein the computer readable program code when executed by the processor causes the service gateway to: receive, from a server of the plurality of servers, a service response to a service request, the service request including a Uniform Resource Locator (URL) of a plurality of URLs, the service response comprising a computing load of the server, each server of the plurality of servers being configured to act as a primary server for the URL, wherein the service gateway is pre-configured to: automatically select, for each URL of the plurality of URLs, the primary server to process service requests associated with the each URL, andselect a secondary server to process the service requests associated with each URL for when the primary server is unavailable to process the service requests;receive a next service request from a host, the next service request including the URL;based on the URL and the computing load of the server, determine whether the server can currently act as the primary server for processing the next service request; andbased on the determination, selectively send the next service request to the server.
  • 2. The system of claim 1, wherein the service gateway is further configured to: receive the service request from the host, the service request including a first Uniform Resource Locator (URL) of a plurality of URLs, the first URL being associated with at least one server of the plurality of servers;determine that the server is the primary server configured to process the first URL and a further server is the secondary server configured to process the first URL;based on the determination that the server is the primary server, select the server to process the service request; andsend the service request to the server.
  • 3. The system of claim 2, wherein each of the plurality of URLs is individually resolved to the primary server configured to process the URL and the secondary server configured to process the URL, each server of the plurality of servers being configured as one of primary servers and one of secondary servers for different URLs.
  • 4. The system of claim 3, wherein the computing load of the server includes a server status of the server.
  • 5. The system of claim 4, wherein the server status indicates one or more of the following: the server is not busy when the server is available for serving as the primary server and is available for serving as the secondary server, the server is busy when the server is available for serving as the primary server and is unavailable for serving as the secondary server, and the server is very busy when the server is available for serving as the primary server with restrictions and is unavailable for serving as the secondary server.
  • 6. The system of claim 4, wherein the determining by the service gateway whether the server is available to process the next service request comprises: determining that the server status indicates the server is not busy;determining that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is not busy and the next service request includes the second URL associated with the server as the primary server, selecting the server to process the next service request.
  • 7. The system of claim 4, wherein the determination by the service gateway whether the server is available to process the next service request comprises: determining that the server status indicates the server is busy;determining that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is busy and the next service request includes the second URL associated with the server as the primary server, selecting the server to process the next service request.
  • 8. The system of claim 4, wherein the determination by the service gateway whether the server is available to process the next service request comprises: determining that the server status indicates the server is busy;determining that the next service request includes a second URL associated with the server as the secondary server to process the second URL; andin response to determination that the server is busy and the next service request includes the second URL associated with the server as the secondary server, selecting a different server to process the next service request.
  • 9. The system of claim 4, wherein the determining by the service gateway whether the server is available to process the next service request comprises: determining that the server status indicates the server is very busy;determining that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is very busy and the next service request includes the second URL associated with the server as the primary server, configuring a timer for a duration; andin response to an expiration of the timer, selecting the server to process the next service request.
  • 10. The system of claim 1, wherein the service response is one of the following: a Hypertext Transport Protocol response, an Extensible Markup Language document, a Session Initiation Protocol packet, and a File Transfer Protocol response.
  • 11. A method for balancing servers based on a server load status, the method comprising: receiving, by a service gateway, from a server of a plurality of servers, a service response to a service request, the service request including a Uniform Resource Locator (URL) of a plurality of URLs, the service response comprising a computing load of the server, each server of the plurality of servers being configured to act as a primary server for the URL, wherein the service gateway: automatically selects, for each URL of the plurality of URLs, the primary server to process service requests associated with the each URL, andselects a secondary server to process the service requests associated with each URL for when the primary server is unavailable to process the service requests;receiving, by the service gateway, a next service request from a host, the next service request including the URL;based on the URL and the computing load of the server, determining, by the service gateway, whether the server can currently act as the primary server for processing the next service request; andbased on the determination, selectively sending, by the service gateway, the next service request to the server.
  • 12. The method of claim 11, further comprising: receiving, by the service gateway, the service request from the host, the service request including a first Uniform Resource Locator (URL) of a plurality of URLs, the first URL being associated with at least one server of the plurality of servers;determining that the server is the primary server configured to process the first URL and a further server is the secondary server configured to process the first URL;based on the determination that the server is the primary server, selecting the server to process the service request; andsending the service request to the server.
  • 13. The method of claim 12, wherein each of the plurality of URLs is individually resolved to the primary server configured to process the URL and the secondary server is configured to process the URL, each server of the plurality of servers being configured as one of primary servers and one of secondary servers for different URLs.
  • 14. The method of claim 13, wherein the computing load of the server includes a server status of the server.
  • 15. The method of claim 14, wherein the server status indicates one or more of the following: the server is not busy when the server is available for serving as the primary server and is available for serving as the secondary server, the server is busy when the server is available for serving as the primary server and is unavailable for serving as the secondary server, and the server is very busy when the server is available for serving as the primary server with restrictions and is unavailable for serving as the secondary server.
  • 16. The method of claim 14, wherein the determining whether the server is available to process the next service request comprises: determining, by the service gateway, that the server status indicates the server is not busy;determining, by the service gateway, that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is not busy and the next service request includes the second URL associated with the server as the primary server, selecting, by the service gateway, the server to process the next service request.
  • 17. The method of claim 14, wherein the determination whether the server is available to process the next service request comprises: determining, by the service gateway, that the server status indicates the server is busy;determining, by the service gateway, that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is busy and the next service request includes the second URL associated with the server as the primary server, selecting, by the service gateway, the server to process the next service request.
  • 18. The method of claim 14, wherein the determining whether the server is available to process the next service request comprises: determining, by the service gateway, that the server status indicates the server is busy;determining, by the service gateway, that the next service request includes a second URL associated with the server as the secondary server to process the second URL; andin response to determination that the server is busy and the next service request includes the second URL associated with the server as the secondary server, selecting, by the service gateway, a different server to process the next service request.
  • 19. The method of claim 14, wherein the determining whether the server is available to process the next service request comprises: determining, by the service gateway, that the server status indicates the server is very busy;determining, by the service gateway, that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determination that the server is very busy and the next service request includes the second URL associated with the server as the primary server, configuring, by the service gateway, a timer for a duration; andin response to an expiration of the timer, selecting, by the service gateway, the server to process the next service request.
  • 20. A system for balancing servers based on a server load status, the system comprising: a plurality of servers configured to process service requests; anda service gateway comprising a processor and a computer readable storage medium having a computer readable program code embodied therewith, wherein the computer readable program code, when executed by the processor causes the service gateway to: receive a service request from the host, the service request including a first Uniform Resource Locator (URL) of a plurality of URLs, the first URL being associated with at least one server of a plurality of servers;determine that a server of the plurality of servers is a primary server configured to process the first URL and a further server of the plurality of servers is a secondary server configured to process the first URL, wherein the service gateway is pre-configured to: automatically select, for each URL of the plurality of URLs, the primary server to process service requests associated with the each URL, andselect the secondary server to process the service requests associated with each URL for when the primary server is unavailable to process the service requests;based on the determination that the server is the primary server, select the server to process the service request;send the service request to the server;receive, from the server, a service response to the service request, the service request including URL, the service response comprising a computing load of the server, each server of the plurality of servers being configured to act as the primary server for the URL;receive a next service request from a host, the next service request including the URL;based on the URL and the computing load of the server, determine whether the server can currently act as the primary server for processing the next service request, wherein the determining whether the server is available to process the next service request comprises: determining that the server status indicates the server is not busy;determining that the next service request includes a second URL associated with the server as the primary server to process the second URL; andin response to determining the server s not busy and the next service request includes the second URL associated with the server as the primary server, selecting the server to process the next service request; and based on the selection, selectively send the next service request to the server.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/956,230, filed Dec. 1, 2015, entitled “System and Method to Balance Servers Based on Server Load Status”, which is a continuation and claims the benefit of U.S. patent application Ser. No. 12/894,142, filed Sep. 30, 2010 and entitled “System and Method to Balance Servers Based on Server Load Status”, now U.S. Pat. No. 9,215,275, issued Dec. 15, 2015. The disclosures of the above-referenced applications are incorporated herein by reference in their entirety for all purposes.

US Referenced Citations (374)
Number Name Date Kind
5218602 Grant et al. Jun 1993 A
5774660 Brendel et al. Jun 1998 A
5935207 Logue et al. Aug 1999 A
5958053 Denker Sep 1999 A
5995981 Wikstrom Nov 1999 A
6003069 Cavill Dec 1999 A
6047268 Bartoli et al. Apr 2000 A
6131163 Wiegel Oct 2000 A
6219706 Fan et al. Apr 2001 B1
6259705 Takahashi et al. Jul 2001 B1
6321338 Porras et al. Nov 2001 B1
6374300 Masters Apr 2002 B2
6459682 Ellesson et al. Oct 2002 B1
6587866 Modi et al. Jul 2003 B1
6748414 Boumas Jun 2004 B1
6772334 Glawitsch Aug 2004 B1
6779017 Lamberton et al. Aug 2004 B1
6779033 Watson et al. Aug 2004 B1
6801949 Bruck et al. Oct 2004 B1
6952728 Alles et al. Oct 2005 B1
7010605 Dharmarajan Mar 2006 B1
7013482 Krumel Mar 2006 B1
7058718 Fontes et al. Jun 2006 B2
7069438 Balabine et al. Jun 2006 B2
7076555 Orman et al. Jul 2006 B1
7143087 Fairweather Nov 2006 B2
7167927 Philbrick et al. Jan 2007 B2
7181524 Lele Feb 2007 B1
7218722 Turner et al. May 2007 B1
7228359 Monteiro Jun 2007 B1
7234161 Maufer et al. Jun 2007 B1
7236457 Joe Jun 2007 B2
7254133 Govindarajan et al. Aug 2007 B2
7269850 Govindarajan et al. Sep 2007 B2
7277963 Dolson et al. Oct 2007 B2
7301899 Goldstone Nov 2007 B2
7308499 Chavez Dec 2007 B2
7310686 Uysal Dec 2007 B2
7328267 Bashyam et al. Feb 2008 B1
7334232 Jacobs et al. Feb 2008 B2
7337241 Boucher et al. Feb 2008 B2
7343399 Hayball et al. Mar 2008 B2
7349970 Clement et al. Mar 2008 B2
7370353 Yang May 2008 B2
7391725 Huitema et al. Jun 2008 B2
7398317 Chen et al. Jul 2008 B2
7423977 Joshi Sep 2008 B1
7430755 Hughes et al. Sep 2008 B1
7463648 Eppstein et al. Dec 2008 B1
7467202 Savchuk Dec 2008 B2
7472190 Robinson Dec 2008 B2
7492766 Cabeca et al. Feb 2009 B2
7506360 Wilkinson et al. Mar 2009 B1
7509369 Tormasov Mar 2009 B1
7512980 Copeland et al. Mar 2009 B2
7533409 Keane et al. May 2009 B2
7552323 Shay Jun 2009 B2
7584262 Wang et al. Sep 2009 B1
7584301 Joshi Sep 2009 B1
7590736 Hydrie et al. Sep 2009 B2
7613193 Swami et al. Nov 2009 B2
7613822 Joy et al. Nov 2009 B2
7673072 Boucher et al. Mar 2010 B2
7675854 Chen et al. Mar 2010 B2
7703102 Eppstein et al. Apr 2010 B1
7707295 Szeto et al. Apr 2010 B1
7711790 Barrett et al. May 2010 B1
7739395 Parlamas et al. Jun 2010 B1
7747748 Allen Jun 2010 B2
7751409 Carolan Jul 2010 B1
7765328 Bryers et al. Jul 2010 B2
7792113 Foschiano et al. Sep 2010 B1
7808994 Vinokour et al. Oct 2010 B1
7826487 Mukerji et al. Nov 2010 B1
7881215 Daigle et al. Feb 2011 B1
7948952 Hurtta et al. May 2011 B2
7965727 Sakata et al. Jun 2011 B2
7970934 Patel Jun 2011 B1
7979585 Chen et al. Jul 2011 B2
7983258 Ruben et al. Jul 2011 B1
7990847 Leroy et al. Aug 2011 B1
7991859 Miller et al. Aug 2011 B1
8032634 Eppstein et al. Oct 2011 B1
8090866 Bashyam et al. Jan 2012 B1
8099492 Dahlin et al. Jan 2012 B2
8122116 Matsunaga et al. Feb 2012 B2
8179809 Eppstein et al. May 2012 B1
8185651 Moran et al. May 2012 B2
8191106 Choyi et al. May 2012 B2
8224971 Miller et al. Jul 2012 B1
8266235 Jalan et al. Sep 2012 B2
8296434 Miller et al. Oct 2012 B1
8312507 Chen et al. Nov 2012 B2
8379515 Mukerji Feb 2013 B1
8499093 Grosser et al. Jul 2013 B2
8539075 Bali et al. Sep 2013 B2
8554929 Szeto et al. Oct 2013 B1
8560693 Wang et al. Oct 2013 B1
8584199 Chen et al. Nov 2013 B1
8595791 Chen et al. Nov 2013 B1
RE44701 Chen et al. Jan 2014 E
8675488 Sidebottom et al. Mar 2014 B1
8681610 Mukerji Mar 2014 B1
8750164 Casado et al. Jun 2014 B2
8782221 Han Jul 2014 B2
8813180 Chen et al. Aug 2014 B1
8826372 Chen et al. Sep 2014 B1
8879427 Krumel Nov 2014 B2
8885463 Medved et al. Nov 2014 B1
8897154 Jalan et al. Nov 2014 B2
8965957 Barros Feb 2015 B2
8977749 Han Mar 2015 B1
8990262 Chen et al. Mar 2015 B2
9094364 Jalan et al. Jul 2015 B2
9106561 Jalan et al. Aug 2015 B2
9154577 Jalan et al. Oct 2015 B2
9154584 Han Oct 2015 B1
9215275 Kannan et al. Dec 2015 B2
9219751 Chen et al. Dec 2015 B1
9253152 Chen et al. Feb 2016 B1
9270705 Chen et al. Feb 2016 B1
9270774 Jalan et al. Feb 2016 B2
9338225 Jalan et al. May 2016 B2
9350744 Chen et al. May 2016 B2
9356910 Chen et al. May 2016 B2
9386088 Zheng et al. Jul 2016 B2
9497201 Chen et al. Nov 2016 B2
9531846 Han et al. Dec 2016 B2
9544364 Jalan et al. Jan 2017 B2
9602442 Han Mar 2017 B2
9609052 Jalan et al. Mar 2017 B2
9705800 Sankar et al. Jul 2017 B2
9843484 Sankar et al. Dec 2017 B2
9900252 Chiong Feb 2018 B2
9906422 Jalan et al. Feb 2018 B2
9906591 Jalan et al. Feb 2018 B2
9942152 Jalan et al. Apr 2018 B2
9942162 Golshan et al. Apr 2018 B2
9960967 Chen et al. May 2018 B2
9961135 Kannan et al. May 2018 B2
9979801 Jalan et al. May 2018 B2
20010049741 Skene et al. Dec 2001 A1
20020032777 Kawata et al. Mar 2002 A1
20020078164 Reinschmidt Jun 2002 A1
20020091844 Craft et al. Jul 2002 A1
20020103916 Chen et al. Aug 2002 A1
20020133491 Sim et al. Sep 2002 A1
20020138618 Szabo Sep 2002 A1
20020143991 Chow et al. Oct 2002 A1
20020178259 Doyle et al. Nov 2002 A1
20020191575 Kalavade et al. Dec 2002 A1
20020194335 Maynard Dec 2002 A1
20020194350 Lu et al. Dec 2002 A1
20030009591 Hayball et al. Jan 2003 A1
20030014544 Pettey Jan 2003 A1
20030023711 Parmar et al. Jan 2003 A1
20030023873 Ben-Itzhak Jan 2003 A1
20030035409 Wang et al. Feb 2003 A1
20030035420 Niu Feb 2003 A1
20030065762 Stolorz et al. Apr 2003 A1
20030091028 Chang et al. May 2003 A1
20030131245 Linderman Jul 2003 A1
20030135625 Fontes et al. Jul 2003 A1
20030195962 Kikuchi et al. Oct 2003 A1
20040062246 Boucher et al. Apr 2004 A1
20040073703 Boucher et al. Apr 2004 A1
20040078419 Ferrari et al. Apr 2004 A1
20040078480 Boucher et al. Apr 2004 A1
20040111516 Cain Jun 2004 A1
20040128312 Shalabi et al. Jul 2004 A1
20040139057 Hirata et al. Jul 2004 A1
20040139108 Tang et al. Jul 2004 A1
20040141005 Banatwala et al. Jul 2004 A1
20040143599 Shalabi et al. Jul 2004 A1
20040187032 Gels et al. Sep 2004 A1
20040199616 Karhu Oct 2004 A1
20040199646 Susai et al. Oct 2004 A1
20040202182 Lund et al. Oct 2004 A1
20040210623 Hydrie et al. Oct 2004 A1
20040210663 Phillips et al. Oct 2004 A1
20040213158 Collett et al. Oct 2004 A1
20040268358 Darling et al. Dec 2004 A1
20050005207 Herneque Jan 2005 A1
20050009520 Herrero et al. Jan 2005 A1
20050021848 Jorgenson Jan 2005 A1
20050027862 Nguyen et al. Feb 2005 A1
20050036501 Chung et al. Feb 2005 A1
20050036511 Baratakke et al. Feb 2005 A1
20050044270 Grove et al. Feb 2005 A1
20050074013 Hershey et al. Apr 2005 A1
20050080890 Yang et al. Apr 2005 A1
20050102400 Nakahara et al. May 2005 A1
20050125276 Rusu Jun 2005 A1
20050163073 Heller et al. Jul 2005 A1
20050198335 Brown et al. Sep 2005 A1
20050213586 Cyganski et al. Sep 2005 A1
20050240989 Kim et al. Oct 2005 A1
20050249225 Singhal Nov 2005 A1
20050259586 Hafid et al. Nov 2005 A1
20050289231 Harada et al. Dec 2005 A1
20060023721 Miyake et al. Feb 2006 A1
20060036610 Wang Feb 2006 A1
20060036733 Fujimoto et al. Feb 2006 A1
20060064478 Sirkin Mar 2006 A1
20060069774 Chen et al. Mar 2006 A1
20060069804 Miyake et al. Mar 2006 A1
20060077926 Rune Apr 2006 A1
20060092950 Arregoces et al. May 2006 A1
20060098645 Walkin May 2006 A1
20060112170 Sirkin May 2006 A1
20060168319 Trossen Jul 2006 A1
20060187901 Cortes et al. Aug 2006 A1
20060190997 Mahajani et al. Aug 2006 A1
20060209789 Gupta et al. Sep 2006 A1
20060230129 Swami et al. Oct 2006 A1
20060233100 Luft et al. Oct 2006 A1
20060251057 Kwon et al. Nov 2006 A1
20060277303 Hegde et al. Dec 2006 A1
20060280121 Matoba Dec 2006 A1
20070019543 Wei et al. Jan 2007 A1
20070086382 Narayanan et al. Apr 2007 A1
20070094396 Takano et al. Apr 2007 A1
20070118881 Mitchell et al. May 2007 A1
20070127381 Oh et al. Jun 2007 A1
20070156919 Potti et al. Jul 2007 A1
20070165622 O'Rourke et al. Jul 2007 A1
20070185998 Touitou et al. Aug 2007 A1
20070203890 Sareen et al. Aug 2007 A1
20070230337 Igarashi et al. Oct 2007 A1
20070245090 King et al. Oct 2007 A1
20070259673 Willars et al. Nov 2007 A1
20070283429 Chen et al. Dec 2007 A1
20070286077 Wu Dec 2007 A1
20070288247 Mackay Dec 2007 A1
20070294209 Strub et al. Dec 2007 A1
20080031263 Ervin et al. Feb 2008 A1
20080101396 Miyata May 2008 A1
20080109452 Patterson May 2008 A1
20080109870 Sherlock et al. May 2008 A1
20080134332 Keohane et al. Jun 2008 A1
20080162679 Maher et al. Jul 2008 A1
20080228781 Chen et al. Sep 2008 A1
20080250099 Shen et al. Oct 2008 A1
20080263209 Pisharody et al. Oct 2008 A1
20080271130 Ramamoorthy Oct 2008 A1
20080282254 Blander et al. Nov 2008 A1
20080291911 Lee et al. Nov 2008 A1
20090049198 Blinn et al. Feb 2009 A1
20090070470 Bauman et al. Mar 2009 A1
20090077651 Poeluev Mar 2009 A1
20090092124 Singhal et al. Apr 2009 A1
20090106830 Maher Apr 2009 A1
20090138606 Moran et al. May 2009 A1
20090138945 Savchuk May 2009 A1
20090141634 Rothstein et al. Jun 2009 A1
20090164614 Christian et al. Jun 2009 A1
20090172093 Matsubara Jul 2009 A1
20090213858 Dolganow et al. Aug 2009 A1
20090222583 Josefsberg et al. Sep 2009 A1
20090227228 Hu et al. Sep 2009 A1
20090228547 Miyaoka et al. Sep 2009 A1
20090262741 Jungck et al. Oct 2009 A1
20090271472 Scheifler et al. Oct 2009 A1
20090313379 Rydnell et al. Dec 2009 A1
20100008229 Bi et al. Jan 2010 A1
20100023621 Ezolt et al. Jan 2010 A1
20100036952 Hazlewood et al. Feb 2010 A1
20100054139 Chun et al. Mar 2010 A1
20100061319 Aso et al. Mar 2010 A1
20100064008 Yan et al. Mar 2010 A1
20100082787 Kommula et al. Apr 2010 A1
20100083076 Ushiyama Apr 2010 A1
20100094985 Abu-Samaha et al. Apr 2010 A1
20100098417 Tse-Au Apr 2010 A1
20100106833 Banerjee et al. Apr 2010 A1
20100106854 Kim et al. Apr 2010 A1
20100128606 Patel et al. May 2010 A1
20100162378 Jayawardena et al. Jun 2010 A1
20100205310 Altshuler et al. Aug 2010 A1
20100210265 Borzsei et al. Aug 2010 A1
20100217793 Preiss Aug 2010 A1
20100223630 Degenkolb et al. Sep 2010 A1
20100228819 Wei Sep 2010 A1
20100228878 Xu et al. Sep 2010 A1
20100235507 Szeto et al. Sep 2010 A1
20100235522 Chen et al. Sep 2010 A1
20100238828 Russell Sep 2010 A1
20100265824 Chao et al. Oct 2010 A1
20100268814 Cross et al. Oct 2010 A1
20100293296 Hsu et al. Nov 2010 A1
20100312740 Clemm et al. Dec 2010 A1
20100318631 Shukla Dec 2010 A1
20100322252 Suganthi et al. Dec 2010 A1
20100330971 Selitser et al. Dec 2010 A1
20100333101 Pope et al. Dec 2010 A1
20110007652 Bai Jan 2011 A1
20110019550 Bryers et al. Jan 2011 A1
20110023071 Li et al. Jan 2011 A1
20110029599 Pulleyn et al. Feb 2011 A1
20110032941 Quach et al. Feb 2011 A1
20110040826 Chadzelek et al. Feb 2011 A1
20110047294 Singh et al. Feb 2011 A1
20110060831 Ishii et al. Mar 2011 A1
20110060840 Susai et al. Mar 2011 A1
20110099403 Miyata et al. Apr 2011 A1
20110110294 Valluri et al. May 2011 A1
20110145324 Reinart et al. Jun 2011 A1
20110153834 Bharrat Jun 2011 A1
20110178985 San Martin Arribas et al. Jul 2011 A1
20110185073 Jagadeeswaran et al. Jul 2011 A1
20110191773 Pavel et al. Aug 2011 A1
20110196971 Reguraman et al. Aug 2011 A1
20110276695 Maldaner Nov 2011 A1
20110276982 Nakayama et al. Nov 2011 A1
20110289496 Steer Nov 2011 A1
20110292939 Subramaian et al. Dec 2011 A1
20110302256 Sureshehandra et al. Dec 2011 A1
20110307541 Walsh et al. Dec 2011 A1
20120008495 Shen et al. Jan 2012 A1
20120023231 Ueno Jan 2012 A1
20120026897 Guichard et al. Feb 2012 A1
20120030341 Jensen et al. Feb 2012 A1
20120066371 Patel et al. Mar 2012 A1
20120084460 McGinnity et al. Apr 2012 A1
20120106355 Ludwig May 2012 A1
20120117571 Davis et al. May 2012 A1
20120144014 Natham et al. Jun 2012 A1
20120151353 Joanny Jun 2012 A1
20120170548 Rajagopalan et al. Jul 2012 A1
20120173759 Agarwal et al. Jul 2012 A1
20120191839 Maynard Jul 2012 A1
20120239792 Banerjee et al. Sep 2012 A1
20120240185 Kapoor et al. Sep 2012 A1
20120290727 Tivig Nov 2012 A1
20120297046 Raja et al. Nov 2012 A1
20130046876 Narayana et al. Feb 2013 A1
20130058335 Koponen et al. Mar 2013 A1
20130074177 Varadhan et al. Mar 2013 A1
20130083725 Mallya et al. Apr 2013 A1
20130124713 Feinberg et al. May 2013 A1
20130148500 Sonoda et al. Jun 2013 A1
20130173795 McPherson Jul 2013 A1
20130176854 Chisu et al. Jul 2013 A1
20130191486 Someya et al. Jul 2013 A1
20130198385 Han et al. Aug 2013 A1
20130250765 Ehsan et al. Sep 2013 A1
20130250770 Zou et al. Sep 2013 A1
20130258846 Damola Oct 2013 A1
20130268646 Doron et al. Oct 2013 A1
20130282791 Kruglick Oct 2013 A1
20130336159 Previdi et al. Dec 2013 A1
20140226658 Kakadia et al. Aug 2014 A1
20140235249 Jeong et al. Aug 2014 A1
20140248914 Aoyagi et al. Sep 2014 A1
20140258465 Li Sep 2014 A1
20140269728 Jalan et al. Sep 2014 A1
20140286313 Fu et al. Sep 2014 A1
20140298091 Carlen et al. Oct 2014 A1
20140325649 Zhang Oct 2014 A1
20140330982 Jalan et al. Nov 2014 A1
20140334485 Jain et al. Nov 2014 A1
20140359052 Joachimpillai et al. Dec 2014 A1
20150098333 Lin et al. Apr 2015 A1
20150156223 Xu et al. Jun 2015 A1
20150215436 Kancherla Jul 2015 A1
20150237173 Virkki et al. Aug 2015 A1
20150312268 Ray Oct 2015 A1
20150350048 Sampat et al. Dec 2015 A1
20150350379 Jalan et al. Dec 2015 A1
20160042014 Jalan et al. Feb 2016 A1
20160044095 Sankar et al. Feb 2016 A1
20160094470 Skog Mar 2016 A1
20160139910 Ramanathan et al. May 2016 A1
20160164792 Oran Jun 2016 A1
Foreign Referenced Citations (113)
Number Date Country
1372662 Oct 2002 CN
1449618 Oct 2003 CN
1473300 Feb 2004 CN
1529460 Sep 2004 CN
1575582 Feb 2005 CN
1714545 Dec 2005 CN
1725702 Jan 2006 CN
1910869 Feb 2007 CN
101004740 Jul 2007 CN
101094225 Dec 2007 CN
101163336 Apr 2008 CN
101169785 Apr 2008 CN
101189598 May 2008 CN
101193089 Jun 2008 CN
101247349 Aug 2008 CN
101261644 Sep 2008 CN
101495993 Jul 2009 CN
101878663 Nov 2010 CN
102143075 Aug 2011 CN
102546590 Jul 2012 CN
102571742 Jul 2012 CN
102577252 Jul 2012 CN
102918801 Feb 2013 CN
103533018 Jan 2014 CN
103944954 Jul 2014 CN
104040990 Sep 2014 CN
104067569 Sep 2014 CN
104106241 Oct 2014 CN
104137491 Nov 2014 CN
104796396 Jul 2015 CN
102577252 Mar 2016 CN
102918801 May 2016 CN
102571742 Jul 2016 CN
104067569 Feb 2017 CN
1209876 May 2002 EP
1770915 Apr 2007 EP
1885096 Feb 2008 EP
2296313 Mar 2011 EP
2577910 Apr 2013 EP
2622795 Aug 2013 EP
2647174 Oct 2013 EP
2760170 Jul 2014 EP
27772026 Sep 2014 EP
2901308 Aug 2015 EP
2772026 Feb 2017 EP
1182560 Nov 2013 HK
1183569 Dec 2013 HK
1183996 Jan 2014 HK
1189438 Jan 2014 HK
1198565 May 2015 HK
1198848 Jun 2015 HK
1199153 Jun 2015 HK
1199779 Jul 2015 HK
1200617 Aug 2015 HK
3764CHN2014 Sep 2015 IN
261CHE2014 Jan 2016 IN
1668CHENP2015 Jul 2016 IN
H0997233 Apr 1997 JP
H1196128 Apr 1999 JP
H11338836 Dec 1999 JP
2000276432 Oct 2000 JP
2000307634 Nov 2000 JP
2001051859 Feb 2001 JP
2001298449 Oct 2001 JP
2002091936 Mar 2002 JP
2003141068 May 2003 JP
2003186776 Jul 2003 JP
2005141441 Jun 2005 JP
2006332825 Dec 2006 JP
2008040718 Feb 2008 JP
2009500731 Jan 2009 JP
2013528330 Jul 2013 JP
2014504484 Feb 2014 JP
2014143686 Aug 2014 JP
2015507380 Mar 2015 JP
5855663 Feb 2016 JP
5906263 Apr 2016 JP
5913609 Apr 2016 JP
5946189 Jul 2016 JP
5963766 Aug 2016 JP
20080008340 Jan 2008 KR
100830413 May 2008 KR
20130096624 Aug 2013 KR
101576585 Dec 2015 KR
101632187 Jun 2016 KR
101692751 Jan 2017 KR
WO2001013228 Feb 2001 WO
WO2001014990 Mar 2001 WO
WO2003103237 Dec 2003 WO
WO2004084085 Sep 2004 WO
WO2006098033 Sep 2006 WO
WO2008053954 May 2008 WO
WO2008078593 Jul 2008 WO
WO2011049770 Apr 2011 WO
WO2011079381 Jul 2011 WO
WO2011149796 Dec 2011 WO
WO2012050747 Apr 2012 WO
WO2012075237 Jun 2012 WO
WO2012083264 Jun 2012 WO
WO2012097015 Jul 2012 WO
WO2013070391 May 2013 WO
WO2013081952 Jun 2013 WO
WO2013096019 Jun 2013 WO
WO2013112492 Aug 2013 WO
WO2013189024 Dec 2013 WO
WO2014031046 Feb 2014 WO
WO2014052099 Apr 2014 WO
WO2014088741 Jun 2014 WO
WO2014093829 Jun 2014 WO
WO2014138483 Sep 2014 WO
WO2014144837 Sep 2014 WO
WO2014179753 Nov 2014 WO
WO2015153020 Oct 2015 WO
Non-Patent Literature Citations (11)
Entry
Abe, et al., “Adaptive Split Connection Schemes in Advanced Relay Nodes,” IEICE Technical Report, 2010, vol. 109 (438), pp. 25-30.
Cardellini, et al., “Dynamic Load Balancing on Web-Server Systems,” IEEE Internet Computing, 1999, vol. 3 (3), pp. 28-39.
FreeBSD, “tcp—TCP Protocal,” Linux Programme□ s Manual [online], 2007, [retrieved on Apr. 13, 2016], Retreived from the Internet: <https://www.freebsd.org/cgi/man.cgi?query=tcp&apropos=0&sektion=7&manpath=SuSe+Linux%2Fi386+11.0&format=asci>.
Gite, “Linux Tune Network Stack (Buffers Size) to Increase Networking Performance,” nixCraft [online], 2009, [retreived on Apr. 13, 2016], Retreived from the Internet: <URL:http://www.cyberciti.biz/faq/linux-tcp-tuning/>.
Goldszmidt, et al., “NetDispatcher: A TCP Connection Router,” IBM Researc Report, RC 20853, 1997, pp. 1-31.
Kjaer, et al., “Resource Allocation and Disturbance Rejection in Web Servers Using SLAs and Virtualized Servers,” IEEE Transactions on Network Service Management, 2009, vol. 6 (4), pp. 226-239.
Koike, et al., “Transport Middleware for Network-Based Control,” IEICE Technical Report, 2000, vol. 100 (53), pp. 13-18.
Sharifian, et al., “An Approximation-Based Load-Balancing Algorithm with Admission Control for Cluster Web Servers with Dynamic Workloads,” The Journal of Supercomputing, 2010, vol. 53 (3), pp. 440-463.
Spatscheck, et al., “Optimizing TCP Forwarder Performance,” IEEE/ACM Transactions on Networking, 2000, vol. 8(2), pp. 146-157.
Search Report and Written Opinion dated Apr. 10, 2012 for PCT Application No. PCT/US2011/052225.
Yamamoto, et al., “Performance Evaluation of Window Size in Proxy-Based TCP for Multi-Hop Wireless Networks,” IPSJ SIG Technical Reports, 2008, vol. 2008 (44), pp. 109-114.
Related Publications (1)
Number Date Country
20180213031 A1 Jul 2018 US
Continuations (2)
Number Date Country
Parent 14956230 Dec 2015 US
Child 15936709 US
Parent 12894142 Sep 2010 US
Child 14956230 US