This invention relates generally to data communications, and more specifically, to a virtual cluster system.
Web services and cloud computing are deployed in an unprecedented pace. New servers are unloaded and installed at datacenters every day. Demands of web services and corporate computing come from many market segments. Consumer oriented services include mobile apps such as iPhone™ apps, iPad™ apps, Android™ apps, mobile applications such as location based services, turn-by-turn navigation services, e-book services such as Kindle™, video applications such as YouTube™ or Hulu™, music applications such as Pandora™ or iTunes™, Internet television services such as Netflix™, and many other fast growing consumer Web services. On the corporate front, cloud computing based services such as Google™ docs, Microsoft™ Office Live and Sharepoint™ software, Salesforce.com™'s on-line software services, tele-presence and Web conferencing services, and many other corporate cloud computing services.
More and more servers are deployed to accommodate the increasing computing needs. These servers are typically managed by service controllers such as server load balancers (SLB) or application delivery controllers (ADC). These controllers manage the load balancing and delivery of service sessions from client host computers to servers. As more servers are deployed, more controllers are deployed accordingly. These network controllers can be pooled together for several significant purposes—to provide scalable services and solutions when dealing with large number of users; to offer high availability of service when dealing with possible unscheduled or scheduled unavailability of a server; to help ease the complexity of administration and management of a larger number of controllers. The controllers in a cluster pool work in tandem with each other in order to provide a consistent service to the users. If there is a configuration change to a controller in a cluster, the change can have a significant adverse effect on operation of other controllers in the cluster if the change conflicts with existing configurations of these controllers. Often, the totality of existing configuration of the cluster is recorded in a configuration file. All controllers in the cluster must adhere to the configuration according to the configuration file for proper operation of the controllers in the cluster. When there is a change of configuration, the configuration file must be updated to include the change. A new configuration file combining the new configuration change and the existing configuration file becomes the most recent configuration file. The new configuration file is to be informed to all controllers in the cluster. If one controller in the cluster becomes unavailable and available again, the newly available controller needs to apply the configuration in the new configuration file in order to work in tandem with other controllers in the cluster.
Therefore, there is a need for a method for a virtual chassis system to synchronize a configuration file in the cluster of controllers.
This summary is provided to introduce a selection of concepts in a simplified form that are further described in the Detailed Description below. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
According to some embodiments, the present technology is directed to a method for synchronizing configuration files by a master blade of a virtual application distribution chassis comprising a plurality of blades, comprises: processing a configuration command received from an administrative host by the master blade; determining that the received configuration command applies changes to at least one network application implemented by one or more blades of the virtual application distribution chassis; updating a first configuration file with the received configuration command and generating an updated tag associated with the updated first configuration file, the updated first configuration file comprising the received configuration command; sending a first configuration message to one or more slave blades of the plurality of blades informing of the updated configuration file; in response to the configuration message informing the one or more slave blades of the updated configuration file, receiving a request for the updated configuration file from at least one slave blade; and sending a second configuration message comprising the updated first configuration file and updated tag.
In one aspect of the present disclosure, the processing a configuration command received from an administrative host by the master blade comprises identifying at least one slave blade from the plurality of blades in the virtual application distribution chassis for the configuration command.
In various aspects of the present disclosure, the updated tag is a time stamp, integer, or revision number.
In various embodiments, the master blade of the virtual application distribution chassis is a server load balancer or application delivery controller.
System and computer readable storage medium corresponding to the above-summarized methods are also described and claimed herein.
Embodiments are illustrated by way of example and not by limitation in the figures of the accompanying drawings, in which like references indicate similar elements.
The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments. Thus, the present invention is not intended to be limited to the embodiment shown but is to be accorded the widest scope consistent with the principles and features described herein.
The present invention can take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment containing both hardware and software elements. In a preferred embodiment, the present invention is implemented in software, which includes but is not limited to firmware, resident software, microcode, etc.
Furthermore, the present invention can take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
The medium can be an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium. Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette, a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk and an optical disk. Current examples of optical disks include compact disk-read only memory (CD-ROM), compact disk-read/write (CD-R/W) and DVD.
A data processing system suitable for storing and/or executing program code will include at least one processor coupled directly or indirectly to memory elements through a system bus. The memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution.
Input/output or I/O devices (including but not limited to keyboards, displays, point devices, etc.) can be coupled to the system either directly or through intervening I/O controllers.
Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified local function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
Administrative host 100 is a computing device with network access capabilities. As shown in
In
Blade 310 includes storage 306 (
Assume that blade 310 is the master blade of the virtual application distribution chassis 300. Master blade 310 additionally includes management interface 307, as illustrated in
In
In another embodiment, configuration command 113 is applied to configure a network application 350 of virtual application distribution chassis 300. For example, the network application 350 is a server load balancing application, and configuration command 113 sets up a server load balancer policy for a server. In one example, the network application 350 is a network traffic management application, and configuration command 113 configures a quality of service parameters for the traffic management application. In another example, the network application 350 is a VPN firewall application, and configuration command 113 configures a company user account for the VPN access. In another example, the network application 350 is a high availability service for another network application, and configuration command 113 configures the high availability service.
In one embodiment, the network application configuration command 113 applies to all the blades, when the network application 350 is implemented by all the blades. In one embodiment, the network application configuration command 113 applies to one or more of the blades implementing the network application 350.
Data network 153 connects the host 100 and the virtual application distribution chassis 300, also referred to as virtual cluster system 300. In various embodiments, data network 153 is an Internet Protocol (IP) network, a corporate data network, a regional corporate data network, an Internet service provider network, or a residential data network. Data network 153 includes a wired network such as Ethernet, and/or a wireless network such as a WiFi network, or cellular network.
Data network 151 connects virtual cluster system 300 to data network 153. In one embodiment, data network 151 resides in a data center, spans across multiple data centers over a wide area network such as optical network, or spreads over multiple buildings in a campus area network.
In one embodiment, switches 412, 414, and 416 reside in a same physical location, such as a data center. In one embodiment, switches 412, 414, and 416 reside in different locations, such as in two or more data centers. Assume that in this embodiment, blades 310, 312, 314 and 316 reside in different data centers.
Blade 310 communicates with the other blades by sending and receiving a virtual cluster system (VCS) message 500. In one embodiment blade 310 sends virtual cluster system message 500 using a broadcast network address such as an Ethernet broadcast address. In one embodiment, virtual cluster system message 500 uses a multicast address, such as an Ethernet multicast address, an IP multicast address, or other multicast network address. In one embodiment, blade 310 establishes a communication session (not shown) with blade 312 and sends virtual cluster system message 500 to blade 312 using the communication session. Blade 310 establishes a separate communication session with each of the other blades to send virtual cluster system message 500 to the other blades. Examples of the communication session include an UDP session, a TCP session, an IP-based communication session, a peer-to-peer communication session, a unicast communication session or other form of data communication session. In one embodiment, blade 310 receives virtual cluster system message 500 using a broadcast, a multicast address or a communication session with blade 312 or one of the other blades.
In one embodiment, blade 310 sends and receives a configuration message 502 which is a type of virtual cluster system message 500. Configuration message 502 communicates a configuration command 113 to the receiving blades. The configuration command 113 in the configuration message 502 may be the actual configuration command sent by the host 100, or may instead be a command which implements the host's configuration command at the receiving slave blade.
Master blade 310 of virtual cluster system 300 receives configuration command 113 from administrative host 100 (800). Master blade 310 determines that configuration command 113 is to be applied by one or more slave blades. For purposes of illustration, assume that master blade 310 determines that configuration command 113 is to be applied by slave blade 314 (801). In one embodiment, administrative host 100 specifies an identity of blade 314 in configuration command 113. In one scenario, configuration command 113 is a configuration of a network interface of blade 314. Administrative host 100 includes blade 314 identity in configuration command 113. Master blade 310 determines that the configuration command 113 is to be applied by blade 314 from the blade 314 identity in configuration command 113. In one embodiment, configuration command 113 includes an identity of the network interface. Master blade 310 determines that the configuration command 113 is to be applied by blade 314 using the network interface identity in configuration command 113. In one embodiment, master blade 310 stores in storage 306 a plurality of network interface identities and identities of the blades in virtual application distribution chassis 300, where each network interface identity is associated with a blade identity. Master blade 310 matches the network interface identity with the plurality of network interface identities in storage 306. Master blade 310 obtains the blade identity of the matching network interface in the storage module 306.
In one embodiment, configuration command 113 applies to network application 350 which is to be applied by all of the blades implementing network application 350. Master blade 310 determines based on configuration command 113 that all of these blades are to apply the configuration command 113. In one embodiment, master blade 310 stores in storage 306 a plurality of configuration commands which are applicable to all of these blades. Master blade 310 matches configuration command 113 with one or more of the plurality of configuration commands in storage 306 and determines that the configuration command 113 is to be applied by all of these blades. For this embodiment, the processing of the configuration command 113 by the master blade 310 with blade 314 is repeated for each of the other blades of the virtual application distribution chassis 300.
In one embodiment, configuration command 113 includes an identity of blade 314 and a configuration for network application 350. Master blade 310 determines that the configuration command 113 is to be applied by blade 314 based on the blade 314 identity from configuration command 113.
In response to determining that the configuration command 113 is to be applied by blade 314, master blade 310 sends a configuration message 502 to slave blade 314 (804). In various embodiments, master blade 310 includes configuration command 113 in configuration message 502. The configuration message 502 is delivered to slave blade 314 is as described above with reference to
In one embodiment, tag 341 is a number. Tag 341 is a number larger than tag 340 to indicate configuration file 331 is more recent than configuration file 330. For example, tag 341 is an integer, and tag 341 is one larger than tag 340. In one embodiment, master blade 310 obtains tag 340 from configuration file 330 and calculates tag 341 by adding one to tag 340.
In one embodiment, tag 341 is a revision number. Master blade 310 obtains tag 340 from configuration file 330 and calculates a newer revision number based on tag 340. Master blade 310 puts the newer revision number as tag 341.
After master blade 310 creates configuration file 331, master blade 310 informs the slave blades of configuration file 331 in order to synchronize configuration file 331 with the configuration files stored locally by the slave blades. Master blade 310 sends a configuration message 505 to slave blades 312, 314 and 316 to inform them of configuration file 331. In one embodiment, master blade 310 includes tag 341 in configuration message 505. Slave blades 312, 314 and 316 receives configuration message 505 and retrieves tag 341.
In one embodiment, master blade 310 includes configuration file 331 in configuration message 505.
In response to determining that tag 341 is more recent than tag 340, slave blade 312 sends a request configuration message 507 to master blade 310 to request a configuration file. In one embodiment, slave blade 312 places tag 341 in request configuration message 507. Master blade 310 receives the request configuration message 507 and retrieves tag 341. Master blade 310 finds a match of tag 341 from configuration message 507 against tag 341 from configuration file 331 in storage 306. Master blade 310 sends configuration file 331 to blade 312. In one embodiment, master blade 310 sends configuration file 331 in a configuration message 509 to slave blade 312. In one embodiment, master blade 310 sends configuration file 331 using a separate file transfer communication session with slave blade 312.
In one embodiment, request configuration message 507 does not include tag 341. After receiving request configuration message 507, master blade 310 automatically sends the most recent configuration file, say configuration file 331, to slave blade 312.
After receiving configuration file 331 from configuration message 509 or a file transfer communication session, slave blade 312 stores configuration file 331 and tag 341 into storage 413 of blade 312. In various embodiments, slave blade 312 replaces configuration file 330 by configuration file 331, or stores configuration file 330 in a different location or using a different file name in storage 413. Configuration file 331 becomes the most recent configuration file for slave blade 312.
In one embodiment, master blade 310 has a configuration file more recent than configuration file 331. For example, between the sending of the configuration message 505 by the master blade 310 and the sending of the reply configuration message 507 by the slave blade 312, another configuration command may have been received from the host 100, resulting in a further configuration change. In this case, master blade 310 sends the more recent configuration file to slave blade 312, despite the fact that slave blade 312 asked for configuration file 331 corresponding to tag 341.
In one embodiment, master blade 310 sends configuration message 502 about tag 341 more than once. In this way, slave blade 312 will receive at least one copy of the configuration message 502 in case slave blade 312 fails to receive the other copies of configuration message 502. In one embodiment, master blade 310 sends configuration message 502 periodically, such as once every 30 seconds, 1 second, 5 minutes, or 100 milliseconds.
Slave blade 316 receives configuration message 505 from master blade 310, and retrieves tag 341 from configuration message 505. Slave blade 316 compares tag 341 with tag 340 of configuration file 330 in storage 417. Slave blade 316 determines tag 341 is more recent than tag 340. Slave blade 316 obtains configuration file 331 from master blade 310, as described above with reference to
Once slave blade 316 obtains configuration file 331, slave blade 316 applies configuration commands in configuration file 331. In one embodiment, slave blade 316 resets itself to clear the configuration based on configuration file 330. After slave blade 316 is reset, slave blade 316 finds configuration file 331 in storage module 416 and applies configuration commands in configuration file 331. In one embodiment, slave blade 316 applies configuration commands in configuration file 331 without clearing the configuration based on the configuration commands in configuration file 330. In one embodiment, slave blade 316 undoes configuration commands based on configuration file 330, and applies configuration commands in configuration file 331.
After slave blade 316 applies configuration commands in configuration file 331, slave blade 316 rejoins virtual application distribution chassis 300. When slave blade 316 receives configuration message 502 and obtains tag 341 from configuration message 505, slave blade 316 compares tag 341 of configuration message 505 with tag 341 from configuration file 331 in storage module 416. Slave blade 316 determines the two tags are the same and configuration file 331 is the most recent configuration file indicated by master blade 310.
Although the present invention has been described in accordance with the embodiments shown, one of ordinary skill in the art will readily recognize that there could be variations to the embodiments and those variations would be within the spirit and scope of the present invention. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.
The present application is a continuation of, and claims the priority benefit of, U.S. Nonprovisional patent application Ser. No. 14/872,937, filed Oct. 1, 2015, titled “Synchronization of Configuration File of Virtual Application Distribution Chassis”, which in turn is a continuation of U.S. Nonprovisional patent application Ser. No. 13/154,399, filed Jun. 6, 2011, titled “Synchronization of Configuration File of Virtual Application Distribution Chassis”, now U.S. Pat. No. 9,154,577, issued on Oct. 6, 2015. The disclosure of the above application is hereby incorporated by reference in its entirety, including all references cited therein.
Number | Name | Date | Kind |
---|---|---|---|
4475192 | Fernow et al. | Oct 1984 | A |
5042032 | Dighe et al. | Aug 1991 | A |
5732041 | Joffe | Mar 1998 | A |
5822512 | Goodrum et al. | Oct 1998 | A |
5859835 | Varma et al. | Jan 1999 | A |
5909686 | Muller et al. | Jun 1999 | A |
5960174 | Dew | Sep 1999 | A |
6047000 | Tsang et al. | Apr 2000 | A |
6058116 | Hiscock et al. | May 2000 | A |
6118768 | Bhatia et al. | Sep 2000 | A |
6134217 | Stiliadis et al. | Oct 2000 | A |
6314501 | Gulick et al. | Nov 2001 | B1 |
6359861 | Sui et al. | Mar 2002 | B1 |
6430156 | Park et al. | Aug 2002 | B1 |
6438134 | Chow et al. | Aug 2002 | B1 |
6470016 | Kalkunte et al. | Oct 2002 | B1 |
6532213 | Chiussi et al. | Mar 2003 | B1 |
6532501 | McCracken | Mar 2003 | B1 |
6560230 | Li et al. | May 2003 | B1 |
6577596 | Olsson et al. | Jun 2003 | B1 |
6594701 | Forin | Jul 2003 | B1 |
6650641 | Albert et al. | Nov 2003 | B1 |
6654374 | Fawaz et al. | Nov 2003 | B1 |
6674721 | Dittia et al. | Jan 2004 | B1 |
6714517 | Fawaz et al. | Mar 2004 | B1 |
6735206 | Oki et al. | May 2004 | B1 |
6735210 | Lindeborg et al. | May 2004 | B1 |
6765915 | Metzger et al. | Jul 2004 | B1 |
6785232 | Kotser et al. | Aug 2004 | B1 |
6813268 | Kalkunte et al. | Nov 2004 | B1 |
6888806 | Miller et al. | May 2005 | B1 |
6891835 | Kalkunte et al. | May 2005 | B2 |
6892309 | Richmond et al. | May 2005 | B2 |
6920109 | Yazaki et al. | Jul 2005 | B2 |
6940861 | Liu et al. | Sep 2005 | B2 |
7139267 | Lu et al. | Nov 2006 | B2 |
7236491 | Tsao et al. | Jun 2007 | B2 |
7353259 | Bakke | Apr 2008 | B1 |
7421478 | Muchow | Sep 2008 | B1 |
7451221 | Basani et al. | Nov 2008 | B2 |
7606867 | Singhal et al. | Oct 2009 | B1 |
7627672 | Lai et al. | Dec 2009 | B2 |
7660824 | Halpern et al. | Feb 2010 | B2 |
7673008 | Kojima | Mar 2010 | B2 |
7738504 | Deaner et al. | Jun 2010 | B1 |
7849178 | Shen et al. | Dec 2010 | B2 |
7949893 | Knaus et al. | May 2011 | B1 |
8122289 | Sargor et al. | Feb 2012 | B2 |
8266235 | Jalan et al. | Sep 2012 | B2 |
8849938 | Jalan et al. | Sep 2014 | B2 |
8897154 | Jalan et al. | Nov 2014 | B2 |
9154577 | Jalan et al. | Oct 2015 | B2 |
9477563 | Jalan et al. | Oct 2016 | B2 |
9596134 | Jalan et al. | Mar 2017 | B2 |
20010043564 | Bloch et al. | Nov 2001 | A1 |
20020012348 | Mizuhara et al. | Jan 2002 | A1 |
20020071387 | Horiguchi et al. | Jun 2002 | A1 |
20020075875 | Dravida et al. | Jun 2002 | A1 |
20020131413 | Tsao et al. | Sep 2002 | A1 |
20030023898 | Jacobs | Jan 2003 | A1 |
20030133406 | Fawaz et al. | Jul 2003 | A1 |
20030140132 | Champagne | Jul 2003 | A1 |
20030158886 | Walls et al. | Aug 2003 | A1 |
20030169734 | Lu et al. | Sep 2003 | A1 |
20030189947 | Beshai | Oct 2003 | A1 |
20040024831 | Yang et al. | Feb 2004 | A1 |
20040059813 | Bolder et al. | Mar 2004 | A1 |
20040228274 | Yazaki et al. | Nov 2004 | A1 |
20040246980 | Balakrishnan | Dec 2004 | A1 |
20040264481 | Darling et al. | Dec 2004 | A1 |
20040268358 | Darling et al. | Dec 2004 | A1 |
20050163049 | Yazaki et al. | Jul 2005 | A1 |
20050243856 | Mishra et al. | Nov 2005 | A1 |
20060101372 | Zhuo et al. | May 2006 | A1 |
20060104230 | Gidwani | May 2006 | A1 |
20060123479 | Kumar et al. | Jun 2006 | A1 |
20060164987 | Ruiz Floriach et al. | Jul 2006 | A1 |
20060206594 | Brown et al. | Sep 2006 | A1 |
20070081527 | Betker et al. | Apr 2007 | A1 |
20070086428 | Lai et al. | Apr 2007 | A1 |
20080104215 | Excoffier | May 2008 | A1 |
20090204699 | Kortright | Aug 2009 | A1 |
20100049836 | Kramer | Feb 2010 | A1 |
20100094967 | Zuckerman et al. | Apr 2010 | A1 |
20100162036 | Linden et al. | Jun 2010 | A1 |
20110066672 | Zamarreno et al. | Mar 2011 | A1 |
20110161405 | He | Jun 2011 | A1 |
20120179770 | Jalan et al. | Jul 2012 | A1 |
20120281540 | Khan et al. | Nov 2012 | A1 |
20120297240 | Jalan et al. | Nov 2012 | A1 |
20120311116 | Jalan et al. | Dec 2012 | A1 |
20130151686 | Takaoka et al. | Jun 2013 | A1 |
20150039674 | Agarwal et al. | Feb 2015 | A1 |
20150195182 | Mathur et al. | Jul 2015 | A1 |
20150312322 | Kamat et al. | Oct 2015 | A1 |
20160020947 | Jalan et al. | Jan 2016 | A1 |
20160062855 | Jalan et al. | Mar 2016 | A1 |
20170013051 | Jalan et al. | Jan 2017 | A1 |
Number | Date | Country |
---|---|---|
1554055 | Dec 2004 | CN |
101019387 | Aug 2007 | CN |
101252506 | Aug 2008 | CN |
101519078 | Sep 2009 | CN |
102708004 | Oct 2012 | CN |
102984194 | Mar 2013 | CN |
2663919 | Nov 2013 | EP |
1183571 | Dec 2013 | HK |
1186802 | Mar 2014 | HK |
H06044090 | Feb 1994 | JP |
2002278787 | Sep 2002 | JP |
2003345640 | Dec 2003 | JP |
2007257023 | Oct 2007 | JP |
2009003923 | Jan 2009 | JP |
5948345 | Jul 2016 | JP |
542960 | Jul 2003 | TW |
I233734 | Jun 2005 | TW |
I277324 | Mar 2007 | TW |
WO2010077222 | Jul 2010 | WO |
WO2012097015 | Jul 2012 | WO |
WO2012170226 | Dec 2012 | WO |
WO2015164027 | Oct 2015 | WO |
Entry |
---|
Parekh et al., “A generalized processor sharing approach to flow control in integrated services networks: the single node case,” IEEE/ACM Trans. Networking, pp. 344-357, Jun. 1993. |
Bennett et al., “WF2Q: Worst-case fair weighted fair queueing,” Proc. IEEE INFOCOM '96, pp. 120-128 San Francisco, CA, Mar. 1996. |
Golestani, S.J., “A self-clocked fair queueing scheme for broadband applications,” Proc. INFOCOM '94, pp. 636-646, Jun. 1994. |
Zhang, L. “Virtual Clock: A new traffic control algorithm for packet-switched networks,” ACM Trans. On Computer Systems, vol. 9, No. 2, pp. 101-124, May 1991. |
Shreedhar et al., “Efficient fair queuing using deficit round-robin,” IEEE/ACM Trans. Networking, vol. 4, No. 3, pp. 375-385, Jun. 1996. |
Stiliadis et al., “Efficient fair queueing algorithms for packet-switched networks,” IEEE/ACM Trans. Networking, vol. 6, No. 2, pp. 175-185, Apr. 1998. |
Suri et al., “Leap forward virtual clock: a new fair queuing scheme with guaranteed delays and throughput fairness,” Proc. INFOCOM '97, pp. 557-565, Apr. 1997. |
Stiliadis et al., “Latency-rate servers: a general model for analysis of traffic scheduling algorithms,” IEEE/ACM Trans. Networking, vol. 6, No. 5, pp. 611-624, Oct. 1998. |
Matsufuru et al. “Efficient fair queueing for ATM networks using uniform round robin,” Proc. INFOCOM '99, pp. 389-397, Mar. 1999. |
Katevenis et al., “Weighted round-robin cell mutliplexing in a general-purpose ATM switch chip,” IEEE Journal on Selected Areas in Communication, vol. 9, No. 8, pp. 1265-1279, Oct. 1991. |
Chaskar et al., “Fair scheduling with tunable latency: A Round Robin approach,” IEEE Globecom '99, pp. 1328-1333, Dec. 1999. |
Bennett et al., “High speed, scalable, and accurate implementation of packet fair queueing algorithms in ATM networks,” Proc. ICNP '97, pp. 7-14, Oct. 1997. |
Nageshwara Rao et al., “Concurrent access of priority queues,” IEEE Trans. On Computers, vol. 37, No. 12, pp. 1657-1665, Dec. 1998. |
Rexford et al., “Hardware-efficient fair queueing architectures for high-speed networks,” Proc. INFOCOM '96, pp. 638-646, Mar. 1996. |
Number | Date | Country | |
---|---|---|---|
20170149606 A1 | May 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14872937 | Oct 2015 | US |
Child | 15426969 | US | |
Parent | 13154399 | Jun 2011 | US |
Child | 14872937 | US |