Not Applicable.
Not Applicable.
The present invention relates generally to communication systems and, more particularly, to packet-based communication systems.
As is known in the art, packet-based communication systems include a series of interconnected devices, such as routers, for enabling point-to-point data exchanges. The failure of network elements and/or traffic surges can cause packet routers to receive traffic at a higher rate than the bandwidth of the link on the output interface to which the traffic is routed. This can result in the build-up of packets in a buffer contained in the router. If the overload condition persists long enough, the buffer can overflow and degrade end-user performance.
So-called overload controls can be used to selectively discard packets so that the “most important” packets are delivered to the extent possible. Two examples of most important packets are those containing network control information and those from end-users who pay a premium for special treatment.
Quality of service (QoS) differentiation of packets is the basis for several known packet-network services. In typical QoS schemes, packets are marked according to their class of service (which is a measure of “value”). An overflow control algorithm uses this information to discard the “least valuable packets” first. It may be desirable to avoid discarding all of the least valuable packets in order to retain end user goodwill. The capability of current routers to maintain throughput-by-class under overload is primarily provided in an algorithm called weighted random early discard (WRED).
However, there are drawbacks to using the WRED algorithm for overload control. More particularly, when WRED is configured to ensure that the most valuable packets are protected from being discarded in the case where the overload is caused by an excess of low-value packets, the low-value packets receive an undesirably small throughput when the overload is caused by high-value packets. That is, class 1 traffic is protected from class 2 overloads but class 2 traffic is not protected from class 1 overloads. While so-called smoothing may avoid preventive discards on small traffic bursts, WRED does not avoid them when the queue empties after a large burst. Further, probabilistic dropping in WRED, which is designed to avoid consecutive packet drops from the same end user, is not needed on backbone routers since these routers serve a relatively large number of users. In addition, the performance of WRED is not easily predicted, so that setting the control parameters to achieve performance objectives is challenging.
Additional known active queue management ways of providing preferential services by class when packets are placed in a single queue include priority queuing (PQ) and Weighted Fair Queuing (WFQ). Deficit Round Robin (DRR) is a scheduling algorithm that provides preferential service by class when packets from different classes are placed in separate queues. However, priority queuing cannot provide bandwidth guarantees to all classes. While Weighted Fair Queuing provides bandwidth guarantees to all classes by controlling the packet processing schedule, its computational requirements grow with the number of connections so it may not scale to backbone routers, which handle a very large number of connections. And Deficit Round Robin applies to router architectures that place packets from different classes into different queues, so it cannot be used in routers with a single queue for all packets. Another scheduling algorithm is described in Clark and Fang, “Explicit Allocation of Best-Effort Packet Delivery Service,” IEEE/ACM Transactions on Networking, Vol. 6, No. 4, August 1998, which is incorporated herein by reference.
These existing overload control algorithms provide only limited levels of performance and do not enable users to configure routers to achieve the limited performance levels with adequate reliability. Without such reliability, systems designers cannot accurately predict the effects of overloads on end-users.
It would, therefore, be desirable to provide a reliable packet-based communication system having enhanced overload performance.
The present invention provides a packet-based communication system having an active queue management feature that discards arriving packets such that an overloaded link interface provides guaranteed bandwidth to QoS classes that share a buffer. With this arrangement, the system provides reliable overload performance and enhanced end-user satisfaction. While the invention is primarily shown and described in conjunction with a router in a network, it is understood that the invention is applicable to packet-based devices and networks in general having QoS differentiation.
In one aspect of the invention, a packet-based communication system includes an overload control mechanism that provides a guaranteed bandwidth to each class of buffered traffic. In one embodiment, a threshold for each class of traffic is computed so that during overload conditions the throughput of a given class is the smaller of its offered load and the guaranteed bandwidth share for that class. Packets in a given class are discarded when the number of buffered packets in that class exceeds the threshold for that class.
The invention will be more fully understood from the following detailed description taken in conjunction with the accompanying drawings, in which:
The Internet and the LAN, e.g., LAN 104b, can include one or more routers 106a-P, 108a-Q for routing traffic from point-to-point to support the end-users 102. In general, one or more of the routers 106, 108 provide overload control for a queue that allocates link bandwidth to various Quality of Service (QoS) classes in fixed proportions. As is known in the art, a connection between routers is referred to as a link. As used herein, where N QoS classes are described, it is understood that class 1 traffic is the most preferred (highest priority) and class N traffic is the least preferred.
As described in detail below, the enhanced overload performance provided by the system 100 is achieved by keeping track of the number of packets in the buffer for each QoS class and discarding new packets from a given class when the current number of packets in that class exceeds a given threshold. The active queue management feature of the present invention discards arriving packets such that an overloaded link interface provides guaranteed bandwidth to QoS classes that share a single buffer.
In one embodiment shown in
Initially, a series of thresholds TH_1, TH_2, . . . , TH_N for a predetermined number N of QoS classes are computed. In general, the thresholds are computed so that, under overload conditions, the throughput of a given class of traffic, e.g., class k, is at least the smaller of its offered load and the guaranteed bandwidth share for class k. The overload control mechanism 202 keeps track of the total number of packets (Qtotal) in the buffer 208 and the number of packets of class k in the buffer (Qk), where k=1, 2, . . . , N−1. class k packets are discarded when Qk>TH_K. class N packets are discarded when Qtotal>TH_N.
Further details of the inventive overload control mechanism are shown and described in
It will be readily apparent to one of ordinary skill in the art that parameters for optimal performance of the inventive overload scheme can be derived in a variety of ways. In general, the parameters should be selected to allocate link bandwidth to QoS classes in fixed proportions and discard packets from a given class based upon an optimal threshold level.
In one particular two-traffic-class embodiment, a lower bound on the class 2 bandwidth during periods of congestion is 1−ξ1, where ξ1 represents the largest throughput that class 1 can use during overload conditions. The second threshold TH_2 can be set to the buffer size minus a predetermined number of packets, e.g., TH_2=BUFFER_SIZE−10. This ensures that, with relatively high probability, the buffer has space for TH_1 class 1 packets. The first threshold TH_1 can then be set to ξ1*TH_2.
Let ρj represent the throughput, as a share of the total bandwidth, that class j traffic offers, θ1 represent the throughput that class 1 achieves, and ξ1 represent the largest throughput that class 1 can use during overload conditions.
Supposing, ρ1+ρ2>1 and ρ1>ξ1, the general overload condition implies that E(Q1)+E(Q2)≅TH_2, where E is the expectation operator. These approximations and Little's Law imply that E(Q1)≅θ1*TH_2. Thus, the first threshold TH_1 can be set such that TH_1=ξ1*TH_2, which implies that θ1≅ξ1.
When ρ1+ρ2<1, the queue lengths are relatively small and the first and second thresholds TH_1, TH_2 do not have any effect as long as the first threshold TH_1 is not “too” small. Thus, θ1=ρ1 and θ2=ρ2. An illustrative preliminary first threshold TH_1 is greater than or equal to about fifty packets.
Now suppose that ρ1=0.4, which is less than ξ1, and ρ2 remains at 0.8. That is, the class 1 traffic is not in an overload condition. As shown in
The overload control feature of the present invention can be readily expanded to included three classes of traffic. For example, where ξj is the throughput bound for class j traffic for j=1, 2, the third threshold TH_3 can be set to BUFFER_SIZE−10. The second threshold can be set as TH_2=ξ2*TH_3, and the first threshold can be set as TH_1=ξ1,*TH_3. In one particular embodiment, ξ1=0.30 and ξ2=0.60. Further traffic classes can added as desired.
Similarly,
The present invention provides an overload control mechanism that enhances overall overload condition performance for a packet-based system and provides reliable operation. By selecting class thresholds, link bandwidth can be allocated to QoS classes in fixed proportions.
One skilled in the art will appreciate further features and advantages of the invention based on the above-described embodiments. Accordingly, the invention is not to be limited by what has been particularly shown and described, except as indicated by the appended claims. All publications and references cited herein are expressly incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
4849968 | Turner | Jul 1989 | A |
5140584 | Suzuki | Aug 1992 | A |
5434848 | Chimento et al. | Jul 1995 | A |
5487061 | Bray | Jan 1996 | A |
5546389 | Wippenbeck et al. | Aug 1996 | A |
5553061 | Waggener et al. | Sep 1996 | A |
5629936 | Lee et al. | May 1997 | A |
5764641 | Lin | Jun 1998 | A |
5764740 | Holender | Jun 1998 | A |
5787081 | Bennett et al. | Jul 1998 | A |
5790524 | Bennett et al. | Aug 1998 | A |
5822300 | Johnson et al. | Oct 1998 | A |
6064651 | Rogers et al. | May 2000 | A |
6072800 | Lee | Jun 2000 | A |
6091709 | Harrison et al. | Jul 2000 | A |
6134218 | Holden | Oct 2000 | A |
6134239 | Heinänen et al. | Oct 2000 | A |
6145010 | Hiscock et al. | Nov 2000 | A |
6160793 | Ghani et al. | Dec 2000 | A |
6167445 | Gai et al. | Dec 2000 | A |
6169740 | Morris et al. | Jan 2001 | B1 |
6208619 | Takeuchi | Mar 2001 | B1 |
6215766 | Ammar et al. | Apr 2001 | B1 |
6215769 | Ghani et al. | Apr 2001 | B1 |
6407999 | Olkkonen et al. | Jun 2002 | B1 |
6463068 | Lin et al. | Oct 2002 | B1 |
6469982 | Henrion et al. | Oct 2002 | B1 |
6570876 | Aimoto | May 2003 | B1 |
6680908 | Gibson et al. | Jan 2004 | B1 |
6826150 | Bhattacharya et al. | Nov 2004 | B1 |
6829224 | Goldman et al. | Dec 2004 | B1 |
6920109 | Yazaki et al. | Jul 2005 | B2 |
7061862 | Horiguchi et al. | Jun 2006 | B2 |
20010026555 | Cnodder et al. | Oct 2001 | A1 |
20020003777 | Miyamoto | Jan 2002 | A1 |
20040071086 | Haumont et al. | Apr 2004 | A1 |