In the area of data networks, there has been much effort concerning distributed rate control to control the sending rate of multiple flows over arbitrary network topologies. Previous efforts have aimed to perform rate control in ways that fairly and fully utilize a network while maintaining low operating congestion (which results in low queuing delay and low congestion induced packet loss), and perhaps smooth changes in transmission rate.
To these ends, there have been many variants of congestion or rate control protocols developed. Examples of rate control algorithms include various congestion control algorithms used in the transmission control protocol (TCP) and variants such as: TCP Reno, TCP New Reno, Compound TCP (CTCP), Binary Increase Congestion TCP (BICTCP), TCP Vegas, Fast TCP, Data Center TCP (DCTCP), and others. The TCP-friendly rate control (TFRC) algorithm has also been used a rate control protocol for video conferencing. Algorithms for co-existence of loss and delay based flows, e.g. PERT (Probabilistic Early Response TCP), and MPERT (Modified PERT), have also been used. Available-bandwidth estimation techniques have been proposed and used for some real-time VoIP/video-conferencing applications. There have also been efforts toward utility-maximization based rate control using delay as the congestion signal (primal-dual formulation).
Some rate control techniques have been specifically directed to applications that require low end-to-end latency. End-to-end latency (to be referred to as “E2E”) refers to the time from when content is generated (at the sender side) to the time when content is consumed (at the receiver side). Examples of low E2E applications (applications sensitive to end-to-end latency) include real-time media delivery systems such as video conferencing and voice over IP, interactive software such as remote desktop applications, online multiplayer games, interactive websites, web-based applications such as Office 365™ and Google Docs™, etc.
With respect to rate control in general, and in particular with respect to rate control with low E2E latency, existing rate control techniques have had shortcomings. High utilization of network bandwidth is often not achieved. Network bandwidth may not be fairly allocated, in particular, amongst multiple conforming flows and amongst additional conforming flows and more aggressive flows such as TCP. Often, there is slow convergence to a steady state of transmission. It may be difficult to dynamically adapt to higher congestion levels when aggressive flows are present. Some rate control algorithms over-oscillate when in a steady state. Moreover, many previous rate control algorithms are not resilient to diversified network conditions such as noise in delay measurements and random packet loss.
Techniques related to bandwidth management are discussed below.
The following summary is included only to introduce some concepts discussed in the Detailed Description below. This summary is not comprehensive and is not intended to delineate the scope of the claimed subject matter, which is set forth by the claims presented at the end.
Described are techniques to use adaptive learning to control bandwidth or rate of transmission of a computer on a network. Congestion observations such as packet delay, packet loss, ECN (Earlier Congestion Notification) marking are used to compute a congestion signal. The congestion signal is adaptively learned to correlate with information about actual congestion on the network, and the transmission rate is adjusted according to the degree of correlation. Transmission rate may not adapt when the congestion signal indicates a low level of congestion, and may decrease when the congestion signal indicates a high level of congestion. The desired congestion level is also adaptively learned. For instance, the relative effects and observation of loss, delay and ECN marking on the congestion signal may change over time. Moreover, a desired operating congestion level may be learnt and minimized by adaptive adjustment.
Many of the attendant features will be explained below with reference to the following detailed description considered in connection with the accompanying drawings.
The present description will be better understood from the following detailed description read in light of the accompanying drawings, wherein like reference numerals are used to designate like parts in the accompanying description.
The learning module 134, in combination with the rate controller 132 may perform various functions for learning-based rate control. Such functions will be briefly mentioned and described in detail below. Parameters for rate control (for instance, as may be used in a utility maximization framework) may be dynamically adjusted to maintain a desired operating congestion level. This may include functionalities such as: increasing the operating congestion level when more aggressive flows (such as TCP) begin; decreasing operating congestion level when such flows end; increasing the operating congestion level for “long flows” with multiple bottlenecks; and decreasing the operating congestion level to levels approaching or at minimal achievable bounds in the presence of conforming flows
Other functionalities may include: using a larger timescale to adjust the parameters when compared to the timescale used to adjust the rate; using a rate control scheme where the desired operating congestion level is independent of transmission rate (and thus independent of number of flows on a link; dynamically adjusting the desired operating congestion level by learning what it should be.
Congestion level signals (e.g., queuing delay, packet loss and ECN marking) may be used as predictors of true congestion level. This may involve determining what are the lowest achievable congestion level signals by looking at statistics of these signals in periods of non-congestion. Congestion level signals may also be used as predictors of congestion by finding correlations between these signals and congestion levels in periods of congestion, which may be accomplished by using the congestion level signals as predictors of sending rate; if a signal is a predictor of sending rate, then it is also a predictor of congestion.
Regarding the rate controller 132, as the learning module 134 refines what is known about congestion and congestion signals, it informs the rate controller 132. For example, if current congestion signals are determined to not correlate with actual congestion, then the learning module 134 may inform the rate controller 132 to not adapt the current rate of transmission.
Note that any arbitrary modularization of functionality can be used to allow the node 120 to perform learning-based rate control functions.
A framework within which learning-based rate control can be implemented will now be described. Consider a network to be a graph G with a set of nodes N and links L that connect the nodes. Each link (an edge in the graph), has a certain propagation delay, channel capacity (bandwidth), and perhaps loss rate. Some subset of the nodes S⊂N comprises source nodes that are originally transmitting content via the links. Corresponding to each source sεS, is a receiver node Rs⊂N, |Rs|=1. Other nodes in the network can be routers, some of which may perform rate control. The links can be wired or wireless, and there may be other underlying physical layer protocols being used (for example in 802.11 type protocols). Note that the underlying link protocols may perform retransmissions on a link which essentially translates random packet loss into noise in delay). For each source sεS, let ⊂L, be the set of links traversed when going from source node “s” to receiver node Rs. The rate at which source S transmits is given by xs.
With reference to formulas herein, “congestion level” over a particular link is defined to mean the current amount of outstanding data in the network over that link. In particular if link lεL has a channel capacity C1, then the congestion level at time t over the link is given by:
Note that the congestion level over a link typically results in queuing delay and packet losses. However, a theoretical exact formulation as to how congestion level translates to queuing delay and packet losses would be a function of various factors, such as the underlying link and its lower layer protocol, as well as the node at the link which is sending data and its rate control policy and buffer size. Note also that not all delay and packet loss increases are actually caused by congestion level and may simply be random. Although rate control aims to react to congestion level on individual links, usually the end-to-end rate control actually sees end-to-end signals such as loss and delay which may or may not be due to congestion.
In this discussion, it is assumed that a source S sees end-to-end signals such as queuing delay, which will be written as δs, and packet loss, which will be written as εs. The function {dot over (f)}(x) will be used to denote the derivate of function f with respect to x.
As an example of how congestion level translates to queuing delay and loss, consider two different rate control policies that may be implemented at individual routers, namely a leaky bucket and token bucket policy.
When a router implements a leaky bucket rate control policy, the router has a queue (buffer) of size B. The router places the packets destined for a particular router port (for example a port with an Ethernet cable connecting to another router) into this queue. The router drains packets from this queue at a constant rate, which is equal to the capacity of that link C1. Thus, if the aggregate incoming rate into the link is larger than the channel capacity, the queuing delay will continue to build up until the buffer size B is filled, after which packets will experience a queuing delay of B and a certain percentage of packets will simply be dropped. That is to say, δs will increase until it is equal to B (where B is in units of delay (i.e. size of buffer in bits/bitrate)), followed by packet loss equal to εs={dot over (Q)}1(t).
As another example, consider a token bucket policy. Such a policy may be used when the underlying physical medium actually has larger capacity than C1, but the network administration limits the rate on a per-port basis. In this policy, for each outgoing port, tokens are inserted into a token bucket at a constant rate C1, where the token bucket has size B. If there are any tokens remaining in the bucket, then packets are immediately sent. If the token bucket is empty, then packets may either be dropped or placed into a queue. Here, if the packets are simply dropped when the token bucket is empty, then packets will only see loss, and there will not be a queuing delay increase upon congestion. In fact the only congestion indication will be whether the congestion level is below or above a certain threshold.
In practice, either or both of the policies may exist along paths between nodes. In addition, routers may implement other control strategies, such as weighted fair queuing, where different priority packets are placed into different queues to give different congestion signals to different flows, thus giving them different rates of operation.
Finally, consider that some links may have underlying packet loss which is random and not due to congestion, such as may occur with wireless links. In this case, the packet loss may be seen, or if the underlying medium attempts a certain number of retransmission, then some random packet loss may be translated into and seen as random noise in delay.
In addition, other networks may also give other signs of congestion such as ECN (explicit congestion notification) markings, which tell nodes whether router buffers are full beyond some threshold.
In sum, it can be difficult to directly translate observed signals such as loss and queuing delay directly into a measurement of congestion level.
To implement an example distributed rate control protocol, in one embodiment the utility maximization framework (described in detail elsewhere) serves as a starting point. In this embodiment, rate control is modeled as the following utility maximization problem, where Us(xs) is the utility obtained when rate xs is used to transmit source S and where x is the vector of source rates for all sources:
It can be shown that this problem can be solved by solving the equivalent maximization problem of the Lagrangian L(x,q), where x is an S-dimensional vector of source rates and q is an L-dimensional vector of the Lagrange multipliers:
This problem can be solved by the iterative Primal-dual algorithm and converges to the optimal solution using:
The point to note here is that the LaGrange multipliers for each of the links q1 is the congestion level over that link as defined in Error! Reference source not found. In addition, ρs=q1 is the sum of congestion levels over all links in the path traversed between source node S and receiver node Rs.
Assuming a utility function such as Us=αs/log(xs), it is possible to arbitrarily multiply all utility functions by a constant and get the same result; the following rate control algorithm results, where t is any time at T1 is any time period:
For the case of a single link, since all flows in theory see the same congestion operating point ρs, in steady state, the following rate allocation (the point at which rate change is zero) is obtained:
Some observations are to be noted. The final rate allocation is independent of ks. The final rate allocation only depends on the ratio of αs/β, which means that if all flows have the same αs, they will all have the same allocation. Similarly, a long flow with two bottleneck links with similar capacity and share on each link will have a congestion level which is twice that of a short flow, and thus will get half the rate. Thus an acceptable rate control strategy is possible when the parameters are chosen carefully and ρs can be accurately predicted as a function of queuing delay δs and loss rate εs.
However, with this scheme the operating congestion point in steady state is given by
Unfortunately, for a given set of parameters α and β, a congestion operating point results which is inversely proportional to the rate, which means that for a fixed capacity link, the operating congestion point increases as the number of flows increases.
Existing TCP has the same issue, since if TCP throughput is viewed as a function of operating congestion point, the following may be observed:
Where L is the packet size, tRTT is the round-trip time, and tRTO is the retransmission time-out value. The variable ε is the loss rate seen by source S. Again, with TCP, the operating congestion point (i.e. the packet loss rate) is inversely proportional to the rate. In view of the above, various issues will be addressed below in implementing bandwidth management protocols, and in particular for low E2E applications. Such issues may include: how to pick an operating congestion point; how to adapt the operating congestion point to network conditions and flows; how to adapt parameters to achieve the desired operating congestion point without giving up fairness; how to correctly estimate the congestion level as defined in Error! Reference source not found. by looking at queuing delay and loss; and how to map desired operating congestion point to desired queuing delay and loss.
Before proceeding further, note that Error! Reference source not found. is in a form known as Inverse-Increase Additive-Decrease (IIAD), because the increasing portion αs/xs is inversely proportional to rate, and the decreasing portion −βρs is additive. Note also that the same results can be achieved by using an Additive-Increase Multiplicative-Decrease (AIMD) method according to the following:
x
s(t+T1)=max(0,xs(t)+ks(αs−βρsxs(t))).
This form will be assumed hereafter, although both forms provide the same steady-state rate point. In addition, the effect of a different αs is to change the final rate allocation. Hereafter, it is assumed that all the flows have the same α=αs. Then, by defining γ=α/β, the equation is rewritten as the following:
x
s(t+T1)=max(0,xs(t)+ks(γ−ρsxs(t))).
This is a reasonable formulation because the operating congestion point is simply a function of γ and the final steady-state rate. The problem becomes to simply find a value for γ. The remainder of the discussion deals with properly and dynamically setting γ as well as correctly estimating ρs from queuing delay δs and loss rate εs.
To approach dynamic parameter adaptation in the utility maximization based framework presented above, it may be assumed that a desired operating congestion point ρd=ρs is known. The subscript d indicates that the operating congestion point is a desired value. Since γ is being adapted in a dynamic manner, γs will be used to denote the value for source S.
One way to perform adaptation is to simply set γs=ρdxs at some interval which is much slower than the interval used to adapt the rate T1. However, this may not result in all flows converging to the same value of γ, which will result in unfair allocation. Another way is to adapt a method similar to that presented in MS Docket #329500.01, which is something like the following three zone AIMD scheme:
This is a classic AIMD type scheme, where all the flows which share the same bottleneck (and thus see the same congestion level) are guaranteed to converge to the same value of γ. This happens because the difference between differing values of γreduces every congestion cycle, since the flow with a larger value of γ has a larger reduction when encountering congestion (multiplicative decrease), whereas the increase is the same when congestion level is below the desired level. Again, if the parameters μmax, μmin, φmax, φmin are the same, the convergence behavior is the same and independent of the scale factor ks
The above piece-wise linear formulation is just one which allows γ to converge. Other adaptation schemes are also possible which allow convergence. In addition, the converged value becomes γs=ρdxs. If all values converge to the same value, then the rate allocation should maintain the same fairness property as for the case with fixed parameters. Of course the desired congestion point may be different for different flows, depending on the bottlenecks encountered in the path.
Maintaining the optimal allocation for the utility maximization framework is also possible even if different flows have differing values of γ. To understand how this is possible, think of breaking up the path between source S and receiver node Rs into individual links along the way, and then:
Where σd,1 becomes the desired operating congestion pointon link 1. So long as all flows which go across this link have the same contribution of desired congestion level from this link, the same rate allocation will result.
If T2 is larger than T1 by a significant amount, then by the timescale separation principle, convergence should occur for both γs as well as xs.
Since γ is actually changing the derivative of xs, convergence can be sped up (and an extra timescale can be avoided) by directly adjusting the rate according to:
This is similar to the proposal in MS Docket #329500.01, with the exception that the desired operating congestion point ρd is not a predetermined constant value and is not the same for all flows. In addition, the congestion level is not simply restricted to be the queuing delay observed. The exact rate change is slightly different from that proposed before (i.e. it is piece-wise linear instead of using the TCP-Illinois style rate adjustment curve). An explicit ks scale factor is available to decouple the operating congestion point from the rate (and congestion level) oscillation in steady state.
Either of these two schemes can be used. That is to say, either rate can be directly adjusted using the desired congestion level, or indirectly by controlling γ using the desired congestion level.
In one embodiment, rather than fix the operating congestion level point, it is dynamically adjusted. Not only is this helpful in the case of multiple bottleneck links (since otherwise a long flow would essentially get zero rate), but it is also useful in determining the lowest possible achievable congestion level while still utilizing full network resources. In addition, this approach allows adaptation in the case when more aggressive flows (defined to be those with a higher operating congestion point) become present or when they depart. A scheme to adapt the desired congestion level may use the following:
ρd(t+T3)=max(1−κT3,0)ρd(t)+min(κT3,1)f(ρobserved,υ)
Where f(ρobserved,υ) is the 100*υth percentile of the observed queuing delay (ρobserved) over a period of length T3 and is chosen to be less than 0.5 (i.e. 50—the percentile value). Essentially, the timescale forgetting factor is κT3; the time period over which ρd changes to the new value. Such an adaptation is carried out so long as the f(ρobserved,υ) (for example the 45-th percentile of congestion level) deviates by more than x % (for example 5%) of the current desired value. In addition, the adaptation need only be carried out when it is determined that the observed periods of congestion level were those periods where congestion was actually present (i.e. observed congestion level was close to or higher than desired congestion level). Longer flows will observe higher congestion levels than shorter flows and thus will increase their desired congestion level. Upon encountering congestion and achieving a desired congestion level, all flows will attempt to decrease their congestion levels to the lowest possible achievable values.
Flows may also choose to reduce their ks values upon decreasing congestion levels so as to reduce oscillation in congestion level values at the expense of slower convergence times.
Determining when there is Congestion
Discussion to this point has concerned congestion level rather than queuing delay or packet loss. These measurable quantities are defined as congestion level signals. This section deals with interpreting those signals as congestion. Note that both the previous and current section may be relevant for low E2E delay scenarios. However, the present section, while generally applicable, may have use for high E2E delay scenarios or in cases where the desired operating “congestion level signals” are within the random noise of the system, for example cases where random loss is inherently present.
Two aspects of how “congestion level signals” relate to “congestion level” will be considered: the lowest observed “congestion level signal” which should be taken to mean congestion; and when congestion is observed, how much to react to it. One method to address these is the following: the lowest observed “congestion level signal” which indicates congestion is occurring can be taken to be the mean plus some constant times the standard deviation of the observed signal in the non-congested state. A non-congested state is defined to be a state where the observed congestion level is significantly lower than the desired congestion level. For example, for queuing delay, it can be said that ρ=0, if δobserved<mean(δnon-congested)+c1stddev(δnon-congested)=δ0. This can be similarly defined for loss.
How much to react each of these signals can also be learned, for example, by trying to compute a linear regression of transmission rate as a function of delay and loss observed (especially looking at those intervals where the desired delay and loss are close to the observed delay and loss). For those intervals, a linear fit can be estimated for rate as xs=λδ+θε+η. The coefficients in this linear are then used fit when changing the rate. If increasing rate results in decreasing delay or loss, for example if λ<0, then set λ to zero using λ:=max (λ, 0) and similarly θ:=max (θ, 0).
Using these two equation along with the previous discussion, a rate adjustment protocol which reacts to “congestion level signals” of loss and delay can be derived as follows:
Note that the parameters αmax, αmin,βmax, βmin will likely be different for the two adaptations for each of the congestion level signals, and the interval over which adaptation is done, T1, T2, will be different. Note also that for both congestion signals a minimum is defined below which t. In addition, parameters λ, θ can be added, which are learned. If no correlation exists between a congestion signal and congestion level, then no rate change takes place due to that signal.
Finally, note that the final rate allocation is “fair” and the same as the original utility maximization problem even if λ, θ go to various values, provided that the correlation is observed. Only the rate of convergence changes as a result. The desired congestion level signals should correctly converge (δd, εd) and that the congestion level signals in the non-congested state are observed to be the same (δ0, ε0).
The loss rate can be determined using a method proposed in TFRC, i.e. using the inverse of a weighted average of interval lengths between loss events. A similar scheme can be used if γ is adapted as in Error! Reference source not found, and rate as adapted as in Error! Reference source not found.
There are several parameters that are to be chosen, for example, δ1, ε1. These parameters are set to be 2× the desired value, for example δ1=2δd. Other values to be chosen are αmax, αmin, βmax, βmin for both the delay and loss adaptation. To prevent large oscillations, set αmin=αmaxmin (1, σδ/9δd), where σδ is the standard deviation in the observed queuing delay. That is, if the standard deviation in the observed delay is small, so is the gap in change in transmission rate. Similarly, set βmin=βmaxmin (1, σδ/δd). Convergence to a fair share is guaranteed while αmin≠0, βmin≠0. Moreover, αmax, βmax can be chosen by heuristic, and perhaps can be made similar to that used in TCP to provide fairness. Also, ks can be chosen to be large initially when operating congestion level (initial value) is high so as to achieve quick convergence to a high rate. It can then be lowered as time progresses and the operating congestion level reduces. Finally, T1, T2 can also be chosen using heuristics and can perhaps be made to be similar to what TCP uses (react to every ACK, NACK) or perhaps a larger timer interval.
If the rate is not directly adjusted and instead γ is adapted according to Error! Reference source not found, then choose μmax, μmin, φmax, φmin in the same manner as αmax, αmin, βmax, βmin.
Although the discussion has related to making the system operate such that the average congestion level is controlled while giving a fair allocation, it is important to note that the variance in congestion level is also important to control. The variance in congestion level is relevant to maintaining full link utilization and to making sure that congestion level signals such as queuing delay and packet loss (which are important to application performance) are controlled. For example,
To prevent variance from being too large, lower ks in case the RTT is large, and lower ks as ρd drops. Also it may be helpful to minimize the discontinuity gap when rate is stabilizing.
To achieve the desired rate, an embodiment may employ something similar to “slow start” in TCP. In particular, quickly increase the rate to some value by doubling the rate upon no-congestion until the first sign of congestion is hit. Since convergence is guaranteed regardless of initial state, this can be used as the initial state for starting the proposed rate control.
Embodiments and features discussed above can be realized in the form of information stored in volatile or non-volatile computer or device readable media. This is deemed to include at least media such as optical storage (e.g., compact-disk read-only memory (CD-ROM)), magnetic media, flash read-only memory (ROM), or any current or future means of storing digital information. The stored information can be in the form of machine executable instructions (e.g., compiled executable binary code), source code, bytecode, or any other information that can be used to enable or configure computing devices to perform the various embodiments discussed above. This is also deemed to include at least volatile memory such as random-access memory (RAM) and/or virtual memory storing information such as central processing unit (CPU) instructions during execution of a program carrying out an embodiment, as well as non-volatile media storing information that allows a program or executable to be loaded and executed. The embodiments and features can be performed on any type of computing device, including portable devices, workstations, servers, mobile wireless devices, and so on.