Multicasting video

Information

  • Patent Grant
  • 6691312
  • Patent Number
    6,691,312
  • Date Filed
    Friday, March 17, 2000
    24 years ago
  • Date Issued
    Tuesday, February 10, 2004
    20 years ago
Abstract
A method of multicasting video to multiple client nodes via intermediate nodes that includes accessing video information descriptive of the video to be multicast, accessing information describing a distribution tree of nodes from a source of the video to the multiple clients nodes via one or more internetwork nodes, accessing rate constraints of nodes in the distribution tree, accessing buffer allocations of the nodes in the distribution tree, and determining one or more smoothed transmission schedules for each node in the distribution tree based on the accessed video information, the accessed information describing the distribution tree, the accessed rate constraints of nodes in the distribution tree, and the accessed buffer allocations of the nodes in the distribution tree, the one or more transmission schedules describing the transmission of video data to one or more children nodes.
Description




FIELD OF THE INVENTION




The invention relates to multicasting video.




BACKGROUND OF THE INVENTION




Development of a high-speed internetworking infrastructure (e.g., the Internet) and digital video technology has led to the emergence of several networked multimedia applications that include streaming video as a media component. Such applications include streaming video broadcasts, distance learning, corporate telecasts, narrowcasts, etc. Digital video traffic typically exhibits high bandwidth requirements and significant burstiness at multiple time scales, owing to the encoding schemes and content variation between and within video scenes. For example, MPEG (Motion Pictures Experts Group) encodes a video in different size frames. Additionally, the information needed to encode each frame can vary, for example, based on how much video information changes from image to image. When combined with other media such as audio, text, HTML (HyperText Markup Language), and images, multi-media transmissions can become even burstier. The high bandwidth requirements coupled with the bursty variable-rate nature of these streams complicates the design of efficient storage, retrieval, and transport mechanisms for such media.




A technique known as workahead smoothing can yield significant reductions in peak rate and rate variability of video transmissions from a server to a single client (i.e. unicast). In smoothing, by transmitting frames early, the sender, or a smoothing node, can coordinate access with the client and can send large video frames at a slower rate without disrupting continuous playback at the client. The frames transmitted ahead of time are temporally stored in buffers present in the server and the client. A characteristic of the smoothed transmission schedule is that the smoothing benefit is a non-decreasing function of buffer sizes present.




SUMMARY OF THE INVENTION




For many applications, streaming video transmission occurs from a server simultaneously to a large number of heterogeneous clients that have different resource capacities (e.g. buffer sizes) and that are scattered geographically over a heterogeneous internetwork that has different resource capacities in different segments. Described herein are techniques that can reduce the bandwidth overhead for such simulcast services.




In general, in one aspect, the invention features a method of multicasting video to multiple client nodes via intermediate nodes. The method includes accessing video information descriptive of the video to be multicast, accessing information describing a distribution tree of nodes from a source of the video to the multiple clients nodes via one or more internetwork nodes, accessing rate constraints of nodes in the distribution tree, accessing buffer allocations of the nodes in the distribution tree, and determining one or more smoothed transmission schedules for each node in the distribution tree based on the accessed video information, the accessed information describing the distribution tree, the accessed rate constraints of nodes in the distribution tree, and the accessed buffer allocations of the nodes, the one or more transmission schedules describing the transmission of video data to one or more children nodes.




Embodiments may include one or more of the following features. Accessing buffer allocations can include determining buffer allocations based on the accessed video information, the accessed information descriptive of a distribution tree, and the accessed rate constraints. Accessing rate constraints may include receiving rate constraints. The determining may include determining an upper constraint for each node, determining a lower constraint for each node, and determining a smoothing schedule based on the determined upper and lower constraints. The determining of a smoothed schedule may use majorization. The method may further include distributing the determined transmission schedules to nodes in the distribution tree. The determined transmission schedules may include schedules that cause at least one node to differentially cache data for distribution to the at least one node's children.




In general, in another aspect, the invention features a computer program product, disposed on a computer readable medium, for multicasting video to multiple client nodes via intermediate nodes. The program includes instructions for causing a processor to access video information descriptive of the video to be multicast, access information describing a distribution tree of nodes from a source of the video to the multiple clients nodes via one or more internetwork nodes, access rate constraints of nodes in the distribution tree, access buffer allocations of the nodes in the distribution tree, and determine one or more smoothed transmission schedules for each node in the distribution tree based on the accessed video information, the accessed information describing the distribution tree, the accessed rate constraints of nodes in the internetwork, the accessed buffer allocations of nodes in the internetwork, and the one or more transmission schedules describing the transmission of video data to one or more children nodes.




Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. Although methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present invention, suitable methods and materials are described below. All publications, patent applications, patents, and other references mentioned herein are incorporated by reference in their entirety. In case of conflict, the present specification, including definitions, will control.




In addition, the materials, methods, and examples are illustrative only and not intended to be limiting.











Other features and advantages of the invention will be apparent from the following detailed description, including the figures, and from the claims.




BRIEF DESCRIPTION OF THE DRAWINGS





FIG. 1

is a diagram of a network server multicasting video to clients.





FIG. 2

is a diagram of network nodes.





FIG. 3

is a diagram of a single link.





FIG. 4

is a graph of a feasible transmission schedule.





FIG. 5

is a graph of a lazy transmission schedule.





FIG. 6

is a listing of pseudo-code for a program that verifies feasible transmission schedule exists.





FIG. 7

is a listing of pseudo-code for a program that determines a smooth transmission schedule for each link.





FIG. 8

is a listing of pseudo-code for a program that determines network node buffer sizes.





FIG. 9

is a flowchart of a process for determining transmission schedules and/or buffer allocation(s).





FIG. 10

is a flowchart of a process for transmitting data in accordance with one or more determined transmission schedules.











DETAILED DESCRIPTION




Presented herein are techniques that integrate workahead smoothing with multicasting to efficiently multicast streamed video from a server to multiple clients over a distribution tree topology. The techniques use “differential caching”, a technique for temporal storing of video frames at intermediate network nodes of the distribution tree. For prerecorded video streaming, integrating smoothing with differential caching is used to compute a set of optimal transmission schedules for a distribution tree, which reduces the peak transmission rate and variability of the video transmission along each link. When buffering at the internal nodes is the constraining resource, techniques can check whether there exists a set of feasible “optimal” multicast schedules to transmit to a particular video. If a feasible “optimal” multicast schedule exists, techniques can determine a set of optimal transmission schedules. When the link bandwidths in the distribution tree are the constraining resource, software can determine the minimal total buffer allocation for all the nodes in the tree, and the corresponding allocation at each node, such that there exists a set of feasible transmission schedules to distribute a particular video. Multicasting smoothed video using schedules computed by the smoothing algorithm can reduce the total transmission bandwidth requirements by more than a factor of three as compared to multicasting of an un-smoothed stream.





FIG. 1

shows a server


102


multicasting streaming video to a number of heterogeneous clients


106




a


-


106




e


that are connected through a heterogeneous internetwork


100


. Segments


108




a


-


108




c


of the network


100


constitute part of a virtual distribution tree over which the video frames are multicast from a sender


102


to the clients


106




a


-


106




e


. The clients


106




a


-


106




e


are leaf nodes of the distribution tree


107


. In addition to the clients, the distribution tree


107


includes internal nodes


104




a


-


104




i


. Each internal node


104




a


-


104




i


receives video frame data from its parents and transmits the video data to each of its children. The nodes also perform other tasks, described below.




The video distribution tree can assume a variety of forms. For example, in an active network, the internal nodes


104




a


-


104




i


can be switches or routers in the network. Alternatively, analogous to the active services approach, the nodes


104




a


-


104




i


can be video gateway or proxy servers that perform special application-specific functions. A recent trend has been the growing deployment of proxies by the ISP's (Internet Service Providers) for caching Web documents. The nodes


104




a


-


104




i


in the video distribution tree can be co-located at some of these Web proxy servers. In a cable network setting, the head-end and mini-fiber nodes in the network could perform distribution tree node tasks.




In general, the root of the distribution tree may be different from the source content server


102


. For example, a proxy or gateway server in a corporate intranet, or a cable head-end may receive streaming video from the remote content server, and simultaneously stream out the video to child nodes on the tree. The distribution tree itself may span multiple network service provider domains. The video service provider can build its distribution tree across multiple domains through cooperative contracts and peering agreements with the various network service providers.




The internetwork is heterogeneous due to different bandwidth capacities, traffic loads, and/or service guarantees that limit capacities for carrying multicast video traffic. The internal nodes


104




a


-


104




i


in the internetwork may also have different amounts of resources, e.g. different buffer sizes for temporal caching of video frames. Similarly, clients


106




a


-


106




f


may have different connectivity to the network, e.g. a client


106


could be connected via a slow modem or a high-speed LAN. A client


106


can be a workstation, PC, hand-held multimedia device or a set-top box connected to a television set, thus possessing varying buffering and computational resources.




The following describes effective multicast smoothing solutions that can efficiently transmit video data over an internetwork to a set of clients, where the internetwork or the clients can be constrained either by buffer size availability or bandwidth capacity availability. The approach described takes advantage of the effectiveness of workahead smoothing in a unicast environment. For prerecorded video such smoothing typically involves computing upper and lower constraints on the amount of data that can be transmitted at any instant, based on prior knowledge of frame sizes and the size of the client playback buffer. The bandwidth smoothing algorithm is then used to construct a transmission schedule that reduces burstiness subject to these constraints.




Generalizing the server to single client bandwidth smoothing solution to multicasting of smoothed streaming video over a distribution tree results in an approach that can reduce burstiness and improve reliability by constructing a global “optimal” set of smoothed transmission schedules, one per link of the distribution tree, depending upon the resource constraint (either buffer or bandwidth). The term “optimal”, as used herein, refers to performance with respect to one or more criteria such as reduction of the peak rates, rate variability, and the effective bandwidth of the video.




I. Differential Caching




The technique described herein uses buffering at the root and intermediate nodes of the distribution tree to smooth the streaming video. Buffer availability at a node allows temporal caching of portions of the video streams. Temporal caching uses a differential caching technique in which different amounts of video data are cached at different nodes for transmission over a heterogeneous internetwork. Caching at the root node allows the root node to smooth an incoming live or stored video stream, and transmit the resultant schedule to a downstream node. The buffers at the internal nodes allow each node to accommodate the difference between the node's incoming and outgoing transmission schedules when the outgoing and incoming link capacities are different. Additionally, by increasing the effective or virtual smoothing buffer size for the upstream node, these buffers provide more aggressive smoothing opportunity along the upstream link. Finally, when the children of a node have different effective buffer capacities, the smoothed transmission schedules are different along the different outgoing links. For a given incoming transmission schedule, the node buffer temporally caches video frames until they are transmitted along each outgoing link. Thus, differential caching allows the transmission schedule to each child to be decoupled to varying degrees depending on the size of the parents buffer cache. This can be extremely beneficial from a network resource requirements point of view. For example, the parent node can smooth more aggressively to a child having a larger effective buffer relative to a child with a smaller buffer. Differential caching allows the more constrained child to be served at its suitable pace, without requiring the frames to be retransmitted from higher up in the tree.




II. Application Aware Multicast




The approach described herein uses application-level information such as video frame sizes and system resource availability to deliver real-time streaming video data to clients for continuous playback. In addition to packet duplication and forwarding, nodes in the distribution tree may also perform transmission schedule computation, differential caching, and real-time streaming of video according to smoothed transmission schedules. Since large parts of the Internet are yet to become native multicast capable, the application-aware approach can be implemented on top of network level multicast primitives, where they exist, and use unicast communication between the nodes in the distribution tree elsewhere. The approach, by integrating smoothing with multicasting can provide far superior performance with respect to reducing the network bandwidth requirements, compared to multicasting the un-smoothed video.




III. Multicast Distribution of Smoothed Video





FIG. 2

shows a directed tree T=(V,E) where V={0, 1, . . . , n} is the set of nodes (or smoothing servers) and E is the set of directed edges within the tree. Consider a node iεV, let p(i) denote the node's parent and s(i) the set of children attached to i, i.e., (p(i),i)εE and s(i)={jεV:(i,j)εE}. Assume node


0


to be the root server in the tree. Let V


1


⊂V be the set of leaves; V


1


={iεV:s(i)=Ø}. Node


0


can be the source server of the video, or, the video may be streaming into the root from a remote source. Also, a leaf can be an end client, or an egress node in the distribution tree. Herein, the terms “root” and “source” are used interchangeably as are “leaf” and “client”.




Associated with node iεV is a buffer capacity 0≦b


i


<∞. Without the loss of generality, assume a discrete time model where one time unit corresponds to the time between successive frames (e.g., 33 msec for 30 frames per second full motion video). Consider an N-frame video stream where the size of frame i is f


i


bits, 1≦i≦N which arrives at node


0


destined for all nodes iεV


1


. Associated with this video stream is an arrival vector A=(A


0


, . . . , A


N


) in which the k-th component corresponds to the cumulative amount of data from the stream which has arrived at node


0


by time k=0,1, . . . , N. It is assumed that A


K


≧A


K−1


, 1≦k≦N. For each node iεV


1


, D=(D


0


, . . . , D


N


) denotes the client playback vector where D


k


represents the cumulative amount of data from this stream that must be removed at leaf i by k=0, 1,2, . . . , N time units since the start of playback. It is assumed that D


0


=0 and D


k


≧D


k−1


, 1≦k≦N. For example, D


k





j=1




k


fj if output from the leaf is the original un-smoothed video. For simplicity, assume here that all clients have the same playback vector. Optimal Multicast Smoothing of Streaming Video Over an Internetwork, Subhabrata Sen, Don Towsley, Zhi-Li Zhang, and Jayanta Dey, UMASS TECHNICAL REPORT 98-77, located at ftp://ftp.cs.umass.edu/pub/techrept/techreport/1998/UM-CS-1998-077.ps, incorporated by reference herein, formulates a more general case where clients have different playback vectors. The case where the root server has an “infinite” buffer, in the sense that b


0


≧D


N


, is referred to herein as the infinite source model, where as the case where b


0


<D


N


is referred to as the finite source model.




Associated with each node iεV\{0} is a schedule S


i


=(S


i,0


, . . . , S


i,N


) in which the k-th component denotes the cumulative amount of data transmitted by node p(i) to node i by time k=0, 1, . . . , N, 1≦i≦n. The notation “\{0}” is used herein to denote “excluding zero”. Note that S


0


≡A, and S


i,k


≧S


i,k−1


, 1≦k≦N. A set of schedules {S


i


}


iεV


is said to be feasible if, simultaneously using the component schedules in the set for video distribution along the tree does not violate any system constraints, and results in lossless, starvation-free playback at each leaf.




In general, the time the root server starts transmitting a video stream (or the time when the video stream begins arriving at the root) may be different from the time a leaf starts playing back the video stream. The difference between these two start points is referred to as startup delay (w). For any given startup delay w≧0, if the time the server starts transmitting a video stream as the reference point (i.e., time


0


), then the playback vector D at the leaf i will be shifted w time units to the right. This shifted playback vector is represented by D(w)=(D


0


(w), D


1


(w), . . . , D


N+w


(w)), where D


i


(w)=0 for 0≦i≦w, and D


i


(w)=D


i−w


for w+1≦i≦N+w. In this case, the root has N+w+1 time units to transmit video stream, namely, S


j


=(S


j,0


, S


j,1


, . . . , S


j,N+w


), jεs(


0


). For ease of notation, the arrival vector A (which starts at time


0


) is extended with w more elements, namely, A=(A


0


, . . . , A


N+w


), where A


i


=A


N


for N+1≦i≦N+w. If time


0


is the instant when the leaf starts the playback, then the arrival vector A is shifted w time units to the left. The corresponding arrival vector is denoted by A(−w)=(A


−w


(w),A


−w+1


(w), . . . , A


N


(w)) where for −w≦i≦N−w, A


i


(w)=A


i+w


, and A


i


(w)=A


N


for N−w+1≦i≦N. Similarly, the root starts transmission at time −w according to schedule S


j


=(S


j,−w


. . . , S


j,0


, . . . , S


j,N


). In the following discussion, the time the root server starts transmission should be assumed to be time


0


unless otherwise stated. Depending on the context, A and D denote either the generic arrival or playback vector, respectively, or the appropriately shifted version.




Buffer Constrained Multicast Smoothing




Given a set of buffer allocations {b


i


}iεV, and consumption vector D at the leaves of the distribution tree, let S(T,{b


i


},A,D) denote the set of feasible schedules. In this context, the set {S


i


}


iεV


is feasible if











max


{



S

p






(
i
)



-

vec






(

b

p






(
i
)



)



,


max

j


g






(
i
)










S
j



}




S
i



min


{




min

j


g






(
i
)










S
j


+

vec






(

b
i

)



,

S

p






(
i
)




}



,

i


V

\


V
l



,




(
1
)









 max{S


p(i)




−vec


(


b




p(i)


),


D}≦S




i


≦min{


D+vec


(


b




i


),


S




p(i)




},iεV




l


,  (2)




where S


0


≡A.




Intuitively, at node i, at any time k, the cumulative incoming data S


i,k


should be sufficient to satisfy the cumulative outgoing data to the child nodes, and should not exceed the cumulative data been received at its parent p(i). Also, S


i


should transmit data fast enough to prevent buffer overflow at p(i), but not fill up the buffer at i so quickly that the transmission to some child {S


j


}


jεs(i)


is unable to transmit data from the buffer in time, before it gets overwritten.




The following inequalities follow:








S




i




≦S




p(i)




≦S




i




+vec


(


b




i


),


iεV\


\{0}.  (3)






Two important considerations in this setting are:




1. Buffer Feasibility: is the buffer allocation feasible, i.e., is the set of feasible schedules S(T,{b


i


},A,D) nonempty? Note that the feasibility question arises as all the buffer allocations are assumed to be finite. Given a particular arrival vector A and playback vector D at the leaves, it is possible that no feasible transmission schedule is possible at one or more edges in the tree, due to buffer overflow at the source or sink of that edge.




2. Optimal smoothed schedules: for a feasible buffer allocation, what is the optimal set of smoothed transmission schedules {S


i


}


iεV?






B. Rate Constrained Multicast Smoothing




Assume that, instead of buffering capacity at the nodes, the link bandwidths in the distribution tree are the constraints. Consider the infinite source model, i.e. b


0


≧D


N


.




Following the notation introduced earlier in this section, the client playback vector at node iεV


1


is D={D


0


, . . . , D


N


}, where D


i


is the cumulative amount of data consumed by a client i. For iεV\{0}, a rate constraint r


i


is associated with the link (p(i),i) which specifies the bandwidth available on the link S


i


=(S


i,0


, S


i,1


, . . . ) denotes a schedule used by node p(i) to transmit data on the link (p(i),i) to node i, where S


i,k


is the cumulative amount of data transmitted by time k. The schedule {S


i


}


iεV


is a feasible schedule if the following conditions are satisfied:











peak






(

S
i

)




r
i


,

i


V

\


{
0
}



,




max

j


g






(
i
)










S
j




S
i




S

p






(
i
)








if





i




V
-

V
l



,




(
4
)









 or


D≦S




i




≦S




p(i)


if


iεV




l


.  (5)




Define S(T,{r


i


},A,D) to be the set of feasible transmission schedule sets for this system. The system determines the following:




1. Minimum Startup Delay: i.e., the minimum (common) playback startup delay w* for the clients for which a feasible transmission schedule exists for this system, i.e., S(T,{r


i


},A,D)≠Ø. Due to the transmission rate constraints, some minimum startup delay may be required to build up sufficient data to guarantee starvation-free playback at each client. Here assume that all the clients start playback at the same time.




2. Optimal buffer allocation: i.e., the minimum buffer allocation b


i


at each node iεV\{0} of the distribution tree, and what is the minimum total buffer allocation b=Σ


iεV\{0}


b


i


among schedules for the system. As explained before, the buffer is used for differential caching. Observe that for a given set of feasible schedules {S


i


}


iεV


, the buffer allocation {b


i


}


iεV\{0}


for the distribution tree is said to be feasible if the following constraints are satisfied: b


i


must be sufficiently large to ensure lossless video transmission, namely, it must be able to accommodate both (a) the maximum difference between the amount of data transmitted from node i to child node l according to a feasible schedule S


1


and that from node i to another child node k according to a schedule S


i


and that transmitted from node i to a child node k according to a schedule S


k


. Formally,











b
i




max

l
,

k


g






(
i
)











{

max


{


S
l

-

S
k


}


}



,




(
6
)







b
i




max

k


g






(
i
)











{

max


{


S
i

-

S
k


}


}

.






(
7
)













IV. Overview of Single Link Smoothing




This section describes the single link smoothing model and some important concepts and results which are used in deriving solutions for the multicast scenario.





FIG. 3

shows a video transmitted across the network via a smoothing server node (which has a b


0


bit buffer) to a smoothing client node (which has a b


1


bit playback buffer). The model for the single link case can be derived from the distribution tree model of

FIG. 2

by setting n=1. Consider next the buffer and rate constrained versions of this problem.




A. Buffer Constrained Single Link Optimal Smoothing




Here the server and client buffers (b


0


and b


1


) are limiting resources. The smoothing problem involves computing transmission schedules which can transmit video from the server to the client in such a way as to reduce the variability of the transmitted video stream, thereby making efficient use of the network bandwidth.




To ensure lossless, continuous playback at the client, the server transmits sufficient data to avoid buffer underflow at the client without overflow in the server buffer. This imposes a lower constraint, L


t


=max {D


t


(w), A


t


−b


0


}, on the cumulative amount of data that the server can transmit by any time t, 0≦t≦N+w. In order to prevent overflow of the client playback buffer, the cumulative amount of data received by the client by time t cannot exceed D


t−1


(w)+b


1


. This leads to the following upper constraint U


t


=min (D


t−1


(w)+b


1


,A


t


}, 0≦t≦N+w. Given these lower (buffer underflow) and upper (buffer overflow) constraint vectors L=(L


0


, . . . , L


N+w


) and U=(U


0


, . . . , U


N+w


), a transmission schedule S=(S


0


, S


1


, . . . S


N+w


) is said to be feasible with respect to L and U if S


0


=L


0


, S


N+w


=L


N+w


, and S neither underflows nor overflows the server or client buffer, i.e. L≦S≦U.





FIG. 4

shows an example of a feasible schedule S that does not violate the lower or upper constraints. In general, for a given pair of constraint vectors (L,U) such that L≦U, multiple feasible transmission schedules S may exist. Among all feasible schedules, the system finds a smooth schedule that minimizes network utilization according to some performance metrics. In J. D. Salehi, Z. -L. Zhang, J. F. Kurose, and D. Towsley, “Supporting stored video: Reducing rate variability and end-to-end resource requirements through optimal smoothing,”


IEEE/ACM Trans. Networking


, vol. 6, pp. 397-410, August 1998, a measure of smoothness based on the theory of majorization is described, and the resulting smoothed schedule minimizes a wider range of bandwidth metrics such as the peak rate, the variability of the transmission rates as well as the empirical effective bandwidth. Henceforth, this “optimal schedule” (linear time construction) is referred to as the majorization schedule. A majorization operation is denoted by the symbol (<).




Given a majorization schedule S*, we say k, 1≦k≦N+w−1, is a change point of the schedule, if the transmission rate changes at time k. Moreover, k is said to be a convex change point if the transmission rate increases at k. Similarly, k is said to be a concave change point if the transmission rate decreases at k. A feature of the majorization schedule is that the convex (concave) change points occur precisely at the times k such that S


k




*


=U


k


(S


k




*


=L


k


). This property leads to the following Lemma which is critical to constructing optimal smoothing schedules for the distribution tree scenario.




Lemma 1: Let L


1


≦U


1


and L


2


≦U


2


such that L


1


≦L


2


and U


1


≦U


2


. Then S


1




*


≦S


2




*


. Moreover, if L


1


=L


2


, then any concave or convex change point of S


1


is also a concave or convex point of S


2


.




Rate Constrained Single Link Smoothing




Next consider the dual problem of the rate constrained single link smoothing, where the bandwidth of the link connecting the server and the client is constrained by a given rate r. Due to constraints, the buffer at the client should be sufficiently large to ensure continuous video playback at the client. Furthermore, it may be necessary for the server to start video transmission sufficiently early. In this context, a server transmission schedule is feasible if the transmission rate of the schedule does not exceed the rate constraints (as well as the arrival vector) and the amount of data needed for the client playback is satisfied.




Assume that the client playback vector starts at time


0


, and as before, the cumulative client playback vector is denoted by D(


0


)=(D


0


(


0


),D


1


(


0


), . . . , D


N


(


0


)). Then, given a startup delay w, assume that the arrival vector starts at time −w instead of time


0


, with the shifted arrival vector A(−w). For the infinite source model, a feasible transmission schedule S


late


can be constructed that transmits data as late as possible, while obeying rate constraint r. We refer to this as the lazy schedule.

FIG. 5

illustrates an example of a lazy schedule.




Define b*(r) to be the minimum buffer required at the client for the transmission schedule S


late


without incurring loss of data, and w*(r,A), the minimum startup delay with respect to which S


late


conforms to the arrival vector, i.e. S


k




late


≦A


K


(−w*), for k≧w*. Then












b
*







(
r
)


=


max

k

0




{


S
k
late

-


D
k







(
0
)



}



,




(
8
)












w


*(


r,A


)=min{


w


≧0


:A




k


(−


w


)−


S




k




late


≧0


,−w≧k≧N}.


  (9)




It can be shown that b* is the minimal buffer requirement and w* is the minimal start-up delay among all feasible schedules with rate constraint r and arrival vector A. Also, b*(r) and w*(r,A) are non-increasing functions of the rate constraint r.




The next theorem relates the rate constrained optimal smoothing problem to its dual buffer constrained optimal smoothing problem. For ease of exposition, time


0


is chosen as the time that A


0


arrives at the server and the server starts video transmission. Corresponding to any playback startup delay w≧w*, the client playback vector is D(w). The new lazy schedule S


late


is then the original S


late


shifted w time units to the right. By the definition of w*, we have w*=min{w>0:A


K


−S


k




late


≧0, 0≦k≦N+w}.




Theorem 1: Let S(r,A,D) be the set of all feasible schedules with respect to rate constraint r and the arrival vector A. Let b*(r) and w*(r,A) be the minimum client buffer requirement and startup delay of S(r,A,D). For any w≧w*(r,A) define L=D(w) and U=min{A,D(w)+vec(b*(r))}. Then the majorization schedule S* with respect to the buffer constraints L, U is also a feasible schedule with respect to the rate constraint r, and peak(S*)=peak(S


late


). In particular, if peak (S


late


)=r, then peak(S*)=r.




As a consequence of theorem 1, note that for w=w*, the majorization schedule S* is majorized (thus “smoothest” under the measure of majorization) by any feasible schedule in S(r,A,D) which has the same client buffer requirement b*(r) and startup delay w*(r,A).




V. Optimal Multicast Smoothing




This section presents solutions to the multicast problems described above. This section uses the results for the single link smoothing problem to build a solution for the multicast problem. A key aspect of the approach involves computing upper and lower constraint curves and exploiting the properties of majorization and lazy transmission schedules.




A. Buffer Constrained Multicast Smoothing




Analogous to the single link case, first compute upper and lower constraint curves at the individual nodes in the distribution tree. Unlike the single link case, the constraints at a node can be affected by both the constraint at its directly connected nodes as well as at remote nodes.




B. Determining an Upper Constraint




For iεV, define a vector U


i




b


recursively as follows.










U
i
b

=

{





D
+

vec






(

b
i

)






for





i




V
l










min

j


g






(
i
)






U
j
b


+

vec






(

b
i

)






for





i




V

\


V
l











(
10
)













U


i




b


can be the buffer overflow (or upper) constraint vector for node i when it is fed by the source of the pre-recorded video, i.e., A


K


=D


N


, k=0, . . . , N. As described below, that a buffer constraint somewhere up stream in the tree can impose a more stringent constraint than U


i




b


on the transmission schedules for the subtree rooted at node i.




For iεV, let P(i) denote the set of nodes on the path from the root, 0, to node i and define










U
i
c

=


min

j


P






(
i
)











U
j
b

.






(
11
)













where U


i




c


is referred to as the effective buffer overflow constraint vector of the subtree rooted at i. Observe that U


i




c


≦U


i




b


. The effective overflow vectors exhibit the following properties.







U




i




c




≦U




p(i)




c




≦U




i




c




+vec


(


b




p(i)


).  (12)




C. Determining a Lower Constraint:




Also associated with node i is an effective buffer underflow constraint vector, L


i




c


defined by the following recurrence relation:










L
i
c

=

{




D
,

i
=
0

,







max






(

D
,


L

p






(
i
)


c

-

vec






(

b

p






(
i
)



)




)


,

i

0.










(
13
)













Now consider a single link system with an arrival vector A in which the source has a buffer capacity of size G


p(i)


≡Σ


jεP(p(i))


b


j


and the receiver has buffer overflow and underflow constraints U


i




c


and L


i




c


. Let S


i




*


denote the majorization schedule for this system. The following lemma shows that the schedules {S


i




*


}


i=1, . . . , n


are feasible transmission schedules for the buffer constrained multicast scenario. This result hinges on a key property of majorization schedules (Lemma 1).








Lemma






(
2
)



:






The





schedule






S
i
*






satisfies





the





following






constraints
.




max



{



S

p






(
i
)


*

-

vec






(

b

p






(
i
)



)



,


max

j


g






(
i
)










S
j
*



}




S
i
*



min


{




min

j


g






(
i
)










S
j
*


+

vec






(

b
i

)



,

S

p






(
i
)


*


}



,

i
=
1

,








n
.












Now the following result regarding whether a feasible set of transmission schedules exists for a given buffer allocation {b


i


}


iεV


and leaf consumption vector D:




Theorem 2: S(T,{b


i


},A,D)⇄≠∅⇄∀iεV


l




*


(L


i




c


≦U


i




c


).




The following theorem regarding the optimality of {S


i




c


}


i=1




n


answers the second question raised earlier.




Theorem 3:




The majorization schedules {S


i




*


}


i =1




n


associated with the finite source single link problems with arrival vector A, source buffers {G


i


}, and buffer overflow and underflow vectors {U


i




c


}


iεV


and {L


i




c


}


iεV


satisfy the following relations.








S




i




*




<S




i




,∀{S




i




εS


(


T,{b




i




},A,D


).






D. Buffer Feasibility Check





FIG. 6

shows a procedure, “check feasibility”, for checking if the buffer allocation is feasible. The procedure returns true if S(T,(b


i


},A,D)≠Ø, and otherwise returns False. Each traversal moves either up the tree starting at the leaves (upward traversal) or down the tree starting at the root node


0


(downward traversal), processing all nodes at the same level before going to the next level.




An upward traversal uses relation (10). A downward traversal uses (11) to compute U


i




c


, and (13) to compute L


i




c


. Theorem 2 is then used to check for feasability.




Given that U


i




c


, and L


i




c


can be computed in O(N) time the complexity of the above algorithm is O(nN).




E. Multicast Smoothing Procedure





FIG. 7

shows a simple procedure for computing the smoothed schedules, “compute smooth”, for the multicast tree, given a feasible buffer allocation to the nodes in the tree. This involves traversing the distribution tree three times. The procedure proceeds as “check feasibility” proceeded, with the additional computation of G


p(i)


using G


p(i)


=G


p(p(i))


+b


p(i)


. However, “compute smooth” computes S


i




*


, the majorization schedule associated with the lower and upper constraints L


i


=max(A−vec(G


p(i)


), L


i




c


) and U


i


=min(A,U


i




c


).




By theorem 3, the set {S


i




*


}


i=1




n


is optimal. Given that S


i




*


can be computed in O(N) time, the complexity of the above algorithm is O(nN). Note that differential caching at intermediate node buffers can enable this smoothing, by temporarily caching differences between, and thereby temporally decoupling to some extent, the transmissions between faster and slower sections of the distribution tree.




VI. Rate Constrained Optimal Multicast Smoothing




A. Determining the Minimum Startup Delay




For each iεV\{0}, consider a rate constrained single link problem with the rate constraint r


i


, the arrival vector A and the client playback vector D. Let b


i




*


and w


i




*


be the minimum buffer allocation and startup delay required for the single link problem. Then, the minimum common startup delay for the clients is given by w*=max


kεV\{0}


w


k




*


.




Given this minimum startup delay w* and assuming that the root server starts video transmission at time


0


, the playback vector at client iεV


l


is then D(w*)




B. Determining Buffer Allocation




The optimal buffer allocation problem mentioned above can be addressed. For iεV\{0}, define the effective buffer requirement b recursively as follows:










b
i
e

=

{





b
i
*

,





i


V
l


,







max


{


b
i
*

,


max

k


g






(
i
)










b
k
e



}


,

i


V

\



V
l

.












(
14
)













Clearly, b


i




*


≦b


i




e


≦b


p(i)




e


. Note that b


i




e


is the largest buffer allocated to any node in the subtree rooted at node i. As will be shown, b


i




e


is a the minimal buffer allocation required for the subtree rooted at node i such that a set of feasible schedule exists for the nodes in the tree.




Now for iεV\{0}, define











b
^

i

=

{





b
i
e

,





i


V
l


,








b
i
e

-


min

k


g






(
i
)










b
k
e



,

i


V

\



V
l

.












(
15
)













Given this set of buffer allocations {{circumflex over (b)}}


iεV\{0}


, define the effective buffer underflow vector at node i, L


i




b


, as L


i




b


=D(w*), and the effective buffer overflow vector at node i, U


i




b


, as










U
i
b

=

{






D






(

w
*

)


+

vec






(

b
i
e

)



,





i


V
l


,









min

k


s






(
i
)










U
k
b


+

vec






(


b
^

i

)



,

i


V

\



V
l

.












(
16
)













Then




Lemma 3: The effective overflow vector has the following property:








U




i




b




≦U




p(i)




b




≦U




i




b




vec


(


{circumflex over (b)}




p(i)


),  (17)








U


i




b




=D


(


w


*)+


vec


(


b




i




e


).  (18)






For iεV\{0}, let S


1




*


be the majorization schedule with respect to the lower and upper constraint vectors (L


i




b


,min{A,U


1




b


}). As in the case of the single link problem, the set of these majorization schedules, {S


i




*


, iεV} (where S


0




*


≠A), is a set of feasible schedules for the rate constrained multicast smoothing system. Namely:




Theorem 4: the schedule S


i




*


, iεV\{0}, satisfies the following constraints:











max


{



S

p






(
i
)


*

-

vec






(


b
^


p






(
i
)



)



,


max

j


g






(
i
)










S
j
*



}




S
i
*



min


{




min

j


g






(
i
)










S
j
*


+

vec






(


b
^

i

)



,

S

p






(
i
)


*


}



,




(
19
)












peak


(


S




i




*


)≦


r




i


.  (20)




Where it is understood that max


jεs(i)


S


j




*


≡D for iεV


l


and S


0




*


=A.




As a consequence of theorem 4, under the same buffer allocation {{circumflex over (b)}


i


}


iεV\{0}


and startup delay w*, the set of majorization schedules {S


i




*


}


iεV


gives us the set of the “smoothest” schedules among all feasible schedules for the rate constrained multicast smoothing problem.




The next theorem establishes the optimality of the buffer allocation {{circumflex over (b)}


i


}


iεV\{0}


.




Theorem 5: the buffer allocation {{circumflex over (b)}


i


}


iεV\{0}


is optimal in the sense that the allocation minimizes, among all the feasible schedules for the system, both the total buffer allocation, Σ


iεV\{0}


{{circumflex over (b)}


i


}, and the buffer allocated for any node in the subtree rooted at node i (namely, the effective buffer allocation b


i




e


at node i), iεV\{0}. As a result, any smaller total buffer allocation will not result in a feasible set of transmission schedules for the system.





FIG. 8

shows a simple procedure, “allocate buffer”, that computes the optimal buffer allocation for a given distribution tree. The procedure involves three traversals through the distribution tree, processing all the nodes at the same level before proceeding to the next one.




The procedure determines b


i




*


and w


i




*


using relations (8) and (9) respectively and determines the common minimum startup delay w*=max iεV\{0} w


i




*


. At node i, b


i




e


is determined using relation (14), and then {circumflex over (b)} is determined late using (15). Since S


i




late


can be computed in time O(N), it is clear that the computation complexity of the above algorithm is O(nN).




Once the optimal buffer allocation is obtained, “compute smooth” can compute the set of optimally smoothed schedules for the multicast tree.




VII. Implementations





FIG. 9

shows a candidate process


900


for implementing the techniques described above. As shown, a selected node, such as a root node, can receive information describing a video (e.g., the frame sizes of a video), the network topology, rate constraints, and determined (e.g., as determined by the buffer allocation procedure of

FIG. 8

) or received buffer allocations. Based on the received and/or determined information, the process


900


can determine


902


smoothed transmission schedules for each node in the network topology, for example, by using the “compute smooth” procedure shown in FIG.


7


. That is, the process


900


can determine a transmission schedule from a node to each of its children nodes. Thereafter, the selected node can send


904


each node the schedules determined for the node. Additionally, the process


900


can distribute any determined buffer allocations to different nodes. This can enable the other nodes to devote excess buffer resources to other uses. Finally, the streaming video begins


906


and each node caches and distributes video data in accordance with the node's transmission schedule.





FIG. 9

is merely one of many possible different processes that take advantage of the techniques described herein. For example, each node may independently receive the information needed to determine its own transmission schedule. Additionally, there is no requirement that the computing device used to determine the schedule be a source node or a particular intermediate node.





FIG. 10

is a flowchart of a process


1000


performed by nodes after receiving or determining a smoothed transmission schedule


1002


. As shown, the node receives


1004


video information in accordance with its parent's transmission schedule and determines


1006


whether to cache and/or transmit the received information to one or more children nodes (or for display) based on its own transmission schedule.




The techniques described here are not limited to any particular hardware or software configuration; they may find applicability in any computing or processing environment. The techniques may be implemented in hardware or software, or a combination of the two. Preferably, the techniques are implemented in computer programs executing on programmable computers that each include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and one or more output devices. Program code is applied to data entered using the input device to perform the functions described and to generate output information. The output information is applied to one or more output devices.




Each program is preferably implemented in a high level procedural or object oriented programming language to communicate with a computer system. However, the programs can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language.




Each such computer program is preferable stored on a storage medium or device (e.g., CD-ROM, hard disk or magnetic diskette) that is readable by a general or special purpose programmable computer for configuring and operating the computer when the storage medium or device is read by the computer to perform the procedures described in this document. The system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer to operate in a specific and predefined manner.




Please incorporate “Optimal Multicast Smoothing of Streaming Video over an Internetwork”, Subhabrata Sen, Don Towsley, Zhi-Li Zhang, and Jayanta Dey, Proc. IEEE Infocom '98, (New York, N.Y., March 1999) by reference, in its entirety, herein.




Other Embodiments




It is to be understood that while the invention has been described in conjunction with the detailed description thereof, the foregoing description is intended to illustrate and not limit the scope of the invention, which is defined by the scope of the appended claims. Other aspects, advantages, and modifications are within the scope of the following claims.



Claims
  • 1. A method of multicasting video to multiple client nodes via intermediate nodes, the method comprising:accessing video information descriptive of the video to be multicast; accessing information describing a distribution tree of nodes from a source of the video to the multiple clients nodes via one or more internetwork nodes; accessing rate constraints of nodes in the distribution tree; accessing buffer allocations of the nodes in the distribution tree; and determining one or more smoothed transmission schedules for each node in the distribution tree based on the accessed video information, the accessed information describing the distribution tree, the accessed rate constraints of nodes in the distribution tree, and the accessed buffer allocations of the nodes in the distribution tree, the one or more transmission schedules describing the transmission of video data to one or more children nodes.
  • 2. The method of claim 1, wherein accessing buffer allocations comprises determining buffer allocations based on the accessed video information, the accessed information descriptive of a distribution tree, and the accessed rate constraints.
  • 3. The method of claim 1, wherein accessing rate constraints comprises receiving rate constraints.
  • 4. The method of claim 1, wherein determining comprises:determining an upper constraint for each node; determining a lower constraint for each node; and determining a smoothing schedule based on the determined upper and lower constraints.
  • 5. The method of claim 4, wherein determining a smoothing schedule comprises determining a majorization schedule.
  • 6. The method of claim 1, further comprising distributing the determined transmission schedules to nodes in the distribution tree.
  • 7. The method of claim 1, wherein the determined transmission schedules comprise schedules that cause at least one node to differentially cache data for distribution to the at least one node's children.
  • 8. A computer program product, disposed on a computer readable medium, for multicasting video to multiple client nodes via intermediate nodes, the program including instructions for causing a processor to:access video information descriptive of the video to be multicast; access information describing a distribution tree of nodes from a source of the video to the multiple clients nodes via one or more internetwork nodes; access rate constraints of nodes in the distribution tree; access buffer allocations of the nodes in the distribution tree; and determine one or more smoothed transmission schedules for each node in the distribution tree based on the accessed video information, the accessed information describing the distribution tree, the accessed rate constraints of nodes in the distribution tree, the accessed buffer allocations of nodes in the distribution tree, and the one or more transmission schedules describing the transmission of video data to one or more children nodes.
  • 9. The computer program of claim 8, wherein the instructions that access buffer allocations comprise instructions that determine buffer allocations based on the accessed video information, the accessed information descriptive of a distribution tree, and the accessed rate constraints.
  • 10. The computer program of claim 8, wherein the instructions accessing rate constraints comprise instructions that receive rate constraints.
  • 11. The computer program of claim 8, wherein the instructions that determine comprises instructions thatdetermine an upper constraint for each node; determine a lower constraint for each node; and determine a smoothing schedule based on the determined upper and lower constraints.
  • 12. The computer program of claim 8, wherein the instructions that determine a smoothing schedule comprise instructions that determine a majorization schedule.
  • 13. The computer program of claim 8, further comprising instructions that distribute the determined transmission schedules to nodes in the distribution tree.
  • 14. The computer program of claim 8, wherein the determined transmission schedules comprise schedules that cause at least one node to differentially cache datal for distribution to the at least one node's children.
CROSS REFERENCE TO RELATED APPLICATIONS

This application claims priority from U.S. Provisional Application Serial No. 60/125,240, filed Mar. 19, 1999.

STATEMENT AS TO FEDERALLY SPONSORED RESEARCH

This invention was made with Government support under National Science Foundation grants NCR-9523807, NCR-9508274, and CDA-9502639. The Government may have certain rights in the invention.

US Referenced Citations (3)
Number Name Date Kind
5940391 Malkin et al. Aug 1999 A
6320865 Agrawala et al. Nov 2001 B1
6377972 Guo et al. Apr 2002 B1
Non-Patent Literature Citations (10)
Entry
Almeroth et al., “The Use of Multicast Delivery to Provide a Scalable and Interactive Video-on-Demand Service,” IEEE Journal of Selected Areas in Communication, vol. 14, No. 6, p. 1110-1122 (1996).
Deering et al., “Multicast Routing in Datagram Internetworks and Extended LANs”, ACM Transactions on Computer Systems, vol. 8, No. 2, p. 85-110 (1990).
Feng et al., “A Comparison of Bandwidth Smoothing Techniques for the Transmission of Prerecorded Compressed Video,” Proc. IEEE Infocom '97 Conf. Comp. Comm., vol. 1, p. 58-66 (1997).
Garrett et al., “Analysis, Modeling and Generation of Self-Similar VBR Video Traffic,” ACM, p. 269-280, (Sep. 1994).
Lakshman et al. “Variable Bit-Rate (VBR) Video Tradeoffs and Potentials,” proc. IEEE, vol. 86, p. 952-973 (May 1998).
Rao et al., “Fast Techniques for the Optimal Smoothing of Stored Video,” Multimedia Systems, vol. 7, p. 222-233 (1999).
Rexford et al., “Online Smoothing of Live, Variable-Bit-Rate Video,” Proceedings of Workshop on Network and Operating System Support . . . , IEEE, p. 235-243 (1997).
Rexford et al., “Smoothing Variable-Bit-Rate Video in an Internetwork,” IEEE/ACM Transations on Networking, vol. 7, No. 2, p. 202-215 (Apr. 1999).
Salehi et al., “Supporting Stored Video: Reducing Rate Variability and End-to-End Resource Requirements Through Optimal Smoothing,” IEEE/ACM Transactions on Networking, vol. 6, No. 4, p. 397-410 (1998).
Sen et al. “Optimal Multicast Smoothing of Streaming Video over an Internetwork,” Proceedings IEEE Infocom '99 Conf. On Comp. Comm., vol. 2, p. 455-463 (1999).
Provisional Applications (1)
Number Date Country
60/125240 Mar 1999 US