Support for Interactive Playback Devices for Performance Aware Peer-to-Peer Content-on Demand Service

Abstract
A method and apparatus are described comprising detecting a playback operation, determining a target playback point for the playback operation and detecting a sub-clip type into which the target playback point falls. The method and apparatus also receive a streamed leading sub-clip, determine a set of needed sub-clips, locate one of the set of needed sub-clips and download the located sub-clip.
Description
FIELD OF THE INVENTION

The present invention relates to support for user interactivity/operations of a video playback device for a performance aware peer-to-peer video-on-demand service.


BACKGROUND OF THE INVENTION

Traditionally, the client-server service model has been used to provide streaming service. A client sends a request to a server, which then streams the content to the client if the server has enough resources to serve the client's request and there is enough bandwidth along the path between the server and the client.


Due to the limited computation and storage resource at the server and limited bandwidth in the network connecting the server and clients, scalability has been an issue with client-server streaming service. Recently, peer-to-peer techniques have been introduced into streaming service. Peers are implemented with the capabilities of clients and servers and contribute to alleviate the workload imposed on the server and distribute the bandwidth requirements across the network by actively caching the content and serving other peers. Studies have shown that peer-to-peer techniques greatly improve system scalability, enabling the system to serve many much more users.


There have been significant efforts to address the scalability issue presented in streaming media service using peer-to-peer networking. These efforts can be classified into two categories notably peer-to-peer live streaming and peer-to-peer stored video streaming or video-on-demand. While both services strive to support a large number of users while offering users good viewing quality, they also face different technical challenges. In peer-to-peer live streaming, minimizing the start-up delay without sacrificing the system scalability is the challenge. In peer-to-peer video-on-demand service, allowing asynchronous users to share is the challenge.


Peer-to-peer streaming schemes also distinguish themselves by the different data dissemination techniques. Two data dissemination methods have been investigated—notably the overlay-based approach and the data-driven approach. In the overlay-based approach, the peers form a mesh or tree structure where parent-child relationships are formed among the peers. A child peer receives data from its parent. In contrast, the peers in the data-driven approach do not have fixed parent-child relationships. The peers look for the missing data, and retrieve the missing data wherever available. While the overlay-based approach is widely used in early peer-to-peer efforts, the data-driven approach is becoming more popular since it addresses the churn and asymmetric bandwidth problem effectively.


While most of the prior art efforts exhibit good scalability and support a greater number of users compared to a traditional client-server service model, the prior art schemes are best-effort in nature and the support of system performance requirements has not been fully investigated. Using a performance aware peer-to-peer video-on-demand service, also means that operations involving user interactivity with a video playback device must be handled differently. To date the prior art has not addressed the handling of such video playback device operations.


SUMMARY OF THE INVENTION

Performance aware peer-to-peer video-on-demand service allows users to select and watch video content over a network whenever they want. In a related application, segmented peer-to-peer video sharing was described. This enabled content sharing in a video-on-demand environment. Performance was addressed by peer-to-peer data downloading and server assisted complementary streaming.


The present invention is directed towards supporting user interactivity/operations for video playback devices for a performance aware peer-to-peer video-on-demand service. Such operations include jump forward/backward, pause/resume, fast forward, and fast reverse.


A method and apparatus for supporting video playback operations for a peer-to-peer video on demand service are described comprising detecting a video playback operation and detecting a sub-clip type. The method and apparatus also receive a streamed leading video sub-clip, determine a set of needed video sub-clips, locate one of the set of needed video sub-clips and downloading the located video sub-clip.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention is best understood from the following detailed description when read in conjunction with the accompanying drawings. The drawings include the following figures briefly described below where like-numbers on the figures represent similar elements:



FIG. 1 is a schematic diagram of a peer-to-peer network in accordance with the present invention.



FIG. 2 is a flowchart of the process to calculate the actual number of streamed sub-clips.



FIG. 3 is a schematic diagram of a video playback device.



FIG. 4 is a flowchart depicting the admission control process from the user/client side.



FIG. 5 is a block diagram of the video playback operations within a video playback device.



FIG. 6 is a flowchart of the operation of an exemplary control module in accordance with the present invention.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Users of video-on-demand service watch different portions of video at any given moment. In order to enable the content sharing among users and maximize the amount of content that is delivered through a peer-to-peer network, it is assumed that each user has the storage capacity to cache a partial copy and/or the entire copy of content that has been played. This is a reasonable assumption given the rapidly increasing storage capacity of video playback devices. It should be noted that a video playback device is any device capable of receiving and playing back video (stored or live) including but not limited to computers, laptops, personal digital assistants (PDAs) and mobile devices. A peer-to-peer network is not limited to a wired line network and may be a wireless or wired line network or a hybrid network employing both wired line and wireless connections.


In the segmented peer-to-peer video-on-demand method and apparatus of the present invention, a video clip is divided into multiple equal length segments, denominated sub-clips. The playback time of the start of the sub-clip is defined as the deadline of this sub-clip. The leading sub-clips are streamed to the video playback device so that the users can start the playback immediately. Meanwhile, a peer-to-peer network is established among users in order to pre-fetch the data of the succeeding sub-clips. In accordance with the system performance aware scheme of the present invention, the data of a sub-clip has to be pre-fetched before its deadline. Once the playback of a sub-clip has started, no peer-to-peer downloading of that sub-clip is allowed since the newly downloaded data may be outdated. Complementary streaming from the original server is initiated from this point on for better system performance. Complementary streaming is described below.


An example is used to illustrate how segmented peer-to-peer video-on-demand serves incoming requests. In this example, it is assumed that users are able to cache the entire copy of the video. The same technique applies even if only a portion of the video copy is cached. It is further assumed that the server only streams the first sub-clip and the data of following sub-clips are downloaded using the peer-to-peer network. The algorithm to compute the number of streamed sub-clips will be presented and described below.


Referring now to FIG. 1, client/user 1 makes a request/demand for video from the server at time t1. The server streams sub-clip 1 (the first video sub-clip) to client 1 immediately, so that client 1 can promptly commence playback. At the same time an attempt is made to locate a peer having/caching sub-clip 2 within the peer-to-peer network. At this moment, the only peer in the peer-to-peer network having/caching sub-clip 2 is the server, which can behave as a peer. Both client 1 and the server (at the least) are members of the peer-to-peer network. At time t2, client 1 is playing back sub-clip 1, while sub-clip 2 is being downloaded (not streamed) from the server. Client/user 2 makes a request/demand for the same video from the server and immediately commences playback of sub-clip 1, which was streamed from the server to client 2. Both the server and client 1 commence downloading (not streaming) sub-clip 2 to client 2. At this moment, the server, client 1, and client 2 are peers in the peer-to-peer network. At time t3, client 3 makes a request/demand for the same video from the server and immediately commences playback of sub-clip 1, which was streamed from the server. By now client 1 is playing back sub-clip 3 and is downloading the data/video of sub-clip 4. Client 2 is playing sub-clip 2 and downloading sub-clip 3. At this time, the server, client 1, client 2 and client 3 (at least) are members of the peer-to-peer network. Client 3 can download sub-clip 2 from the server, client 1, and client 2. As time progresses, peer video playback continues. The peer-to-peer downloading pre-fetches the data of the sub-clip that follows the sub-clip currently being played, as show in FIG. 1 with current time of t4. By time t5, client 1 has already finished its video playback and exited the system. Client 2 is playing back the last sub-clip and client 3 is playing back sub-clip 4 and downloading sub-clip 5. The server and the second client are peers in the peer-to-peer network for purposes of downloading sub-clip 5. At last, client 2 has also exited the system after finishing the video playback. Client 3 is watching/playing back sub-clip 5 and will exit the system at the end of sub-clip 5.


As discussed above, although extra care is taken to address the performance issues (timely arrival of the sub-clips at/by the user), some data may still be missing by the time of deadline (or shortly before the deadline) when peer-to-peer downloading ceases. How to use the server to stream the missing data so as to further improve the peer video playback performance is now described. This is called complementary streaming herein. As the deadline approaches, the peer client prepares a missing data vector Vmissing, which is a bit map that uses a first flag, for example “1” to indicate that a block is received, and a second flag, for example “0” to indicate a block is still missing. The missing data vector is sent to the server (signaling). The server starts to stream out the missing data at the normal playback rate as the deadline approaches so that the missing data/video can be filled in time for the peer video playback.


Supporting user interactivity, i.e., video device playback operations, is an important aspect of video-on-demand service. Users invoking jump forward/backward operations want to playback the video from an arbitrary point inside the clip at the normal playback rate. Letting the target playback point, or TPP denote the intended new playback point. If TPP is later than the current playback point, it is a jump forward operation. Otherwise, it is a jump backward operation.


As shown above in FIG. 1, there exist four types of sub-clips in performance aware peer-to-peer video-on-demand service: empty sub-clips, downloaded sub-clips, downloading in-process sub-clips, and streamed sub-clips. Depending on the type of sub-clip the target playback point (TPP) falls into, the jump forward/backward operations are handled differently. In the following, how to compute the maximum number of sub-clips that may need to be streamed from the server directly is described first. The actual number of streamed sub-clips, based on the type of sub-clip that TPP falls into and the types of following sub-clips are also described. Finally, how to modify the sub-clip deadlines in order to support jump forward/backward operations is described.


The number of streamed sub-clips is maximized when no data after the target playback point (TPP) is cached in the buffer. In such scenario, supporting jump forward/backward operations is similar to starting a new video from the TPP. Suppose the TPP falls into sub-clip i. Below the method to compute the maximum number of streamed sub-clips is described, assuming no data is cached for the sub-clip k, for k>i.


The time interval from the TPP to the end of sub-clip i is denoted by tleftover, and the maximum number of streamed sub-clips is denoted by n*. Thus, the relationship





(rdownlink−rplayback)(tleftover+n*L)≧rplaybackL


where rdownlink is the downlink speed, rplayback is the playback rate, and L is the duration of the sub-clip is derived. Since n* has to be an integer, and has to be no greater than the total number of sub-clips after the sub-clip i,










n
*

=

min


{






(






r
playback


L

-







(


r
downlink

-

r
playback


)



t

leftover










)

+



(


r
downlink

-

r
playback


)


L




,

N
-
i


}






(

Equation





1

)







where (.)+ is a non-negative function, and N is the total number of sub-clips.


Next the actual number of sub-clips needed to be streamed is investigated. There are four types of sub-clips in performance aware peer-to-peer video-on-demand service: empty sub-clip, downloaded sub-clip, streamed sub-clip, and downloading-in-process sub-clip. Both streamed sub-clips and downloading-in-process sub-clips cache portions of the sub-clip data so are treated herein the same as empty sub-clips. This simplifies the problem and a conservative number of streamed sub-clips is, thus, computed. Applying complementary streaming if a portion of the data is available in the buffer is recommended.



FIG. 2 is a flowchart of the process to calculate the actual number of streamed sub-clips for fast forward, jump forward and jump backward operations. The return value is initialized at 205. The current sub-clip counter is initialized to the next sub-clip at 210. A determination is made at 215 if the current sub-clip is fully cached. If the current sub-clip is fully cached then return with the return value at 220. If the current sub-clip is not fully cached then increment the return value at 225 and then increment the current sub-clip counter at 230. The process essentially loops through the n sub-clips. If the sub-clip k is fully cached, the entire downlink bandwidth is available to download the data of the next empty sub-clip. Since downlink speed is always larger than the playback rate (otherwise no video-on-demand service can be provided in the first place), the empty sub-clip will be fully downloaded before its deadline. Therefore, any empty sub-clips after the sub-clip k can be downloaded before their deadlines, and no direct streaming is required.


Next, how to adjust the sub-clip deadlines to accommodate jump forward/backward operations is considered. Let t denote the current time, and dk denote the new deadline of k-th sub-clip. Then, the following equation is derived







d

k
=t+t
leftover+(k−i−1)L  (Equation 2)


for i≦k≦N, where L is the duration of a sub-clip.


Supporting the pause operation in performance aware peer-to-peer video-on-demand service is straightforward—just stop the video playback and pause any ongoing streaming or downloading. The sub-clip deadlines are changed to be infinite to postpone the downloading process. Note that the pause operation only occurs in streamed sub-clips and downloaded sub-clips.


The resume operation is handled differently depending upon the type of sub-clip. Suppose that the video is paused at sub-clip i. If the current sub-clip is a streamed sub-clip, the user via the video playback device signals the server to resume the playback. The server also resumes streaming the sub-clips that had been designated to be streamed before the pause operation was invoked. The rationale behind this is that all the actions are temporarily suspended by the pause operation, and the resume operation re-starts the streaming without changing the system status. Meanwhile, the sub-clip deadlines are modified to reflect the elapsed time incurred by the pause operation. Suppose that the playback time from the resumption point to the beginning of the following sub-clip i+1 is tleftover, and the current time is t, equation (2) is used to compute the new deadlines.


If the pausing point falls in a downloaded sub-clip, the user via the video playback device resumes the playing of current sub-clip immediately, resumes the downloading process, and the deadlines of following sub-clips are modified using equation (2) as well.


For the fast forward operation, the video is played back at a higher rate than the normal playback rate. The speedup factor of the fast forward operation is denoted by δ. The playback rate for the fast forward operation is δ·rplayback.


In the following, the maximum number of streamed sub-clips, n* is calculated first. Then the actual number of streamed sub-clips is calculated. Finally, modification of the sub-clip deadlines is performed. It will become obvious that there is a strong correlation between jump forward operation and fast forward operation.


The time interval from the starting playback point to the end of current sub-clip is denoted by tleftover. The playback rate for the fast forward operation is denoted as r′playback, i.e., r′playback=δ·rplayback. Thus, the following relationship is derived





(rdownlink−rplayback)(tleftover−n*L)/δ≧rplaybackL  (Equation 3)


It should be noted that the second term on the left hand side of equation (3) is divided by δ to reflect the fact that the playback rate is sped up by a factor of δ. Since n* has to be an integer and no larger than N−i, the following relationship can be derived










n
*

=

min


{






(






r
playback


L





δ

-







(


r
downlink

-

r
playback



)



t

leftover










)

+



(


r
downlink

-

r
playback



)


L




,

N
-
i


}






(

Equation





4

)







The method of FIG. 2 can be applied to find the actual number of streamed sub-clips. The adjusted deadline for sub-clip k, i≦k≦N, is thus







d

k
=t+(tleftover+(k−i−1)L)/δ  (Equation 5)


The leftover time for the fast reverse operation is the time interval from the current playback point to the beginning of current sub-clip. Using tleftover to denote the leftover time for the fast reverse operation, the maximum number of streamed sub-clips is:










n
*

=

min


{






(






r
playback


L





δ

-







(


r
downlink

-

r
playback



)



t

leftover










)

+



(


r
downlink

-

r
playback



)


L




,

i
-
1


}






(

Equation





6

)







Equation 4 and equation 6 are similar except for the last term, which can be explained by the opposite playback directions. It should also be noted that both fast forward and fast reverse operations can be implemented for multiple speeds in either direction.


The method illustrated in FIG. 2 only needs to be modified slightly to compute the actual number of streamed sub-clips for the fast reverse operation. More specifically, k is looped backward (decremented) from i−1 to i−n* in step 230.


The deadline for sub-clip k, 1<k<i, can be calculated as follows:







d

k
=t+(tleftover+(i−k−1)L)/δ  (Equation 7)



FIG. 3 is a schematic diagram of a video playback device. Video playback operations are supported through a button on the video playback device or via a button activated by a user of a remote control device or a software command activated, for example, by a voice command. The means for invoking the video playback operations are not limited to those suggested above, which are merely illustrative embodiments of such means for invoking video playback operations.



FIG. 4 is a flowchart depicting the admission control process from the user/client side. At the user/client side, once the client receives the response from the server, a check is made to determine if the request/demand has been admitted by testing the value of N (the leading number of sub-clips to be streamed to the client from the server) at 405. If the value of N is greater than zero then the request/demand has been admitted and the user starts to receive the streamed N sub-clips at 410. It should be noted that setting N=−1 on the server side and testing for N>0 on the user/client side is but one possible implementation. The request admission test could, for example, be implemented by means of a flag or any other suitable means. The current sub-clip number, Nc, is set to N+1 (the next sub-clip—the sub-clip to be downloaded) at 415. A test is then performed at 420 to determine if there are more sub-clips to download. If either the request/demand was not admitted or all sub-clips of the video have been received by the user then the process ends. Meanwhile, a peer within the peer-to-peer network having/caching the (N+1)-st sub-clip is located and starts uploading the (N+1)-st sub-clip to the peer needing the (N+1)-st sub-clip at 425. If it is determined that the deadline, d, for the download (measured against the current time, t) has not been reached at 430 then the downloading continues at 435. If it is determined that the deadline, d, has been reached (as measured against the current time, t) at 430 then a missing data vector is prepared at 440. Actually the missing data vector is prepared slightly before the deadline or when it is determined that the download cannot be completed before the deadline, d. If the download cannot be completed before the deadline then a test is performed to determine if complementary streaming is necessary at 445. Complementary streaming will be described in greater detail below. Meanwhile, the current sub-clip counter is incremented at 455. If complementary streaming is required to ensure system performance (arrival of sub-clips by the user before the deadlines) is met then complementary streaming is invoked at 450. As the deadline, d, for downloading the (N+1)-st sub-clip is reached, the user starts to playback the (N+1)-st sub-clip and a peer within the peer-to-peer network having/caching the following sub-clip is located and starts the video sub-clip downloading process.



FIG. 5 is a block diagram of the video playback operations within a video playback device, where the operational modules are directed by a control module. The control module calls/invokes the corresponding video playback functional/operation modules for any playback requests. It should be noted that a straight playback operation is not discussed herein as it is the subject of a related operation.



FIG. 6 is a flowchart of the operation of an exemplary control module in accordance with the present invention. The control module determines the sub-clip type at 605. The type of operation is determined by 610, 620, 630, 640, 650 and 660. If the user has pressed a button (or spoken a command) to invoke a function not supported then it is ignored. Instead of a series of decision points a single “case-type” decision point could be implemented. If the video playback operation is a jump forward (JF) then at 615 a determination of the maximum number of sub-clip necessary to stream is made; a determination of the actual number of sub-clips to be streamed is made based on what is already fully cached; streaming of the sub-clip containing the TPP is started (if the sub-clip containing the TPP is not fully cached); new sub-clip deadlines are calculated; playback is started; and downloading of sub-clips beyond the streamed sub-clips is started. The same procedure applies for the jump backward operation.


If the video playback operation is a pause operation then at 635 video playback is stopped; streaming is stopped; downloading is stopped; and the sub-clip deadlines are set to infinity. If the video playback operation is a resume operation then at 645 new sub-clip deadlines are calculated as a result of the pause operation; if sub-clip(s) were being streamed then resume streaming sub-clips; start video playback; and resume downloading any sub-clips that were scheduled to be downloaded.


If the video playback operation is a fast forward operation then at 655 a calculation is made of the new sped-up playback rate; a determination of the maximum number of sub-clip necessary to stream is made; a determination of the actual number of sub-clips to be streamed is made based on what is already fully cached; streaming of the sub-clip containing the TPP is started (if the sub-clip containing the TPP is not fully cached); new sub-clip deadlines are calculated; playback is started; and downloading of sub-clips beyond the streamed sub-clips is started. The similar procedure applies for the fast reverse operation, except Equation 6 replaces equation 4 in computing the actual number of streamed sub-clips. It should be remembered that both fast forward and fast reverse can be implemented for multiple speeds in either direction.


It is to be understood that the present invention may be implemented in various forms of hardware, software, firmware, special purpose processors, or a combination thereof. Preferably, the present invention is implemented as a combination of hardware and software. Moreover, the software is preferably implemented as an application program tangibly embodied on a program storage device. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (CPU), a random access memory (RAM), and input/output (I/O) interface(s). The computer platform also includes an operating system and microinstruction code. The various processes and functions described herein may either be part of the microinstruction code or part of the application program (or a combination thereof), which is executed via the operating system. In addition, various other peripheral devices may be connected to the computer platform such as an additional data storage device and a printing device.


It is to be further understood that, because some of the constituent system components and method steps depicted in the accompanying figures are preferably implemented in software, the actual connections between the system components (or the process steps) may differ depending upon the manner in which the present invention is programmed. Given the teachings herein, one of ordinary skill in the related art will be able to contemplate these and similar implementations or configurations of the present invention.

Claims
  • 1. A method for supporting playback operations for a peer-to-peer content-on-demand service, said method comprising: detecting a playback operation;determining a target Playback point for said playback operation; anddetecting a sub-clip type into which said target playback point falls.
  • 2. The method according to claim 1, further comprising: receiving a streamed leading sub-clip;determining a set of needed sub-clips;locating one of said set of needed sub-clips; anddownloading said located sub-clip.
  • 3. The method according to claim 1, wherein said peer-to-peer content-on-demand service is performance aware.
  • 4. The method according to claim 2, wherein said playback operation is one of jump forward and further comprising: determining a maximum number of sub-clips to stream;determining an actual number of sub-clips to stream;streaming a sub-clip containing said target playback point;calculating sub-clip deadlines;starting playback;downloading sub-clips following said streamed sub-clip.
  • 5. The method according to claim 2, wherein said playback operation is one of jump backward and further comprising: determining a maximum number of sub-clips to stream;determining an actual number of sub-clips to stream;streaming a sub-clip containing a target playback point;starting playback at a sub-clip containing a target playback point;calculating sub-clip deadlines; anddownloading sub-clips following said streamed sub-clip.
  • 6. The method according to claim 2, wherein said playback operation is one of pause and further comprising: stopping playback;stopping sub-clip streaming;stopping sub-clip downloading; andsetting sub-clip deadlines to infinity.
  • 7. The method according to claim 6, wherein said playback operation is one of resume and further comprising: calculating new sub-clip deadlines;resuming sub-clip streaming, if sub-clips were being streamed;starting playback; andresuming sub-clip downloading of any sub-clips that were being downloaded when said pause operation was invoked.
  • 8. The method according to claim 2, wherein said playback operation is one of fast forward and further comprising: calculating a sped-up playback rate;determining a maximum number of sub-clips to stream;determining an actual number of sub-clips to stream;streaming sub-clips;calculating sub-clip deadlines;starting playback;downloading sub-clips following said streamed sub-clip.
  • 9. The method according to claim 2, wherein said playback operation is one of fast reverse and further comprising: calculating a sped-up playback rate;determining a maximum number of sub-clips to stream;determining an actual number of sub-clips to stream;streaming sub-clips;starting playback;calculating sub-clip deadlines; anddownloading sub-clips preceding said streamed sub-clip.
  • 10. An apparatus for supporting playback operations for a peer-to-peer content-on-demand service, comprising: means for detecting a playback operation;means for determining a target playback point for said playback operation; andmeans for detecting a sub-clip type into which said target playback point falls.
  • 11. The apparatus according to claim 10, further comprising: means for receiving a streamed leading sub-clip;means for determining a set of needed sub-clips;means for locating one of said set of needed sub-clips; andmeans for downloading said located sub-clip.
  • 12. The apparatus according to claim 10, wherein said peer-to-peer content-on-demand service is performance aware.
  • 13. The apparatus according to claim 11, wherein said playback operation is one of jump forward and further comprising: means for determining a maximum number of sub-clips to stream;means for determining an actual number of sub-clips to stream;means for streaming a sub-clip containing said target playback point;means for calculating sub-clip deadlines;means for starting playback;means for downloading sub-clips following said streamed sub-clip.
  • 14. The apparatus according to claim 11, wherein said playback operation is one of jump backward and further comprising: means for determining a maximum number of sub-clips to stream;means for determining an actual number of sub-clips to stream;means for streaming a sub-clip containing a target playback point;means for starting playback at a sub-clip containing a target playback point;means for calculating sub-clip deadlines; andmeans for downloading sub-clips following said streamed sub-clip.
  • 15. The apparatus according to claim 11, wherein said playback operation is one of pause and further comprising: means for stopping playback;means for stopping sub-clip streaming;stopping sub-clip downloading; andsetting sub-clip deadlines to infinity.
  • 16. The apparatus according to claim 15, wherein said playback operation is one of resume and further comprising: means for calculating new sub-clip deadlines;means for resuming sub-clip streaming, if sub-clips were being streamed;means for starting playback; andmeans for resuming sub-clip downloading of any sub-clips that were being downloaded when said pause operation was invoked.
  • 17. The apparatus according to claim 11, wherein said playback operation is one of fast forward and further comprising: means for calculating a sped-up playback rate;means for determining a maximum number of sub-clips to stream;means for determining an actual number of sub-clips to stream;means for streaming sub-clips;means for calculating sub-clip deadlines;means for starting playback;means for downloading sub-clips following said streamed sub-clip.
  • 18. The apparatus according to claim 11, wherein said playback operation is one of fast reverse and further comprising: means for calculating a sped-up playback rate;means for determining a maximum number of sub-clips to stream;means for determining an actual number of sub-clips to stream;means for streaming sub-clips;means for starting playback;means for calculating sub-clip deadlines; andmeans for downloading sub-clips preceding said streamed sub-clip.
PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/US2006/024975 6/27/2006 WO 00 12/3/2008