Use of objective quality measures of streamed content to reduce streaming bandwidth

Information

  • Patent Grant
  • RE49990
  • Patent Number
    RE49,990
  • Date Filed
    Friday, September 24, 2021
    2 years ago
  • Date Issued
    Tuesday, May 28, 2024
    a month ago
Abstract
A method includes encoding video at a first bitrate to produce first encoded sections, and at a second bitrate that is less than the first bitrate to produce second encoded sections. The method further includes measuring levels of objective quality, indicative of perceptual quality, of the first and second encoded sections. The method includes determining, based on the measured objective quality levels, replaceable sections among the first encoded sections, and replacement sections among the second encoded sections, each of the replacement sections to replace a corresponding one of the replaceable sections. The method includes streaming the first encoded sections except for the replaceable sections, and the replacement sections instead of their corresponding replaceable sections.
Description
BACKGROUND

Distribution of multimedia (also referred to herein as “media” and/or “program(s)”), such as movies and the like, from network services to a client device may be achieved through adaptive bitrate streaming of the media. Typically, the media may be encoded at different bitrates and resolutions into multiple bitrate streams that are stored in the network services. Conventional adaptive bitrate streaming of media includes determining streaming conditions, e.g., an available streaming bandwidth at the client device, and then streaming a selected one of the different bitrate streams from the network services to the client device based on the determined conditions.


From the perspective of the network service, streaming media includes transmitting the media in response to requests from the client device. From the perspective of the client device, streaming media includes continuously requesting and receiving the media from the network services, and storing the received media in a buffer for subsequent presentation or playback, essentially, in near real-time. The buffered media may be presented, i.e., played back, in audio-visual form, for example.


The human visual system perceives a perceptual or subjective quality of streamed, presented media, and is able to detect small changes in the perceptual quality. The perceptual quality generally increases and decreases as the encoded bitrate of the streamed program (i.e., “streaming bitrate”) increases and decreases. Therefore, more or less available streaming bandwidth may translate to more or less perceptual quality, respectively.


Given the ever pressing need to conserve bandwidth at the client device, conventional streaming techniques tend to select a streaming bitrate deemed to be sufficiently high to meet an acceptable level of perceptual quality, based on the streaming bandwidth conditions determined at the client device, as mentioned above. This bandwidth-centric determination and selection at the client device does not take into consideration changes or variations in the content contained in the media itself over time as the media is streamed and, therefore, often results in unnecessarily high, and therefore, bandwidth-wasteful, streaming bitrates.





BRIEF DESCRIPTION OF THE DRAWINGS/FIGURES


FIG. 1 is a block diagram of an example network environment in which adaptive streaming of multimedia (also referred to herein as “media” and “program(s)”) from network services to a client device may be implemented.



FIG. 2 is an illustration of an example encoded video program generated by a media distributor.



FIG. 3 is an illustration of a container file that encodes a single audio stream.



FIG. 4 is an illustration of a container file that encodes multiplexed audio and video streams.



FIG. 5 is an illustration of a container file that encodes multiplexed video, audio, text, and metadata streams.



FIG. 6 is an illustration of an example of swapping-while-streaming from the container files of FIG. 2.



FIG. 7 is an illustration of an example of pre-swapped streaming using the container files of FIG. 2.



FIG. 8 is an illustration of example level-drop (or swap) information that identifies swappable, co-located clusters, as determined in cluster swapping examples of FIGS. 6 and 7.



FIG. 9 is a sequence diagram of example high-level interactions between a distributor and a client device in the network environment of FIG. 1.



FIG. 10 is a flowchart of an example method of streaming a program, which may be performed at a client device of FIG. 1.



FIG. 11 is a flowchart of an example method of adaptive streaming of programs from network services to a client device, which may be performed in a distributor of the network services.



FIG. 12 is a flowchart of a method expanding on the streaming in the method of FIG. 11, corresponding to a swapping-while-streaming embodiment, in which streaming is from multiple files.



FIG. 13 is a flowchart of a method expanding on the streaming of the method of FIG. 11, corresponding to a pre-swapped embodiment, in which streaming is from a single file constructed from multiple different files.



FIG. 14 is a block diagram of an example computer system corresponding to any network services, including a distributor in the network services.



FIG. 15 is a block diagram of an example system representing a client device.



FIG. 16 is a block diagram of a computer system configured to perform processing of media/programs and adaptive streaming.



FIG. 17 is a block diagram of an example computer program hosted in a client-side computer system (e.g., client device) similar to the computer system of FIG. 14.





In the drawings, the leftmost digit(s) of a reference number identifies the drawing in which the reference number first appears.


DETAILED DESCRIPTION

Embodiments described herein are directed to adaptive streaming of a video program from network services to a client device that utilize encoded video section swapping to reduce streaming bandwidth, while minimally impacting a perceptual/presentation quality of the streamed program at the client device. The adaptive streaming with section swapping embodiments reduce streaming bandwidth at client the device based on characteristics or properties of content in the video program, i.e., measures of objective quality indicative of perceptual quality, that change over time while the program is being streamed. More specifically, the embodiments determine swappable, corresponding/co-located sections of video encoded at different bitrates based on the measured objective quality levels of the co-located sections and the predetermined criterion of objective quality. The swappable, co-located sections include a section encoded at a first bitrate and a section encoded at a second bitrate that is less than the first bitrate. During streaming of the video program, the section encoded at the first bitrate is replaced with the co-located section encoded at the lesser second bitrate, thereby conserving streaming bandwidth by a substantial amount approximately equally to a difference between the first and second bitrates.


Network Environment



FIG. 1 is a block diagram of an example network environment 100 in which adaptive streaming of programs from network services to a client device may be implemented. Network environment 100 includes a collection of server-side services 102 that interact and cooperate to originate, manage, and distribute, e.g., stream, programs to a user operated client device 104 over one or more networks 106, such as the Internet. Such programs include, but are not limited to, entertainment programs (e.g., television shows, movies, cartoons, news programs, etc.), educational programs (e.g., classroom video, adult education video, learning programs, etc.), and advertising programs (e.g., commercials, infomercials, or marketing content). Network services 102 communicate with each other and with client device 104 using any suitable communication protocol, such as an Internet protocol, which may include Transmission Control Protocol/Internet Protocol (TCP/IP), Hypertext Transfer Protocol (HTTP), etc.


Client device 104 may be capable of wireless and/or wired communication with networks 106. Client device 104 includes processing, storage, communication, and user interface capabilities sufficient to provide all of the client device functionality described herein. Such functionality may be provided, at least in part, by one or more applications, such as computer programs, that execute on client device 104. Applications executed on client device 104 may include a client-side application, which presents Graphical User Interfaces (GUIs) through which a user of the client device may interact with and request services from corresponding server-side applications hosted in services 102. Accordingly, under user control, client device 104 may request/select programs from services 102, stream the selected programs from the services, and then present the streamed programs, in other words, playback the streamed programs.


Network services 102 include, but are not limited to, one or more media provider(s) 112 to originate source programs, and one or more media distributor(s) 114 to encode, store, and distribute the programs to client device 104. Each of the services 102 may be implemented as one or more computer servers that execute one or more associated server-side computer program applications suited to the given service. Media providers 112, such as Netflix®, HBO®, cable networks, and so on, utilize network services 102 to manage and deliver their revenue bearing programs to client device 104. Media providers 112 download their source programs to distributor 114, which encodes, stores, and then streams the encoded programs to client device 104 when requested to do so.


Distributor 114 includes an encoder 122 and a stream manager 126. Encoder 122 may encode each program into a number of alternative streams to support adaptive bitrate streaming of the program. The alternative streams encode the same program in different ways, such as at one or more of different bitrates, one or more different resolutions, and/or one of more different frame rates. As will be described more fully below in connection with FIG. 2, each of the encoded streams is typically stored in one or more container files. Encoder 122 also generates a program index file for the container files associated with the encoded program. Stream manager 126 manages streaming of the encoded program from the container files to client device 104 when the client device requests the program. Stream manager 126 cooperates with the requesting client device 104 to support adaptive bitrate streaming of the program from the container files to the client device.


Distributor 114 may also store auxiliary streams which contain information associated with the program streams mentioned above. The auxiliary streams are encoded at low bitrates, e.g., at bitrates of 200 kbps or much less. The auxiliary streams may include metadata synchronized in time with and descriptive of the content in associated main program streams. The metadata may include cues indicating or bracketing, e.g., commercial segments, or other non-program segments/content, such as level-drop information for encoded section swapping as will be described below, interspersed throughout the program streams. Typically, such auxiliary streams would be streamed simultaneously with their associated program streams and handled appropriately at the client device. However, the auxiliary streams may be streamed before the program streams.


As discussed above, client-side GUIs provide a user with access to services and program offerings. Such client-side GUIs typically include easily navigable program guides, and may present programs and channel selection options, program descriptions, advertisements, programming/user control options, and other typical programming features, as would be appreciated by those of ordinary skill in the relevant arts. The client-side GUIs accept user selections/requests, such as a request to view a program. In response to such GUI selections/requests, the client-side application sends appropriate requests to a counterpart server-side application residing in distributor 114, to initiate the appropriate actions among services 102 that will satisfy the client selections/requests, e.g., enable a client device to stream the selected program from the distributor for presentation to the user.


Container Files—Streaming Sources


As described above, distributor 114 encodes source programs from providers 112. To support adaptive bitrate streaming, distributor 122 may encode the source programs at multiple bitrates to produce multiple streams for each source program, as will be described more fully below in connection with FIG. 2. While streaming such encoded programs, client device 104 may switch between streams (and thus between encoded bitrates and corresponding streaming rates) according to conditions at the client device.



FIG. 2 is an illustration of an example encoded video program 200 generated by distributor 112. Encoded video program 200 includes multiple (encoded) video streams 1-4 encoded at multiple corresponding bitrates Rate 1-Rate 4. Encoded video streams 1-4 encode video at multiple video resolutions Res 1-Res 4, which may be equal to or different from each other. Encoded video program 200 includes a program stream index 204 and multiple container files 208(1)-208(4) corresponding to streams 1-4.


Program stream index 204 includes pointers 210(1)-(4), e.g., Uniform Resource Locators (URLs), to corresponding container files 208(1)-(4), and lists encoding parameters used to encode each of the streams 1-4, including, but not limited to, encoded bitrates Rate 1-Rate 4, encoding resolutions Res 1-Res 4, frame rates, and encoding techniques/standards. Exemplary, non-limiting, bitrates may range from below 125 kilo-bits-per-second (kbps) up to 15,000 kbps, or even higher, depending on the type of encoded media.


Each of container files 208 comprises sequential clusters 212 of a larger media sector (not shown in FIG. 2), and sequential blocks 214 of encoded media (which may also include audio, text, multimedia, etc., in addition to video) within each of the clusters. Each cluster 212, and each block 214, includes a time code TC indicating a start time for the media encoded in the blocks of that cluster, and encodes a fixed duration of media. For example, each cluster 212 of container file 208(1) encodes two seconds of video. In other embodiments, each cluster may encode a different duration of media, which may vary from two seconds. Each cluster 212 is a self-contained unit of media that may be decoded and presented on client devices 204 without reference to any other clusters. Clusters 212 may also include successive cluster numbers identifying a streaming sequence of the clusters.


Each cluster/block 212/214 in a given one of container files 208 encodes the same content (e.g., video content) as corresponding clusters in the other ones of the container files. For example, the cluster/block indicated at A in container file 208(1) has encoded therein the same video as that encoded in the clusters/blocks indicated at B, C, and D of container files 208(2), 208(3), and 208(4), respectively. Corresponding clusters/blocks are also referred to herein as “co-located” clusters/blocks because they encode the same video and share the same time code TC, i.e., they are aligned or coincide in time.


Container files may encode a single stream, such as a video stream (as depicted in FIG. 2), an audio stream, or a text stream (e.g., subtitles). Alternatively, each container file may encode multiple multiplexed streams, such as a mix of video, audio, and text streams. FIGS. 3-5 are further illustrations of diverse container files.



FIG. 3 is an illustration of a container file 300 that encodes a single audio stream.



FIG. 4 is an illustration of a container file 400 that encodes multiplexed audio and video streams.



FIG. 5 is an illustration of a container file 500 that encodes multiplexed video, audio, text, and metadata streams.


In addition, a container file may encode only a metadata stream at a relatively low bitrate.


The encoded container files depicted in FIGS. 2-5 support adaptive streaming to client device 104. If conditions change while streaming, then client device 104 may switch between container files to stream at rates best suited to the conditions.


In embodiments: the container files may be Matroska containers based on Extensible Binary Meta Language (EBML), which is a derivative of Extensible Binary Meta Language (XML), or files encoded in accordance with the Moving Picture Experts Group (MPEG) standard; the program index may be provided in a Synchronized Multimedia Integration Language (SMIL) format; and client device 104 may implement adaptive streaming from distributor 114 over networks 106 using the HTTP protocol.


The container files described above may support adaptive streaming of encoded video programs across an available spectrum bandwidth that is divided into multiple, i.e., n, levels. Video having a predetermined video resolution for each level may be encoded at a bitrate corresponding to the bandwidth associated with the given level. For example, in DivX® Plus Streaming, by Rovi Corporation, the starting bandwidth is 125 kbps and the ending bandwidth is 8400 kbps, and the number n of bandwidth levels is eleven (11). Each bandwidth level encodes a corresponding video stream, where the maximum encoded bitrate of the video stream (according to a hypothetical reference decoder model of the video coding standard H.264) is set equal to the bandwidth/bitrate of the given level. In DivX® Plus Streaming, the 11 levels are encoded according to 4 different video resolution levels, in the following way: mobile (2 levels), standard definition (4 levels), 720p (2 levels), and 1080p (3 levels).


Section Swapping: Objective Quality vs. Perceptual Quality


Adaptive streaming embodiments described herein take into consideration certain characteristics in the content in video programs (also referred to herein simply as “video”) to be streamed, to reduce streaming bitrates, as will now be described. The human visual system is able to detect small changes in the perceptual or presentation quality of presented video. However, perceptual quality and changes in perceptual quality are difficult to measure directly. Instead, certain characteristics or properties of encoded video, that are indicative of perceptual quality video as actually presented (i.e., once the encoded video is decoded and presented), may be determined, e.g., measured directly, in a straight forward manner. Such measured properties represent an objective quality of the video. As the content of encoded video varies across successive sections of the encoded video, and between co-located sections of encoded video in different video streams, the objective quality correspondingly varies, and may be measured to determine a level of objective quality corresponding to each of the aforementioned sections. The term “section” as used herein refers to a number of successive frames of video, including, but not limited to, multimedia audio-visual content, which may be collected into successive blocks and clusters in container files for streaming, as described above.


The difference in objective quality levels between co-located sections of encoded video (e.g., between co-located clusters from two container files that encode the same video but at different bitrates) may be used to determine a corresponding difference in the perceptual quality of the two video sections. Specifically, the difference in objective quality levels may be used to determine whether that difference is sufficiently large as to cause a visually noticeable difference in the perceptual quality of the two sections (once decoded and presented). With this in mind, there exists a criterion of objective quality (referred to as a predetermined “swap criterion” of objective quality), for the two objective quality levels, that translates to a virtually imperceptible difference in perceptual quality. Two co-located sections having objective quality levels that meet this criterion are considered interchangeable or swappable with each other for purposes of streaming to a client device because of their imperceptible or nearly imperceptible difference in perceptual quality. “Co-located” sections are also referred to herein as “corresponding” sections.


Swappable, co-located, encoded sections may be swapped to reduce streaming bandwidth in cases where co-located sections include a first section encoded at a relatively high bitrate and a (co-located) second section encoded at a relatively low bitrate. Specifically, streaming the (lower bitrate) second section instead of (i.e., in place of) its co-located (higher bitrate) first section reduces streaming bandwidth, while maintaining perceptual quality. The first section is referred to herein as a “replaceable section” and the second section that is streamed instead of the first section is referred to herein as the “replacement section.” The more often high bitrate sections are replaced with their co-located low bitrate sections while streaming, i.e., the more often bitrate swapping occurs, the more streaming bandwidth is conserved.


Measures of objective quality of a section of encoded video (e.g., of a cluster/block of encoded video) include, but are not limited to, a signal-to-noise ratio (SNR) of the section, a peak SNR (PSNR) of the section, a structural similarity index (SSIM) that measures a similarity between sections, and so on.


PSNR is a commonly used measure for assessing a quality of reconstructed video after compression (encoding). PSNR is measured on a logarithmic scale and depends on the Mean Squared Error (MSE) between an original source image and an impaired image or video frame. A higher PSNR generally indicates better quality, while a lower PSNR generally indicates poorer quality. Formulas for PSNR are given below:







MSE
=


1
mn






i
=
0


m
-
1




[


l

(

i
,
j

)

-

K
(

i
,
j

)


]

2







PSNR
=

10
·


log
10

(


MAX
i
2

/
MSEi

)








Where: m & n are dimensions of an image; I & K are components of an image (Y, U, V); and MAXi is a maximum possible pixel value.


Each section of encoded video may be assigned one or more scores based on the different possible objective quality measures. Multiple successive encoded sections may have their objective quality levels combined into and, therefore, associated with, a single measurement of objective quality, which may include, e.g., a mean or variance of the SNR (or PSNR) of each of the multiple sections.


A criterion of objective quality that indicates co-located sections may be swapped with little or no impact on perceptual quality—if the co-located section have levels of objective quality that each meet that criterion—may be based on one of:

    • a. an absolute level of objective quality, e.g., co-located clusters are declared swappable if the PSNR of each cluster is above a threshold PSNR; and
    • b. a difference between respective levels objective quality of co-located encoded sections, e.g., co-located clusters are declared swappable if a difference between their respective PSNR values is less than a certain PSNR difference threshold.


For relative criterion (b), a variable scale may be defined using as a base PSNR a highest level PSNR among encoded sections under consideration; e.g., for encoded section PSNR levels in a range of 29 dB and 31 dB, an acceptable difference value may be 1.5 dB, and for encoded section PSNR levels in a range of 27 dB and 29 dB, an acceptable PSNR difference level may be only 1 dB. The variable scale may be extended to cover all encoded section PSNR levels that are expected as part of an encoding process, and may also extend to cover other objective quality metrics such as a mean, a variance, a difference between highest and lowest PSNR levels in a cluster, or a formula combining some or all of these or other metrics. Such assignments may be altered based on the type of objective quality that is used.


Embodiments directed to adaptive streaming using section swapping described below include (i) a “swapping-while-streaming” embodiment in which swappable co-located sections from different files are swapped while streaming, and (ii) a “pre-swapped streaming” embodiment in which a single file containing pre-swapped sections is constructed from multiple encoded files before streaming is initiated, and then all of the sections in the single file, including the pre-swapped sections, are streamed from that file. Examples of these embodiments are described below.


Swapping-While-Streaming


In swapping-while-streaming, swappable co-located sections are swapped while streaming from multiple different container files in real-time, such as from multiple container files 208 in FIG. 2.



FIG. 6 is an illustration of an example of swapping-while-streaming from container files 208 of FIG. 2. With reference to FIG. 6, the example assumes the following:

    • a. Separate container files 208(1), 208(2), and 208(3) represent three levels N, N−1, and N−2, respectively, of the n levels available for, e.g., DivX® Plus Streaming;
    • b. Levels N, N−1, and N−2 are each encoded at successively decreasing bitrates but may be at the same resolution. In other examples, the Levels are each encoded at successively decreasing bitrates but may be at different resolutions;
    • c. Swappable, co-located clusters (including a replaceable section and its corresponding replacement section) have been determined and identified across the container files 208(1)-208(3). As described above, the swappable, co-located clusters are determined based on determined/measured objective quality levels of the clusters in the container files and a comparison of the determined/measured objective quality levels (of co-located sections) against a predetermined criterion of objective quality indicative of a minimal difference in perceptual quality; and
    • d. Level N is initially selected for streaming.


Traversing left-to-right in FIG. 6, initially, encoded sections from level N are streamed from container file 208(1). Encoded sections are streamed predominantly from level N. However, while the sections from Level N are streamed, replaceable (swappable) sections thereof are encountered and replaced with co-located replacement (swappable) sections from level N−1, or level N−2, as indicated at LD1, or LD2, respectively. That is, the lower level, co-located replacement sections are streamed instead of the higher level replaceable sections, resulting in conserved streaming bandwidth. Each replacement section is streamed at the same time its corresponding co-located replaceable section would have been streamed had it not been replaced, so as to maintain a proper encoded video sequence. A section swap is referred to as a “level-drop” because the source level (n) from which the replacement section is streamed is lower than the current level, e.g., in a swap, the level drops from Level N to Level N−1, or Level N−2. In swapping-while-streaming, client device 104 and distributor 114 cooperate with each other to swap sections from different container files.


Pre-Swapped Streaming


In pre-swapped streaming, a single file containing swappable sections, i.e., level-drops, is constructed from multiple files before streaming. Then, the sections in the single file, including the pre-swapped sections, are streamed from that file.



FIG. 7 is an illustration of an example of pre-swapped streaming using container files 208 of FIG. 2. Before streaming, swappable, co-located clusters are determined and identified across levels N, N−1, and N−2, similar to the swapping-while-streaming embodiment. The determined swaps, or level-drops, are incorporated into Level N. In other words, determined and identified replaceable sections of Level N are replaced with co-located replacement sections of Levels N−1 and N−2, as identified. As a result, a new, pre-swapped Level N file is constructed that incorporates original sections of Level N that were not indicated as replaceable, and the determined/identified replacement sections (i.e., level-drops) instead of the replaceable sections. Then, sections are streamed only from the new, pre-swapped Level N file. In pre-swapped streaming, streaming of the pre-swapped sections is transparent to client device 104, which simply requests streaming sections from the single file. The examples of FIGS. 6 and 7 may be extended over more container files at different bitrates, and thus more levels and corresponding level-drops.


Level-Drop Information



FIG. 8 is an illustration of example level-drop (or swap) information 800 that identifies swappable, co-located clusters, as determined in the cluster swapping examples of FIGS. 6 and 7. In the manner described above, and further below, distributor 114 measures objective quality levels of sections, e.g., clusters in container files 208, and then determines swappable co-located clusters between the files based on the measured objective quality levels and a predetermined swap criterion. Distributor 114 generates level-drop information 800 as records 802, 804, and 806 that identify the determined, swappable, co-located clusters in the different files. Each record identifies swappable clusters including a starting cluster and its contiguous following clusters (if any), and also blocks within clusters. Each record may include one or more of cluster/block numbers, time codes TC, locations/addresses of the clusters in the file. Level-drop information 800 includes:

    • a. records 802 linked vertically to indicate determined, time-ordered, replaceable clusters among the clusters of level N, e.g., in container file 208(1);
    • b. records 804 linked vertically to indicate determined, time-ordered replacement clusters among the clusters of level N−1, e.g., in container file 208(2); and
    • c. records 806 to indicate a determined replacement cluster among the clusters of level N−2, e.g., in container file 208(3).
    • Horizontal links 810 between records indicating co-located, replaceable and replacement clusters, e.g. between Level N and Levels N−1, N−2.


Distributor 114 may embed level-drop information 800 as in container files as metadata. Alternatively, or additionally, level-drop information may be stored as an auxiliary/metadata, level-drop file that is stored and indexed separately from the container files. Therefore, in the swapping-while-streaming embodiment, level-drop information 800 may be streamed from distributor 114, to client device 104, with streamed clusters. Alternatively, level-drop information may be streamed from the auxiliary file separately from the clusters.


In the pre-swapped embodiment, distributor 114 uses level-drop information 800 to construct a single pre-swapped file and need not be streamed to client device 104. However, client device 104 does not need the level-drop information.


Server-Client Sequence Diagram



FIG. 9 is a sequence diagram of example high-level interactions 900 between distributor 114 and client device 104 in network environment 100. Interactions 900 progress in time from top-to-bottom in FIG. 9, as a sequence of phases beginning with an initialization/set-up phase 904 indicated in the top left-hand corner of FIG. 9.


During initialization/set-up phase 904, distributor 114 encodes video programs and stores the encoded video programs in container files for subsequent streaming to client device 104. Distributor 114 determines swappable co-located sections among the container files associated with each program based on determined objective quality measures and a predetermined swap criterion for the objective quality measures of co-located clusters in different files, as discussed above. Distributor 114 embeds drop-level information, such as information 900, in the container files, or stores the information in a separate drop-level file. In the pre-swapped embodiment, distributor 114 constructs a single file for streaming using the drop-level information and the multiple container files.


During a user interaction phase 920, client device 104 presents client-side GUIs to a user. At 924, the user selects a program from the GUIs, and, in response, client device 104 sends a request for the program to distributor 114.


During a pre-streaming phase 930, in response to the client device request (at 924), at 932, the distributor 114 sends an index of URLs associated with the requested program to client device 104. In the swapping-while-streaming embodiment, distributor 114 may also sends drop-level information if such information is provided in a separate file. Alternatively, distributor 114 may stream the drop-level information to client 104 in subsequent pre-streaming and streaming phases 930, 940, described below. The URLs may include a first URL directed to a program index (e.g., index 204 in FIG. 2) for the encoded program corresponding to the requested program and stored in distributor 114, and a second URL directed to a drop-level file, if available.


During streaming phase 940, streaming of the requested program from distributor 114 to client device 104 commences. Client device 104 determines a streaming bandwidth available at the client device and selects a stream from among the multi-bitrate streams, as indicated in the program index, that best matches the determined bandwidth. Client device 104 continually requests encoded stream sections from container files in distributor 114 based on the index information and the level-drop information (in the swapping-while-streaming embodiment), and receives the requested blocks from the distributor. In the swapping-while-streaming embodiment, the level-drop information may have been received during pre-streaming phase 930, may be streamed from a separate level-drop file, and/or may be embedded as records in the encoded video sections streamed during stage 940. Client device 104 buffers the received sections, decodes the buffered sections, and then presents the decoded sections. As streaming conditions change, client device 104 may switch to a new stream, i.e., request sections from another stream having a bitrate better suited to the changed conditions.


In yet another embodiment, level-drop information may be calculated dynamically at client device 104 instead of distributor 114 based on objective quality levels embedded in streamed sections. In this embodiment, distributor 114 embeds measured objective quality levels in corresponding encoded video sections, and streams the sections to the client device 104. Client device 104 calculates level-drop information based on the received objective quality levels, and performs swapping-while-streaming based on the dynamically calculated objective quality levels.


In both the swapping-while-streaming and pre-swapped embodiments, client device 104 intermittently requests replacement blocks having a lower encoded bitrate than the co-located replaceable block, which advantageously conserves streaming bandwidth at the client device. Each replacement block is streamed at the same its corresponding or co-located replaceable block would have been streamed had it not been replaced.


Client-Side Method



FIG. 10 is a flowchart of an example summary method 1000 of streaming a video program with swapped sections, which may be performed at client device 104.

    • 1005 includes requesting to stream the video program.
    • 1010 includes receiving an index to multiple files containing the video encoded at different bitrates, including at least a first bitrate and a second bitrate that is less than the first bit rate.
    • 1015 includes determining an available streaming bandwidth.
    • 1020 includes selecting from the received index the file encoded at the first bitrate based on the determined streaming bandwidth.
    • 1025 includes initiating streaming of the selected file.
    • 1030 includes streaming the selected file, including requesting successive sections of encoded video and receiving the requested sections. The swapping-while-streaming embodiment also includes receiving level-drop information in any number of ways, including streaming the level-drop information from an auxiliary file before or while streaming the encoded video, or as metadata embedded with the streamed encoded sections. The streaming in the swapping-while-streaming embodiment includes requesting encoded sections from multiple levels (e.g., levels N, N−1, and N−2, etc.), i.e., requesting swappable sections, based on the level-drop information, so as to conserve bandwidth.


Server/Network-Side Method



FIG. 11 is a flowchart of an example summary method 1100 of preparing, and adaptive streaming of, a video program with swapped sections from services 102 to client device 104, which may be performed in distributor 114.

    • 1110 includes encoding video at a first bitrate to produce first encoded sections.
    • 1115 includes encoding the video at a second bitrate that is less than the first bitrate to produce second encoded sections, each having encoded therein the video of a corresponding one of the first encoded sections. The first and second encoded sections may encode video that has the same resolution for each of the first and second encoded sections, or may encode video having different resolutions.
    • 1120 includes measuring levels of objective quality, indicative of perceptual quality, of the first and second encoded sections.
    • 1125 includes determining, based on the measured objective quality levels and a predetermined criterion of objective quality indicative of a relatively small difference in perceptual quality, swappable sections, including replaceable sections among the first encoded sections, and corresponding replacement sections among the second encoded sections, each of the replacement sections to replace a corresponding/co-located one of the replaceable sections. This can be thought of as declaring or identifying certain co-located sections as swappable sections if they meet the predetermined criterion. A result of the determining in 1125 is to identify sections in lower levels (e.g., N−1, N−2) which may be swapped with higher levels (e.g., N) with little or no impact to the subjective/perceptual quality of the video.
    • 1130 includes generating, from the determining in 1125, level-drop information identifying the replaceable sections and their corresponding replacement sections in the first and second container files, respectively.
    • 1135 includes streaming the first encoded sections except for the replaceable sections, and the replacement sections instead of their corresponding replaceable sections, thereby conserving streaming bandwidth. Each replacement section is streamed at the same time its corresponding or co-located replaceable section would have been streamed had it not been replaced.



FIG. 12 is a flowchart of a method 1200 expanding on the streaming at 1140, corresponding to the swapping-while-streaming embodiment described above, in which streaming is from multiple files.

    • 1205 includes streaming the replacement sections instead of the replaceable sections from a second file containing the second encoded sections. Each replacement section is streamed from the second file at the same time its corresponding replaceable section would have been streamed from the first file.
    • 1210 includes streaming the replacement sections from a second file containing the second encoded sections.



FIG. 13 is a flowchart of a method 1300 expanding on the streaming at 1140, corresponding to the pre-swapped embodiment described above, in which streaming is from a single file constructed from multiple different bitrate files.

    • 1305 includes constructing a file containing (i) the first encoded sections, except for the replaceable sections, and (ii) the replacement sections instead of their corresponding replaceable sections.
    • 1310 includes streaming the first encoded sections, except for the replacement section, and the replacement sections from the constructed file.


Methods and systems disclosed herein may be implemented with respect to one or more of a variety of systems including one or more consumer systems, such as described below with reference to FIGS. 13 and 14. Methods and systems disclosed herein are not, however, limited to the examples of FIGS. 13 and 14.


Prototype Example

In a prototype example, a full length feature movie (video program) (2 hours and 25 minutes) was encoded at three bitrates: 8400 kbps, 5300 kbps, and 3500 kbps utilizing a 2-pass encoding process at a resolution of 1920×816 pixels and 24 frames per second. The PSNR for each frame was determined against the original source video, and the frames were grouped into clusters of 48 frames, corresponding to 2 seconds of video frames. From here, the lowest PSNR of the frames in a cluster was determined for each cluster, and a minimum PSNR of 40 dB was chosen as the least acceptable objective quality metric for a cluster. If the minimum PSNR for a cluster at a lower level was available, that cluster was swapped for the higher level cluster. This approach was used for 8400 kbps (using 8400 kbps, 5300 kbps, and 3500 kbps), and for 5300 kbps (using 5300 kbps and 3500 kbps). The results were an accumulative bandwidth savings of 36% for an 8400 kbps stream and 18% for a 5300 kbps stream, with minimal impact on perceptual quality.


System and Computer Block Diagrams



FIG. 14 is a block diagram of an example computer system 1400 corresponding to any of services 102, including distributor 114. Computer system 1400, which may be, e.g., a server, includes one or more processors 1405, a memory 1410 in which instruction sets and databases for computer program applications are stored, a mass storage 1420 for storing, e.g., encoded programs and drop-level information, and an input/output (I/O) module 1415 through which components of computer system 1400 may communicate with networks 106.



FIG. 15 is a block diagram of an example system 1500 representing, e.g., client device 104, and may be implemented, and configured to operate, as described in one or more examples herein.


System 1500 or portions thereof may be implemented within one or more integrated circuit dies, and may be implemented as a system-on-a-chip (SoC).


System 1500 may include one or more processors 1504 to execute client-side application programs stored in memory 1505.


System 1500 may include a communication system 1506 to interface between processors 1504 and communication networks, such as networks 106. Communication system 1506 may include a wired and/or wireless communication system.


System 1500 may include a stream processor 1507 to process program streams, received over channel 1508 and through communication system 1506, for presentation at system 1500. Stream processor 1507 includes a buffer 1507a to buffer portions of received, streamed programs, and a decoder 1507b to decode and decrypt the buffered programs in accordance with encoding and encryption standards, and using decryption keys. In an alternative embodiment, decoder 1507b may be integrated with a display and graphics platform of system 1500. Stream processor 1507 together with processors 1504 and memory 1505 represent a controller of system 1500. This controller includes modules to perform the functions of one or more examples described herein, such as a streaming module to stream programs through communication system 1506.


System 1500 may include a user interface system 1510.


User interface system 1510 may include a monitor or display 1532 to display information from processor 1504, such as client-side storefront GUIs.


User interface system 1510 may include a human interface device (HID) 1534 to provide user input to processor 1504. HID 1534 may include, for example and without limitation, one or more of a key board, a cursor device, a touch-sensitive device, and or a motion and/or image sensor. HID 1534 may include a physical device and/or a virtual device, such as a monitor-displayed or virtual keyboard.


User interface system 1510 may include an audio system 1536 to receive and/or output audible sound.


System 1500 may correspond to, for example, a computer system, a personal communication device, and/or a television set-top box.


System 1500 may include a housing, and one or more of communication system 1506, processors 1504, memory 1505, user interface system 1510, or portions thereof may be positioned within the housing. The housing may include, without limitation, a rack-mountable housing, a desk-top housing, a lap-top housing, a notebook housing, a net-book housing, a set-top box housing, a portable housing, and/or other conventional electronic housing and/or future-developed housing. For example, communication system 1502 may be implemented to receive a digital television broadcast signal, and system 1500 may include a set-top box housing or a portable housing, such as a mobile telephone housing.


Accordingly, system 1500 may include, but is not limited to, stand-alone equipment, such as personal computers, laptops, ultrabooks, and tablets, and mobile phones and smartphones/Personal Digital Assistants (PDAs). System 150 may also represent and include a suite of interconnected devices, such a set-top box/video game console device, a remote to operate such a device, and an audio-visual display and/or computer. System 1500 may also represent and include (digital video disk) DVD and Blu-ray players, and televisions.



FIG. 16 is a block diagram of a computer system 1600, configured to perform processing of media/programs and adaptive streaming as described herein.


Computer system 1600 includes one or more computer instruction processing units and/or processor cores, illustrated here as processor 1602, to execute computer readable instructions, also referred to herein as computer program logic.


Computer system 1600 may include memory, cache, registers, and/or storage, illustrated here as memory 1604, which may include a non-transitory computer readable medium encoded with computer programs, illustrated here as computer program 1606.


Memory 1604 may include data 1608 to be used by processor 1602 in executing computer program 1606, and/or generated by processor 1602 during execution of computer program 1606. Data 1608 includes a database 1608a of container files and generated drop-level information for use in the methods described herein.


Computer program 1606 may include the following server-side instructions:

    • a. encoding instructions 1610 to cause processor 1602 to encode programs at different bitrates into different container files;
    • b. measurer instructions 1612 to measure objective quality levels of encoded video sections;
    • c. determiner instructions 1614 to cause processor 1602 to determine, based on the measured objective quality levels and a predetermined criterion of objective quality indicative of a relatively small difference in perceptual quality, swappable, co-located sections among the different container files, e.g., to determine replaceable sections and corresponding replacement sections;
    • d. generating instructions 1616 to generate level-drop information identifying the determined co-located swappable sections;
    • e. streaming instructions 1618 to stream the encoded programs, and swap the swappable sections as appropriate in the swapping-while-streaming embodiment; and
    • f. file constructing instructions 1620 to construct a file from multiple files containing pre-swapped sections in the pre-swapped embodiment.



FIG. 17 is a block diagram of an example computer program 1706 hosted in a client-side computer system similar to computer system 1600. Computer program 1606 may include the following client-side instructions:

    • a. client-side application instructions 1710 to cause a client-side processor to communicate with corresponding server-side distributors, present corresponding client-side navigable GUIs, permit a user to select programs for presentation, and present streamed programs;
    • b. bandwidth determiner instructions 1712 to cause the processor to determine an available streaming bandwidth;
    • c. streaming instructions 1714 to cause the processor to initiate and maintain streaming of programs. Streaming instructions 1714 include instructions to cause the processor to identify swappable co-located sections from drop-level information received from a server-side peer and to stream the identified swappable sections accordingly; and
    • d. decoding instructions 1716 to cause the processor to decode streamed programs.


Methods and systems disclosed herein may be implemented in hardware, software, firmware, and combinations thereof, including discrete and integrated circuit logic, application specific integrated circuit (ASIC) logic, and microcontrollers, and may be implemented as part of a domain-specific integrated circuit package, and/or a combination of integrated circuit packages. Software may include a computer readable medium encoded with a computer program including instructions to cause a processor to perform one or more functions in response thereto. The computer readable medium may include a transitory and/or non-transitory medium. The processor may include a general purpose instruction processor, a controller, a microcontroller, and/or other instruction-based processor.


Methods and systems are disclosed herein with the aid of functional building blocks illustrating functions, features, and relationships thereof. At least some of the boundaries of these functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternate boundaries may be defined so long as the specified functions and relationships thereof are appropriately performed.


A method embodiment which may be performed at a client-side or a network/server-side comprises:


identifying

    • a. replaceable sections among first sections of video encoded at a first bitrate, and
    • b. replacement sections of the video each encoded at a bitrate that is less than the first bitrate and having encoded therein the video of a corresponding one of the replaceable sections, each replacement section and the corresponding replaceable section having respective measured levels of objective quality that meet a predetermined criterion of objective quality; and


streaming the first sections of video except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections.


The method may further comprise:


decoding the streamed first sections and the replacement sections to recover the video encoded therein; and


presenting the recovered video.


In the method, the predetermined criterion of objective quality, if met, may translate to an acceptable difference in perceptual quality levels of the video when decoded from the corresponding replaceable and replacement sections, and then presented.


In the method, the replacement sections may include:


first replacement sections encoded at a second bit rate that is less than the first bitrate; and


second replacement sections encoded at a third bitrate that is less than the second bitrate.


In the method, the streaming may include:


streaming the first sections except for the replaceable sections from a first file containing the first sections and the replaceable sections; and


streaming the replacement sections instead of the replaceable sections from a second file containing the replacement sections.


The method may further comprise:


requesting to stream the video;


receiving an index to multiple files containing the video encoded at different bitrates, including at least the first bitrate and the bitrate that is less than the first bit rate;


determining an available streaming bandwidth;


selecting the file encoded at the first bitrate based on the determined streaming bandwidth,


wherein


the identifying may include receiving level-drop information identifying the replaceable and replacement sections, and


the streaming may include streaming the first sections except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections, based on the identifying information.


The streaming may include receiving requests for the sections of video and transmitting the requested sections of video.


The identifying may include streaming information identifying the replaceable and replacement sections in streaming sources.


In the method, each of the measured levels of objective quality may be based on a signal to noise ratio (SNR) and the predetermined criterion may be based on a one of an absolute SNR level and a relative SNR between the measured levels of objective quality.


A computer program product embodiment corresponding to the above-recited method embodiment comprises a non-transitory computer readable medium encoded with a computer program, including instructions to cause a processor to:


identify

    • a. replaceable sections among first sections of video encoded at a first bitrate, and
    • b. replacement sections of the video each encoded at a bitrate that is less than the first bitrate and having encoded therein the video of a corresponding one of the replaceable sections, each replacement section and the corresponding replaceable section having respective measured levels of objective quality that meet a predetermined criterion of objective quality; and


stream the first sections of video except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections.


The predetermined criterion of objective quality, if met, may translate to an acceptable difference in perceptual quality levels of the video decoded from the corresponding replaceable and replacement sections, and then presented.


The replacement sections may include:


first replacement sections encoded at a second bit rate that is less than the first bitrate; and


second replacement sections encoded at a third bitrate that is less than the second bitrate.


The instruction to cause the processor to stream may include instructions to cause the processor to:


stream the first sections except for the replaceable sections from a first file containing the first sections and the replaceable sections; and


stream the replacement sections instead of the replaceable sections from a second file containing the replacement sections.


The instructions may further include instructions to cause the processor to:


request to stream the video;


receive an index to multiple files containing the video encoded at different bitrates, including at least the first bitrate and the bitrate that is less than the first bit rate;


determine an available streaming bandwidth;


select the file encoded at the first bitrate based on the determined streaming bandwidth; and


receive the information identifying the replaceable and replacement sections,


wherein


the instructions to cause the processor to identify may include instructions to cause the processor to receive level-drop information that identifies the replaceable and replacement sections, and


the instructions to cause the processor to stream may include instructions to cause the processor to stream the first sections except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections, based on the identifying information.


The instructions to cause the processor to stream may include instructions to cause the processor to receive requests for the sections of video and transmit the requested sections of video.


The instructions to cause the processor to identify may include instructions to cause the processor to stream information identifying the replaceable and replacement sections in streaming sources.


Each of the measured levels of objective quality may be based on a signal to noise ratio (SNR) and the predetermined criterion is based on a one of an absolute SNR level and a relative SNR between the measured levels of objective quality.


An apparatus embodiment corresponding to the above method and computer program product embodiments comprises:


a processor and memory configured to:


identify

    • a. replaceable sections among first sections of video encoded at a first bitrate, and
    • b. replacement sections of the video each encoded at a bitrate that is less than the first bitrate and having encoded therein the video of a corresponding one of the replaceable sections, each replacement section and the corresponding replaceable section having respective measured levels of objective quality that meet a predetermined criterion of objective quality; and


stream the first sections of video except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections.


The predetermined criterion of objective quality, if met, may translate to an acceptable difference in perceptual quality levels of the video decoded from the corresponding replaceable and replacement sections, and then presented.


The replacement sections may include:


first replacement sections encoded at a second bit rate that is less than the first bitrate; and


second replacement sections encoded at a third bitrate that is less than the second bitrate.


The processor and memory, configured to stream, may be configured to:


stream the first sections except for the replaceable sections from a first file containing the first sections and the replaceable sections; and


stream the replacement sections instead of the replaceable sections from a second file containing the replacement sections.


The processor and memory may be further configured to:


request to stream the video;


receive an index to multiple files containing the video encoded at different bitrates, including at least the first bitrate and the bitrate that is less than the first bit rate;


determine an available streaming bandwidth;


select the file encoded at the first bitrate based on the determined streaming bandwidth,


wherein


the processor and memory, configured to identify, may be further configured to receive level-drop information identifying the replaceable and replacement sections, and


the processor and memory, configured to stream, may be further configured to stream the first sections except for the replaceable sections, and the replacement sections instead of the corresponding replaceable sections, based on the identifying information.


The processor and memory, configured to stream, may be further configured to receive requests for the sections of video and transmit the requested sections of video.


The processor and memory, configured to identify, may be further configured to identify the replaceable and replacement sections from streamed information.


The apparatus may further comprise:


a user interface system;


a communication system to communicate with a network; and


a housing to house the processor and memory, the communication system, and the user interface system.


The communication system may include a wireless communication system; and


the housing includes a mobile hand-held housing to receive the processor and memory, the user interface system, the communication system, and a battery.


While various embodiments are disclosed herein, it should be understood that they have been presented by way of example only, and not limitation. It will be apparent to persons skilled in the relevant art that various changes in form and detail may be made therein without departing from the spirit and scope of the methods and systems disclosed herein. Thus, the breadth and scope of the claims should not be limited by any of the examples disclosed herein.

Claims
  • 1. A method of encoding video, the method performed by an encoding system, comprising: encoding video at a first bitrate to produce first encoded sections using an encoding system;encoding the video at a second bitrate that is less than the first bitrate to produce second encoded sections using the encoding system, each having encoded therein the video of a corresponding one of the first encoded sections;measuring levels of objective quality using the encoding system, indicative of perceptual quality, of the first and second encoded sections;determining, based on the measured objective quality levels, replaceable sections among the first encoded sections, and replacement sections among the second encoded sections using the encoding system, each to replace a corresponding one of the replaceable sections;generating identifying information that identifies the replaceable sections and their corresponding replacement sections in the first and second container files, respectively, using the encoding system;streaming the identifying information using the encoding system; andin response to requests from a streaming client device that are dependent on the streamed identifying information, streaming the first encoded sections except for the replaceable sections using the encoding device, and streaming the replacement sections instead of their corresponding replaceable sections at a same time as the corresponding replaceable section would have been streamed had the corresponding replaceable section not been replaced.
  • 2. The method of claim 1, wherein the determining includes: determining, using the encoding device, if corresponding first and second encoded sections have respective measured objective quality levels that each meet a predetermined criterion of objective quality; anddeclaring, using the encoding device, the corresponding first and second encoded sections as corresponding replaceable and replacement sections, respectively, if the respective measured objective quality levels each meet the predetermined criterion.
  • 3. The method of claim 2, wherein the predetermined criterion of objective quality is indicative of an acceptable difference in the perceptual quality levels of the corresponding replaceable and replacement sections when decoded and presented.
  • 4. The method of claim 2, wherein the predetermined criterion of objective quality is one of: a threshold level of objective quality that must be exceeded by each of the respective measured objective quality levels; anda maximum difference between the respective measured objective quality levels that must not be exceeded.
  • 5. The method of claim 1, wherein: the measuring includes measuring a signal-to-noise ratio (SNR) as the objective quality level; andthe determining includes determining whether the measured objective quality levels meet a predetermined criterion that is based on one of an absolute objective quality level and a relative objective quality level.
  • 6. The method of claim 1, further comprising: constructing a first file containing the first encoded sections, including the replaceable sections;constructing a second file containing the second encoded sections, including the replacement sections; andwherein the streaming includes:streaming the first encoded sections, except for the replaceable sections, from the first container file; andstreaming the replacement sections from the second container file.
  • 7. The method of claim 1, further comprising constructing a file containing the first encoded sections, except for the replaceable sections which are replaced by their corresponding replacement sections, wherein the streaming includes streaming from the file the first encoded sections, except for the replaceable sections, and the replacement sections.
  • 8. A non-transitory computer readable medium encoded with a computer program, including instructions to cause a processor to: encode video at a first bitrate to produce first encoded sections;encode the video at a second bitrate that is less than the first bitrate to produce second encoded sections, each having encoded therein the video of a corresponding one of the first encoded sections;measure objective quality levels, indicative of perceptual quality levels, of the first and second encoded sections;determine, based on the measured objective quality levels, replaceable sections among the first encoded sections, and replacement sections among the second encoded sections, each to replace a corresponding one of the replaceable sections;generate identifying information that identifies the replaceable sections and their corresponding replacement sections in the first and second container files, respectively, using the encoding system;stream the identifying information; andin response to requests from a playback device that are dependent on the streamed identifying information, stream the first encoded sections except for the replaceable sections, and stream the replacement sections instead of their corresponding replaceable sections at a same time as the corresponding replaceable section would have been streamed had the corresponding replaceable section not been replaced.
  • 9. The computer readable medium of claim 8, wherein the instructions to cause the processor to determine include instructions to cause the processor to: determine if corresponding first and second encoded sections have respective measured objective quality levels that each meet a predetermined criterion of objective quality; anddeclare the corresponding first and second encoded sections as corresponding replaceable and replacement sections, respectively, if the respective measured objective quality levels each meet the predetermined criterion.
  • 10. The computer readable medium of claim 9, wherein the predetermined criterion of objective quality is indicative of an acceptable difference in the perceptual quality levels of the corresponding replaceable and replacement sections when decoded and presented.
  • 11. The computer readable medium of claim 9, wherein the predetermined criterion of objective quality is one of: a threshold level of objective quality that must be exceeded by each of the respective measured objective quality levels; anda maximum difference between the respective measured objective quality levels that must not be exceeded.
  • 12. The computer readable medium of claim 8, wherein: the instructions to cause the processor to measure include instructions to cause the processor to measure a signal-to-noise ratio (SNR) as the objective quality level; andthe instructions to cause the processor to determine include instructions to cause the processor to determine whether the measured objective quality levels meet a predetermined criterion that is based on one of an absolute objective quality level and a relative objective quality level.
  • 13. The computer readable medium of claim 8, wherein the instructions further include instructions to cause the processor to: construct a first file containing the first encoded sections, including the replaceable sections;construct a second file containing the second encoded sections, including the replacement sections; andwherein the instructions to cause the processor to stream include instructions to cause the processor to:stream the first encoded sections, except for the replaceable sections, from the first container file; andstream the replacement sections from the second container file.
  • 14. The computer readable medium of claim 8, further comprising constructing a file containing the first encoded sections, except for the replaceable sections which are replaced by their corresponding replacement sections, wherein the streaming includes streaming from the file the first encoded sections, except for the replaceable sections, and the replacement sections.
  • 15. An apparatus, comprising: a processor and memory configured to:encode video at a first bitrate to produce first encoded sections, andencode the video at a second bitrate that is less than the first bitrate to produce second encoded sections, each having encoded therein the video of a corresponding one of the first encoded sections;measure objective quality levels, indicative of perceptual quality levels, of the first and second encoded sections;determine, based on the measured objective quality levels, replaceable sections among the first encoded sections, and replacement sections among the second encoded sections, each to replace a corresponding one of the replaceable sections; andgenerate identifying information that identifies the replaceable sections and their corresponding replacement sections in the first and second container files, respectively, using the encoding system;stream the identifying information; andin response to requests from a playback device that are dependent on the streamed identifying information stream the first encoded sections except for the replaceable sections, and the replacement sections instead of their corresponding replaceable sections at a same time as the corresponding replaceable section would have been streamed had the corresponding replaceable section not been replaced.
  • 16. The apparatus of claim 15, wherein the processor and memory, configured to determine, are further configured to: determine if corresponding first and second encoded sections have respective measured objective quality levels that each meet a predetermined criterion of objective quality; anddeclare the corresponding first and second encoded sections as corresponding replaceable and replacement sections, respectively, if the respective measured objective quality levels each meet the predetermined criterion.
  • 17. The apparatus of claim 16, wherein the predetermined criterion of objective quality is indicative of an acceptable difference in the perceptual quality levels of the corresponding replaceable and replacement sections when decoded and presented.
  • 18. The apparatus of claim 16, wherein the predetermined criterion of objective quality is one of: a threshold level of objective quality that must be exceeded by each of the respective measured objective quality levels; anda maximum difference between the respective measured objective quality levels that must not be exceeded.
  • 19. The apparatus of claim 16, wherein: the processor and memory, configured to measure, are further configured to measure a signal-to-noise ratio (SNR) as the objective quality level; andthe processor and memory, configured to determine, are further configured to determine whether the measured objective quality levels meet a predetermined criterion that is based on one of an absolute objective quality level and a relative objective quality level.
  • 20. The apparatus of claim 16, wherein: the processor and memory are further configured toconstruct a first file containing the first encoded sections, including the replaceable sections, andconstruct a second file containing the second encoded sections, including the replacement sections; andthe processor and memory, configured to stream, are further configured tostream the first encoded sections, except for the replaceable sections, from the first container file, andstream the replacement sections from the second container file.
  • 21. The apparatus of claim 16, wherein: the processor and memory are further configured to construct a file containing the first encoded sections, except for the replaceable sections which are replaced by their corresponding replacement sections; andthe processor and memory, configured to stream, are further configured to stream from the file the first encoded sections, except for the replaceable sections, and the replacement sections.
  • 22. The apparatus of claim 15, further comprising: a communication system to communicate with a network; anda housing to house the processor and memory, and the communication system.
  • 23. A method of encoding video, the method performed by an encoding system, comprising: encoding a portion of video at a first bitrate to produce a first encoded section using an encoding system;encoding the same portion of video at a second bitrate that is less than the first bitrate to produce a second encoded section using the encoding system;measuring objective quality levels using the encoding system with respect to each of the first and second encoded sections;selecting one of the first and second encoded sections based on the measured objective quality levels of the first and second encoded sections using the encoding system;constructing a file containing the selected one of the first and second encoded sections, wherein the file contains at least one encoded section encoded at the first bitrate and at least one encoded section encoded at the second bitrate using the encoding system;storing a video stream comprising the file containing the selected one of the first and second encoded sections using the encoding system; andstreaming, using the encoding system, the stored video stream in response to a request from a playback device.
  • 24. The method of claim 23, wherein the first and second sections encode video at a same resolution.
  • 25. The method of claim 23, wherein the first and second encoded sections are each a self-contained unit of media capable of being decoded without reference to any other encoded section.
  • 26. The method of claim 25, wherein the video stream comprises a plurality of encoded sections including the selected one of the first and second encoded sections and each of the plurality of encoded sections encodes a fixed duration of media.
  • 27. The method of claim 26, wherein the video stream comprises a plurality of encoded sections including the selected one of the first and second encoded sections and at least some of the plurality of encoded sections encode different durations of media.
  • 28. The method of claim 23, wherein each of the first and second encoded sections comprises a number of successive frames of video.
  • 29. The method of claim 23, wherein a measured objective quality of a section of video is at least one measure selected from the group consisting of: a signal-to-noise ratio (SNR) of the section, a peak SNR (PSNR) of the section, and a structural similarity index (SSIM) that measures a similarity between sections.
  • 30. The method of claim 23, further comprising: determining when the second encoded section has a respective measured objective quality level that meets a predetermined criterion for objective quality; andselecting the corresponding second encoded section when the measured objective quality level meets the predetermined criterion.
  • 31. The method of claim 30, wherein the predetermined criterion of objective quality is indicative of an acceptable difference in the perceptual quality levels of the corresponding first and second sections when decoded and presented.
  • 32. The method of claim 30, wherein the measuring includes measuring a signal-to-noise ratio (SNR) as the objective quality level; and the determining includes determining whether the measured objective quality level meet a predetermined criterion that is based on one of an absolute objective quality level and a relative objective quality level.
  • 33. A method of encoding video, the method performed by an encoding system, comprising: encoding a portion of video at a first bitrate to produce a first encoded section using an encoding system, where the first encoded section is a self-contained unit of media capable of being decoded without reference to any other encoded section;encoding the same portion of video at a second bitrate that is less than the first bitrate to produce a second encoded section using the encoding system, where the second encoded section is a self-contained unit of media capable of being decoded without reference to any other encoded section;measuring objective quality levels with respect to each of the first and second encoded sections using the encoding system;determining when the second encoded section has a respective measured objective quality level that meets a predetermined criterion for objective quality using the encoding system;selecting one of the first and second encoded sections based on the measured objective quality levels of the first and second encoded sections using the encoding system, where the second encoded section is selected when the measured objective quality level of the second encoded section meets the predetermined criterion for objective quality;constructing a file containing the selected one of the first and second encoded sections using the encoding system, wherein the file contains at least one encoded section encoded at the first bitrate and at least one encoded section encoded at the second bitrate;storing a video stream comprising the file containing the selected one of the first and second encoded sections using the encoding system; andstreaming the stored video stream in response to a request from a playback device using the encoding system.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation reissue of application Ser. No. 15/950,950, which is an application for reissue of U.S. Pat. No. 9,313,510.

US Referenced Citations (1183)
Number Name Date Kind
3609227 Kuljian Sep 1971 A
4009331 Goldmark et al. Feb 1977 A
4694357 Rahman et al. Sep 1987 A
4802170 Trottier Jan 1989 A
4964069 Ely Oct 1990 A
5119474 Beitel et al. Jun 1992 A
5132992 Yurt et al. Jul 1992 A
5274758 Beitel et al. Dec 1993 A
5341474 Gelman et al. Aug 1994 A
5361332 Yoshida et al. Nov 1994 A
5396497 Veltman Mar 1995 A
5400401 Wasilewski et al. Mar 1995 A
5404436 Hamilton Apr 1995 A
5420801 Dockter et al. May 1995 A
5420974 Morris et al. May 1995 A
5471576 Yee Nov 1995 A
5477263 Ocallaghan et al. Dec 1995 A
5479303 Suzuki et al. Dec 1995 A
5487167 Dinallo et al. Jan 1996 A
5502766 Boebert et al. Mar 1996 A
5509070 Schull Apr 1996 A
5533021 Branstad et al. Jul 1996 A
5537408 Branstad et al. Jul 1996 A
5539908 Chen et al. Jul 1996 A
5541662 Adams et al. Jul 1996 A
5544318 Schmitz et al. Aug 1996 A
5550863 Yurt et al. Aug 1996 A
5574785 Ueno et al. Nov 1996 A
5583652 Ware Dec 1996 A
5589993 Naimpally et al. Dec 1996 A
5600721 Kitazato Feb 1997 A
5614940 Cobbley et al. Mar 1997 A
5621794 Matsuda et al. Apr 1997 A
5627936 Prasad May 1997 A
5630005 Ort May 1997 A
5633472 DeWitt et al. May 1997 A
5642171 Baumgartner et al. Jun 1997 A
5642338 Fukushima et al. Jun 1997 A
5655117 Goldberg et al. Aug 1997 A
5664044 Ware Sep 1997 A
5675382 Bauchspies Oct 1997 A
5675511 Prasad et al. Oct 1997 A
5684542 Tsukagoshi Nov 1997 A
5715403 Stefik Feb 1998 A
5717816 Boyce et al. Feb 1998 A
5719786 Nelson et al. Feb 1998 A
5745643 Mishina Apr 1998 A
5751280 Abbott May 1998 A
5751358 Suzuki et al. May 1998 A
5754648 Ryan et al. May 1998 A
5757968 Ando May 1998 A
5761417 Henley et al. Jun 1998 A
5763800 Rossum et al. Jun 1998 A
5765164 Prasad et al. Jun 1998 A
5794018 Vrvilo et al. Aug 1998 A
5805700 Nardone et al. Sep 1998 A
5813010 Kurano et al. Sep 1998 A
5819160 Foladare et al. Oct 1998 A
5822524 Chen et al. Oct 1998 A
5828370 Moeller et al. Oct 1998 A
5838791 Torii et al. Nov 1998 A
5841432 Carmel et al. Nov 1998 A
5844575 Reid Dec 1998 A
5848217 Tsukagoshi et al. Dec 1998 A
5852664 Iverson et al. Dec 1998 A
5854873 Mori et al. Dec 1998 A
5867625 McLaren Feb 1999 A
5874986 Gibbon et al. Feb 1999 A
5878135 Blatter et al. Mar 1999 A
5887110 Sakamoto et al. Mar 1999 A
5892900 Ginter et al. Apr 1999 A
5892915 Duso et al. Apr 1999 A
5903261 Walsh et al. May 1999 A
5907597 Mark May 1999 A
5907658 Murase et al. May 1999 A
5923869 Kashiwagi et al. Jul 1999 A
5946446 Yanagihara Aug 1999 A
5956729 Goetz et al. Sep 1999 A
5959690 Toebes, VIII et al. Sep 1999 A
5970147 Davis Oct 1999 A
5973679 Abbott et al. Oct 1999 A
5999812 Himsworth Dec 1999 A
6002834 Hirabayashi et al. Dec 1999 A
6009237 Hirabayashi et al. Dec 1999 A
6016381 Taira et al. Jan 2000 A
6031622 Ristow et al. Feb 2000 A
6038257 Brusewitz et al. Mar 2000 A
6038316 Dwork et al. Mar 2000 A
6044469 Horstmann Mar 2000 A
6046778 Nonomura et al. Apr 2000 A
6047100 McLaren Apr 2000 A
6057832 Lev et al. May 2000 A
6058240 McLaren May 2000 A
6064794 McLaren et al. May 2000 A
6065050 DeMoney May 2000 A
6018611 Nogami et al. Jun 2000 A
6079566 Eleftheriadis et al. Jun 2000 A
6097877 Katayama et al. Aug 2000 A
6108422 Newby et al. Aug 2000 A
6141754 Choy Oct 2000 A
6151634 Glaser et al. Nov 2000 A
6155840 Sallette Dec 2000 A
6169242 Fay et al. Jan 2001 B1
6175921 Rosen Jan 2001 B1
6192075 Jeng et al. Feb 2001 B1
6195388 Choi et al. Feb 2001 B1
6199107 Dujari Mar 2001 B1
6204883 Tsukagoshi Mar 2001 B1
6222981 Rijckaert Apr 2001 B1
6266483 Okada et al. Jul 2001 B1
6282320 Hasegawa et al. Aug 2001 B1
6282653 Berstis et al. Aug 2001 B1
6289450 Pensak et al. Sep 2001 B1
6292621 Tanaka et al. Sep 2001 B1
6308005 Ando et al. Oct 2001 B1
6320905 Konstantinides Nov 2001 B1
6330286 Lyons et al. Dec 2001 B1
6347145 Kato et al. Feb 2002 B2
6351538 Uz Feb 2002 B1
6373803 Ando et al. Apr 2002 B2
6374144 Viviani et al. Apr 2002 B1
6389218 Gordon et al. May 2002 B2
6389473 Carmel et al. May 2002 B1
6395969 Fuhrer May 2002 B1
6397230 Carmel et al. May 2002 B1
6415031 Colligan et al. Jul 2002 B1
6418270 Steenhof et al. Jul 2002 B1
6441754 Wang et al. Aug 2002 B1
6445877 Okada et al. Sep 2002 B1
6449719 Baker Sep 2002 B1
6453115 Boyle Sep 2002 B1
6453116 Ando et al. Sep 2002 B1
6466671 Maillard et al. Oct 2002 B1
6466733 Kim Oct 2002 B1
6504873 Vehvilaeinen Jan 2003 B1
6510513 Danieli Jan 2003 B1
6510554 Gordon et al. Jan 2003 B1
6512883 Shim et al. Jan 2003 B2
6516064 Osawa et al. Feb 2003 B1
6532262 Fukuda et al. Mar 2003 B1
6535920 Parry et al. Mar 2003 B1
6578200 Takao et al. Jun 2003 B1
6587506 Noridomi et al. Jul 2003 B1
6594699 Sahai et al. Jul 2003 B1
6621979 Eerenberg et al. Sep 2003 B1
6625320 Nilsson et al. Sep 2003 B1
6628713 Kojima et al. Sep 2003 B1
6642967 Saunders Nov 2003 B1
6654933 Abbott et al. Nov 2003 B1
6658056 Duruöz et al. Dec 2003 B1
6665835 Gutfreund et al. Dec 2003 B1
6671408 Kaku Dec 2003 B1
6690838 Zhou Feb 2004 B2
6697568 Kaku Feb 2004 B1
6714909 Gibbon et al. Mar 2004 B1
6721794 Taylor et al. Apr 2004 B2
6724944 Kalevo et al. Apr 2004 B1
6725281 Zintel Apr 2004 B1
6742082 Lango et al. May 2004 B1
6751623 Basso et al. Jun 2004 B1
6771703 Oguz et al. Aug 2004 B1
6807306 Girgensohn et al. Oct 2004 B1
6810031 Hegde et al. Oct 2004 B1
6810131 Nakagawa et al. Oct 2004 B2
6810389 Meyer Oct 2004 B1
6813437 Ando et al. Nov 2004 B2
6819394 Nomura et al. Nov 2004 B1
6850252 Hoffberg Feb 2005 B1
6856997 Lee et al. Feb 2005 B2
6859496 Boroczky et al. Feb 2005 B1
6868525 Szabo Mar 2005 B1
6871006 Oguz et al. Mar 2005 B1
6912253 Li et al. Jun 2005 B1
6912513 Candelore Jun 2005 B1
6917652 Lyu Jul 2005 B2
6931531 Takahashi Aug 2005 B1
6931543 Pang et al. Aug 2005 B1
6944621 Collart Sep 2005 B1
6944629 Shioi et al. Sep 2005 B1
6956901 Boroczky et al. Oct 2005 B2
6957350 Demos Oct 2005 B1
6965646 Firestone Nov 2005 B1
6965724 Boccon-Gibod et al. Nov 2005 B1
6965993 Baker Nov 2005 B2
6970564 Kubota et al. Nov 2005 B1
6983079 Kim Jan 2006 B2
6985588 Glick et al. Jan 2006 B1
6988144 Luken et al. Jan 2006 B1
7006757 Ando et al. Feb 2006 B2
7007170 Morten Feb 2006 B2
7020287 Unger Mar 2006 B2
7023924 Keller et al. Apr 2006 B1
7023992 Kubota et al. Apr 2006 B1
7043021 Graunke et al. May 2006 B2
7043473 Rassool et al. May 2006 B1
7051110 Hagai et al. May 2006 B2
7054968 Shrader et al. May 2006 B2
7058177 Trimberger et al. Jun 2006 B1
7073191 Srikantan et al. Jul 2006 B2
7103906 Katz et al. Sep 2006 B1
7110542 Tripathy Sep 2006 B1
7120250 Candelore Oct 2006 B2
7124303 Candelore et al. Oct 2006 B2
7127155 Ando et al. Oct 2006 B2
7139868 Parry et al. Nov 2006 B2
7143289 Denning et al. Nov 2006 B2
7150045 Koelle et al. Dec 2006 B2
7151832 Fetkovich et al. Dec 2006 B1
7151833 Candelore et al. Dec 2006 B2
7165175 Kollmyer et al. Jan 2007 B1
7167560 Yu Jan 2007 B2
7181438 Szabo Feb 2007 B1
7185363 Narin et al. Feb 2007 B1
7188183 Paul et al. Mar 2007 B1
7191335 Maillard Mar 2007 B1
7197234 Chatterton Mar 2007 B1
7203313 England et al. Apr 2007 B2
7206940 Evans et al. Apr 2007 B2
7209892 Galuten et al. Apr 2007 B1
7212726 Zetts May 2007 B2
7231132 Davenport Jun 2007 B1
7231516 Sparrell et al. Jun 2007 B1
7233669 Candelore Jun 2007 B2
7233948 Shamoon et al. Jun 2007 B1
7237061 Boic Jun 2007 B1
7242772 Tehranchi Jul 2007 B1
7243346 Seth et al. Jul 2007 B1
7274861 Yahata et al. Sep 2007 B2
7295673 Grab et al. Nov 2007 B2
7302490 Gupta et al. Nov 2007 B1
7315829 Tagawa et al. Jan 2008 B1
7328345 Morten et al. Feb 2008 B2
7330875 Parasnis et al. Feb 2008 B1
7340528 Noblecourt et al. Mar 2008 B2
7346163 Pedlow, Jr. et al. Mar 2008 B2
7349886 Morten et al. Mar 2008 B2
7349976 Glaser et al. Mar 2008 B1
7352956 Winter et al. Apr 2008 B1
7356143 Morten Apr 2008 B2
7356245 Belknap et al. Apr 2008 B2
7363647 Fakharzadeh Apr 2008 B1
7366788 Jones et al. Apr 2008 B2
7376233 Candelore et al. May 2008 B2
7376831 Kollmyer et al. May 2008 B2
7382879 Miller Jun 2008 B1
7397853 Kwon et al. Jul 2008 B2
7400679 Kwon et al. Jul 2008 B2
7406174 Palmer Jul 2008 B2
7406176 Zhu et al. Jul 2008 B2
7418132 Hoshuyama Aug 2008 B2
7421411 Kontio et al. Sep 2008 B2
7443449 Momosaki et al. Oct 2008 B2
7454780 Katsube et al. Nov 2008 B2
7457359 Mabey et al. Nov 2008 B2
7457415 Reitmeier et al. Nov 2008 B2
7460668 Grab et al. Dec 2008 B2
7472280 Giobbi Dec 2008 B2
7478325 Foehr et al. Jan 2009 B2
7484103 Woo et al. Jan 2009 B2
7493018 Kim Feb 2009 B2
7499930 Naka et al. Mar 2009 B2
7499938 Collart Mar 2009 B2
7515710 Russell et al. Apr 2009 B2
7526450 Hughes et al. Apr 2009 B2
7539213 Guillemot et al. May 2009 B2
7546641 Robert et al. Jun 2009 B2
7577980 Kienzle et al. Aug 2009 B2
7594271 Zhuk et al. Sep 2009 B2
7610365 Kraft et al. Oct 2009 B1
7623759 Shimoda Nov 2009 B2
7624337 Sull et al. Nov 2009 B2
7627750 Chan Dec 2009 B1
7627888 Ganesan et al. Dec 2009 B2
7639921 Seo et al. Dec 2009 B2
7640358 Deshpande Dec 2009 B2
7640435 Morten Dec 2009 B2
7644172 Stewart et al. Jan 2010 B2
7653686 Yoneda Jan 2010 B2
7664262 Haruki Feb 2010 B2
7664872 Osborne et al. Feb 2010 B2
7676555 Bushee et al. Mar 2010 B2
7689510 Lamkin et al. Mar 2010 B2
7697686 Puiatti et al. Apr 2010 B2
7702925 Hanko et al. Apr 2010 B2
7711052 Hannuksela et al. May 2010 B2
7711647 Gunaseelan et al. May 2010 B2
7720352 Belknap et al. May 2010 B2
7734806 Park Jun 2010 B2
7747853 Candelore Jun 2010 B2
7756270 Shimosato et al. Jul 2010 B2
7756271 Zhu et al. Jul 2010 B2
7761892 Ellis et al. Jul 2010 B2
7779097 Lamkin et al. Aug 2010 B2
7787622 Sprunk Aug 2010 B2
7788271 Soma et al. Aug 2010 B2
7797720 Gopalakrishnan et al. Sep 2010 B2
7817608 Rassool et al. Oct 2010 B2
7840693 Gupta et al. Nov 2010 B2
7853980 Pedlow et al. Dec 2010 B2
7864186 Robotham et al. Jan 2011 B2
7869691 Kelly et al. Jan 2011 B2
7873740 Sitaraman et al. Jan 2011 B2
7877002 Ikeda et al. Jan 2011 B2
7881478 Derouet Feb 2011 B2
7882034 Hug et al. Feb 2011 B2
7885405 Bong Feb 2011 B1
7895311 Juenger Feb 2011 B1
7907833 Lee Mar 2011 B2
7913277 Rahrer Mar 2011 B1
7945143 Yahata et al. May 2011 B2
7949703 Matsuzaki et al. May 2011 B2
7962942 Craner Jun 2011 B1
7970835 St Jun 2011 B2
7974714 Hoffberg Jul 2011 B2
7984513 Kyne et al. Jul 2011 B1
7991156 Miller Aug 2011 B1
8001471 Shaver et al. Aug 2011 B2
8015491 Shaver et al. Sep 2011 B2
8023562 Zheludkov et al. Sep 2011 B2
8046453 Olaiya Oct 2011 B2
8054880 Yu et al. Nov 2011 B2
8065708 Smyth et al. Nov 2011 B1
8069260 Speicher et al. Nov 2011 B2
8073900 Guedalia et al. Dec 2011 B2
8074083 Lee et al. Dec 2011 B1
8078644 Hannuksela Dec 2011 B2
8082442 Keljo et al. Dec 2011 B2
8131875 Chen Mar 2012 B1
8135041 Ramaswamy Mar 2012 B2
8160157 Lamy-Bergot et al. Apr 2012 B2
8169916 Pai et al. May 2012 B1
8170210 Manders et al. May 2012 B2
8195714 Mckibben et al. Jun 2012 B2
8201264 Grab et al. Jun 2012 B2
8213607 Rose et al. Jul 2012 B2
8213768 Morioka et al. Jul 2012 B2
8218439 Deshpande Jul 2012 B2
8225061 Greenebaum Jul 2012 B2
8233768 Soroushian et al. Jul 2012 B2
8243924 Chen et al. Aug 2012 B2
8245124 Gupta Aug 2012 B1
8249168 Graves Aug 2012 B2
8261356 Choi et al. Sep 2012 B2
8265168 Masterson et al. Sep 2012 B1
8270473 Chen et al. Sep 2012 B2
8270819 Vannier Sep 2012 B2
8275871 Ram et al. Sep 2012 B2
8286213 Seo Oct 2012 B2
8286621 Halmone Oct 2012 B2
8289338 Priyadarshi et al. Oct 2012 B2
8290157 Candelore Oct 2012 B2
8291460 Peacock Oct 2012 B1
8296434 Miller et al. Oct 2012 B1
8311094 Kamariotis et al. Nov 2012 B2
8311111 Xu et al. Nov 2012 B2
8311115 Gu et al. Nov 2012 B2
8312079 Newsome et al. Nov 2012 B2
8321556 Chatterjee et al. Nov 2012 B1
8325800 Holcomb et al. Dec 2012 B2
8327009 Prestenback et al. Dec 2012 B2
8341715 Sherkin et al. Dec 2012 B2
8346753 Hayes Jan 2013 B2
8365235 Hunt et al. Jan 2013 B2
8369421 Kadono et al. Feb 2013 B2
8380041 Barton et al. Feb 2013 B2
8386621 Park Feb 2013 B2
8396114 Gu et al. Mar 2013 B2
8397265 Henocq et al. Mar 2013 B2
8401188 Swaminathan Mar 2013 B1
8401900 Cansler et al. Mar 2013 B2
8407753 Kuo Mar 2013 B2
8412841 Swaminathan et al. Apr 2013 B1
8423889 Zagorie et al. Apr 2013 B1
8452110 Shoham et al. May 2013 B2
8456380 Pagan Jun 2013 B2
8464066 Price et al. Jun 2013 B1
8472792 Butt Jun 2013 B2
8473630 Galligan Jun 2013 B1
8484368 Robert et al. Jul 2013 B2
8510303 Soroushian et al. Aug 2013 B2
8510404 Carmel et al. Aug 2013 B2
8514926 Ro et al. Aug 2013 B2
8515265 Kwon et al. Aug 2013 B2
8516529 Lajoie et al. Aug 2013 B2
8526610 Shamoon et al. Sep 2013 B2
8527645 Proffit et al. Sep 2013 B1
8543842 Ginter et al. Sep 2013 B2
8555329 Fröjdh et al. Oct 2013 B2
8571993 Kocher et al. Oct 2013 B2
8595378 Cohn et al. Nov 2013 B1
8606069 Okubo et al. Dec 2013 B2
8630419 Mori Jan 2014 B2
8631247 O'loughlin et al. Jan 2014 B2
8640166 Craner et al. Jan 2014 B1
8649669 Braness et al. Feb 2014 B2
8650599 Shindo et al. Feb 2014 B2
8656183 Russell et al. Feb 2014 B2
8677428 Lewis et al. Mar 2014 B2
8681866 Jia Mar 2014 B1
8683066 Hurst et al. Mar 2014 B2
8689267 Hunt Apr 2014 B2
8726264 Allen et al. May 2014 B1
8731193 Farkash et al. May 2014 B2
8731369 Li et al. May 2014 B2
RE45052 Li Jul 2014 E
8767825 Wang et al. Jul 2014 B1
8774609 Drake et al. Jul 2014 B2
8781122 Chan et al. Jul 2014 B2
8782268 Pyle et al. Jul 2014 B2
8804956 Hiriart Aug 2014 B2
8805109 Shoham et al. Aug 2014 B2
8806188 Braness et al. Aug 2014 B2
8818896 Candelore Aug 2014 B2
8819116 Tomay et al. Aug 2014 B1
8832434 Apostolopoulos et al. Sep 2014 B2
8843586 Pantos et al. Sep 2014 B2
8849950 Stockhammer et al. Sep 2014 B2
8850205 Choi et al. Sep 2014 B2
8850498 Roach et al. Sep 2014 B1
8856218 Inskip Oct 2014 B1
8908984 Shoham et al. Dec 2014 B2
8909922 Kiefer et al. Dec 2014 B2
8914534 Braness et al. Dec 2014 B2
8914836 Shivadas et al. Dec 2014 B2
8918533 Chen et al. Dec 2014 B2
8918535 Ma et al. Dec 2014 B2
8918636 Kiefer Dec 2014 B2
8918908 Ziskind et al. Dec 2014 B2
8948249 Sun et al. Feb 2015 B2
8964977 Ziskind et al. Feb 2015 B2
8997161 Priyadarshi et al. Mar 2015 B2
8997254 Amidei et al. Mar 2015 B2
9014471 Shoham et al. Apr 2015 B2
9015782 Acharya et al. Apr 2015 B2
9025659 Soroushian et al. May 2015 B2
9038116 Knox et al. May 2015 B1
9038121 Kienzle et al. May 2015 B2
9042670 Carmel et al. May 2015 B2
9049497 Chen et al. Jun 2015 B2
9060207 Scherkus et al. Jun 2015 B2
9094737 Shivadas et al. Jul 2015 B2
9098335 Muthiah et al. Aug 2015 B2
9111098 Smith et al. Aug 2015 B2
9124773 Chan et al. Sep 2015 B2
9125073 Oyman et al. Sep 2015 B2
9184920 Grab et al. Nov 2015 B2
9191151 Luby et al. Nov 2015 B2
9191457 Van der Schaar Nov 2015 B2
9197685 Soroushian Nov 2015 B2
9201922 Soroushian et al. Dec 2015 B2
9203816 Brueck et al. Dec 2015 B2
9210481 Braness et al. Dec 2015 B2
9215466 Zhai et al. Dec 2015 B2
9247311 Kiefer Jan 2016 B2
9247312 Braness et al. Jan 2016 B2
9247317 Shivadas et al. Jan 2016 B2
9253178 Blom et al. Feb 2016 B2
9264475 Shivadas et al. Feb 2016 B2
9294531 Zhang et al. Mar 2016 B2
9313510 Shivadas et al. Apr 2016 B2
9343112 Amidei et al. May 2016 B2
9344517 Shivadas et al. May 2016 B2
9344721 Dikvall May 2016 B2
9380096 Luby et al. Jun 2016 B2
9386064 Luby et al. Jul 2016 B2
9479805 Rothschild et al. Oct 2016 B2
9485469 Kahn et al. Nov 2016 B2
9485546 Chen et al. Nov 2016 B2
9571827 Su et al. Feb 2017 B2
9584557 Panje et al. Feb 2017 B2
9584847 Ma et al. Feb 2017 B2
9615061 Carney et al. Apr 2017 B2
9621522 Kiefer et al. Apr 2017 B2
9628536 Luby et al. Apr 2017 B2
9667684 Ziskind et al. May 2017 B2
9672286 Soroushian et al. Jun 2017 B2
9674254 Pare et al. Jun 2017 B2
9686332 Binns et al. Jun 2017 B1
9706259 Chan et al. Jul 2017 B2
9712890 Shivadas et al. Jul 2017 B2
9761274 Delpuch et al. Sep 2017 B2
9798863 Grab et al. Oct 2017 B2
9813740 Panje et al. Nov 2017 B2
9866878 van der Schaar et al. Jan 2018 B2
9883204 Braness et al. Jan 2018 B2
9906785 Naletov et al. Feb 2018 B2
9967189 Patel et al. May 2018 B2
9967305 Braness May 2018 B2
9967521 Kahn et al. May 2018 B2
10169094 Iyer Jan 2019 B2
10171873 Krebs Jan 2019 B2
10212486 Chan et al. Feb 2019 B2
10225299 van der Schaar et al. Mar 2019 B2
10225588 Kiefer et al. Mar 2019 B2
10244272 Kiefer et al. Mar 2019 B2
10264255 Naletov et al. Apr 2019 B2
10321168 van der Schaar et al. Jun 2019 B2
10341698 Kiefer et al. Jul 2019 B2
10368096 Braness et al. Jul 2019 B2
10382785 Braness et al. Aug 2019 B2
10437896 Soroushian et al. Oct 2019 B2
10462537 Shivadas et al. Oct 2019 B2
10715806 Naletov et al. Jul 2020 B2
10805368 van der Schaar et al. Oct 2020 B2
10856020 Kiefer et al. Dec 2020 B2
10893305 van der Schaar et al. Jan 2021 B2
10992955 Braness et al. Apr 2021 B2
RE48761 Shivadas et al. Sep 2021 E
11638033 Braness et al. Apr 2023 B2
20010021276 Zhou Sep 2001 A1
20010030710 Werner Oct 2001 A1
20010036355 Kelly et al. Nov 2001 A1
20010046299 Wasilewski et al. Nov 2001 A1
20010052077 Fung et al. Dec 2001 A1
20010052127 Seo et al. Dec 2001 A1
20010053222 Wakao et al. Dec 2001 A1
20010055337 Matsuzaki et al. Dec 2001 A1
20020026560 Jordan et al. Feb 2002 A1
20020034252 Owen et al. Mar 2002 A1
20020048450 Zetts Apr 2002 A1
20020051494 Yamaguchi et al. May 2002 A1
20020057739 Hasebe et al. May 2002 A1
20020057898 Normile May 2002 A1
20020062313 Lee et al. May 2002 A1
20020067432 Kondo et al. Jun 2002 A1
20020075572 Boreczky et al. Jun 2002 A1
20020076112 Devara Jun 2002 A1
20020087569 Fischer et al. Jul 2002 A1
20020091665 Beek et al. Jul 2002 A1
20020093571 Hyodo Jul 2002 A1
20020107802 Philips Aug 2002 A1
20020110193 Yoo et al. Aug 2002 A1
20020114330 Cheung et al. Aug 2002 A1
20020116481 Lee Aug 2002 A1
20020118953 Kim Aug 2002 A1
20020120934 Abrahams et al. Aug 2002 A1
20020135607 Kato et al. Sep 2002 A1
20020136298 Anantharamu et al. Sep 2002 A1
20020141503 Kobayashi et al. Oct 2002 A1
20020143413 Fay et al. Oct 2002 A1
20020143547 Fay et al. Oct 2002 A1
20020147980 Satoda Oct 2002 A1
20020154779 Asano et al. Oct 2002 A1
20020159528 Graziani et al. Oct 2002 A1
20020159598 Rubinstein et al. Oct 2002 A1
20020161462 Fay Oct 2002 A1
20020161797 Gallo et al. Oct 2002 A1
20020164024 Arakawa et al. Nov 2002 A1
20020169926 Pinckney et al. Nov 2002 A1
20020169971 Asano et al. Nov 2002 A1
20020180929 Tseng et al. Dec 2002 A1
20020184159 Tadayon et al. Dec 2002 A1
20020184515 Oho et al. Dec 2002 A1
20020191112 Akiyoshi et al. Dec 2002 A1
20020191959 Lin et al. Dec 2002 A1
20020191960 Fujinami et al. Dec 2002 A1
20030001964 Masukura et al. Jan 2003 A1
20030002577 Pinder Jan 2003 A1
20030002578 Tsukagoshi et al. Jan 2003 A1
20030005442 Brodersen et al. Jan 2003 A1
20030021296 Wee et al. Jan 2003 A1
20030031178 Haeri Feb 2003 A1
20030035488 Barrau Feb 2003 A1
20030035545 Jiang Feb 2003 A1
20030035546 Jiang et al. Feb 2003 A1
20030041257 Wee et al. Feb 2003 A1
20030043847 Haddad Mar 2003 A1
20030044080 Frishman et al. Mar 2003 A1
20030051237 Sako et al. Mar 2003 A1
20030053541 Sun et al. Mar 2003 A1
20030061305 Copley et al. Mar 2003 A1
20030061369 Aksu et al. Mar 2003 A1
20030063675 Kang et al. Apr 2003 A1
20030065777 Mattila et al. Apr 2003 A1
20030077071 Lin et al. Apr 2003 A1
20030078891 Capitant Apr 2003 A1
20030078930 Surcouf et al. Apr 2003 A1
20030079222 Boykin et al. Apr 2003 A1
20030081776 Candelore May 2003 A1
20030093799 Kauffman et al. May 2003 A1
20030123855 Okada et al. Jul 2003 A1
20030128296 Lee Jul 2003 A1
20030133506 Haneda Jul 2003 A1
20030135633 Dror et al. Jul 2003 A1
20030135742 Evans Jul 2003 A1
20030142594 Tsumagari et al. Jul 2003 A1
20030152224 Candelore et al. Aug 2003 A1
20030152370 Otomo et al. Aug 2003 A1
20030163824 Gordon et al. Aug 2003 A1
20030165328 Grecia Sep 2003 A1
20030174844 Candelore Sep 2003 A1
20030185302 Abrams Oct 2003 A1
20030185542 McVeigh et al. Oct 2003 A1
20030206558 Parkkinen et al. Nov 2003 A1
20030206717 Yogeshwar et al. Nov 2003 A1
20030210821 Yogeshwar et al. Nov 2003 A1
20030216922 Gonzales et al. Nov 2003 A1
20030229900 Reisman Dec 2003 A1
20030231863 Eerenberg et al. Dec 2003 A1
20030231867 Gates et al. Dec 2003 A1
20030233464 Walpole et al. Dec 2003 A1
20030236836 Borthwick Dec 2003 A1
20030236907 Stewart et al. Dec 2003 A1
20040001594 Krishnaswamy et al. Jan 2004 A1
20040003008 Wasilewski et al. Jan 2004 A1
20040006701 Kresina Jan 2004 A1
20040021684 Millner Feb 2004 A1
20040022391 Obrien Feb 2004 A1
20040024688 Bi et al. Feb 2004 A1
20040025180 Begeja et al. Feb 2004 A1
20040028227 Yu Feb 2004 A1
20040031058 Reisman Feb 2004 A1
20040037421 Truman Feb 2004 A1
20040039916 Aldis et al. Feb 2004 A1
20040047592 Seo et al. Mar 2004 A1
20040047607 Seo et al. Mar 2004 A1
20040047614 Green Mar 2004 A1
20040049690 Candelore et al. Mar 2004 A1
20040049694 Candelore Mar 2004 A1
20040052501 Tam Mar 2004 A1
20040071453 Valderas Apr 2004 A1
20040073917 Pedlow et al. Apr 2004 A1
20040076237 Kadono et al. Apr 2004 A1
20040081333 Grab et al. Apr 2004 A1
20040081434 Jung et al. Apr 2004 A1
20040084035 Newton May 2004 A1
20040088557 Malcolm et al. May 2004 A1
20040093494 Nishimoto et al. May 2004 A1
20040093618 Baldwin et al. May 2004 A1
20040101059 Joch et al. May 2004 A1
20040101142 Nasypny May 2004 A1
20040105549 Suzuki et al. Jun 2004 A1
20040107356 Shamoon et al. Jun 2004 A1
20040114687 Ferris et al. Jun 2004 A1
20040117347 Seo et al. Jun 2004 A1
20040136698 Mock Jul 2004 A1
20040139335 Diamand et al. Jul 2004 A1
20040143760 Alkove et al. Jul 2004 A1
20040146276 Ogawa Jul 2004 A1
20040158878 Ratnakar et al. Aug 2004 A1
20040184534 Wang Sep 2004 A1
20040184616 Morten et al. Sep 2004 A1
20040202320 Amini et al. Oct 2004 A1
20040213094 Suzuki Oct 2004 A1
20040217971 Kim Nov 2004 A1
20040243488 Yamamoto et al. Dec 2004 A1
20040243714 Wynn et al. Dec 2004 A1
20040255115 DeMello et al. Dec 2004 A1
20040255236 Collart Dec 2004 A1
20040267952 He et al. Dec 2004 A1
20050004875 Kontio et al. Jan 2005 A1
20050005025 Harville et al. Jan 2005 A1
20050005143 Lang et al. Jan 2005 A1
20050013494 Srinivasan et al. Jan 2005 A1
20050015509 Sitaraman et al. Jan 2005 A1
20050015797 Noblecourt et al. Jan 2005 A1
20050038826 Bae et al. Feb 2005 A1
20050055399 Savchuk Mar 2005 A1
20050063541 Candelore Mar 2005 A1
20050066063 Grigorovitch et al. Mar 2005 A1
20050071280 Irwin et al. Mar 2005 A1
20050076232 Kawaguchi Apr 2005 A1
20050102371 Aksu May 2005 A1
20050108320 Lord et al. May 2005 A1
20050114534 Lee May 2005 A1
20050114896 Hug May 2005 A1
20050120132 Hutter Jun 2005 A1
20050132208 Hug et al. Jun 2005 A1
20050138655 Zimler et al. Jun 2005 A1
20050144468 Northcutt Jun 2005 A1
20050149450 Stefik et al. Jul 2005 A1
20050177741 Chen et al. Aug 2005 A1
20050180641 Clark Aug 2005 A1
20050183120 Jain et al. Aug 2005 A1
20050190911 Pare et al. Sep 2005 A1
20050192904 Candelore Sep 2005 A1
20050193070 Brown et al. Sep 2005 A1
20050193322 Lamkin et al. Sep 2005 A1
20050196147 Seo et al. Sep 2005 A1
20050198364 Val et al. Sep 2005 A1
20050204289 Mohammed et al. Sep 2005 A1
20050207442 Zoest et al. Sep 2005 A1
20050207578 Matsuyama et al. Sep 2005 A1
20050216752 Hofmeyr et al. Sep 2005 A1
20050227773 Lu et al. Oct 2005 A1
20050243912 Kwon et al. Nov 2005 A1
20050254508 Aksu et al. Nov 2005 A1
20050262257 Major et al. Nov 2005 A1
20050265555 Pippuri Dec 2005 A1
20050273695 Schnurr Dec 2005 A1
20050275656 Corbin et al. Dec 2005 A1
20060013568 Rodriguez Jan 2006 A1
20060015580 Gabriel et al. Jan 2006 A1
20060020825 Grab Jan 2006 A1
20060026294 Virdi et al. Feb 2006 A1
20060026302 Bennett et al. Feb 2006 A1
20060026654 An et al. Feb 2006 A1
20060036549 Wu Feb 2006 A1
20060037057 Xu Feb 2006 A1
20060052095 Vazvan Mar 2006 A1
20060053080 Edmonson et al. Mar 2006 A1
20060059223 Klemets et al. Mar 2006 A1
20060064605 Giobbi Mar 2006 A1
20060078301 Ikeda et al. Apr 2006 A1
20060093318 Cohen et al. May 2006 A1
20060093320 Hallberg et al. May 2006 A1
20060095472 Krikorian et al. May 2006 A1
20060109856 Deshpande May 2006 A1
20060120378 Usuki et al. Jun 2006 A1
20060129909 Butt et al. Jun 2006 A1
20060165163 Burazerovic et al. Jul 2006 A1
20060165233 Nonaka et al. Jul 2006 A1
20060168298 Aoki et al. Jul 2006 A1
20060168639 Gan et al. Jul 2006 A1
20060173887 Breitfeld et al. Aug 2006 A1
20060179239 Fluhr et al. Aug 2006 A1
20060181965 Collart Aug 2006 A1
20060210245 Mccrossan et al. Sep 2006 A1
20060212370 Shear et al. Sep 2006 A1
20060218251 Tanabe Sep 2006 A1
20060235880 Qian Oct 2006 A1
20060235883 Krebs Oct 2006 A1
20060245727 Nakano et al. Nov 2006 A1
20060259588 Lerman et al. Nov 2006 A1
20060263056 Lin et al. Nov 2006 A1
20060267986 Bae Nov 2006 A1
20060274835 Hamilton et al. Dec 2006 A1
20060294164 Armangau et al. Dec 2006 A1
20070005333 Setiohardjo et al. Jan 2007 A1
20070031110 Rijckaert Feb 2007 A1
20070033419 Kocher et al. Feb 2007 A1
20070044010 Sull et al. Feb 2007 A1
20070047645 Takashima Mar 2007 A1
20070047901 Ando et al. Mar 2007 A1
20070053513 Hoffberg Mar 2007 A1
20070055982 Spilo Mar 2007 A1
20070058928 Naito et al. Mar 2007 A1
20070067472 Maertens et al. Mar 2007 A1
20070067622 Nakano et al. Mar 2007 A1
20070083467 Lindahl et al. Apr 2007 A1
20070083617 Chakrabarti et al. Apr 2007 A1
20070086528 Mauchly et al. Apr 2007 A1
20070100757 Rhoads May 2007 A1
20070101271 Hua et al. May 2007 A1
20070101387 Hua et al. May 2007 A1
20070106863 Bonwick et al. May 2007 A1
20070133603 Weaver Jun 2007 A1
20070136817 Nguyen Jun 2007 A1
20070140647 Kusunoki et al. Jun 2007 A1
20070154165 Hemmeryckx-Deleersnijder et al. Jul 2007 A1
20070156770 Espelien Jul 2007 A1
20070157267 Lopez-Estrada Jul 2007 A1
20070162568 Gupta et al. Jul 2007 A1
20070162981 Morioka et al. Jul 2007 A1
20070166000 Nallur et al. Jul 2007 A1
20070168541 Gupta et al. Jul 2007 A1
20070168542 Gupta et al. Jul 2007 A1
20070178933 Nelson Aug 2007 A1
20070180051 Kelly et al. Aug 2007 A1
20070180125 Knowles et al. Aug 2007 A1
20070185982 Nakanowatari et al. Aug 2007 A1
20070192810 Pritchett et al. Aug 2007 A1
20070201502 Abramson Aug 2007 A1
20070201695 Saarikivi Aug 2007 A1
20070204003 Abramson Aug 2007 A1
20070204011 Shaver et al. Aug 2007 A1
20070204115 Abramson Aug 2007 A1
20070217339 Zhao Sep 2007 A1
20070217759 Dodd Sep 2007 A1
20070220118 Loyer Sep 2007 A1
20070234391 Hunter et al. Oct 2007 A1
20070239839 Buday et al. Oct 2007 A1
20070250536 Tanaka et al. Oct 2007 A1
20070255940 Ueno Nov 2007 A1
20070256141 Nakano et al. Nov 2007 A1
20070271317 Carmel Nov 2007 A1
20070271385 Davis et al. Nov 2007 A1
20070274679 Yahata et al. Nov 2007 A1
20070277219 Toebes et al. Nov 2007 A1
20070277234 Bessonov et al. Nov 2007 A1
20070280298 Hearn et al. Dec 2007 A1
20070288745 Kwan Dec 2007 A1
20070292107 Yahata et al. Dec 2007 A1
20070297422 Matsuo et al. Dec 2007 A1
20080005175 Bourke et al. Jan 2008 A1
20080008319 Poirier Jan 2008 A1
20080008455 De Lange et al. Jan 2008 A1
20080022005 Wu et al. Jan 2008 A1
20080043832 Barkley et al. Feb 2008 A1
20080046718 Grab et al. Feb 2008 A1
20080046925 Lee et al. Feb 2008 A1
20080066099 Brodersen et al. Mar 2008 A1
20080066181 Haveson et al. Mar 2008 A1
20080077592 Brodie et al. Mar 2008 A1
20080086456 Rasanen et al. Apr 2008 A1
20080086570 Dey et al. Apr 2008 A1
20080086747 Rasanen et al. Apr 2008 A1
20080101466 Swenson et al. May 2008 A1
20080101718 Yang et al. May 2008 A1
20080104633 Noblecourt et al. May 2008 A1
20080120330 Reed et al. May 2008 A1
20080120342 Reed et al. May 2008 A1
20080120389 Bassali et al. May 2008 A1
20080120637 Deiss May 2008 A1
20080126248 Lee et al. May 2008 A1
20080131078 Jeong et al. Jun 2008 A1
20080134043 Georgis Jun 2008 A1
20080137541 Agarwal et al. Jun 2008 A1
20080137736 Richardson et al. Jun 2008 A1
20080137847 Candelore et al. Jun 2008 A1
20080151817 Fitchett Jun 2008 A1
20080155615 Craner et al. Jun 2008 A1
20080160911 Chou et al. Jul 2008 A1
20080162949 Sato et al. Jul 2008 A1
20080168516 Flick et al. Jul 2008 A1
20080172441 Speicher et al. Jul 2008 A1
20080177793 Epstein et al. Jul 2008 A1
20080184119 Eyal et al. Jul 2008 A1
20080187283 Takahashi Aug 2008 A1
20080192818 DiPietro et al. Aug 2008 A1
20080195664 Maharajh et al. Aug 2008 A1
20080195744 Bowra et al. Aug 2008 A1
20080201705 Wookey Aug 2008 A1
20080205860 Holtman Aug 2008 A1
20080209534 Keronen et al. Aug 2008 A1
20080219449 Ball et al. Sep 2008 A1
20080229025 Plamondon Sep 2008 A1
20080240144 Kruse et al. Oct 2008 A1
20080256105 Nogawa et al. Oct 2008 A1
20080263354 Beuque Oct 2008 A1
20080271102 Kienzle et al. Oct 2008 A1
20080279535 Haque et al. Nov 2008 A1
20080294453 Baird-Smith et al. Nov 2008 A1
20080298358 John et al. Dec 2008 A1
20080310454 Bellwood et al. Dec 2008 A1
20080310496 Fang Dec 2008 A1
20080313541 Shafton et al. Dec 2008 A1
20080320100 Pantos et al. Dec 2008 A1
20080320160 Sitaraman et al. Dec 2008 A1
20090010429 Kim et al. Jan 2009 A1
20090010622 Yahata et al. Jan 2009 A1
20090013195 Ochi et al. Jan 2009 A1
20090031220 Tranchant et al. Jan 2009 A1
20090037959 Suh et al. Feb 2009 A1
20090048852 Burns et al. Feb 2009 A1
20090055546 Jung et al. Feb 2009 A1
20090060452 Chaudhri Mar 2009 A1
20090064341 Hartung et al. Mar 2009 A1
20090066839 Jung et al. Mar 2009 A1
20090067367 Buracchini et al. Mar 2009 A1
20090077143 Macy, Jr. Mar 2009 A1
20090097644 Haruki Apr 2009 A1
20090106082 Senti et al. Apr 2009 A1
20090132599 Soroushian et al. May 2009 A1
20090132721 Soroushian et al. May 2009 A1
20090132824 Terada et al. May 2009 A1
20090136216 Soroushian et al. May 2009 A1
20090138570 Miura et al. May 2009 A1
20090150406 Giblin Jun 2009 A1
20090150557 Wormley et al. Jun 2009 A1
20090165148 Frey et al. Jun 2009 A1
20090168795 Segel et al. Jul 2009 A1
20090169001 Tighe et al. Jul 2009 A1
20090169181 Priyadarshi et al. Jul 2009 A1
20090172201 Carmel et al. Jul 2009 A1
20090178090 Oztaskent Jul 2009 A1
20090196139 Bates et al. Aug 2009 A1
20090201988 Gazier et al. Aug 2009 A1
20090217317 White et al. Aug 2009 A1
20090226148 Nesvadba et al. Sep 2009 A1
20090228395 Wegner et al. Sep 2009 A1
20090249081 Zayas Oct 2009 A1
20090265737 Issa et al. Oct 2009 A1
20090268905 Matsushima et al. Oct 2009 A1
20090276636 Grab et al. Nov 2009 A1
20090282162 Mehrotra et al. Nov 2009 A1
20090290706 Amini et al. Nov 2009 A1
20090290708 Schneider et al. Nov 2009 A1
20090293116 DeMello Nov 2009 A1
20090303241 Priyadarshi et al. Dec 2009 A1
20090307258 Priyadarshi et al. Dec 2009 A1
20090307267 Chen et al. Dec 2009 A1
20090310819 Hatano Dec 2009 A1
20090310933 Lee Dec 2009 A1
20090313544 Wood et al. Dec 2009 A1
20090313564 Rottler et al. Dec 2009 A1
20090316783 Au et al. Dec 2009 A1
20090328124 Khouzam et al. Dec 2009 A1
20090328228 Schnell Dec 2009 A1
20100005393 Tokashiki et al. Jan 2010 A1
20100040351 Toma et al. Feb 2010 A1
20100057928 Kapoor et al. Mar 2010 A1
20100058405 Ramakrishnan et al. Mar 2010 A1
20100074324 Qian et al. Mar 2010 A1
20100074333 Au et al. Mar 2010 A1
20100083322 Rouse Apr 2010 A1
20100094969 Zuckerman et al. Apr 2010 A1
20100095121 Shetty et al. Apr 2010 A1
20100106968 Mori et al. Apr 2010 A1
20100107260 Orrell et al. Apr 2010 A1
20100111192 Graves May 2010 A1
20100138903 Medvinsky Jun 2010 A1
20100142915 Mcdermott et al. Jun 2010 A1
20100142917 Isaji Jun 2010 A1
20100158109 Dahlby et al. Jun 2010 A1
20100161825 Ronca et al. Jun 2010 A1
20100166060 Ezure et al. Jul 2010 A1
20100185854 Burns et al. Jul 2010 A1
20100186092 Takechi et al. Jul 2010 A1
20100189183 Gu et al. Jul 2010 A1
20100198943 Harrang et al. Aug 2010 A1
20100218208 Holden Aug 2010 A1
20100228795 Hahn Sep 2010 A1
20100235472 Sood et al. Sep 2010 A1
20100235528 Bocharov et al. Sep 2010 A1
20100250532 Soroushian et al. Sep 2010 A1
20100290761 Drake et al. Nov 2010 A1
20100299522 Khambete et al. Nov 2010 A1
20100306249 Hill et al. Dec 2010 A1
20100313225 Cholas et al. Dec 2010 A1
20100313226 Cholas et al. Dec 2010 A1
20100316126 Chen et al. Dec 2010 A1
20100319014 Lockett et al. Dec 2010 A1
20100319017 Cook Dec 2010 A1
20100332595 Fullagar et al. Dec 2010 A1
20110002381 Yang et al. Jan 2011 A1
20110010466 Fan et al. Jan 2011 A1
20110016225 Park et al. Jan 2011 A1
20110035517 Minnick et al. Feb 2011 A1
20110047209 Lindholm et al. Feb 2011 A1
20110055585 Lee Mar 2011 A1
20110058675 Brueck et al. Mar 2011 A1
20110060808 Martin et al. Mar 2011 A1
20110066673 Outlaw Mar 2011 A1
20110067057 Karaoguz et al. Mar 2011 A1
20110069757 Ammu et al. Mar 2011 A1
20110078440 Feng et al. Mar 2011 A1
20110080940 Bocharov Apr 2011 A1
20110082914 Robert et al. Apr 2011 A1
20110082924 Gopalakrishnan Apr 2011 A1
20110083009 Shamoon et al. Apr 2011 A1
20110096828 Chen et al. Apr 2011 A1
20110103374 Lajoie et al. May 2011 A1
20110107379 Lajoie et al. May 2011 A1
20110116772 Kwon et al. May 2011 A1
20110126191 Hughes et al. May 2011 A1
20110129011 Cilli et al. Jun 2011 A1
20110135090 Chan et al. Jun 2011 A1
20110138018 Raveendran et al. Jun 2011 A1
20110142415 Rhyu Jun 2011 A1
20110145726 Wei et al. Jun 2011 A1
20110145858 Philpott et al. Jun 2011 A1
20110149753 Bapst et al. Jun 2011 A1
20110150100 Abadir Jun 2011 A1
20110153785 Minborg et al. Jun 2011 A1
20110153835 Rimac et al. Jun 2011 A1
20110158470 Martin et al. Jun 2011 A1
20110170687 Hyodo et al. Jul 2011 A1
20110173345 Knox et al. Jul 2011 A1
20110179185 Wang et al. Jul 2011 A1
20110184738 Kalisky et al. Jul 2011 A1
20110191439 Dazzi et al. Aug 2011 A1
20110191803 Baldwin et al. Aug 2011 A1
20110197237 Turner Aug 2011 A1
20110197261 Dong et al. Aug 2011 A1
20110197267 Gravel et al. Aug 2011 A1
20110213827 Kaspar et al. Sep 2011 A1
20110222786 Carmel et al. Sep 2011 A1
20110225302 Park et al. Sep 2011 A1
20110225315 Wexler et al. Sep 2011 A1
20110225417 Maharajh et al. Sep 2011 A1
20110238789 Luby et al. Sep 2011 A1
20110239078 Luby et al. Sep 2011 A1
20110246657 Glow Oct 2011 A1
20110246659 Bouazizi Oct 2011 A1
20110246661 Manzari et al. Oct 2011 A1
20110252118 Pantos et al. Oct 2011 A1
20110264530 Santangelo et al. Oct 2011 A1
20110268178 Park Nov 2011 A1
20110276555 Fiero Nov 2011 A1
20110276695 Maldaner et al. Nov 2011 A1
20110283012 Melnyk Nov 2011 A1
20110291723 Hashimoto Dec 2011 A1
20110296048 Knox et al. Dec 2011 A1
20110302319 Ha et al. Dec 2011 A1
20110305273 He et al. Dec 2011 A1
20110314130 Strasman Dec 2011 A1
20110314176 Frojdh et al. Dec 2011 A1
20110314500 Gordon et al. Dec 2011 A1
20120005312 Mcgowan et al. Jan 2012 A1
20120005368 Knittle Jan 2012 A1
20120017282 Kang et al. Jan 2012 A1
20120023251 Pyle et al. Jan 2012 A1
20120036365 Kyslov et al. Feb 2012 A1
20120036544 Chen et al. Feb 2012 A1
20120042090 Chen et al. Feb 2012 A1
20120047542 Lewis et al. Feb 2012 A1
20120066360 Ghosh Mar 2012 A1
20120093214 Urbach Apr 2012 A1
20120110120 Willig et al. May 2012 A1
20120114302 Randall May 2012 A1
20120124191 Lyon May 2012 A1
20120134496 Farkash et al. May 2012 A1
20120137336 Applegate et al. May 2012 A1
20120144117 Weare et al. Jun 2012 A1
20120144445 Bonta et al. Jun 2012 A1
20120147958 Ronca et al. Jun 2012 A1
20120166633 Baumback et al. Jun 2012 A1
20120167132 Mathews et al. Jun 2012 A1
20120170642 Braness et al. Jul 2012 A1
20120170643 Soroushian et al. Jul 2012 A1
20120170906 Soroushian et al. Jul 2012 A1
20120170915 Braness et al. Jul 2012 A1
20120173751 Braness et al. Jul 2012 A1
20120177101 van der Schaar Jul 2012 A1
20120179834 Van Der Jul 2012 A1
20120188069 Colombo et al. Jul 2012 A1
20120189069 Iannuzzelli et al. Jul 2012 A1
20120201475 Carmel et al. Aug 2012 A1
20120201476 Carmel et al. Aug 2012 A1
20120233345 Hannuksela Sep 2012 A1
20120240176 Ma et al. Sep 2012 A1
20120254455 Adimatyam et al. Oct 2012 A1
20120257678 Zhou et al. Oct 2012 A1
20120260277 Kosciewicz Oct 2012 A1
20120263434 Wainner et al. Oct 2012 A1
20120265562 Daouk et al. Oct 2012 A1
20120278496 Hsu Nov 2012 A1
20120288015 Zhang et al. Nov 2012 A1
20120289147 Raleigh et al. Nov 2012 A1
20120294355 Holcomb et al. Nov 2012 A1
20120297039 Acuna et al. Nov 2012 A1
20120307883 Graves Dec 2012 A1
20120311094 Biderman et al. Dec 2012 A1
20120311174 Bichot et al. Dec 2012 A1
20120314778 Salustri et al. Dec 2012 A1
20120317235 Nguyen et al. Dec 2012 A1
20120331167 Hunt Dec 2012 A1
20130007223 Luby et al. Jan 2013 A1
20130013730 Li et al. Jan 2013 A1
20130013803 Bichot et al. Jan 2013 A1
20130019107 Grab et al. Jan 2013 A1
20130019273 Ma et al. Jan 2013 A1
20130041808 Pham et al. Feb 2013 A1
20130044821 Braness et al. Feb 2013 A1
20130046849 Wolf Feb 2013 A1
20130046902 Villegas Nuñez et al. Feb 2013 A1
20130051554 Braness et al. Feb 2013 A1
20130054958 Braness et al. Feb 2013 A1
20130055084 Soroushian et al. Feb 2013 A1
20130058480 Ziskind et al. Mar 2013 A1
20130061040 Kiefer et al. Mar 2013 A1
20130061045 Kiefer et al. Mar 2013 A1
20130064466 Carmel et al. Mar 2013 A1
20130066838 Singla et al. Mar 2013 A1
20130080267 McGowan Mar 2013 A1
20130094565 Yang et al. Apr 2013 A1
20130097309 Ma et al. Apr 2013 A1
20130114944 Soroushian et al. May 2013 A1
20130124859 Pestoni et al. May 2013 A1
20130128962 Rajagopalan et al. May 2013 A1
20130152767 Katz et al. Jun 2013 A1
20130159633 Lilly Jun 2013 A1
20130166580 Maharajh Jun 2013 A1
20130166765 Kaufman Jun 2013 A1
20130166906 Swaminathan et al. Jun 2013 A1
20130170561 Hannuksela Jul 2013 A1
20130170764 Carmel et al. Jul 2013 A1
20130173513 Chu et al. Jul 2013 A1
20130179199 Ziskind et al. Jul 2013 A1
20130179589 McCarthy et al. Jul 2013 A1
20130179992 Ziskind et al. Jul 2013 A1
20130182952 Carmel et al. Jul 2013 A1
20130196292 Brennen et al. Aug 2013 A1
20130212228 Butler et al. Aug 2013 A1
20130223812 Rossi Aug 2013 A1
20130226578 Bolton et al. Aug 2013 A1
20130226635 Fisher Aug 2013 A1
20130227081 Luby et al. Aug 2013 A1
20130227111 Wright et al. Aug 2013 A1
20130227122 Gao et al. Aug 2013 A1
20130297602 Soroushian et al. Nov 2013 A1
20130301424 Kotecha et al. Nov 2013 A1
20130311670 Tarbox et al. Nov 2013 A1
20130329781 Su et al. Dec 2013 A1
20140003516 Soroushian Jan 2014 A1
20140019592 Arana et al. Jan 2014 A1
20140019593 Reznik et al. Jan 2014 A1
20140037620 Ferree et al. Feb 2014 A1
20140052823 Gavade et al. Feb 2014 A1
20140059156 Freeman, II et al. Feb 2014 A1
20140096171 Shivadas et al. Apr 2014 A1
20140096269 Amidei et al. Apr 2014 A1
20140101722 Moore Apr 2014 A1
20140114951 Sasaki et al. Apr 2014 A1
20140115650 Zhang et al. Apr 2014 A1
20140119432 Wang et al. May 2014 A1
20140140253 Lohmar et al. May 2014 A1
20140140396 Wang et al. May 2014 A1
20140140417 Shaffer et al. May 2014 A1
20140143301 Watson et al. May 2014 A1
20140143431 Watson et al. May 2014 A1
20140143440 Ramamurthy et al. May 2014 A1
20140149557 Lohmar et al. May 2014 A1
20140177734 Carmel et al. Jun 2014 A1
20140189065 Schaar et al. Jul 2014 A1
20140201382 Shivadas et al. Jul 2014 A1
20140211840 Butt et al. Jul 2014 A1
20140211859 Carmel et al. Jul 2014 A1
20140241420 Orton-jay et al. Aug 2014 A1
20140241421 Orton-jay et al. Aug 2014 A1
20140247869 Su et al. Sep 2014 A1
20140250473 Braness et al. Sep 2014 A1
20140258714 Grab Sep 2014 A1
20140269927 Naletov et al. Sep 2014 A1
20140269936 Shivadas et al. Sep 2014 A1
20140280763 Grab et al. Sep 2014 A1
20140297804 Shivadas et al. Oct 2014 A1
20140297881 Shivadas et al. Oct 2014 A1
20140355668 Shoham et al. Dec 2014 A1
20140359678 Shivadas et al. Dec 2014 A1
20140359679 Shivadas et al. Dec 2014 A1
20140359680 Shivadas et al. Dec 2014 A1
20140376720 Chan et al. Dec 2014 A1
20150006662 Braness Jan 2015 A1
20150019550 Maharajh et al. Jan 2015 A1
20150026677 Stevens et al. Jan 2015 A1
20150043554 Meylan et al. Feb 2015 A1
20150049957 Shoham et al. Feb 2015 A1
20150063693 Carmel et al. Mar 2015 A1
20150067715 Koat et al. Mar 2015 A1
20150104153 Braness et al. Apr 2015 A1
20150117836 Amidei et al. Apr 2015 A1
20150117837 Amidei et al. Apr 2015 A1
20150139419 Kiefer et al. May 2015 A1
20150188758 Amidei et al. Jul 2015 A1
20150188842 Amidei et al. Jul 2015 A1
20150188921 Amidei et al. Jul 2015 A1
20150189017 Amidei et al. Jul 2015 A1
20150189373 Amidei et al. Jul 2015 A1
20150281310 Ziskind et al. Oct 2015 A1
20150288530 Oyman Oct 2015 A1
20150288996 Van Der Schaar et al. Oct 2015 A1
20150334435 Shivadas et al. Nov 2015 A1
20150373421 Chan et al. Dec 2015 A1
20160048593 Soroushian et al. Feb 2016 A1
20160070890 Grab et al. Mar 2016 A1
20160112382 Kiefer et al. Apr 2016 A1
20160149981 Shivadas et al. May 2016 A1
20160219303 Braness et al. Jul 2016 A1
20160323342 Luby et al. Nov 2016 A1
20170011055 Pitts Jan 2017 A1
20170083474 Meswani et al. Mar 2017 A1
20170214947 Kiefer et al. Jul 2017 A1
20170223389 Soroushian et al. Aug 2017 A1
20170238030 Ziskind et al. Aug 2017 A1
20170280203 Chan et al. Sep 2017 A1
20180007451 Shivadas et al. Jan 2018 A1
20180046949 Kahn et al. Feb 2018 A1
20180060543 Grab et al. Mar 2018 A1
20180081548 Barzik et al. Mar 2018 A1
20180131980 Van Der Schaar et al. May 2018 A1
20180220153 Braness et al. Aug 2018 A1
20180255366 Lockett et al. Sep 2018 A1
20180262757 Naletov et al. Sep 2018 A1
20180285261 Mandal et al. Oct 2018 A1
20180332094 Braness Nov 2018 A1
20190020907 Kiefer et al. Jan 2019 A1
20190020928 Chan et al. Jan 2019 A1
20190045219 Braness et al. Feb 2019 A1
20190045220 Braness et al. Feb 2019 A1
20190045234 Kiefer et al. Feb 2019 A1
20190158553 Van Der Schaar et al. May 2019 A1
20190268596 Naletov et al. Aug 2019 A1
20190297364 van der Schaar et al. Sep 2019 A1
20190342587 Kiefer et al. Nov 2019 A1
20190356928 Braness et al. Nov 2019 A1
20200059706 Shivadas et al. Feb 2020 A1
20200396454 Naletov et al. Dec 2020 A1
20210076082 Kiefer et al. Mar 2021 A1
20210099504 Van Der Schaar et al. Apr 2021 A1
20210136429 Van Der Schaar et al. May 2021 A1
20210250608 Braness et al. Aug 2021 A1
Foreign Referenced Citations (259)
Number Date Country
2010203605 May 2015 AU
2237293 Jul 1997 CA
2749170 Jul 2010 CA
2749170 Jun 2016 CA
2823829 Jan 2019 CA
1169229 Dec 1997 CN
1221284 Jun 1999 CN
1235473 Nov 1999 CN
1629939 Jun 2005 CN
1723696 Jan 2006 CN
101252401 Aug 2008 CN
102549557 Jul 2012 CN
102549557 Sep 2015 CN
105072454 Nov 2015 CN
105072454 Apr 2019 CN
757484 Feb 1997 EP
813167 Dec 1997 EP
0818111 Jan 1998 EP
0936812 Aug 1999 EP
0818111 Jan 2000 EP
1056273 Nov 2000 EP
1158799 Nov 2001 EP
1187483 Mar 2002 EP
1420580 May 2004 EP
1453319 Sep 2004 EP
1536646 Jun 2005 EP
1553779 Jul 2005 EP
1657835 May 2006 EP
1283640 Oct 2006 EP
1718074 Nov 2006 EP
2180664 Apr 2010 EP
2360923 Aug 2011 EP
2384475 Nov 2011 EP
2486517 Aug 2012 EP
2486727 Aug 2012 EP
2507995 Oct 2012 EP
2564354 Mar 2013 EP
2616991 Jul 2013 EP
2616991 Jul 2013 EP
2617192 Jul 2013 EP
2617192 Jul 2013 EP
2661696 Nov 2013 EP
2661875 Nov 2013 EP
2661895 Nov 2013 EP
2486727 Mar 2014 EP
2564354 Mar 2014 EP
2616991 Mar 2014 EP
2617192 Mar 2014 EP
2716048 Apr 2014 EP
2721826 Apr 2014 EP
2486517 Jun 2014 EP
2486517 Jun 2014 EP
2751990 Jul 2014 EP
2751990 Jul 2014 EP
2807821 Dec 2014 EP
2807821 Dec 2014 EP
2751990 Apr 2015 EP
2661875 Nov 2019 EP
2661696 May 2020 EP
3697096 Aug 2020 EP
3700219 Aug 2020 EP
3742740 Nov 2020 EP
3975574 Mar 2022 EP
3742740 May 2022 EP
2360923 Aug 2011 FR
2398210 Aug 2004 GB
1125765 Aug 2009 HK
08046902 Feb 1996 JP
08111842 Apr 1996 JP
08163488 Jun 1996 JP
08287613 Nov 1996 JP
09037225 Feb 1997 JP
H1175178 Mar 1999 JP
11164307 Jun 1999 JP
11275576 Oct 1999 JP
11328929 Nov 1999 JP
2000201343 Jul 2000 JP
2001043668 Feb 2001 JP
2001209726 Aug 2001 JP
2001346165 Dec 2001 JP
2002164880 Jun 2002 JP
2002170363 Jun 2002 JP
2002518898 Jun 2002 JP
2002218384 Aug 2002 JP
2003179597 Jun 2003 JP
2003250113 Sep 2003 JP
2004013823 Jan 2004 JP
2004515941 May 2004 JP
2004172830 Jun 2004 JP
2004187161 Jul 2004 JP
2004234128 Aug 2004 JP
2004304767 Oct 2004 JP
2004328218 Nov 2004 JP
2005027153 Jan 2005 JP
2005504480 Feb 2005 JP
2005080204 Mar 2005 JP
2005173241 Jun 2005 JP
2005284041 Oct 2005 JP
2005286881 Oct 2005 JP
2006155500 Jun 2006 JP
2006521035 Sep 2006 JP
2006524007 Oct 2006 JP
2007036666 Feb 2007 JP
2007174375 Jul 2007 JP
2007235690 Sep 2007 JP
2007535881 Dec 2007 JP
2008235999 Oct 2008 JP
2009522887 Jun 2009 JP
2009530917 Aug 2009 JP
4516082 May 2010 JP
2012514951 Jun 2012 JP
2013513298 Apr 2013 JP
5200204 Jun 2013 JP
2014506430 Mar 2014 JP
5681641 Jan 2015 JP
5723888 May 2015 JP
2015167357 Sep 2015 JP
6038805 Dec 2016 JP
6078574 Feb 2017 JP
2017063453 Mar 2017 JP
2018160923 Oct 2018 JP
6453291 Jan 2019 JP
6657313 Feb 2020 JP
202080551 May 2020 JP
2021158694 Oct 2021 JP
7000475 Dec 2021 JP
100221423 Sep 1999 KR
2002013664 Feb 2002 KR
1020020064888 Aug 2002 KR
20040039852 May 2004 KR
20060030164 Apr 2006 KR
20060106250 Oct 2006 KR
20060116967 Nov 2006 KR
100669616 Jan 2007 KR
20070005699 Jan 2007 KR
20070020727 Feb 2007 KR
20090016282 Feb 2009 KR
20100106418 Oct 2010 KR
20110133024 Dec 2011 KR
1020130133830 Dec 2013 KR
101635876 Jul 2016 KR
101874907 Jul 2018 KR
101917763 Nov 2018 KR
10-1988877 Jun 2019 KR
10-2072839 Jan 2020 KR
10-2122189 Jun 2020 KR
10-2195414 Dec 2020 KR
102191317 Dec 2020 KR
10-2274290 Jul 2021 KR
10-2352043 Jan 2022 KR
102408120 Jun 2022 KR
102445689 Sep 2022 KR
2011007344 Feb 2012 MX
316584 Dec 2013 MX
2328040 Jun 2008 RU
146026 Dec 2010 SG
1995015660 Jun 1995 WO
1996013121 May 1996 WO
199800973 Jan 1998 WO
9731445 Apr 1998 WO
1997031445 Apr 1998 WO
199834405 Aug 1998 WO
1998047290 Oct 1998 WO
9910836 Mar 1999 WO
1999010836 Mar 1999 WO
1999065239 Dec 1999 WO
2000049762 Aug 2000 WO
2000049763 Aug 2000 WO
0104892 Jan 2001 WO
2001031497 May 2001 WO
2001050732 Jul 2001 WO
2001065762 Sep 2001 WO
2002001880 Jan 2002 WO
2002008948 Jan 2002 WO
200223315 Mar 2002 WO
2003028293 Apr 2002 WO
2002035832 May 2002 WO
2002037210 May 2002 WO
2002054196 Jul 2002 WO
2002054776 Jul 2002 WO
2002073437 Sep 2002 WO
2002087241 Oct 2002 WO
2003030000 Apr 2003 WO
2003046750 Jun 2003 WO
2003047262 Jun 2003 WO
2003061173 Jul 2003 WO
2003096136 Nov 2003 WO
2004012378 Feb 2004 WO
2004054247 Jun 2004 WO
2004097811 Nov 2004 WO
2004100158 Nov 2004 WO
2004102571 Nov 2004 WO
2005008385 Jan 2005 WO
2005015935 Feb 2005 WO
2005050373 Jun 2005 WO
2005057906 Jun 2005 WO
2005109224 Nov 2005 WO
2005125214 Dec 2005 WO
2006018843 Feb 2006 WO
20060012398 Feb 2006 WO
2006018843 Dec 2006 WO
2007044590 Apr 2007 WO
2007072257 Jun 2007 WO
2007073347 Jun 2007 WO
2007093923 Aug 2007 WO
2007101182 Sep 2007 WO
2007113836 Oct 2007 WO
2008010275 Jan 2008 WO
2008032908 Mar 2008 WO
2008042242 Apr 2008 WO
2008086313 Jul 2008 WO
2008090859 Jul 2008 WO
2007113836 Nov 2008 WO
2008135932 Nov 2008 WO
2007113836 Dec 2008 WO
2009006302 Jan 2009 WO
2009065137 May 2009 WO
2009070770 Jun 2009 WO
2009109976 Sep 2009 WO
2010005673 Jan 2010 WO
2010060106 May 2010 WO
2010080911 Jul 2010 WO
2010089962 Aug 2010 WO
2010108053 Sep 2010 WO
2010111261 Sep 2010 WO
2010122447 Oct 2010 WO
2010147878 Dec 2010 WO
2011042898 Apr 2011 WO
2011042900 Apr 2011 WO
2011068668 Jun 2011 WO
2011086190 Jul 2011 WO
2011087449 Jul 2011 WO
2011101371 Aug 2011 WO
2011103364 Aug 2011 WO
2011132184 Oct 2011 WO
2011135558 Nov 2011 WO
2012035533 Mar 2012 WO
2012035534 Mar 2012 WO
2012035534 Jul 2012 WO
2012094171 Jul 2012 WO
2012094181 Jul 2012 WO
2012094189 Jul 2012 WO
20120094181 Jul 2012 WO
20120094189 Jul 2012 WO
2012035533 Aug 2012 WO
2012162806 Dec 2012 WO
2012171113 Dec 2012 WO
2013030833 Mar 2013 WO
2013032518 Mar 2013 WO
2013103986 Jul 2013 WO
2013111126 Aug 2013 WO
2013111126 Aug 2013 WO
2013032518 Sep 2013 WO
2013144942 Oct 2013 WO
2014145901 Sep 2014 WO
2014193996 Dec 2014 WO
2014193996 Feb 2015 WO
2015031982 Mar 2015 WO
2013111126 Jun 2015 WO
Non-Patent Literature Citations (408)
Entry
3GPP TS 26.247, V10.1.0, 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects Transparent end-to-end Packet-switches Streaming Services (PSS); Nov. 2011, 112 pages.
Chinese Patent Application 201180060590.1 office action dated Aug. 6, 2015, 11 pgs.
Examination report for GB1308663.2, dated May 18, 2016, 3 pgs.
Extended European Search Report for European Application No. 21208230.9, Search completed Feb. 18, 2022, dated Mar. 1, 2022, 15 Pgs.
Filed Application and Filing Receipt for U.S. Appl. No. 61/359,748, Application filed Jun. 29, 2010, Receipt mailed Jul. 13, 2010, 38 pages.
Great Britain Application GB1308663.2 search report dated Jan. 5, 2017, 1 pg.
ISO/IEC 14496-12 Information technology—Coding of audio-visual objects—Part 12: ISO base media file format, Amendment 3: DASH support and RTP reception hint track processing, 2011, 44 pgs.
ISO/IEC CD 23001-6 MPEG systems technologies Part 6: Dynamic adaptive streaming over HTTP (DASH), Oct. 15, 2010, 70 pgs.
ISO/IEC DIS 23009-1, Information technology—Dynamic adaptive streaming over HTTP (DASH)—Part 1: Media presentation description and segment formats, dated Aug. 30, 2011, 132 pgs.
ISO/IEC JTC1/SC29/WG11, MPEG/M18620, Oct. 2010, Text of ISO/IEC 23001-6: Dynamic adaptive streaming over http (DASH), 72 pgs.
ISO/IEC JTC1/SC29/WG11, MPEG/N11578, Text of ISO/IEC 23001-6: Dynamic adaptive streaming over HTTP (DASH), Oct. 2010, 70 pgs.
ISO/IEC JTC1/SC29-WG11—Coding of Moving Pictures and Audio, MPEG2010/M18692, Jan. 2010, 10 pgs.
Search Report for Canadian patent application 2,816,621, dated Oct. 30, 2014, 6 pgs.
Search report for European Patent Application 11838186.2, dated Jul. 13, 2017, 6 pgs.
“Pixel aspect ratio—Wikipedia”, Nov. 24, 2010, pp. 1-8.
Hunt, “Encoding for streaming”, The Netflix Blog, Nov. 6, 2008, printed from https://web.archive.org/web/20081216044437/http:/blog.netflix.com/2008/11/encoding-for-streaming.htm., retrieved on Feb. 8, 2022, 28 pgs.
Pereira, “Security on Over the Top TV Services”, Thesis, Nov. 2011, 114 pgs.
Stockhammer, “MPEG's Dynamic Adaptive Streaming over HTTP (DASH)—An Enabling Standard for Internet TV”, Qualcomm Incorporated, Apr. 11, 2015, Retrieved from the Internet, https://www.w3.org/2011/09/webtv/slides/W3C-Workshop.pdf, 30 pgs.
“IBM Closes Cryptolopes Unit,” Dec. 17, 1997, CNET News, Printed on Apr. 25, 2014 from http://news.cnet.com/IBM-closes-Cryptolopes-unit/2100-1001_3206465.html, 3 pages.
Written Opinion for International Application No. PCT/US2007/063950 filed Mar. 14, 2007, report completed Mar. 1, 2008; report dated Mar. 19, 2008, 6 pgs.
U.S. Appl. No. 13/905,804, “Notice of Allowance,” dated Aug. 12, 2015, 8 pgs.
Broadq—The Ultimate Home Entertainment Software, printed May 11, 2009 from ittp://web.srchive.org/web/20030401122010/www.broadq.com/qcasttuner/, 1 pg.
Cloakware Corporation, “Protecting Digital Content Using Cloakware Code Transformation Technology”, Version 1.2, May 2002, pp. 1-10.
EP11774529 Supplementary European Search Report, completed Jan. 31, 2014, 2 pgs.
European Search Report Application No. EP 08870152, Search Completed May 19, 2011, dated May 26, 2011, 9 pgs.
European Supplementary Search Report for Application EP09759600, completed Jan. 25, 2011, 11 pgs.
Extended European Search Report for European Application EP10821672, completed Jan. 30, 2014, 3 pgs.
Extended European Search Report for European Application EP11824682, completed Feb. 6, 2014, 4 pgs.
Extended European Search Report for European Application No. 14763140.2, Search completed Sep. 26, 2016, dated Oct. 5, 2016, 9 Pgs.
Extended European Search Report for European Application No. 20172313.7 Search completed Aug. 19, 2020 dated Aug. 27, 2020, 11 Pgs.
Final draft ETSI ES 202 109, V1.1.1, ETSI Standard, Terrestrial Trunked Radio (TETRA); Security; Synchronization mechanism for end-to-end encryption, Oct. 2002, 17 pgs.
IBM Corporation and Microsoft Corporation, “Multimedia Programming Interface and Data Specifications 1.0”, Aug. 1991, printed from http://www.kk.iij4u.or.jp/˜kondo/wave/mpidata.txt on Mar. 6, 2006, 100 pgs.
Information Technology—MPEG Systems Technologies—Part 7: Common Encryption in ISO Base Media File Format Files (ISO/IEC 23001-7), Apr. 2015, 24 pgs.
InformationWeek, “Internet on Wheels”, InformationWeek: Front End: Daily Dose, Jul. 20, 1999, Printed on Mar. 26, 2014, 3 pgs.
International Preliminary Report for Application No. PCT/US2011/066927, Filed Dec. 22, 2011, Report dated Jul. 10, 2013, 13 pgs.
International Preliminary Report for International Application No. PCT/US2011/067243, International Filing Date Dec. 23, 2011, dated Jul. 10, 2013, 7 pgs.
International Preliminary Report on Patentability for International Application PCT/US14/30747, Report dated Sep. 15, 2015, dated Sep. 24, 2015, 6 Pgs.
International Preliminary Report on Patentability for International Application No. PCT/US2007/063950, Report Completed Dec. 18, 2009, 3 pgs.
International Preliminary Report on Patentability for International Application No. PCT/US2008/083816, dated May 18, 2010, 6 pgs.
International Preliminary Report on Patentability for International Application No. PCT/US2010/56733, dated Jun. 5, 2012, 5 pgs.
International Preliminary Report on Patentability for International Application No. PCT/US2011/068276, dated Mar. 4, 2014, 23 pgs.
International Preliminary Report on Patentability for International Application PCT/US2011/067167, dated Feb. 25, 2014, 8 pgs.
International Preliminary Report on Patentability for International Application PCT/US2013/043181, dated Dec. 31, 2014, dated Jan. 8, 2015, 11 Pgs.
International Preliminary Report on Patentability for International Application PCT/US2014/039852, dated Dec. 1, 2015, dated Dec. 5, 2015, 8 Pgs.
International Search Report and Written Opinion for International Application No. PCT/US07/63950, completed Feb. 19, 2008; dated Mar. 19, 2008, 9 pgs.
International Search Report and Written Opinion for International Application No. PCT/US08/87999, completed Feb. 7, 2009, dated Mar. 19, 2009, 6 pgs.
International Search Report and Written Opinion for International Application No. PCT/US09/46588, completed Jul. 13, 2009, dated Jul. 23, 2009, 7 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2004/041667, completed May 24, 2007, dated Jun. 20, 2007, 6 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2005/025845, completed Feb. 5, 2007 and dated May 10, 2007, 8 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2008/083816, completed Jan. 10, 2009, dated Jan. 22, 2009, 7 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2010/020372, Completed Feb. 10, 2009, dated Mar. 1, 2010, 8 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2010/56733, Completed Jan. 3, 2011, dated Jan. 14, 2011, 9 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2011/067243, International Filing Date Dec. 23, 2011, Search Completed Apr. 24, 2012, dated May 8, 2012, 8 pgs.
International Search Report and Written Opinion for International Application No. PCT/US2013/043181, completed Nov. 27, 2013, dated Dec. 6, 2013, 12 pgs.
International Search Report and Written Opinion for International Application PCT/US14/30747, completed Jul. 30, 2014, dated Aug. 22, 2014, 7 Pgs.
International Search Report and Written Opinion for International Application PCT/US14/39852, completed Oct. 21, 2014, dated Dec. 5, 2014, 11 pgs.
International Search Report and Written Opinion for International Application PCT/US2011/066927, completed Apr. 3, 2012, dated Apr. 20, 2012, 14 pgs.
International Search Report and Written Opinion for International Application PCT/US2011/067167, completed Jun. 19, 2012, dated Jul. 2, 2012, 11 pgs.
International Search Report and Written Opinion for International Application PCT/US2011/068276, completed Jun. 19, 2013, dated Jul. 8, 2013, 24 pgs.
International Search Report and Written Opinion for PCT/US2013/020572, International Filing Date Jan. 7, 2013, Search Completed Mar. 19, 2013, dated Apr. 29, 2013, 10 pgs.
International Telecommunication Union, Telecommunication Standardization Sector of ITU, H.233, Line Transmission of Non-Telephone Signals, Confidentiality System for Audiovisual Services, ITU-T Recommendation H.233, Mar. 1993, 18 pgs.
ISO/IEC 14496-12 Information technology—Coding of audio-visual objects—Part 12: ISO base media file format, Feb. 2004 (“MPEG-4 Part 12 Standard”), 62 pgs.
ISO/IEC 14496-12:2008(E) Informational Technology—Coding of Audio-Visual Objects Part 12: ISO Base Media File Format, Oct. 2008, 120 pgs.
ISO/IEC FCD 23001-6 MPEG systems technologies Part 6: Dynamic adaptive streaming over HTTP (DASH), Jan. 28, 2011, 86 pgs.
ITS International, “Fleet System Opts for Mobile Server”, Aug. 26, 1999, Printed on Oct. 21, 2011 from http://www.itsinternational.com/News/article.cfm?recordID=547, 2 pgs.
Lifehacker—Boxqueue Bookmarklet Saves Videos for Later Boxee Watching, printed Jun. 16, 2009 from http://feeds.gawker.com/˜r/lifehacker/full/˜3/OHvDmrlgZZc/boxqueue-bookmarklet-saves-videos-for-late-boxee-watching, 2 pgs.
Linksys Wireless-B Media Adapter Reviews, printed May 4, 2007 from http://reviews.cnet.com/Linksys_Wireless_B_Media_Adapter/4505-6739_7-30421900.html?tag=box, 5 pgs.
Linksys, Kiss DP-500, printed May 4, 2007 from http://www.kiss-technology.com/?p=dp500, 2 pgs.
Linksys®: “Enjoy your digital music and pictures on your home entertainment center, without stringing wires!”, Model No. WMA 11B, printed May 9, 2007 from http://www.linksys.com/servlet/Satellite?c=L_Product_C2&childpagename=US/Layout&cid=1115416830950&p.
Microsoft Corporation, “Chapter 8, Multimedia File Formats” 1991, Microsoft Windows Multimedia Programmer's Reference, 3 cover pgs, pp. 8-1 to 8-20.
Microsoft Corporation, Advanced Systems Format (ASF) Specification, Revision 01.20.03, Dec. 2004, 121 pgs.
Microsoft Media Platform: Player Framework, “Microsoft Media Platform: Player Framework v2.5 (formerly Silverlight Media Framework)”, May 3, 2011, 2 pages.
Microsoft Media Platform: Player Framework, “Silverlight Media Framework v1.1”, Jan. 2010, 2 pages.
Microsoft Windows® XP Media Center Edition 2005, Frequently asked Questions, printed May 4, 2007 from http://www.microsoft.com/windowsxp/mediacenter/evaluation/faq.mspx.
Microsoft Windows® XP Media Center Edition 2005: Features, printed May 9, 2007, from http://www.microsoft.com/windowsxp/mediacenter/evaluation/features.mspx, 4 pgs.
MPEG-DASH presentation at Streaming Media West 2011, Nov. 2011, 14 pgs.
Office Action for Chinese Patent Application No. CN200880127596.4, dated May 6, 2014, 8 pgs.
Office Action for U.S. Appl. No. 13/223,210, dated Apr. 30, 2015, 14 pgs.
Office Action for U.S. Appl. No. 14/564,003, dated Apr. 17, 2015, 28 pgs.
Open DML AVI-M-JPEG File Format Subcommittee, “Open DML AVI File Format Extensions”, Version 1.02, Feb. 28, 1996, 29 pgs.
PC world.com, Future Gear: PC on the HiFi, and the TV, from http://www.pcworld.com/article/id, 108818-page, 1/article.html, printed May 4, 2007, from IDG Networks, 2 pgs.
Pomelo, LLC Tech Memo, Analysis of Netflix's Security Framework for ‘Watch Instantly’ Service, March-Apr. 2009, 18 pgs.
Qtv—About BroadQ, printed May 11, 2009 from http://www.broadq.com/en/about.php, 1 pg.
Server-Side Stream Repackaging (Streaming Video Technologies Panorama, Part 2), Jul. 2011, 15 pgs.
Supplementary European Search Report for Application No. EP 04813918, Search Completed Dec. 19, 2012, 3 pgs.
Supplementary European Search Report for Application No. EP 10729513, completed Dec. 9, 2013, 4 pgs.
Supplementary European Search Report for EP Application 11774529, completed Jan. 31, 2014, 2 pgs.
Supplementary European Search Report for European Application No. 07758499.3, Report Completed Jan. 25, 2013, 8 pgs.
Text of ISO/IEC 23001-6: Dynamic adaptive streaming over HTTP (DASH), Oct. 2010, 71 pgs.
Universal Mobile Telecommunications System (UMTS), ETSI TS 126 233 V9.1.0 (Jun. 2011) 3GPP TS 26.233 version 9.1.0 Release 9, 18 pgs.
Universal Mobile Telecommunications Systems (UMTS); ETSI TS 126 244 V9.4.0 (May 2011) 3GPP TS 26.244 version 9.4.0 Release 9, 58 pgs.
Windows Media Center Extender for Xbox, printed May 9, 2007 from http://www.xbox.com/en-US/support/systemuse/xbox/console/mediacenterextender.htm, 2 pgs.
Windows® XP Media Center Edition 2005, “Experience more entertainment”, retrieved from http://download.microsoft.com/download/c/9/a/c9a7000a-66b3- 455b-860b-1c16f2eecfec/MCE.pdf on May 9, 2007, 2 pgs.
3GPP TS 26.247, V1.3.0, 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects Transparent end-to-end Packet-switches Streaming Services (PSS);, Progressive Download and Dynamic Adaptive Streaming over http (3GP-DASH) (Release 10), Mar. 2011, 72 pgs.
Written Opinion for International Application No. PCT/US2008/083816, Opinion completed Jan. 10, 2009, dated Jan. 22, 2009, 5 pgs.
Written Opinion for International Application No. PCT/US2009/046588, completed Jul. 14, 2009, dated Jul. 23, 2009, 5 pgs.
Written Opinion of the International Searching Authority for International Application No. PCT/US 08/87999, date completed Feb. 7, 2009, dated Mar. 19, 2009, 4 pgs.
International Search Report and Written Opinion for International Application PCT/US2010/020372, Search Report dated Mar. 1, 2010, 7 pgs.
Written Opinion for International Application No. PT/US2005/025845 filed Jul. 21, 2005, report completed Feb. 5, 2007 and dated May 10, 2007, 5 pgs.
Written Opinion for International Application No. PCT/US2004/041667, Filing Date Dec. 8, 2004, Report Completed May 24, 2007, dated Jun. 20, 2007, 4 pgs.
Written Opinion for International Application No. PCT/US2005/025845, completed Feb. 5, 2007 and dated May 10, 2007, 6 pgs.
Innovation of technology arrived, I-O Data, Nov. 2004, Retrieved from http://www.iodata.com/catalogs/AVLP2DVDLA_Flyer200505.pdf on May 30, 2013, 2 pgs., I-O Data, 2 pgs.
“Adaptive Streaming Comparison”, Jan. 28, 2010, 5 pgs.
“Adobe Flash Video File Format Specification”, Aug. 2010, Version 10.1, 89 pages.
“Apple HTTP Live Streaming specification”, Aug. 2017, 60 pgs.
“Best Practices for Multi-Device Transcoding”, Kaltura Open Source Video, Printed on Nov. 27, 2013 from knowledge.kaltura.com/best-practices-multi-device-transcoding, 13 pgs.
“Container format (digital)”, printed Aug. 22, 2009 from http://en.wikipedia.org/wiki/Container_format_(digital), 4 pgs.
“Data Encryption Decryption using AES Algorithm, Key and Salt with Java Cryptography Extension”, Available at https://www.digizol.com/2009/10/java-encrypt-decrypt-jce-salt.html, Oct. 200, 6 pgs.
“Delivering Live and On-Demand Smooth Streaming”, Microsoft Silverlight, 2009, 28 pgs.
“Diagram | Matroska”, Dec. 17, 2010, Retrieved from http://web.archive.org/web/201 01217114656/http://matroska.org/technical/diagram/index.html on Jan. 29, 2016, 5 pages, Dec. 17, 2010.
“Draft CR: Trick Mode for HTTP Streaming”, 3GPP TSG-SA4 Meeting #58, Apr. 26-30, 2010, Vancouver, Canada, S4-100237, 3 pgs.
“DVD—MPeg differences”, printed Jul. 2, 2009 from http://dvd.sourceforge.net/dvdinfo/dvdmpeg.html, 1 pg.
“DVD subtitles”, sam.zoy.org/writings/dvd/subtitles, dated Jan. 9, 2001, printed Jul. 2, 2009, 4 pgs.
“Final Committee Draft of MPEG-4 streaming text format”, International Organisation for Standardisation, Feb. 2004, 22 pgs.
“HTTP Based Adaptive Streaming over HSPA”, Apr. 2011, 73 pgs.
“HTTP Live Streaming”, Mar. 2011, 24 pgs.
“HTTP Live Streaming”, Sep. 2011, 33 pgs.
“IBM Spearheading Intellectual Property Protection Technology for Information on the Internet; Cryptolope Containers Have Arrived”, May 1, 1996, Business Wire, Printed on Aug. 1, 2014 from http://www.thefreelibrary.com/IBM+Spearheading+Intellectual+Property+Protection+Technology+for...-a018239381, 6pg.
“Information Technology—Coding of audio-visual objects—Part 14: MP4 file format”, International Standard, ISO/IEC 14496-14, First Edition, Nov. 15, 2003, 18 pages.
“Information Technology—Coding of audio-visual objects—Part 17: Streaming text”, International Organisation for Standardisation, Feb. 2004, 22 pgs.
“Information technology—Coding of audio-visual objects—Part 18: Font compression and streaming”, ISO/IEC 14496-18, First edition Jul. 1, 2004, 26 pgs.
“Information technology—Generic coding of moving pictures and associated audio information: Systems”, International Standard ISO/IEC 13818-1, Second Edition, Dec. 1, 2000, 174 pages (presented in two parts).
“Information Technology—Coding of Audio Visual Objects—Part 2: Visual”, International Standard, ISO/IEC 14496-2, Third Edition, Jun. 1, 2004, pp. 1-724. (presented in three parts).
“Information-Technology—Generic coding of moving pictures and associated audio: Systems, Recommendation H.222.0”, International Standard, ISO/IEC 13818-1, Draft 1209, Apr. 25, 1995, 151 pages.
“Information-Technology—Generic coding of moving pictures and associated audio: Systems, Recommendation H.222.0”, International Standard, ISO/IEC 13818-1, Draft 1540, Nov. 13, 1994, 161 pages.
“Java Cryptography Architecture API Specification & Reference”, Available at https://docs.oracle.com/javase/1.5.0/docs/guide/security/CryptoSpec.html, Jul. 25, 2004, 68 pgs.
“Java Cryptography Extension, javax.crypto.Cipher class”, Available at https://docs.oracle.com/javase/1.5.0/docs/api/javax/crypto/Cipher.html, 2004, 24 pgs.
“JCE Encryption—Data Encryption Standard (DES) Tutorial”, Available at https://mkyong.com/java/jce-encryption-data-encryption-standard-des-tutorial/, Feb. 25, 2009, 2 pgs.
“Kiss Players, Kiss DP-500”, retrieved from http://www.kiss-technology.com/?p=dp500 on May 4, 2007, 1 pg.
“Live and On-Demand Video with Silverlight and IIS Smooth Streaming”, Microsoft Silverlight, Windows Server Internet Information Services 7.0, Feb. 2010, 15 pgs.
“Matroska”, Wikipedia, Jul. 10, 2017, retrieved from https://en.wikipedia.org/wiki/Matroska on Jul. 20, 2017, 3 pages.
“Matroska Streaming | Matroska”, Retrieved from the Internet: URL:http://web.archive.org/web/201 0121711431 O/http://matroska.org/technical!streaming/index.html [retrieved on Jan. 29, 2016], Dec. 17, 2010.
“Microsoft Smooth Streaming specification”, Jul. 22, 2013, 56 pgs.
“MovieLabs Specification for Next Generation Video—Version 1.0”, Motion Picture Laboratories, Inc., 2013, Retrieved from: http://movielabs.com/ngvideo/MovieLabs%20Specification%20for%20Next%20Generation%20Video%20v1.0.pdf, 5 pgs.
“MPEG-2”, Wikipedia, Jun. 13, 2017, retrieved from https://en.wikipedia.org/wiki/MPEG-2 on Jul. 20, 2017, 13 pages.
“MPEG-4 File Format, Version 2”, Sustainability of Digital Formats: Planning for Library of Congress Collections, Retrieved from: https://www.loc.gov/preservation/digital/formats/fdd/fdd000155.shtml, Last updated Feb. 21, 2017, 8 pgs.
“MPEG-4 Part 14”, Wikipedia, Jul. 10, 2017, retrieved from https://en.wikipedia.org/wiki/MPEG-4_Part_14 on Jul. 20, 2017, 5 pages.
“Netflix turns on subtitles for PC, Mac streaming”, Yahoo! News, Apr. 21, 2010, Printed on Mar. 26, 2014, 3 pgs.
“OpenDML AVI File Format Extensions”, ** check for duplicate** OpenDML AVIM-JPEG File Format Subcommittee, retrieved from www.the-labs.com/Video/odmlff2-avidef.pdf, Sep. 1997, 42 pgs.
“OpenDML AVI File Format Extensions Version 1.02”, OpenDMLAVI MJPEG File Format Subcommittee. Last revision: Feb. 28, 1996. Reformatting: Sep. 1997, 42 pgs.
“QCast Tuner for PS2”, printed May 11, 2009 from http://web.archive.org/web/20030210120605/www.divx.com/software/detail.php?ie=39, 2 pgs.
“Series H: Audiovisual and Multimedia Systems Infrastructure of audiovisual services—Coding of moving video; High efficiency video coding”, International Telecommunication Union, ITU-T H.265, Apr. 2015, 634 pages (presented in six parts).
“Single-Encode Streaming for Multiple Screen Delivery”, Telestream Wowza Media Systems, 2009, 6 pgs.
“Smooth Streaming Client”, The Official Microsoft IIS Site, Sep. 24, 2010, 4 pages.
“Specifications | Matroska”, Retrieved from the Internet: URL:http://web.archive.org/web/201 00706041303/http:/1www.matroska.org/technical/specs/index.html [retrieved on Jan. 29, 2016, Jul. 6, 2010.
“Specifications Matroska”, Dec. 17, 2010, [retrieved on Mar. 2, 2018], 12 pages.
“Supplementary European Search Report for Application No. EP 10834935, International Filing Date Nov. 15, 2010, Search Completed May 27, 2014, 9 pgs.”
“Supported Media Formats”, Supported Media Formats, Android Developers, Printed on Nov. 27, 2013 from developer.android.com/guide/appendix/media-formats.html, 3 pgs.
“SWF and FLV File Format Specification”, Adobe, Jun. 2007, Version 9, 298 pages.
“Text of ISO/IEC 14496-18/COR1, Font compression and streaming”, ITU Study Group 16—Video Coding Experts Group—ISO/IEC MPEG & ITU-T VCEG(ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 06), No. N8664, Oct. 27, 2006, 8 pgs.
“Text of ISO/IEC 14496-18/FDIS, Coding of Moving Pictures and Audio”, ITU Study Group 16—Videocoding Experts Group—ISO/IEC MPEG & ITU-T VCEG(ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 06), No. N6215, Dec. 2003, 26 pgs.
“The MPEG-DASH Standard for Multimedia Streaming Over the Internet”, IEEE MultiMedia, vol. 18, No. 4, 2011, 7 pgs.
“Thread: SSME (Smooth Streaming Medial Element) config.xml review (Smooth Streaming Client configuration file)”, Printed on Mar. 26, 2014, 3 pgs.
“Transcoding Best Practices”, From movideo, Printed on Nov. 27, 2013 from code.movideo.com/Transcoding_Best_Practices, 5 pgs.
“Using HTTP Live Streaming”, iOS Developer Library, http://developer.apple.com/library/ios/#documentation/networkinginternet/conceptual/streamingmediaguide/UsingHTTPLiveStreaming/UsingHTTPLiveStreaming.html#//apple_ref/doc/uid/TP40008332-CH102-SW1, Feb. 11, 2014, 10 pgs.
“Video File Format Specification”, Adobe, Apr. 2008, Version 9, 46 pages.
“Video Manager and Video Title Set IFO file headers”, printed Aug. 22, 2009 from http://dvd.sourceforge.net/dvdinfo/ifo.htm, 6 pgs.
“What is a DVD?”, printed Aug. 22, 2009 from http://www.videohelp.com/dvd, 8 pgs.
“What is a VOB file”, http://www.mpucoder.com/DVD/vobov.html, printed on Jul. 2, 2009, 2 pgs.
“What's on a DVD?”, printed Aug. 22, 2009 from http://www.doom9.org/dvd-structure.htm, 5 pgs.
“Windows Media Player 9”, Microsoft, Mar. 23, 2017, 3 pgs.
“DVD-Mpeg differences”, printed on Jul. 2, 2009, http://dvd.sourceforge.net/dvdinfo/dvdmpeg.html, 1 pg.
Casares, Juan et al., “Simplifying Video Editing Using Metadata”.
Long et al., “Silver: Simplifying Video Editing with Metadata”, CHI 2003: New Horizons, Apr. 5-10, 2003, pp. 628-629.
U.S. Appl. No. 13/224,298, “Final Office Action”, dated May 19, 2014, 26 pgs.
U.S. Appl. No. 13/905,804, “Non-Final Office Action Received”, U.S. Appl. No. 13/905,804, “Non-Final Office Action”, dated Jul. 25, 2014, 15 pgs.
Abomhara et al., “Enhancing Selective Encryption for H.264/AVC Using Advanced Encryption Standard”, International Journal of computer Theory and Engineering, Apr. 2010, vol. 2, No. 2, pp. 223-229.
Adams et al., “Will http adaptive streaming become the dominant mode of video delivery in cable networks?”, https://www.nctatechnicalpapers.com/Paper/2011/2011-will-http-adaptive-streaming-become-the-dominant-mode-of-video-delivery-in-cable-networks-, 10 pgs.
Adhikari et al., “Unreeling Netflix: Understanding and Improving Multi-CDN Movie Delivery”, 2012 Proceedings IEEE InfoCom, Mar. 25-30, 2012, Orlando, Florida, 9 pgs.
Adzic et al., “Optimized Adaptive HTTP Streaming for Mobile Devices”, International Society for Optics and Photonics, Applications of Digital Image Processing XXXIV, vol. 8135, Sep. 2011, p. 81350T.
Akhshabi et al. “An Experimental Evaluation of Rate-Adaptation Algorithms in Adaptive Streaming over HTTP”, MMSys'11, Feb. 23-25, 2011, 12 pgs.
Alattar et al., A.M, “Improved selective encryption techniques for secure transmission of MPEG video bit-streams”, In Proceedings 1999 International Conference on Image Processing (Cat. 99CH36348), vol. 4, IEEE, 1999, pp. 256-260.
Anonymous, “Method for the encoding of a compressed video sequence derived from the same video sequence compressed at a different bit rate without loss of data”, ip.com, ip.com No. IPCOM000008165D, May 22, 2002, pp. 1-9.
Antoniou et al., “Adaptive Methods for the Transmission of Video Streams in Wireless Networks”, 2015, 50 pgs.
Apostolopoulos et al., “Secure Media Streaming and Secure Transcoding”, Multimedia Security Technologies for Digital Rights Management, 2006, 33 pgs.
Asai et al., “Essential Factors for Full-Interactive VOD Server: Video File System, Disk Scheduling, Network”, Proceedings of Globecom '95, Nov. 14-16, 1995, 6 pgs.
Author Unknown, “Blu-ray Disc—Blu-ray Disc—Wikipedia, the free encyclopedia”, printed Oct. 30, 2008 from http://en.wikipedia.org/wiki/Blu-ray_Disc, 11 pgs.
Author Unknown, “Blu-ray Movie Bitrates Here—Blu-ray Forum”, printed Oct. 30, 2008 from http://forum.blu-ray.com/showthread.php?t=3338, 6 pgs.
Author Unknown, “MPEG-4 Video Encoder: Based on International Standard ISO/IEC 14496-2”, Patni Computer Systems, Ltd., printed Jan. 24, 2007, USA, pp. 1-15.
Author Unknown, “O'Reilly—802.11 Wireless Networks: The Definitive Guide, Second Edition”, printed Oct. 30, 2008 from http://oreilly.com/catalog/9780596100520, 2 pgs.
Author Unknown, “Turbo-Charge Your Internet and PC Performance”, printed Oct. 30, 2008 from Speedtest.net—The Global Broadband Speed Test, 1 pg.
Author Unknown, “White paper, The New Mainstream Wireless LAN Standard”, Broadcom Corporation, Jul. 2003, 12 pgs.
Beker et al., “Cipher Systems, The Protection of Communications”, 1982, 40 pgs.
Blasiak, “Video Transrating and Transcoding: Overview of Video Transrating and Transcoding Technologies”, Ingenient Technologies, TI Developer Conference, Aug. 6-8, 2002, 22 pgs.
Bloom et al., “Copy Protection for DVD Video”, Proceedings of the IEEE, vol. 87, No. 7, Jul. 1999, pp. 1267-1276.
Bocharov et al., “Portable Encoding of Audio-Video Objects, The Protected Interoperable File Format (PIFF)”, Microsoft Corporation, First Edition Sep. 8, 2009, 30 pgs.
Bross et al., “High Efficiency Video Coding (HEVC) text specification draft 10 (for FDIS & Last Call)”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11, Document: JCTVC-L1003_v34, 12th Meeting: Geneva, CH, Jan. 14-23, 2013 (presented in three parts).
Bulterman et al. “Synchronized Multimedia Integration Language (SMIL 3.0)”, W3C Recommendation, Dec. 1, 2008, https://www.w3.org/TR/2008/REC-SMIL3-20081201/, 321 pgs. (presented in five parts).
Cahill et al., “Locally Adaptive Deblocking Filter for Low Bit Rate Video”, Proceedings 2000 International Conference on Image Processing, Sep. 10-13, 2000, Vancouver, BC, Canada, 4 pgs.
Candelore, U.S. Appl. No. 60/372,901, filed Apr. 17, 2002, 5 pgs.
Casares et al., “Simplifying Video Editing Using Metadata”, DIS2002, 2002, pp. 157-166.
Chaddha et al., “A Frame-work for Live Multicast of Video Streams over the Internet”, Proceedings of 3rd IEEE International Conference on Image Processing, Sep. 19, 1996, Lausanne, Switzerland, 4 pgs.
Cheng, “Partial Encryption for Image and Video Communication”, Thesis, Fall 1998, 95 pgs.
Cheng et al., “Partial encryption of compressed images and videos”, IEEE Transactions on Signal Processing, vol. 48, No. 8, Aug. 2000, 33 pgs.
Cheung et al., “On the Use of Destination Set Grouping to Improve Fairness in Multicast Video Distribution”, Proceedings of IEEE INFOCOM'96, Conference on Computer Communications, vol. 2, IEEE, 1996, 23 pgs.
Collet, “Delivering Protected Content, An Approach for Next Generation Mobile Technologies”, Thesis, 2010, 84 pgs.
Concolato et al., “Live HTTP Streaming of Video and Subtitles within a Browser”, MMSys 2013, Feb. 26-Mar. 1, 2013, Oslo, Norway, 5 pgs.
De Cock et al., “Complexity-Based Consistent-Quality Encoding in the Cloud”, IEEE International Conference on Image Processing (ICIP), Date of Conference Sep. 25-28, 2016, Phoenix, AZ, pp. 1484-1488.
Deutscher, “IIS Transform Manager Beta—Using the MP4 to Smooth Task”, Retrieved from: https://web.archive.org/web/20130328111303/http://blog.johndeutscher.com/category/smooth-streaming, Blog post of Apr. 29, 2011, 14 pgs.
Diamantis et al., “Real Time Video Distribution using Publication through a Database”, Proceedings SIBGRAPI'98. International Symposium on Computer Graphics, Image Processing, and Vision (Cat. No. 98EX237), Oct. 1990, 8 pgs.
Dworkin, “Recommendation for Block Cipher Modes of Operation: Methods and Techniques”, NIST Special Publication 800-38A, 2001, 66 pgs.
Eskicioglu et al., “An Integrated Approach to Encrypting Scalable Video”, Proceedings IEEE International Conference on Multimedia and Expo, Aug. 26-29, 2002, Lausanne, Switzerland, 4 pgs.
Fang et al., “Real-time deblocking filter for MPEG-4 systems”, Asia-Pacific Conference on Circuits and Systems, Oct. 28-31, 2002, Bail, Indonesia, pp. 541-544.
Fecheyr-Lippens, “A Review of HTTP Live Streaming”, **check for duplicate** Internet Citation, Jan. 25, 2010, pp. 1-37.
Fecheyr-Lippens, “A Review of HTTP Live Streaming”, Jan. 2010, 38 pgs.
Fielding et al., “Hypertext Transfer Protocol—HTTP1.1”, Network Working Group, RFC 2616, Jun. 1999, 114 pgs.
Fukuda et al., “Reduction of Blocking Artifacts by Adaptive DCT Coefficient Estimation in Block-Based Video Coding”, Proceedings 2000 International Conference on Image Processing, Sep. 10-13, 2000, Vancouver, BC, Canada, pp. 969-972.
Garg et al., “An Experimental Study of Throughput for UDP and VolP Traffic in IEEE 802.11b Networks”, Wireless Communications and Networkings, Mar. 2003, pp. 1748-1753.
Gast, “When is 54 Not Equal to 54? A Look at 802.11a, b and g Throughput”, Aug. 8, 2003, printed Oct. 30, 2008 from www.oreillynet.com/pub/a/wireless/2003/08/08/wireless_throughput.html, 4 pgs.
Ghosh, “Enhancing Silverlight Video Experiences with Contextual Data”, Retrieved from: http://msdn.microsoft.com/en-us/magazine/ee336025.aspx, 2010, 15 pgs.
Griffith, Eric, “The Wireless Digital Picture Frame Arrives”, Wi-Fi Planet, printed May 4, 2007 from http://www.wi-fiplanet.com/news/article.php/3093141, Oct. 16, 2003, 3 pgs.
Hartung et al., “DRM Protected Dynamic Adaptive HTTP Streaming”, MMSys 2011 Proceedings of the Second Annual ACM Conference on Multimedia Systems, San Jose, California, Feb. 23-25, 2011, pp. 277-282.
Huang, U.S. Pat. No. 7,729,426, U.S. Appl. No. 11/230,794, filed Sep. 20, 2005, 143 pgs.
Huang et al., “Adaptive MLP post-processing for block-based coded images”, IEEE Proceedings—Vision, Image and Signal Processing, vol. 147, No. 5, Oct. 2000, pp. 463-473.
Huang et al., “Architecture Design for Deblocking Filter in H.264/JVT/AVC”, 2003 International Conference on Multimedia and Expo., Jul. 6-9, 2003, Baltimore, MD, 4 pgs.
Hurtado Guzman, “Development and Implementation of an Adaptive HTTP Streaming Framework for H264/MVC Coded Media”, Politecnico di Torino, Nov. 2010, 108 pgs.
Hwang et al., “Efficient and User Friendly Inter-domain Device Authentication/Access control for Home Networks”, Proceedings of the 2006 International Conference on Embedded and Ubiquitous Computing, Seoul, Korea, Aug. 1-4, 2006, pp. 131-140.
Inlet Technologies, “Adaptive Delivery to iDevices”, 2010, 2 pages.
Inlet Technologies, “Adaptive delivery to iPhone 3.0”, 2009, 2 pgs.
Inlet Technologies, “HTTP versus RTMP”, 2009, 3 pages.
Inlet Technologies, “The World's First Live Smooth Streaming Event: The French Open”, 2009, 2 pages.
I-O Data, “Innovation of technology arrived”, Nov. 2004, Retrieved from http://www.iodata.com/catalogs/AVLP2DVDLA_Flyer200505.pdf, 2 pgs.
Jain et al., U.S. Appl. No. 61/522,623, filed Aug. 11, 2011, 44 pgs.
Jung et al., “Design and Implementation of an Enhanced Personal Video Recorder for DTV”, IEEE Transactions on Consumer Electronics, vol. 47, No. 4, Nov. 2001, 6 pgs.
Kalva, Hari, “Delivering MPEG-4 Based Audio-Visual Services”, 2001, 113 pgs.
Kang et al., “Access Emulation and Buffering Techniques for Steaming of Non-Stream Format Video Files”, IEEE Transactions on Consumer Electronics, vol. 43, No. 3, Aug. 2001, 7 pgs.
Kaspar et al., “Using HTTP Pipelining to Improve Progressive Download over Multiple Heterogeneous Interfaces”, IEEE ICC proceedings, 2010, 5 pgs.
Kim et al., “A Deblocking Filter with Two Separate Modes in Block-Based Video Coding”, IEEE transactions on circuits and systems for video technology, vol. 9, No. 1, 1999, pp. 156-160.
Kim et al., “Tree-Based Group Key Agreement”, Feb. 2004, 37 pgs.
Kozintsev et al., “Improving last-hop multicast streaming video over 802.11”, Workshop on Broadband Wireless Multimedia, Oct. 2004, pp. 1-10.
Kurzke et al., “Get Your Content Onto Google TV”, Google, Retrieved from: http://commondatastorage.googleapis.com/io2012/presentations/live%20to%20website/1300.pdf, 2012, 58 pgs.
Lang, “Expression Encoder, Best Practices for live smooth streaming broadcasting”, Microsoft Corporation, 2010, retrieved from http://www.streamingmedia.com/conferences/west2010/presentations/SMWest-12010-Expression-Encoder.pdf, 20 pgs.
Laukens, “Adaptive Streaming—A Brief Tutorial”, EBU Technical Review, 2011, 6 pgs.
Legault et al., “Professional Video Under 32-bit Windows Operating Systems”, SMPTE Journal, vol. 105, No. 12, Dec. 1996, 10 pgs.
Levkov, “Mobile Encoding Guidelines for Android Powered Devices”, Adobe Systems Inc., Addendum B, Dec. 22, 2010, 42 pgs.
Li et al., “Content-Aware Playout and Packet Scheduling for Video Streaming Over Wireless Links”, IEEE Transactions on Multimedia, vol. 10, No. 5, Aug. 2008, pp. 885-895.
Li et al., “Layered Video Multicast with Retransmission (LVMR): Evaluation of Hierarchical Rate Control”, Proceedings of IEEE INFOCOM'98, the Conference on Computer Communications. Seventeenth Annual Joint Conference of the IEEE Computer and Communications Societies. Gateway to the 21st Century, Cat. No. 98, vol. 3, 1998, 26 pgs.
Lian et al., “Efficient video encryption scheme based on advanced video coding”, Multimed. Tools Appl. vol. 38, 2008, pp. 75-89.
Lin et al., “Multipass Encoding for Reducing Pulsing Artifacts in Cloud Based Video Transcoding”, IEEE International Conference on Image Processing (ICIP), Date of Conference Sep. 27, 30, 2015, Quebec City, QC, Canada, pp. 907-911.
List et al., “Adaptive deblocking filter”, IEEE transactions on circuits and systems for video technology, vol. 13, No. 7, Jul. 2003, pp. 614-619.
Liu et al., “A Formal Framework for Component Deployment”, OOPSLA 2006, Proceedings of the 21st Annual ACM SIGPLAN Conference on Object-Oriented Programming Systems, Language, and Applications, Portland, Oregon, Oct. 22-26, 2006, pp. 325-344.
Long et al., “Silver: Simplifying Video Editing with Metadata”, Demonstrations, CHI 2003: New Horizons, Apr. 5-10, 2003, pp. 628-629.
Massoudi et al., “Overview on Selective Encryption of Image and Video: Challenges and Perspectives”, EURASIP Journal on Information Security, Nov. 2008, 18 pgs.
Matroska, “Diagram”, Matroska, Technical/Info, Diagram, 2016, retrieved from https://www.matroska.org/technical/diagram/index.html on Jul. 20, 2017, 9 pages.
Matroska, “Specifications”, Matroska, Technical/Info, Specifications, Jun. 25, 2017, retrieved from https://www.matroska.org/technical/specs/index.html on Jul. 20, 2017, 20 pages.
McCanne et al., “Receiver-driven Layered Multicast”, Conference proceedings on Applications, technologies, architectures, and protocols for computer communications, Aug. 1996, 14 pgs.
Meier, “Reduction of Blocking Artifacts in Image and Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 9, No. 3, Apr. 1999, pp. 490-500.
Miras, “On Quality Aware Adaptation of Internet Video”, University of London, PhD dissertation, May 2004, 181 pgs.
Morrison, “EA IFF 85 Standard for Interchange Format Files”, Jan. 14, 1985, printed from http://www.dcs.ed.ac.uk/home/mxr/gfx/2d/IFF.txt on Mar. 6, 2006, 24 pgs.
Moscoso, Pedro Gomes, “Interactive Internet TV Architecture Based on Scalable Video Coding”, Instituto Superior Techico, Universidad Technica de Lisboa, May 2011, 103 pgs.
MSDN, “Adaptive streaming, Expression Studio 2.0”, Apr. 23, 2009, 2 pgs.
Nelson, “Arithmetic Coding + Statistical Modeling = Data Compression: Part 1—Arithmetic Coding”, Doctor Dobb's Journal, Feb. 1991, USA, pp. 1-12.
Nelson, Mark, “Arithmetic Coding + Statistical Modeling = Data Compression: Part 1—Arithmetic Coding”, Doctor Dobb's Journal, Feb. 1991, printed from http://www.dogma.net/markn/articles/arith/part1.htm; printed Jul. 2, 2003, 12 pgs.
Nelson, Michael, “IBM's Cryptolopes”, Complex Objects in Digital Libraries Course, Spring 2001, Retrieved from http://www.cs.odu.edu/˜mln/teaching/unc/inls210/?method=display&pkg_name=cryptolopes.pkg&element_name=cryptolopes.ppt, 12 pages.
Newton et al., “Preserving Privacy by De-identifying Facial Images”, Carnegie Mellon University School of Computer Science, Technical Report, CMU-CS-03-119, Mar. 2003, 26 pgs.
Noboru, “Play Fast and Fine Video on Web! codec”, Co.9 No. 12, Dec. 1, 2003, pp. 178-179.
Noe, Alexander, “AVI File Format”, http://www.alexander-noe.com/video/documentation/avi.pdf, Dec. 14, 2006, pp. 1-26.
Noe, Alexander, “Definitions”, Apr. 11, 2006, retrieved from http://www.alexander-noe.com/video/amg/definitions.html on Oct. 16, 2013, 2 pages.
O'Brien, U.S. Appl. No. 60/399,846, filed Jul. 30, 2002, 27 pgs.
Ooyala, “Widevine Content Protection”, Ooyala Support Center for Developers. Ooyala, Inc., 2013. Jun. 3, 2013. http://support.ooyala.com/developers/documentation/concepts/player_v3_widevine_integration.html.
O'Rourke, “Improved Image Decompression for Reduced Transform Coding Artifacts”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 5, No. 6, Dec. 1995, pp. 490-499.
Oyman et al., “Quality of Experience for HTTP Adaptive Streaming Services”, IEEE Communications Magazine, Apr. 2012, vol. 50, No. 4, pp. 20-27, DOI: 10.1109/MCOM.2012.6178830.
Ozer, Jan, “Adaptive Streaming in the Field”, Streaming Media, Dec. 2010-Jan. 2011, pp. 36-47.
Padiadpu, Rashmi, “Towards Mobile Learning: A Scorm Player for the Google Android Platform”, Master Thesis, Hamburg University of Applied Sciences, 2008, 66 pgs.
Papagiannaki et al., “Experimental Characterization of Home Wireless Networks and Design Implications”, INFOCOM 2006, 25th IEEE International Conference of Computer Communications, Proceedings, Apr. 2006, 13 pgs.
Park et al., “A postprocessing method for reducing quantization effects in low bit-rate moving picture coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 9, No. 1, Feb. 1999, pp. 161-171.
Peek, David, “Consumer Distributed File Systems”, Dissertation, Doctor of Philosophy, Computer Science and Engineering, The University of Michigan, 2009, 118 pgs.
Phamdo, “Theory of Data Compression”, printed on Oct. 10, 2003, 12 pgs.
Richardson, “H.264 and MPEG-4 Video Compression”, Wiley, 2003, 306 pgs. (presented in 2 parts).
Rosenberg et al., “Indicating User Agent Capabilities in the Session Initiation Protocol (SIP)”, Network Working Group, RFC 3840, Aug. 2004, 36 pgs.
Schulzrinne, H, “Real Time Streaming Protocol 2.0 (RTSP): draft-ietfmmusic-rfc2326bis-27”, MMUSIC Working Group of the Internet Engineering Task Force (IETF), 296 pgs. (presented in two parts), Mar. 9, 2011, 296 pages.
Silvia, “Adaptive HTTP Streaming for Open Codecs”, Oct. 9, 2010, [retrieved on: Mar. 2, 2018.
Sima et al., “An Efficient Architecture for Adaptive Deblocking Filter of H.264 AVC Video Coding”, IEEE Transactions on Consumer Electronics, vol. 50, No. 1, Feb. 2004, pp. 292-296.
Spanos et al., “Performance Study of a Selective Encryption Scheme for the Security of Networked, Real-Time Video”, Proceedings of the Fourth International Conference on Computer Communications and Networks, IC3N'95, Sep. 20-23, 1995, Las Vegas, NV, pp. 2-10.
Srinivasan et al., “Windows Media Video 9: overview and applications”, Signal Processing: Image Communication, 2004, 25 pgs.
Stockhammer, “Dynamic Adaptive Streaming over HTTP—Standards and Design Principles”, Proceedings of the second annual ACM conference on Multimedia, Feb. 2011, pp. 133-145.
Tan, Yap-Peng et al., “Video transcoding for fast forward/reverse video playback”, IEEE ICIP, 2002, pp. I-713 to I-716.
Taxan, “A Vel LinkPlayer2 for Consumer”, I-O Data USA—Products—Home Entertainment, printed May 4, 2007 from http://www.iodata.com/usa/products/products.php?cat=HNP&sc=AVEL&pld=AVLP2/DVDLA&ts=2&tsc, 1 pg.
Timmerer et al., “HTTP Streaming of MPEG Media”, Proceedings of Streaming Day, 2010, 4 pgs.
Tiphaigne et al., “A Video Package for Torch”, Jun. 2004, 46 pgs.
Trappe et al., “Key Management and Distribution for Secure Multimedia Multicast”, IEEE Transaction on Multimedia, vol. 5, No. 4, Dec. 2003, pp. 544-557.
Tripathi et al., “Improving Multimedia Streaming with Content-Aware Video Scaling”, Retrieved from: http://digitalcommons.wpi.edu/computerscience-pubs/96, 2001, 17 pgs.
Unknown, “AVI RIFF File Reference (Direct X 8.1 C++ Archive)”, printed from http://msdn.microsoft.com/archive/en-us/dx81_c/directx_cpp/htm/avirifffilereference.asp?fr . . . on Mar. 6, 2006, 7 pgs.
Unknown, “Entropy and Source Coding (Compression)”, TCOM 570, 1999-9, pp. 1-22.
Unknown, “MPEG-4 Video Encoder: Based on International Standard ISO/IEC 14496-2”, Patni Computer Systems, Ltd., publication date unknown, 15 pgs.
Van Deursen et al., “On Media Delivery Protocols in the Web”, 2010 IEEE International Conference on Multimedia and Expo, Jul. 19-23, 2010, 6 pgs.
Venkatramani et al., “Securing Media for Adaptive Streaming”, Multimedia 2003 Proceedings of the Eleventh ACM International Conference on Multimedia, Nov. 2-8, 2003, Berkeley, California, 4 pgs.
Ventura, Guillermo Albaida, “Streaming of Multimedia Learning Objects”, AG Integrated Communication System, Mar. 2003, 101 pgs.
Waggoner, “Compression for Great Digital Video”, 2002, 184 pgs.
Wang et al., “Image Quality Assessment: From Error Visibility to Structural Similarity”, IEEE Transactions on Image Processing, Apr. 2004, vol. 13, No. 4, pp. 600-612.
Watanabem et al., “MPEG-2 decoder enables DTV trick plays”, esearcher System LSI Development Lab, Fujitsu Laboratories Ltd., Kawasaki, Japan, Jun. 2001, 2 pgs.
Wiegand, “Joint Video Team (JVT) of ISO/IEC MPEG and ITU-T VCEG”, Jan. 2002, 70 pgs.
Willig et al., U.S. Appl. No. 61/409,285, filed Nov. 2, 2010, 43 pgs.
Yang et al., “Projection-Based Spatially Adaptive Reconstruction of Block-Transform Compressed Images”, IEEE Transactions on Image Processing, vol. 4, No. 7, Jul. 1995, pp. 896-908.
Yang et al., “Regularized Reconstruction to Reduce Blocking Artifacts of Block Discrete Cosine Transform Compressed Images”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 3, No. 6, Dec. 1993, pp. 421-432.
Yu et al., “Video deblocking with fine-grained scalable complexity for embedded mobile computing”, Proceedings 7th International Conference on Signal Processing, Aug. 31-Sep. 4, 2004, pp. 1173-1178.
Zakhor, “Iterative Procedures for Reduction of Blocking Effects in Transform Image Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 2, No. 1, Mar. 1992, pp. 91-95.
Prosecution File History for U.S. Pat. No. 10,225,588, IPR filed Feb. 15, 2020, 2937 pgs., presented in 29 parts.
U.S. Appl. No. 61/530,305, filed Sep. 1, 2011, 6 pgs.
Wayback Machine, Grooveshark—Features, All Your Music In One Place, printed Aug. 15, 2016 from https://web.archive.org/web/20081013115837/http://www,grooveshark.com/features, 6 pgs.
“Broadcom BCM7413 Product Brief”, Dec. 11, 2008, 2 pgs.
“Common Interface Specification for Conditional Access and other Digital Video Broadcasting Decoder Applications”, European Standard, EN 50221, Feb. 1997, 86 pgs.
“Free music was never so cool before Grooveshark”, Wayback Machine, Grooveshark, Startup Meme, May 31, 2008, printed Aug. 15, 2016 from https://web.archive.org/web/20080601173852/http://startupmeme.com/2008/05/31/free-music-was-never-so-wool-before-grooveshark/, 2 pgs.
“HTTP Live Streaming on the Leading Media CDN”, Akamai website, retrieved from http://www.akamai.com/html/resources/http-live-streaming.html, 2015, accessed May 11, 2015, 5 pgs.
“Instantly convert songs into tiny URLs with TinySong”, Wayback Machine, Startup Memo Technology Blog, printed Aug. 15, 2016 from https://seb.archive.org/web/2008919133853/http://startupmeme.com/instantly-convert-songs-into-tiny-urls-with-tinysong/, 4pgs.
“Media Delivery Solutions for Streaming Video and Software Delivery”, Akamai website, Retrieved from http://www.akamai.com/html/solutions/media-delivery-solutions.html, 2015, Accessed May 11, 2015, 5 pgs.
“Microsoft Announces Breakthrough Technology Enabling Simple Access to Broad Set of Digital Content, Including Music, Games, Video, Ring Tones and Pictures”, Microsoft, Feb. 12, 2017, Retrieved from https://news.microsoft.com/2007/02/12/microsoft-announces-breakthrough-technology-enabling-simple-access-to-broad-set-of-digital-content-including-music-games-video-ring-tones-and-pictures/, 5 pgs.
“SDMI Secure Digital Music Initiative”, SDMI Portable Device Specification, Part 1, Version 1.0, Jul. 8, 1999, pp. 1-35.
“Server ‘Trick Play’ support for MPEG-2 Transport Stream Files”, www.live555.com/liveMedia/transport-stream-trick-play.html, 2006, Dec. 31, 2020, 1 pg.
“The Live555 Media Server”, www.live555.com/mediaServer/#about, 2006, printed Dec. 31, 2020, 3 pgs.
“Twitpic's Future”, Twitpic, Oct. 25, 2014, Retrieved from: https://web.archive.org/web/20150521043642/https://blog.twitpic.com/index.html, 12 pgs.
“What is Fliggo?”, Wayback Machine, printed Aug. 15, 2016 from https://web.archive.org/web/20080623065120/http://www.fliggo.com/about, 3 pgs.
ADB, “ADB-3800W Datasheet”, 2007, 2 pgs.
Agi et al., “An Empirical Study of Secure MPEG Video Transmissions”, IEEE, Mar. 1996, 8 pgs., DOI: 10.1109/NDSS.1996.492420.
Ahmed et al., “An Efficient Chaos-Based Feedback Stream Cipher (ECBFSC) for Image Encryption and Decryption”, Informatica, Mar. 2007, vol. 31, No. 1, pp. 121-129.
Arachchi et al., “Adaptation-aware encryption of scalable H.264/AVC for content security”, Signal Processing: Image Communication, Jul. 2009, vol. 24, pp. 468-483, doi:10.1016/j.image.2009.02.004.
Bell et al., “The BellKor 2008 Solution to the Netflix Prize”, Netflix Prize, 2008, 21 pgs.
Catone, Josh “10 Ways to Share Music on Twitter”, Mashable, May 29, 2009, Retrieved from: https://mashable.com/2009/05/29/twitter-music/#vJCdrVzNOOqx, 5 pgs.
Chesler, Oliver “TinySong is like TinyURL for music”, wire to the ear, Jun. 30, 2008, printed Aug. 15, 2016 from https://web.archive.org/web/20080907100459/http://www.wiretotheear.com/2008/06/30/tinysongis-like-tinyurl-for-music, 8 pgs.
Conklin et al., “Video coding for streaming media delivery on the Internet”, IEEE Transactions on Circuits and Systems for Video Technology, Mar. 2001, vol. 11, No. 3, pp. 269-281.
Deshpande et al., “Scalable Streaming of JPEG2000 Images Using Hypertext Transfer Protocol”, Multimedia '01: Proceedings of the Ninth ACM International Conference on Multimedia, Oct. 2001, pp. 372-381. https://doi.org/10.1145/500141.500197.
Entone, “Amulet High Definition IP Television Receiver User's Guide”, 2008, 28 pgs.
Entone, “Hydra HD IP Video Gateway”, 2008, 2 pgs.
International Standard, Information technology—Generic coding of moving pictures and associated audio information: Systems, ISO/IEC 13818-1:2000(E), Dec. 1, 2000 174 pgs.
Decision Granting Institution of Inter Partes Review 35 U.S.C. § 314, IPR2020-00558, U.S. Pat. No. 10,225,588, 46 pgs.
Decision Granting Petitioner's Request on Rehearing 37 C.F.R. § 42.71(d) Granting Institution of Inter Partes Review 35 U.S.C. § 314, IPR2020-00614 U.S. Pat. No. 7,295,673, 29 pgs.
Declaration of Patrick McDaniel, Ph.D., Inter Partes Review of U.S. Pat. No. 10,225,588, IPR filed Feb. 15, 2020, 211 pgs.
Extended European Search Report for European Application No. 19211286.0, Search completed Jul. 3, 2020, dated Jul. 13, 2020, 9 Pgs.
Extended European Search Report for European Application No. 19211291.0, Search completed Jul. 6, 2020, dated Jul. 14, 2020, 12 Pgs.
First Amended Complaint for Patent Infringement, DivX, LLC v. Netflix, Inc., No. 2:19-cv-1602-PSG, Am. Compl. (C.D. Cal Aug. 21, 2019), 229 pgs., IPR filed Feb. 15, 2020.
International Preliminary report on Patentability for International Application No. PCT/US2005/025845, report dated Jun. 19, 2007, 6 pgs.
International Preliminary Report on Patentability for International Application PCT/US2010/020372, Completed Oct. 6, 2011, 6 Pgs.
Petition for Inter Partes Review of U.S. Pat. No. 10,225,588, IPR2020-00558, 96 pgs., IPR filed Feb. 15, 2020.
Power of Attorney—Hulu, LLC (IPR2020-00558), 4 pgs., IPR filed Feb. 15, 2020.
Power of Attorney—Netflix, Inc. (IPR2020-00558), 4 pgs., IPR filed Feb. 15, 2020.
Proceedings of the Second KDD Workshop on Large-Scale Recommender Systems and the Netflix Prize Competition, Las Vegas, Nevada, Aug. 24, 2008, 34 pgs.
Prosecution File History for U.S. Appl. No. 13/340,623 to Kiefer et al. (“Kiefer”), IPR filed Feb. 15, 2020, 1249 pgs., presented in 6 parts.
ETSI, “Digital Video Broadcasting (DVB) Support for use of scrambling and Conditional Access (CA) within digital broadcasting systems”, Oct. 1996, 13 pgs.
ETSI, “Digital Video Broadcasting (DVB); Implementation guidelines for the use of Video and Audio Coding in Contribution and Primary Distribution Applications based on the MPEG-2 Transport Stream”, ETSI TS 102 154 V1.2.1, May 2004, 73 pgs.
Fahmi et al., “Proxy Servers for Scalable Interactive Video Support”, Computer, Sep. 2001, vol. 45, No. 9, pp. 54-60, https://doi.org/10.1109/2.947092.
Fitzek et al., “A Prefetching Protocol for Continuous Media Streaming in Wireless Environments”, IEEE Journal on Selected Areas in Communications, Oct. 2001, vol. 19, No. 10, pp. 2015-2028, DOI: 10.1109/49.957315.
Ho, “Digital Video Broadcasting Conditional Access Architecture”, Report prepared for CS265-Section 2, Fall 2002, Prof Stamp, 7 pgs.
INCITS/ISO/IEC, “Information Technology—Generic Coding of Moving Pictures And Associated Audio Information: Video (Formerly ANSI/ISO/IEC 13818-2-2000)”, Second edition, Dec. 15, 2000, 220 pgs., (presented in two parts).
ISMA, “ISMA Encryption and Authentication, Version 1.1, Area / Task Force: DRM”, Internet Streaming Media Alliance, Sep. 15, 2006, pp. 1-64.
ITU-T, “Series J: Cable Networks and Transmission of Television, Sound Programme and Other Multimedia Signals”, Technical method for ensuring privacy in long-distance international MPEG-2 television transmission conforming to ITU-T J.89, ITU-T Recommendation J.96, Mar. 2001, 34 pgs.
Kabir, “Scalable and Interactive Multimedia Streaming Over the Internet”, Thesis, 2005, 207 pgs.
Krikor et al., “Image Encryption Using DCT and Stream Cipher”, European Journal of Scientific Research, Jan. 2009, vol. 32, No. 1, pp. 48-58.
Lew et al., “Content-Based Multimedia Information Retrieval: State of the Art and Challenges”, ACM Transactions on Multimedia Computing, Communications and Applications, Feb. 2006, vol. 2, No. 1, pp. 1-19.
Lian et al., “Recent Advances in Multimedia Information System Security”, Informatica, Jan. 2009, vol. 33, pp. 3-24.
Lian et al., “Selective Video Encryption Based on Advanced Video Coding”, PCM, Nov. 2005, Part II, LNCS 3768, pp. 281-290.
Lievaart, “Characteristics that differentiate CA Systems”, Irdeto access, Nov. 2001, 5 pgs.
Lloyd, “Supporting Trick Mode Playback Universally Across the Digital Television Industry”, Thesis, 2005, 111 pgs.
Lomas et al., “Educause Learning Initiative, Collaboration Tools”, Educause Learning Initiative, Aug. 2008, ELI Paper 2: 2008, 11 pgs.
Macaulay et al., “Whitepaper—IP Streaming of MPEG-4: Native RTP vs MPEG-2 Transport Stream”, Envivio, Oct. 2005, 12 pgs.
Martin et al., “Privacy Protected Surveillance Using Secure Visual Object Coding”, IEEE Transactions on Circuits and Systems for Video Technology, Aug. 2008, vol. 18, No. 8, pp. 1152-1162, DOI: 10.1109/TCSVT.2008.927110.
Meyer et al., “Security mechanisms for Multimedia-Data with the Example MPEG-I-Video”, SECMPEG, 1992, 10 pgs.
Molavi et al., “A Security Study of Digital TV Distribution Systems”, Thesis, Jun. 2005, 112 pgs.
Montes, “Muusic: mashup de servicios web musicales”, Ingenieria Tecnica en Informatica de Gestion, Nov. 2008, 87 pgs.
Nelson, “The Data Compression Book”, M&T Publishing, 1992, 533 pgs., (presented in two parts).
OIPF Open Forum, “OIPF Release 1 Specification, vol. 3, Content Metadata”, OIPF, Oct. 8, 2009, vol. 1.1, 47 pgs.
OIPF Open Forum, “OIPF Release 1 Specification, vol. 7—Authentication, Content Protection and Service Protection”, OIPF, Oct. 8, 2009, vol. 1.1, 88 pgs.
Open IPTV Forum, “Functional Architecture”, Jan. 16, 2008, vol. 1.1, 141 pgs.
Open IPTV Forum, “OIPF Release 1 Specification, vol. 1—Overview”, OIPF, Oct. 8, 2009, vol. 1.1, 48 pgs.
Open IPTV Forum, “OIPF Release 1 Specification, vol. 2, Media Formats”, OIPF, Oct. 8, 2009, vol. 1.1, 22 pgs.
Park et al., “An Efficient Encryption and Key Management Scheme for Layered Access Control of H.265/Scalable Video Coding”, IEICI Trans. Inf. & Syst., May 2009, vol. E92-D, No. 5, pp. 851-858, DOI: 1031587/transinf.E92.D.851.
Park et al., “Combined Scheme of Encryption and Watermarking in H.264/Scalable Video Coding”, New Directions in Intelligent Interactive Multimedia, SCI 142, 2008, pp. 351-361.
Qiao et al., “Comparison of MPEG Encryption Algorithms”, Comput. & Graphics, 1998, vol. 22, No. 4, pp. 437-448.
Raju et al., “Fast and Secure Real-Time Video Encryption”, Sixth Indian Conference on Computer Vision, Graphics & Image Processing, Jan. 2009, pp. 257-264, doi:10.1109/ACVGIP.2008.100.
Senoh et al., “DRM Renewability & Interoperability”, IEEE Xplore, Conference: Consumer Communications and Networking Conference, 2004, Feb. 2004, pp. 424-429, DOI: 10.1109/CCNC.2004.1286899 Conference: Consumer Communications and Networking Conference, 2004. CCNC 2004. First IEEE.
Sheu, Tsang-Ling et al., “Dynamic layer adjustments for SVC segments in P2P streaming networks”, Computer Symposium (ICS), 2010, 2010 International, Tainan, Taiwan, R.O.C., pp. 793-798, 2010.
Shojania et al., “Experiences with MPEG-4 Multimedia Streaming”, CiteSeer, Jan. 2001, 3 pgs., DOI: 10.1145/500141.500221.
Symes, “Video Compression Demystified”, McGraw-Hill, 2001, 353 pgs., (presented in two parts).
Taymans et al., “GStreamer Application Development Manual (1.6.0)”, 2007, 159 pgs.
Thomas et al., “A Novel Secure H.264 Transcoder Using Selective Encryption”, Proceedings in International Conference on Image Processing, Jan. 2007, vol. 4, pp. IV-85-IV-88, DOI: 10.1109/ICIP.2007.4379960.
Toscher et al., “The BigChaos Solution to the Netflix Prize 2008”, Netflix Prize, Nov. 25, 2008, 17 pgs.
Tosun et al., “Efficient multi-layer coding and encryption of MPEG video streams”, 2000 IEEE International Conference on Multimedia and Expo. ICME2000. Proceedings. Latest Advances in the Fast Changing World of Multimedia (Cat. No. 00TH8532), July 30-Aug. 2, 2000, pp. 119-122, DOI: 10.1109/ICME.2000.869559.
Um, “Selective Video Encryption of Distributed Video Coded Bitstreams and Multicast Security over Wireless Networks”, Thesis, Aug. 2006, 142 pgs.
Van Grove, Jennifer “Top 5 Ways to Share Videos on Twitter”, Mashable, May 23, 2009, Retrieved from: https://mashable.com/2009/05/23/video-for-twitter/#Jvn9IIYy6qqA, 6 pgs.
Wang, “Lightweight Encryption in Multimedia”, Thesis, Jun. 2005, 184 pgs.
Weng, “A Multimedia Socail-Networking Community for Mobile Devices”, 2007, 30 pgs.
Wong, “Web Client Programming with Perl”, 1997, printed Jan. 8, 2021 from: https://www.oreilly.com/openbook-webclientch03.html, 31 pgs.
Wu, “A Fast MPEG Encryption Algorithm and Implementation of AES on CAM”, Thesis, Oct. 6, 2003, 91 pgs.
Yuksel, “Partial Encryption of Video for Communication and Storage”, Thesis, Sep. 2003, 78 pgs.
“Supported Media Formats”, Supported Media Formats, Android Developers, Nov. 27, 2013, 3 pgs.
European Search Report for Application 11855103.5, search completed Jun. 26, 2014, 9 pgs.
European Search Report for Application 11855237.1, search completed Jun. 12, 2014, 9 pgs.
Federal Computer Week, “Tool Speeds Info to Vehicles”, Jul. 25, 1999, 5 pages.
HTTP Live Streaming Overview, Networking & Internet, Apple, Inc., Apr. 1, 2011, 38 pages.
Informationweek: Front End: Daily Dose, “Internet on Wheels”, Jul. 20, 1999, 3 pgs.
International Search Report and Written Opinion for International Application PCT/US2011/066927, International Filing Date Dec. 22, 2011, Report Completed Apr. 3, 2012, Mailed Apr. 20, 2012, 14 pgs.
International Search Report and Written Opinion for International Application PCT/US2011/067167, International Filing Date Dec. 23, 2011, Report Completed Jun. 19, 2012, Mailed Jul. 2, 2012, 11 pgs.
The Official Microsoft IIS Site, Smooth Streaming Client, 4 pages.
Author Unknown, “Tunneling QuickTime RTSP and RTP over HTTP”, Published by Apple Computer, Inc.: 1999 (month unknown), 6 pages.
“Netflix turns on subtitles for PC, Mac streaming”, 3 pgs.
Deutscher, “IIS Transform Manager Beta-Using the MP4 to Smooth Task”, Retrieved from: https://web.archive.org/web/20130328111303/http://blog.johndeutscher.com/category/smooth-streaming, Blog post of Apr. 17, 2010, 14 pgs.
Gannes, “The Lowdown on Apple's HTTP Adaptive Bitrate Streaming”, GigaOM, Jun. 10, 2009, 12 pgs.
Kim, Kyuheon, “MPEG-2 ES/PES/TS/PSI”, Kyung-Hee University, Oct. 4, 2010, 66 pages.
Nelson, “Smooth Streaming Deployment Guide”, Microsoft Expression Encoder, Aug. 2010, 66 pgs.
Noe, A., “Matroska File Format (under construction!)”, Retrieved from the Internet: URL: http://web.archive.org web/20070821155146/www.matroska.org/technical/specs/matroska.pdf, retrieved on Jan. 19, 2011, Jun. 24, 2007, 1-51.
Ozer, “The 2012 Encoding and Transcoding Buyers' Guide”, Streamingmedia.com, Retrieved from: http://www.streamingmedia.com/Articles/Editorial/Featured-Articles/The-2012-Encoding-and-Transcoding-Buyers-Guide-84210.aspx, 2012, 8 pgs.
Pantos, “HTTP Live Streaming, draft-pantos-http-live-streaming-10”, IETF Tools, Oct. 15, 2012, Retrieved from: http://tools.ietf.org/html/draft-pantos-http-live-streaming-10, 37 pgs.
Pantos, R, “HTTP Live Streaming: draft-pantos-http-live-streaming-06”, Published by the Internet Engineering Task Force (IETF), Mar. 31, 2011, 24 pages.
RGB Networks, “Comparing Adaptive HTTP Streaming Technologies”, Nov. 2011, Retrieved from: http://btreport.net/wp-content/up-loads/2012/02/RGB-Adaptive-HTTP-Streaming-Comparison-1211-01.pdf, 20 pgs.
Schulzrinne, H et al., “Real Time Streaming Protocol 2.0 (RTSP): draft-ietfmmusic-rfc2326bis-27”, MMUSIC Working Group of the Internet Engineering Task Force (IETF), Mar. 9, 2011, 296 pages.
Siglin, “HTTP Streaming: What You Need to Know”, streamingmedia.com, 2010, 15 pages.
Siglin, “Unifying Global Video Strategies, MP4 File Fragmentation for Broadcast, Mobile and Web Delivery”, Nov. 16, 2011, 16 pgs.
Wu, Feng et al., “Next Generation Mobile Multimedia Communications: Media Codec and Media Transport Perspectives”, In China Communications, Oct. 2006, pp. 30-44.
Zambelli, Alex “IIS Smooth Streaming Technical Overview”, Microsoft Corporation, Mar. 2009.
Continuations (1)
Number Date Country
Parent 15950950 Apr 2018 US
Child 17485109 US
Reissues (2)
Number Date Country
Parent 13731949 Dec 2012 US
Child 15950950 US
Parent 13731949 Dec 2012 US
Child 17485109 US