The present invention relates to digital watermarking and more particularly relates to client-side digital watermarking.
Watermarking of digital content is an effective means for identifying copyright information. A digital watermark is data that is encoded into digital content in a manner that may or may not be humanly perceptible. In general, watermarks may be encoded into the digital content in either the spatial domain or the frequency domain. Watermarks provide a way to add non-removable or editable information to content with no or minimal degradation to the original content. Depending on the specific watermark technology, it may be difficult or even impossible to remove the watermark without severely degrading content quality. Companies such as Digimark and Verimatrix have implemented successful digital watermarking technologies for still photos and video imaging respectively.
In a broadcast or multicast video transport system, digital watermarking can most easily be accomplished at the source of the video broadcast. This approach delivers video content having a common watermark to each termination or client receiving the broadcast and can clearly provide a non-removable label identifying copyright restrictions. However, this common watermark provides no deterrent to the user against anonymous redistribution. Such redistribution can occur through public Peer-to-Peer (P2P) networks, darknets, or postings to video sharing sites. As such, it is desirable to use client-side watermarking at the termination of the multicast to clearly identify the end user. If the user then redistributes the content illegally, the watermark may be used to trace the content to the user.
In order to apply a high fidelity, robust watermark, it is desirable to apply the watermark to non-compressed content or content that has been transmitted in a lossless format. Thus, for client-side watermarking in a video distribution system, an issue arises due to the fact that the distributed video content is highly compressed. Traditionally, the client decompresses the compressed video content, applies the watermark, and then re-compresses the watermarked video content. However, as a result of the decompression, watermarking, and re-compression of the watermarked content, the quality of the video content may be significantly reduced. Recent advances in technology enabling increased compression ratios further compound this issue. For example, the H.264 (or MPEG4 Part 10) standard has a compression ratio of 1:32 which is twice that of an MPEG2. The increased compression ratio of H.264 makes effective client-side watermarking increasingly more difficult due to the lossy nature of the compression.
Thus, there is a need for a system and method for providing client-side watermarking in a manner that does not significantly reduce the quality of the digital content.
The present invention relates to client-side watermarking of digital content using hybrid Intra-Frames (I-Frames). In general, a content source provides a compressed video stream and a hybrid I-Frame stream to a client device via a network. The hybrid I-Frame stream includes a number of low-loss I-Frames corresponding to select ones of the I-Frames in the compressed video stream to be used for client-side watermarking. The client device watermarks the I-Frames in the hybrid I-Frame stream, optionally compresses the watermarked I-Frames, and replaces the select ones of the I-Frames in the compressed video stream with the watermarked and optionally compressed I-Frames to provide a watermarked version of the compressed video stream.
Those skilled in the art will appreciate the scope of the present invention and realize additional aspects thereof after reading the following detailed description of the preferred embodiments in association with the accompanying drawing figures.
The accompanying drawing figures incorporated in and forming a part of this specification illustrate several aspects of the invention, and together with the description serve to explain the principles of the invention.
The embodiments set forth below represent the necessary information to enable those skilled in the art to practice the invention and illustrate the best mode of practicing the invention. Upon reading the following description in light of the accompanying drawing figures, those skilled in the art will understand the concepts of the invention and will recognize applications of these concepts not particularly addressed herein. It should be understood that these concepts and applications fall within the scope of the disclosure and the accompanying claims.
The content source 12 may be one or more servers operating to distribute digital video content to the client devices 14-1 through 14-N. For example, the content source 12 may be one or more servers providing an Internet Protocol Television (IPTV) service. In general, the content source 12 includes a source-side encoder 18. The source-side encoder 18 may be implemented in hardware, software, or a combination of hardware and software. In operation, the source-side encoder 18 receives a raw digital video input and processes the raw digital video input to provide a compressed video stream such as, for example, an MPEG2 or MPEG4 (H.264) video stream or any applicable video compression standard. In addition, the source-side encoder 18 provides a hybrid I-Frame stream including a number of low-loss I-Frames corresponding to select ones of the I-Frames in the compressed video stream that are to be used for client-side watermarking. As used herein, “low-loss” means that the hybrid I-Frames in the hybrid I-Frame stream are encoded or compressed with an essentially lossless algorithm or with an algorithm having a compression factor that is relatively low-loss as compared to a compression factor of the algorithm used to generate the compressed video stream.
In this embodiment, the source-side encoder 18 multicasts the compressed video stream and the hybrid I-Frame stream to the client devices 14-1 through 14-N as layered multicast streams. However, while multicasting is used in this embodiment, the present invention is not limited thereto. The content source 12 may alternatively unicast the compressed video content and the hybrid I-Frames to the client devices 14-1 through 14-N using a common transmission channel or separate transmission channels. For example, the content source 12 may provide the compressed video stream to the client device 14-1 via a satellite-based television network and provide the hybrid I-Frame stream to the client device 14-1 via a land-based network such as a cable network or Digital Subscriber Line (DSL) network.
The client devices 14-1 through 14-N may be, for example, set-top boxes, personal computers, mobile devices such as Personal Digital Assistants (PDAs) or mobile phones, or the like. The client devices 14-1 through 14-N generally include client-side re-encoders 20-1 through 20-N. The client-side re-encoders 20-1 through 20-N may be implemented in hardware, software, or a combination of hardware and software.
Using the client-side re-encoder 20-1 as an example, the client-side re-encoder 20-1 operates to receive the compressed video stream and the hybrid I-Frame stream from the content source 12. The client-side re-encoder 20-1 then watermarks the hybrid I-Frames and optionally compresses the watermarked I-Frames to a level appropriate or required for insertion into the compressed video stream. The information included in the watermark may vary depending on the particular implementation. As an example, the watermark may include information such as, but not limited to, the name and address of a user of the client device 14-1, a credit card number of a credit card issued to the owner of the client device 14-1, a device identifier (ID) of the client device 14-1, an Internet Protocol (IP) address of the client device 14-1, or the like or any combination thereof. The client-side re-encoder 20-1 then replaces corresponding I-Frames in the compressed video stream with the watermarked and optionally compressed I-Frames generated from the hybrid I-Frame stream, thereby providing a watermarked version of the compressed video stream. Further, by applying the watermark to the low-loss I-Frames in the hybrid I-Frame stream, a high fidelity, robust watermark is provided.
The source-side encoder 18 then processes the compressed video stream to identify one or more select I-Frames to be used for client-side watermarking (step 102). The manner in which the select I-Frames are identified may vary depending on the particular implementation. Two exemplary processes for identifying the select I-Frames are discussed below in detail. The source-side encoder 18 then identifies segments of the raw digital video input corresponding to the select I-Frames to be used for client-side watermarking and generates hybrid I-Frames corresponding to the select I-Frames using the identified segments of the raw digital video input (step 104). The source-side encoder 18 then sends the compressed video stream and the hybrid I-Frames to one or more of the client devices 14-1 through 14-N (step 106).
Again, while the discussion herein focuses on watermarking and compressing the hybrid I-Frames, compression of the watermarked hybrid I-Frames may be optional in some implementations. For example, the watermarked hybrid I-Frames may not be compressed and used to replace the select I-Frames in the compressed video stream at an appropriate point either during or after decompression of the compressed video stream.
An I-Frame detection and selection function 30 operates to monitor the compressed video stream output by the industry standard video encoder 26 to detect I-Frames. Upon detecting the I-Frames, the I-Frame detection and selection function 30 provides an I-Frame synchronization signal or message to a variable delay buffer 32 such that segments of the raw digital video input corresponding to the I-Frames in the compressed video stream are provided to a hybrid I-Frame generator 34.
The I-Frame detection and selection function 30 also identifies the select I-Frames to be used for client-side watermarking and provides an I-Frame selection signal or message to the hybrid I-Frame generator 34 to identify the select I-Frames. In response, the hybrid I-Frame generator 34 processes segments of the raw digital video input corresponding to the select I-Frames to provide the hybrid I-Frame stream. More specifically, the hybrid I-Frame generator 34 may perform tagging, encapsulation, and compression. Tagging may be used to associate each of the hybrid I-Frames with a corresponding I-Frame in the compressed video stream. The compression for the hybrid I-Frames is preferably lossless or nearly lossless. A lossless compression algorithm is a compression algorithm that allows the exact original data to be reconstructed from the compressed data during decompression. Exemplary lossless compression algorithms include the Huffman coding and arithmetic coding. Alternatively, rather than being lossless, the hybrid I-Frames may be compressed according to a lossy compression scheme that is relatively low-loss as compared to the encoding and compression scheme used by the industry standard video encoder 26. For example, the hybrid I-Frames may have a compression factor of 10 or less. In contrast, in an MPEG4 or H.264 system, the compressed video stream may have a compression factor of 32 for the I-Frame.
In addition, since in this example the compressed video stream is encrypted by the encryption function 28, a decryption key to be used on the client-side to decrypt the compressed video stream may optionally be embedded in one or more of the hybrid I-Frames. As an example, the decryption key may be embedded into one or more of the hybrid I-Frames as a watermark.
The hybrid I-Frame generator 34 may also limit the number of hybrid I-Frames based on a maximum I-Frame ratio. The maximum I-Frame ratio may be input to the hybrid I-Frame generator 34 or be embedded in the hybrid I-Frame generator 34. In general, the maximum I-Frame ratio defines a maximum number of hybrid I-Frames with respect to time. For example, the maximum I-Frame ratio may define the maximum number of hybrid I-Frames per second. Thus, if the I-Frames selected by the I-Frame detection and selection function 30 exceeds the maximum I-Frame ratio, then the hybrid I-Frame generator 34 may limit the number of generated I-Frames such that the maximum I-Frame ratio is not exceeded.
At this point, a synchronization function 36 may optionally control variable delay buffers 38 and 40 to synchronize the encrypted version of the compressed video stream and the hybrid I-Frame stream. In this example, the encrypted version of the compressed video stream and the hybrid I-Frame stream are then multicast to one or more of the client devices 14-1 through 14-N by a layered multicast streaming function 42. For example, the encrypted version of the compressed video stream and the hybrid I-Frame stream may be multicast to the client devices 14-1 through 14-N as provided by Internet Protocol version 4 (IPv4) or Internet Protocol version 6 (IPv6).
A watermarking function 52 operates to watermark and then compress the hybrid I-Frames. In one embodiment, the hybrid I-Frames may be watermarked with watermarking instructions, which may identify what information, or watermarking data, is to be included in the watermark added to the hybrid I-Frames. For example, the watermarking instructions may provide that the name and address of the user of the client device 14-1, a credit card number of a credit card owned by the user of the client device 14-1, a device ID of the client device 14-1, an IP address of the client device 14-1, or the like is to be included in the watermark. If the hybrid I-Frames are watermarked with watermarking instructions, the watermark extraction function 44 extracts the watermarking instructions and provides the watermarking instructions to the watermarking function 52. Alternatively, the watermarking data may be predetermined and known by the watermarking function 52.
In the illustrated embodiment, the watermarking instructions are provided in the hybrid I-Frames and identify the watermarking data to be included in the watermark. As such, the watermarking data is obtained from, for example, a control system of the client device 14-1 or the user of the client device 14-1. The watermarking function 52 then watermarks the hybrid I-Frames with a watermark including the watermarking data. The particular watermarking technique used by the watermarking function 52 may be any type of watermarking technique such as a spatial domain or frequency domain watermarking technique. An exemplary spatial domain watermarking technique is the Fredrich Algorithm. An exemplary frequency domain watermarking technique is the Khao Kotch Algorithm. However, various other watermarking techniques may be used as will be apparent to one of ordinary skill in the art upon reading this disclosure. Once watermarked, the watermarked hybrid I-Frames are compressed to a level that is appropriate for insertion into the compressed video stream.
The synchronization function 50 controls variable delay buffers 54 and 56 such that the watermarked and compressed hybrid I-Frames are synchronized to the corresponding select I-Frames in the compressed video stream to be replaced at an input of an I-Frame replacement function 58. The I-Frame replacement function 58 then replaces the select I-Frames of the compressed video stream with the watermarked and compressed hybrid I-Frames, thereby providing a watermarked version of the compressed video stream. The watermarked version of the compressed video stream may then be presented to the user via appropriate hardware and/or software. In addition or alternatively, the watermarked version of the compressed video stream may be stored in a digital storage device associated with the client device 14-1.
More specifically, the compressed video stream from the industry standard video encoder 26 (
A watermarking function 64 then watermarks the entropy decoded I-Frame with the sample watermark using a frequency domain watermarking technique. Preferably, the frequency domain watermarking technique is the same frequency domain watermarking technique to be used by the client devices 14-1 through 14-N for client-side watermarking. Entropy encoder 66, which may be referred to as a re-encoding function, then re-encodes the watermarked I-Frame and provides the watermarked I-Frame to an industry standard video decoder 68. Note that the entropy decoder 62, the watermarking function 64, and the entropy encoder 66 may generally be referred to herein as a watermarking system. Based on the watermarked I-Frame from the entropy encoder 66, the industry standard video decoder 68 decodes the I-Frame and the associated P-Frames and B-Frames for the GOP to provide decoded video frames. The decoded video frames are then provided to an error calculation function 70 and compared to corresponding segments of the raw digital video input in order to calculate an error for the GOP.
More specifically, for each decoded video frame, the decoded video frame is compared to a corresponding segment of the raw digital video input in order to calculate an error for that frame. The comparison may be, for example, pixel by pixel. However, numerous methods for determining an error value between the decoded video frame and the corresponding segment of the raw digital video input will be apparent to one of ordinary skill in the art upon reading this disclosure. The errors for each of the decoded video frames may be combined and optionally averaged to provide the error for the GOP. A decision function 72 then compares the error for the GOP to a predetermined error threshold range. In one embodiment, the error threshold range is defined by a predetermined maximum error threshold value. If the error is greater than the predetermined maximum error threshold value, then the I-Frame is not selected as an I-Frame that may be used for client-side watermarking. If the error is less than the predetermined maximum error threshold value, then the I-Frame is selected as an I-Frame that may be used for client-side watermarking. In addition to the error, the decision function 72 may consider GOP size, or the number of frames in the GOP. Note that it may be desirable to restrict I-Frame selection to those I-Frames associated with GOPs having less than a predetermined maximum number of frames.
While the I-Frame detection and selection function 30 of
The I-Frame detection and selection function 30 preferably does not evaluate two successive I-Frames. More specifically, when decoding the frames in the GOP for the I-Frame being evaluated, the I-Frame in the next GOP may be referenced by one or more B-Frames in the GOP. At the point of decoding the frames in the GOP under evaluation, the I-Frame in the next GOP is not watermarked. As such, the error value for the GOP depends on using the non-watermarked I-Frame in the next GOP. If the I-Frame in the next GOP were then evaluated and selected for watermarking, the I-Frame in the next GOP would be watermarked at the client-side, and the watermarked I-Frame would be used to decode the GOP. As a result, the error calculated for the GOP using the non-watermarked I-Frame in the next GOP is no longer a valid indicator of the error that will be introduced in the GOP due to client-side watermarking. Therefore, it is preferable that no two successive I-Frames be evaluated for watermarking.
Before proceeding to
The process then proceeds to
At some point, assuming that a frame in the GOP references the I-Frame for the next GOP, the I-Frame for the next GOP will be detected prior to the end of the GOP. Since the I-Frame detection and selection function 30 is in evaluation mode, the I-Frame for the next GOP is provided to the industry standard video decoder 68 and decoded to provide a decoded video frame (step 126). The decoded video frame is thereafter used to decode frames in the GOP that reference the I-Frame for the next GOP.
Since the last frame in the GOP has still not been detected, steps 100, 102, 122, 124, and 118 are repeated for each subsequent frame until the last frame in the GOP is detected in step 118. In this embodiment, once the last frame in the GOP is detected in step 118, the decision function 72 determines whether the error for the GOP is greater than an error threshold (step 128). If the error for the GOP is not greater than the error threshold, then the I-Frame under evaluation is selected as an I-Frame that may be used for client-side watermarking (step 130). If the error for the GOP is greater than the error threshold, then the I-Frame under evaluation is not selected. At this point, the I-Frame detection and selection function 30 transitions out of evaluation mode and evaluation is complete (step 132).
The I-Frame detector 60 operates to detect I-Frames in the compressed video stream from the industry standard video encoder 26 (
In an alternative embodiment, the error calculation function 70 may obtain the watermarked segment of the raw digital video input output by the watermarking function 74 and calculate the error for the I-Frame based on a comparison of the watermarked segment of the raw digital video input output by the watermarking function 74 and the corresponding non-watermarked segment of the raw digital video input.
Note that, rather than using the industry standard video decoder 68, a custom video decoder may be used. If so, the custom video decoder may be designed such that custom video decoder may receive the watermarked segment from the watermarking function 74 such that the encoder 76 is not needed.
As another note, while the I-Frame detection and selection function 30 of
The process then proceeds to
At some point, assuming that one or more frames in the GOP references the I-Frame for the next GOP, the I-Frame for the next GOP will be detected prior to the end of the GOP. Since the I-Frame detection and selection function 30 is in evaluation mode, the I-Frame for the next GOP is provided to the industry standard video decoder 68 and decoded to provide a decoded video frame (step 226). The decoded video frame is thereafter used to decode frames in the GOP that reference the I-Frame for the next GOP.
Assuming that the last frame in the GOP has still not been detected, steps 200, 202, 222, 224, and 218 are repeated for each subsequent frame until the last frame in the GOP is detected in step 218. In this embodiment, once the last frame in the GOP is detected in step 218, the decision function 72 determines whether the error for the GOP is greater than an error threshold (step 228). If the error for the GOP is not greater than the error threshold, then the I-Frame under evaluation is selected as an I-Frame that may be used for client-side watermarking (step 230). If the error for the GOP is greater than the error threshold, then the I-Frame under evaluation is not selected. At this point, the I-Frame detection and selection function 30 transitions out of evaluation mode and evaluation is complete (step 232).
The present invention provides substantial opportunity for variation without departing from the spirit or scope of the present invention. For example, while client devices 14-1 through 14-N are preferably devices such as set-top boxes, personal computers, mobile devices such as PDAs or mobile phones, or the like, the present invention is not limited thereto. The client devices 14-1 through 14-N may be intermediary nodes between the content source 12 and end user nodes. For example, the intermediary nodes may be servers associated with content distributors. Thus, as used herein, “client-side watermarking” should not be limited to watermarking at an end user device. Rather, “client-side watermarking” may occur on any node downstream of the content source 12. Similarly, a “client” may be any node downstream of the content source.
Those skilled in the art will recognize improvements and modifications to the preferred embodiments of the present invention. All such improvements and modifications are considered within the scope of the concepts disclosed herein and the claims that follow.
The present application is a continuation of U.S. patent application Ser. No. 13/668,934, entitled CLIENT-SIDE WATERMARKING USING HYBRID I-FRAMES, which was filed on Nov. 5, 2012; which was a continuation of U.S. Pat. No. 8,320,610, which issued on Nov. 27, 2012; which was a continuation of U.S. Pat. No. 7,983,444, which issued on Jul. 19, 2011; which was a continuation of U.S. Pat. No. 7,738,676, which issued on Jun. 15, 2010; which are incorporated by reference herein in its entirety for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
2280083 | Reeves | Apr 1942 | A |
5278834 | Mazzola | Jan 1994 | A |
5613004 | Cooperman et al. | Mar 1997 | A |
5687236 | Moskowitz et al. | Nov 1997 | A |
5758257 | Herz et al. | May 1998 | A |
5790935 | Payton | Aug 1998 | A |
5809139 | Girod et al. | Sep 1998 | A |
5818838 | Backes et al. | Oct 1998 | A |
5905800 | Moskowitz et al. | May 1999 | A |
6128736 | Miller | Oct 2000 | A |
6141753 | Zhao et al. | Oct 2000 | A |
6282299 | Tewfik et al. | Aug 2001 | B1 |
6389541 | Patterson | May 2002 | B1 |
6510234 | Cox et al. | Jan 2003 | B1 |
6567107 | Stannard | May 2003 | B1 |
6721282 | Motley | Apr 2004 | B2 |
6735699 | Sasaki et al. | May 2004 | B1 |
6738493 | Cox et al. | May 2004 | B1 |
6751670 | Patterson | Jun 2004 | B1 |
6774926 | Ellis et al. | Aug 2004 | B1 |
6804779 | Carroni et al. | Oct 2004 | B1 |
6975743 | Venkatesan et al. | Dec 2005 | B2 |
6987985 | Purkayastha et al. | Jan 2006 | B2 |
7003131 | Watson et al. | Feb 2006 | B2 |
7016668 | Vaidyanathan et al. | Mar 2006 | B2 |
7020304 | Alattar et al. | Mar 2006 | B2 |
7036024 | Watson | Apr 2006 | B2 |
7065607 | England et al. | Jun 2006 | B2 |
7167599 | Diehl | Jan 2007 | B1 |
7382879 | Miller | Jun 2008 | B1 |
7409556 | Wu et al. | Aug 2008 | B2 |
7430302 | Thorwirth | Sep 2008 | B2 |
7983440 | Roberts et al. | Jul 2011 | B1 |
7987490 | Ansari et al. | Jul 2011 | B2 |
8102863 | Evans | Jan 2012 | B1 |
8320610 | Evans et al. | Nov 2012 | B2 |
8630450 | Evans et al. | Jan 2014 | B2 |
20010051996 | Cooper et al. | Dec 2001 | A1 |
20020010759 | Hitson et al. | Jan 2002 | A1 |
20020013812 | Krueger et al. | Jan 2002 | A1 |
20020049580 | Kutaragi et al. | Apr 2002 | A1 |
20020054578 | Zhang et al. | May 2002 | A1 |
20020057799 | Kohno | May 2002 | A1 |
20020061029 | Dillon | May 2002 | A1 |
20020104003 | Iwamura | Aug 2002 | A1 |
20020104015 | Barzilai et al. | Aug 2002 | A1 |
20020104099 | Novak | Aug 2002 | A1 |
20020110123 | Shitama | Aug 2002 | A1 |
20020114336 | Chow | Aug 2002 | A1 |
20020129367 | Devara | Sep 2002 | A1 |
20020144267 | Gutta et al. | Oct 2002 | A1 |
20020156842 | Signes et al. | Oct 2002 | A1 |
20020168082 | Razdan | Nov 2002 | A1 |
20020172395 | Foote et al. | Nov 2002 | A1 |
20020194151 | Fenton et al. | Dec 2002 | A1 |
20020199203 | Duffy et al. | Dec 2002 | A1 |
20030009769 | Hensgen et al. | Jan 2003 | A1 |
20030012403 | Rhoads et al. | Jan 2003 | A1 |
20030050055 | Ting et al. | Mar 2003 | A1 |
20030058836 | Even | Mar 2003 | A1 |
20030081580 | Vaidyanathan et al. | May 2003 | A1 |
20030093665 | Cooper et al. | May 2003 | A1 |
20030138127 | Miller et al. | Jul 2003 | A1 |
20030145325 | Finster et al. | Jul 2003 | A1 |
20030152096 | Chapman | Aug 2003 | A1 |
20030161268 | Larsson et al. | Aug 2003 | A1 |
20040008864 | Watson et al. | Jan 2004 | A1 |
20040010692 | Watson | Jan 2004 | A1 |
20040010694 | Collens et al. | Jan 2004 | A1 |
20040030798 | Andersson et al. | Feb 2004 | A1 |
20040042421 | Mahany | Mar 2004 | A1 |
20040049787 | Maissel et al. | Mar 2004 | A1 |
20040070593 | Neely et al. | Apr 2004 | A1 |
20040083487 | Collens et al. | Apr 2004 | A1 |
20040086122 | Watson | May 2004 | A1 |
20040088549 | Ukai et al. | May 2004 | A1 |
20040088557 | Malcolm et al. | May 2004 | A1 |
20040117824 | Karaoguz et al. | Jun 2004 | A1 |
20040131184 | Wu et al. | Jul 2004 | A1 |
20040139047 | Rechsteiner et al. | Jul 2004 | A1 |
20040156528 | Joo et al. | Aug 2004 | A1 |
20040187005 | Molaro | Sep 2004 | A1 |
20040234099 | Venkatesan et al. | Nov 2004 | A1 |
20040248615 | Purkayastha et al. | Dec 2004 | A1 |
20040252651 | Karlsen et al. | Dec 2004 | A1 |
20040263941 | Chen et al. | Dec 2004 | A1 |
20040264372 | Huang | Dec 2004 | A1 |
20050008017 | Datta et al. | Jan 2005 | A1 |
20050034001 | Pontarelli | Feb 2005 | A1 |
20050050103 | Kesteloot et al. | Mar 2005 | A1 |
20050065891 | Lee et al. | Mar 2005 | A1 |
20050081042 | Venkatesan et al. | Apr 2005 | A1 |
20050086069 | Watson et al. | Apr 2005 | A1 |
20050097331 | Majidimehr et al. | May 2005 | A1 |
20050108769 | Arnold et al. | May 2005 | A1 |
20050120127 | Bradley et al. | Jun 2005 | A1 |
20050125405 | Watson et al. | Jun 2005 | A1 |
20050169632 | Song et al. | Aug 2005 | A1 |
20050182989 | Zarnke et al. | Aug 2005 | A1 |
20050183120 | Jain et al. | Aug 2005 | A1 |
20050192987 | Marsh | Sep 2005 | A1 |
20050201340 | Wang et al. | Sep 2005 | A1 |
20050201726 | Malcolm et al. | Sep 2005 | A1 |
20050216942 | Barton | Sep 2005 | A1 |
20050220321 | Langelaar | Oct 2005 | A1 |
20050239497 | Bahl et al. | Oct 2005 | A1 |
20050251491 | Medina et al. | Nov 2005 | A1 |
20050286438 | Rajkotia | Dec 2005 | A1 |
20060010274 | Olson | Jan 2006 | A1 |
20060015735 | Kudo et al. | Jan 2006 | A1 |
20060047967 | Akhan et al. | Mar 2006 | A1 |
20060048185 | Alterman | Mar 2006 | A1 |
20060048186 | Alterman | Mar 2006 | A1 |
20060050880 | Taylor et al. | Mar 2006 | A1 |
20060053452 | Lee et al. | Mar 2006 | A1 |
20060056349 | Nakatugawa et al. | Mar 2006 | A1 |
20060059560 | Montulli | Mar 2006 | A1 |
20060072786 | Watson et al. | Apr 2006 | A1 |
20060075243 | Lakamp et al. | Apr 2006 | A1 |
20060080707 | Laksono | Apr 2006 | A1 |
20060085830 | Bruck et al. | Apr 2006 | A1 |
20060095401 | Krikorian et al. | May 2006 | A1 |
20060095471 | Krikorian et al. | May 2006 | A1 |
20060095472 | Krikorian et al. | May 2006 | A1 |
20060095792 | Hurtado et al. | May 2006 | A1 |
20060123235 | Vanstone | Jun 2006 | A1 |
20060127037 | Van Hoff et al. | Jun 2006 | A1 |
20060129672 | Mayer | Jun 2006 | A1 |
20060133644 | Wells et al. | Jun 2006 | A1 |
20060156003 | Zhang et al. | Jul 2006 | A1 |
20060161776 | Van Der Veen et al. | Jul 2006 | A1 |
20060165169 | Ng et al. | Jul 2006 | A1 |
20060173794 | Sellars et al. | Aug 2006 | A1 |
20060174128 | Yuval | Aug 2006 | A1 |
20060200416 | White et al. | Sep 2006 | A1 |
20060206933 | Molen et al. | Sep 2006 | A1 |
20070002742 | Krishnaswamy et al. | Jan 2007 | A1 |
20070061488 | Alagappan et al. | Mar 2007 | A1 |
20070061580 | Venkatesan et al. | Mar 2007 | A1 |
20070133673 | Imaizumi | Jun 2007 | A1 |
20080044087 | Levy et al. | Feb 2008 | A1 |
20090037388 | Cooper et al. | Feb 2009 | A1 |
20090067328 | Morris et al. | Mar 2009 | A1 |
20090133116 | Waisbard et al. | May 2009 | A1 |
20090137009 | Lai et al. | May 2009 | A1 |
20090287837 | Felsher | Nov 2009 | A1 |
Number | Date | Country |
---|---|---|
2306869 | May 1997 | GB |
0040021 | Jul 2000 | WO |
0069163 | Nov 2000 | WO |
2006046099 | May 2006 | WO |
Entry |
---|
Degermark, Mikael, et al., “Low-loss TCP/IP Header Compression for Wireless Networks,” Mobicom '96, Proceedings of the Second Annual International Conference on Mobile Computing and Networking, Rye, New York, USA, Nov. 10-12, 1996, ACM 1996, 14 pages. |
Dittman, J., et al., “Robust MPEG Video Watermarking Technologies,” Proceedings of the Sixth ACM international Conference on Multimedia, 1998, http://www.ipsi.fraunhofer.de/mobile/publications/fullpapers/ACM/acm—ab.doc, 10 pages. |
Fridrich, J., “Applications of Data Hiding in Digital Images,” Tutorial for the ISPACS '98 Conference in Melbourne, Australia, Nov. 4-6, 1998, 33 pages. |
Fridrich, J., “Methods for data hiding,” Center for Intelligent Systems & Department of Systems Science and Industrial Engineering, SUNY Binghamton, 1997, 33 pages, http://74.125.47.132/search?q=cache:unDoOhDSrQYJ: www.ws.binghamton.edu/fridrich/Research/meth—ps.ps+Fridrich+Methods+for+Data+Hiding&hl=en&ct=clnk&cd=5&gl=us. |
Judge, P., et al., “Whim: Watermarking Multicast Video with a Hierarchy of Intermediaries,” Computer Networks: The International Journal of Computer and Telecommunications Networking Archive, 2002, vol. 39, Issue 6, pp. 699-712. |
Koch, E., et al., “Towards Robust and Hidden Image Copyright Labeling,” Proceedings of 1995 IEEE Workshop on Nonlinear Signal and Image Processing, Jun. 20-22, 1995, 4 pages. |
No Author, “Double 108Mbps Wireless PC Card (WG511U), Netgear, Inc.,” PC Connection, 2006, 2 pages, http://http://www.pcconnection.com/ProductDetail?sku=5373172&SourcelD=k40132. |
No Author, “Gefen HDMI CAT-5 Extreme User Manual”, 2006, 16 pages. |
No Author, “Global-Link Product Manual,” Logical Solutions Inc., Jan. 2005, 68 pages. |
No Author, “Global-Link: KVM over IP”, http://www.thinklogical.com/products/globalLink.php, accessed Jul. 13, 2007, 1 page. |
No Author, “HDTV Extender”, Network Technologies Incorporated, http://www.nti1.com/hdtv-extend.html, 2005, 2 pages. |
No Author, “Introducting the World's First ‘Legal’ HDMI Distribution Amp with HDCP on Every Port!”, PureLink Digital Extender, Dtrovision LLC, http://www.dtrovision.com, accessed Jul. 13, 2007, 2 pages. |
No Author, “Kaleidescape Home Page”, Kaleidescape, Inc., http://www.kaleidescape.com, accessed Jul. 13, 2007, 1 page. |
No Author, “MediaMax Overview”, http://www.axonix.com/mediamax/, accessed Jul. 13, 2007, 2 pages. |
No Author, “Xtendex Series ST-C5DVI-150 150 Foot DVI Video Extender Installation and Operation Manual,” Network Technologies Incorporated, revised Oct. 4, 2006, 11 pages. |
Raisinghani, V., “ECLAIR: An Efficient Cross Layer Architecture for Wireless Protocol Stacks,” 5th World Wireless Congress, San Francisco, CA, May 25-28, 2004, 6 pages. |
Setton, E., et al., “Cross-Layer Design of Ad Hoc Networks for Real-Time Video Streaming,” IEEE Wireless Communications, Aug. 2005, vol. 12, Issue 4, pp. 59-65. |
Srivastava, V., et al., “Cross-Layer Design: A Survey and the Road Ahead,” IEEE Communications Magazine, Dec. 2005, pp. 112-119. |
Notice of Allowance for U.S. Appl. No. 11/555,707, mailed Jan. 4, 2010, 15 pages. |
Non-Final Office Action for U.S. Appl. No. 12/772,374, mailed Sep. 27, 2010, 6 pages. |
Notice of Allowance for U.S. Appl. No. 12/772,374, mailed Mar. 10, 2011, 8 pages. |
Notice of Allowance for U.S. Appl. No. 13/162,673, mailed Feb. 13, 2012, 10 pages. |
Notice of Allowance for U.S. Appl. No. 13/162,673, mailed Jul. 13, 2012, 8 pages. |
Non-Final Office Action for U.S. Appl. No. 13/668,934, mailed Mar. 21, 2013, 6 pages. |
Notice of Allowance for U.S. Appl. No. 13/668,934, mailed Aug. 6, 2013, 9 pages. |
Number | Date | Country | |
---|---|---|---|
20140105451 A1 | Apr 2014 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13668934 | Nov 2012 | US |
Child | 14133755 | US | |
Parent | 13162673 | Jun 2011 | US |
Child | 13668934 | US | |
Parent | 12772374 | May 2010 | US |
Child | 13162673 | US | |
Parent | 11555707 | Nov 2006 | US |
Child | 12772374 | US |