The invention relates to a method and arrangement for embedding a watermark in motion image signals such as movies projected in cinemas. The invention also relates to a method and arrangement for detecting a watermark embedded in such motion image signals.
Watermark embedding is an important aspect of copy protection strategies. Although most copy protection schemes deal with protection of electronically distributed contents (broadcasts, storage media), copy protection is also desired for movies shown in theaters. Nowadays, illegal copying of cinema material by means of a handheld video camera is already common practice. Although the quality is usually low, the economical impact of illegal VHS tapes, CD-Videos and DVDs can be enormous. For this reason, cinema owners are obliged to prevent the presence of video cameras on their premises. Not following this rule may be sanctioned with a ban on the future availability of content. In view hereof, it is envisioned to provide that a watermark will be added during show time. The watermark is to identify the cinema, the presentation time, operator, etc.
Robustness to geometric distortions is a key requirement for such watermark embedding schemes. A handheld camera will not only seriously degrade the video by filtering (the optical path from the screen to the camera, transfer to tape, etc.) but also seriously geometrically distort the video (shifting, scaling, rotation, shearing, changes in perspective, etc.). In addition, these geometrical distortions can change from frame to frame.
A prior-art method of embedding a watermark in cinema movies, which meets the robustness requirements, is disclosed in Jaap Haitsma and Ton Kalker: A Watermarking Scheme for Digital Cinema; Proceedings ICIP, Vol. 2, 2001, pp. 487-489. The robustness to geometric distortions is achieved by exploiting only the temporal axis to embed the watermark. The watermark is a periodic pseudo-random sequence of watermark samples having two distinct values, e.g. ‘1’ and ‘−1’. One watermark sample is embedded in each image. The value ‘1’ is embedded in an image by increasing a global property (e.g. the mean luminance) of the image, the value ‘−1’ is embedded by decreasing said global property.
The prior-art watermark embedding method actually embeds flicker. By embedding the same watermark sample in a number of consecutive images, the flicker is made imperceptible (the human eye is less sensitive to low-frequency flicker).
Flicker of the recorded movie is also caused by a) the typical mismatch between the cinema projector's frame rate (24 frames per second) and the camcorder's frame rate (25 fps for PAL, 29.97 fps for NTSC), and b) the difference between the two display scan formats (progressive vs. interlace). This kind of flicker is so annoying that de-flickering tools have been made widely available to the public. For example, a de-flicker plug-in for the video capturing and processing application “Virtualdub” has been found on the Internet.
A problem of the prior-art watermark embedding scheme is that de-flicker tools also remove the embedded watermark.
It is an object of the invention to further improve the prior-art watermark embedding and detection method. It is a particular object of the invention to provide a watermark embedding and detection scheme, which is robust to de-flickering operations.
To this end, the method of embedding a watermark in a motion image signal according to the invention includes dividing each image into at least a first and a second image area. One value of a watermark sample is embedded in an image by increasing the global property (e.g. the mean luminance) of its first area and decreasing the global property of its second area. The other value of the watermark sample is embedded in the opposite way, i.e. by decreasing the global property of the first image area and increasing the global property of the second image area.
The invention exploits the insight that de-flickering tools remove flicker by adjusting the mean luminance of successive images to exhibit a low-pass character. The mean luminance is adjusted in practice by multiplying all pixels of an image by the same factor. Because this operation does not affect the (sign of the) modifications applied to the distinct image areas, the watermark information will be retained. In view hereof, the global property of an image area being modified to embed the watermark is the mean luminance of said image area.
In a preferred embodiment of the method, the first and the second image area are the upper and lower half of an image. In general, there are more horizontal than vertical movements in a movie. The horizontal movements influence the mean luminance values to a smaller extent.
In the simplest embodiment of the watermark embedder, the sequence w(n) is directly applied to embedding stages 11 and 12. In such an embodiment, embedding stage 11 adds one applied watermark sample w(n) to every pixel of the first image area, whereas embedding stage 12 subtracts the same watermark sample from every pixel of the second image area. Clipping is performed where necessary. The mean luminances of the first and second image areas are thus oppositely modulated by the watermark.
Other examples of global image properties that can be modulated by the watermark are picture histograms (a list of relative frequencies of luminance values in the picture), or features derived therefrom such as high order moments (average of luminance values to a power k). The mean luminance is a specific example of the latter (k=1).
Since the Human Visual System (HVS) is sensitive to flicker in low spatial frequencies, this simple embodiment may suffer from artifacts in especially non-moving flat areas. These artifacts are significantly reduced by lowering the flicker frequency of the watermark. This is performed by a repetition stage 13, which repeats each watermark sample during a predetermined number K of consecutive images. The same watermark sample is thus embedded in K consecutive frames. The watermark repeats itself every N=1024 frames. The watermark sample w(n) being embedded in frame k can be mathematically denoted by w(└k/K┘mod N). For simplicity, this expression will hereinafter be abbreviated to w(k).
The preferred embodiment of the embedder which is shown in
Fw,1(n,k)=F1(n,k)+CF,1(n,k)w(k)
Fw,2(n,k)=F2(n,k)−CF,2(n,k)w(k)
It will be appreciated that both embedding operations can be carried out in a time-sequential manner by a single processing circuit under appropriate software control.
The two image areas are subsequently combined by a combining stage 18 into a single watermarked image Fw(n,k).
In practice, the mean luminance values of a movie exhibit a low-pass nature as a function of the frame number k (i.e. as a function of time). The detector estimates the mean luminance values of the original (unwatermarked) movie by low-pass filtering (25,27) the respective mean luminance values fw,1(k) and fw,2(k). Estimations of the mean luminance modifications as introduced by the embedder are subsequently obtained by subtracting (26,27) the low-pass filtered mean values from the unfiltered mean luminance values. The detector estimates the embedded watermark sample by subtracting (23) both estimations, followed by a sign operation (29). The estimated watermark sample being embedded in frame k is denoted v(k).
The arrangement thus generates a sequence of estimated watermark samples. In a correlation stage 3, the estimated sequence of watermark samples is correlated with the watermark being looked for. The detector receives this watermark being looked for in the form of a pseudo-random sequence w(n) of length N, where w(n)ε[−1, 1]. The detector comprises a repetition stage 24, which is identical to same repetition stage 13 of the embedder. The repetition stage repeats each watermark sample during K consecutive images. The watermark repeats itself every N=1024 frames. The watermark samples being applied to the correlation stage 3 for frame are denoted w(k). Again, w(k) is an abbreviation for the mathematically more correct expression w(└k/K┘modN).
It should be noted that the low-pass filter/subtracter combinations 25,26 and 27,28, as well as the sign operation 29 are optional.
The watermark is detected by determining the similarity of the contents of each buffer with the reference watermark w(n) being looked for. Each watermark can identify, for instance, one theatre. A well-known example of similarity is cross-correlation, but other measures are possible. The contents of each buffer are cross-correlated with the reference watermark in respective correlators 33,34, . . . The correlation is preferably performed using Symmetrical Phase Only Matched Filtering (SPOMF). For a description of SPOMF, reference is made to International Patent application WO 99/45706. In said document, the correlation is performed in the two-dimensional spatial domain. Blocks of N×N image pixels are correlated with an N×N reference watermark. The result of the SPOMF operation is an N×N pattern of correlation values exhibiting one or more peaks if a watermark has been embedded.
The K correlators 33,34, . . . operate in the one-dimensional time domain. The output of each correlator is a series of N correlation values which is stored in a corresponding one of K buffers 35,36, . . . A peak detector 37 searches the highest correlation value in the K buffers, and applies said peak value to a threshold circuit 38. If the peak value of at least one of the buffers is larger than a given threshold value, it is decided that the watermark is present. Otherwise, the content will be classified as not watermarked.
A payload can be encoded in the signal by embedding shifted versions of the watermark w(n) in a manner similar to the one disclosed in International Patent Application WO-A-99/45705. It should further be noted that, although parallel correlators are shown in
In order to describe the invention in even more details, a mathematical analysis of the prior-art watermarking scheme, an analysis of the publicly available de-flickering tool, and the operation of the watermarking scheme in accordance with the invention will now be given.
The watermark w is a periodic pseudo-random sequence containing only ‘1’ and ‘−1’ sample values with period M. A watermark sample w(n) is embedded in K consecutive frames k, k+1, . . . , k+K−1. By embedding one watermark sample in K consecutive frames, the frequency of the flickering due to the embedding is decreased. A ‘1’ is embedded in an image by increasing the luminance value of each pixel with a value CF(n,k). A ‘−1’ is embedded by decreasing the luminance value of each pixel with CF(n,k). Herein, n is the spatial coordinate of a pixel within frame k. More mathematically, we have:
where F is the frame to be embedded and Fw is the embedded frame. The change CF is chosen to be such that the watermark is not visible, and therefore depends on F. In [2], a texture detector and a motion detector are used to determine CF. As a result, the mean luminance values of the watermarked video fw
with N being the number of pixels per frame, will exhibit a change with respect to the original mean luminance values
fw(k)=f(k)+cF(k)w(└k/T┘) (1)
Herein, cF is the local depth of the watermark w, which is directly related to CF:
Due to the watermark embedding, the mean luminance values will decrease or increase with respect to the original mean luminance values in time. See Eq. (1). In practice, the mean luminance values of a movie exhibit a low-pass nature. Therefore, the detector estimates these luminance values of the original unwatermarked movie by low-pass filtering the mean luminance values of the watermarked movie fw. The detector estimates the watermark v by subtracting these low-pass filtered means from the mean luminance values of the watermarked movie fw, followed by a sign operation. More mathematically,
v(k)=sign{fw(k)−(fw {circle around (x)}g)(k)} (2)
where {circle around (x)} denotes a (one-dimensional) convolution, and g is a low-pass filter. Since a watermark sample is embedded in K consecutive frames, this operation yields K estimates {tilde over (w)}1 of the watermark w:
{tilde over (w)}1(k)=v(1+kK),0≦1<K (3)
Each of these K estimated watermarks {tilde over (w)}l is correlated with the original watermark w. If the absolute correlation value dl =|<w,{tilde over (w)}l >| is larger than a threshold value, it is decided that the video sequence is watermarked.
A movie is projected progressively at a frame rate of 24 frames per second (fps), however, a standard video camera records at 25 fps (PAL) or 29.97 fps (NTSC) interlaced. Due to this interlacing, the luminance will not be the same throughout recording of one frame, as the shutter may just be opening or just be closing. Since the video camera and the projector are not synchronized, this problem is difficult to be solved for a camera man. In addition, since the frame rates of the projector and the video camera are not matched, a similar problem reveals itself; at some points in time, a frame is recorded when the shutter is half-open or even fully shut. The result of these mismatches is a flickering in the recorded movie.
A ‘de-flicker’ plug-in for Virtualdub can be found on the Internet. This plug-in removes the flickering in four steps:
Note that β(k) is non-negative for all k, since {circumflex over (f)}w (k)≧0 and h is a low-pass filter. If we ignore the rounding and the clipping for the moment, the result of these multiplications in the last step is that the means of the new constructed video {tilde over (f)}w,deflic resembles the low-pass filtered means of {circumflex over (f)}w:
{circumflex over (f)}w,deflic =β(k){circumflex over (f)}(k)=({circumflex over (f)}w{circle around (x,)}h)(k)
Perceptually, this means that the new constructed video exhibits less flickering, because flickering can be seen as a high-frequency component in the mean luminance values of the frames, which are now filtered out.
Unfortunately, the watermarking scheme is actually a flickering, although imperceptible. As a direct consequence, this ‘de-flicker’ plug-in removes the watermark. The watermark embedding scheme must thus be modified in such a way that it is robust to de-flickering. This is all the more true as this ‘de-flicker’ tool is widely used to undo the pirate copies from the flickering.
To this end, each frame is divided into two parts (e.g. left/right or top/bottom), and the watermark sample is embedded in these parts in an opposite way. To embed a watermark sample, the mean luminance of one part is increased and the mean luminance of the other part is decreased. Consequently, the mean luminance values fw of the watermarked movie now consist of two parts
fw(k)=fw,l(k)+fw,2(k)
with (cf. Eq. (1))
fw,1(k)=f1(k)+cF,1(k)w(└k/T┘) and
fw,2(k)=f2(k)−cF,2(k)w(└k/T┘) (4)
After capturing with a camera, the ‘de-flicker’ tool removes the flickering by low-pass filtering the mean luminance values {circumflex over (f)}w:
{circumflex over (f)}w,deflic(k)=β(k)└{circumflex over (f)}w,1(k)+{circumflex over (f)}w,2(k)┘
The detection of the watermark for the modified watermarking scheme is similar to the detection method described above. First, the detector estimates the luminance values of the original unwatermarked movie for both parts by low-pass filtering the mean luminance values of both parts. Then it subtracts the result of both operations from the luminance values of the corresponding parts. Finally, it makes an estimate of the watermark {tilde over (v)} by subtraction followed by a sign operation (cf. Eq. (2)):
{tilde over (v)}(k)=sign{(fw,1,defic(k)−(fw,1,deflic {circle around (x)}g)(k)−(fw,2,defic(k)−(fw,2,deflic{circle around (x)}g)(k))} (5)
The K estimations {tilde over (w)}1 (see Eq. (3)) are obtained in a similar way and correlated with the watermark w.
The embedded watermark survives the de-flicker operation because the de-flicker tool multiplies all the pixels of an image by the same factor β(k), thereby leaving the luminance differences between the two image areas substantially intact The effect of the invention can also be explained more mathematically. It is assumed that the original unwatermarked movie exhibits a low-pass nature. After capturing of the movie with a camera, the mean luminance values of the watermarked parts {circumflex over (f)}w,1 and {circumflex over (f)}w,2 exhibit a flickering
{circumflex over (f)}w,1,(k)=γ(k)fw,1(k) and
{circumflex over (f)}w,2(k)=γ(k)fw,2(k)
Herein, γ(k)>0 corresponds to the change in the mean luminance value (the flickering) of frame k. The ‘de-flicker’ plug-in removes this flickering by low-pass filtering the mean luminance values
{circumflex over (f)}w,deflic(k)=β(k){circumflex over (f)}w(k)=β(k)γ(k)[f(k)+{cF,1(k)−cf,2(k)}w(└k/T┘)]≈f(k).
From this expression it follows that
Since in practice {cF,1(k)−cF,2(k)}w(└k/T┘) is relatively small compared to f(k), we can approximate β(k)γ(k) by 1. By using this approximation, we see that
{circumflex over (f)}w,l,deflic (k)=β(k)γ(k)fw,l(k)≈fw,l(k)=fl(k)+cF,l(k)w(└k/T┘)
For the other part, we obtain a similar result
{circumflex over (f)}w,2,deflic(k)≈f2(k)−cF,2(k)w(└k/T┘)
Using these results, we finally obtain the following expression for {tilde over (v)}(see Eq. (5))
where it is assumed that the low-pass filter g completely filters out the watermark. Note that cF,l(k)+cF,2(k) does not influence the sign of the expression, because it is non-negative for all k. It can be seen from this expression that the watermark indeed survives the ‘de-flicker’ operation after the modification.
Methods and arrangements are disclosed for embedding and detecting a watermark in a cinema movie, such that the watermark can be detected in a copy made by a handheld video camera. The watermark embedder divides each image frame into two areas. A watermark bit ‘+1’ is embedded in a flame by increasing the luminance of the first part and decreasing the luminance of the second part. A watermark bit ‘−1’ is embedded by decreasing the luminance of the first part and increasing the luminance of the second part. It is achieved with the invention that the embedded watermark survives ‘de-flicker’ operations that are often used to remove flicker caused by the different frame rates of cinema projection equipment and consumer camcorders.
Number | Date | Country | Kind |
---|---|---|---|
03100109 | Jan 2003 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB2004/050010 | 1/13/2004 | WO | 00 | 7/20/2005 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2004/066626 | 8/5/2004 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5991426 | Cox et al. | Nov 1999 | A |
6477431 | Kalker et al. | Nov 2002 | B1 |
6553127 | Kurowski | Apr 2003 | B1 |
6590996 | Reed et al. | Jul 2003 | B1 |
6807285 | Iwamura | Oct 2004 | B1 |
6970575 | Echizen et al. | Nov 2005 | B2 |
7188248 | Watson | Mar 2007 | B2 |
7191334 | Kalker | Mar 2007 | B1 |
7197164 | Levy | Mar 2007 | B2 |
7369677 | Petrovic et al. | May 2008 | B2 |
7430302 | Thorwirth | Sep 2008 | B2 |
7706569 | Iwamura | Apr 2010 | B2 |
7748051 | Morimoto et al. | Jun 2010 | B2 |
20040250079 | Kalker et al. | Dec 2004 | A1 |
20050105726 | Neubauer et al. | May 2005 | A1 |
20060129566 | Van Leest | Jun 2006 | A1 |
Number | Date | Country |
---|---|---|
0933919 | Apr 1999 | EP |
11289255 | Oct 1999 | JP |
9945705 | Sep 1999 | WO |
9945706 | Sep 1999 | WO |
WO 9945705 | Sep 1999 | WO |
0038097 | Jun 2000 | WO |
WO03001813 | Jan 2003 | WO |
Number | Date | Country | |
---|---|---|---|
20060129566 A1 | Jun 2006 | US |