This application is the U.S. National Phase under 35 U.S.C. §371 of International Application No. PCT/CN2008/071028, filed on May 21, 2008, the disclosure of which Application is incorporated by reference herein.
The present invention relates to a method for facilitating the search of video content.
The so called term “video clip” appearing in this specification means a finite duration video content along with associated audio tracks, whether in digital or analog formats. Since video content consists of time-consecutive frames of video images. A video clip consists of a finite number of time-consecutive number of video images, along with the associated audio tracks of the same duration.
The so called term “fingerprint” appearing in this specification means a series of dot information, in which each dot information is selected from a frame of pattern of television signals, and a plurality of frames can be selected from the television signals, and one or more dot data can be selected from one frame of pattern of television signals, so that the so called “fingerprint” can be used to uniquely identify said television signals.
The so called term “visually identical” appearing in this specification means that two video content segment are visually identical if they are obtained from a single video image capture or recording device at the same time. In other words, they originate from a single video source and at the same time, i.e., a single time-space video source. For example, two copies of a single video tape are visually identical because they are from the same source. Two versions of compressed video data streams are visually identical if they are encoded and/or re-encoded from the same video content source, despite the fact that they may have different compression formats, bit rates or resolutions. On the other hand, two different video recordings of the same scene, but shot from two different cameras, or two different video recordings of the same scene but shot at different times from the same camera, are NOT visually identical because they are not created from a single time-space video source. When there are two different video recordings of different durations, a section of each recording may still be visually identical.
The fingerprint is used to seek out visually identical video segments between two different video content pieces. The content pieces may be in analog recording format, or in digital compressed format, or in digital uncompressed format. When two video clips are to be compared to determine if they are visually identical, in stead of comparing the two video contents, an automatic procedure can be deployed to compare the fingerprints obtained from each of the video clips. If the fingerprints match each other, then it is to say that the video clips are visually identical to each other.
A typical application is to use the technique to perform fingerprint based identification of video content. Specifically, known video clips are first registered into a fingerprint database, and when new video clips are obtained, their fingerprints are compared with the fingerprints already in the database to determine if the new content is visually identical to a previously registered video clip.
In this document, the terms “video”, “video content”, and “video signals” generically represent the same concept, i.e., visual information that can be displayed on television or computer monitors. The terms “video frames”, “digital video images”, and “video image frames” generically represent digitized video images, i.e., time-consecutive images that together form the motion video content. The video images, as part of the same video content, have the same number of video samples formed in rows and columns. The number of samples in a row is the width or horizontal resolution of the image, and the number of samples in a column is the height or the vertical resolution of the image.
In addition, the term “fingerprint” or “fingerprint data” represent the data formed by sampling consecutive video frames. The fingerprint or fingerprint data can be used to determine if two video contents are visually identical or not. Continuous samples of video frames form fingerprint data streams, or fingerprint streams. To better organize the fingerprint stream, sometimes, it is necessary to partition a continuous fingerprint stream into multiple segments. These segments are so called “fingerprint data segments” or just “fingerprint segments”.
With digital compression technology, it is possible to carry many television channels in the spectrum space of a single analog channel. Because of this, more television channels are transmitted over the same spectrum, and there are more channels for viewers to watch. Digitally compressed video and audio signals are binary data streams that can be transmitted, stored and played out as computer data files or streams. Therefore, digital video/audio data are typically handled in digital forms during production, transmission and storage phases.
Most of the video content in television broadcast, cable broadcast and on the internet originates in digital formats. Most of the content is already produced and stored in digital storage devices before it is distributed to consumers. For the purpose of managing advertising activities and marketing campaigns, government regulation enforcement, market research, and broadcasting signal monitoring, there is a need to continuously monitor the video signals as they are distributed to viewers.
In prior art, the operator may be given a video clip and be asked to search through the archived recordings to see where and when the video clip has shown up in video distributions in the past. In other words, the operator may be asked to search through the archived recordings to seek video content that is visually identical to the given video clip. For example, advertisers may want to determine if a particular commercial video has been distributed properly over the last year in certain geographic areas, so that they can track the effectiveness of their advertising campaign.
There are several problems with the above. The first problem is the fact that video content typically consumes massive amount of storage capacity. For example, a single channel of television content broadcast for 24 hours would consume at least 10 GB. For the purpose of monitoring thousands of television channels for a year or more, the storage capacity requirement can easily get into the Petabyte (1 PB=1,000,000,000,000,000 bytes) ranges.
The other problem is that even if the content is saved in storage systems, the cost to maintain, search and archive the content data can be too expensive for most users. This cost is reflected in terms of storage, computation and network hardware systems needed to perform the tasks. In many scenarios, the information to be archived and searched is not on the specific content itself, but more specifically related to when and where the content is distributed. This information can be valuable for content owners, marketers and relevant government regulators to track the coverage of specific video content. The purpose of this invention is to provide a method to facilitate this capability.
Therefore, there is a need to provide a method for facilitating the archiving and search of video content without a huge storage capacity required, and to be able to search the information easily at low hardware cost. There is also a need to collect statistics and extraction additional information from the archived video information automatically.
It is object of the present invention to provide a method for facilitating the search of video content, which can collect statistics and extraction additional information from the archived video information automatically.
It is another object of the present invention to provide a method for facilitating the search of video content, which can extract fingerprint information from video content for the purpose of archiving without the huge storage capacity required.
It is another object of the present invention to provide a method for facilitating the search of video content, which can search through video fingerprint data for identifying information and statistics about historical recording of video content easily at low hardware cost.
It is another object of the present invention to provide a method for facilitating the search of video content, which can extract fingerprint information from video content across geographically different remote locations and over long period of time for the purpose of archiving without the huge storage capacity required.
Therefore, according to the present invention, there is provided a method for facilitating the search of video content, comprising the steps of submitting a video content clip for specific time duration of video content; extracting a fingerprint associated with the video clip by a fingerprint extractor; selecting database entries matching the specific time of video content from a fingerprint archive database; delivering the database entries matching the specific time of video content to a search module; reconstructing the searched database entries into continuous fingerprint streams by the search module, in which the searched database entries from the same location and channel are grouped together, and the fingerprint data sections are stripped out of the entries and concatenated according to the segment ID and time-stamps contained within each entry; searching for a match with the fingerprint associated with the video clip by a matcher from the continuous fingerprint streams to find a match; and combining a matching result with additional information obtained from the archive entries to form a search report.
Preferably, the additional information obtained from the archive entries is time, location, channel, and/or content types.
Preferably, the video content clip is from an analog video tape player, an analog video output of a digital set-top or personal video recorder (PVR) receiver, an analog output of a DVD player, or an analog video output of a video tuner receiver module.
Preferably, a converter is used to take the video clip as input, to perform analog to digital conversion and to pass the resulting digitized video frame data to the fingerprint extractor, which obtains the fingerprint samples for the video frames transferred from the converter.
Preferably, the video content clip is in digital compressed video data format, and it is made to first go through the receiver converter to perform decompression and deliver the digital decompressed video signals to the fingerprint extractor.
Preferably, the input to the extractor is the digitized video frame data, and it is stored into a frame buffer, and a sub-sampler obtains selected samples from each video frame data.
Preferably, said fingerprint entries are searched from said archive database according to location, time, channel, and/or programming content.
Preferably, said fingerprint stream is advanced relative to said searching fingerprint by one frame and the matching operation is repeated, if a match is not found.
Preferably, if a match is found, the additional information associated with the entries is combined with the information on the searching video clip to product a single report message.
Preferably, multiple searching operations, each corresponding a different searching video clip, can be performed concurrently with each other.
By using fingerprint in stead of the video content itself, it is possible to archive huge amount of video content without the associated storage capacity required. For example, it is possible to maintain a fingerprint archive for 1000 television channels, on a 24 hours basis, for 10 years, with a fingerprint archive database of no more than 40 terabytes (1 TB=1,000,000,000,000 bytes). However, to store video of the same duration at 1 million bits per second, the storage required would be 40 petabytes (1 PB=1000 TB), which is 1000 times higher than maintaining a fingerprint archive.
In addition, the fingerprint archive can be automatically compared for searching applications. For example, a user may have a video clip of 15 seconds, and wants to know if this same video clip has ever appeared in the past 10 years in any of the 1000 television channels. By using the technique from this invention, it would be possible to obtain information on: time of its appearance; location of the distribution (where the video content was originally recorded); and channel number, i.e., which television channel actually carried this signal, so that it is possible to know whether the video has appeared anywhere in the recorded archive.
By using the invention described in this document, it's possible to perform a search like the above across a huge archive which would be impossible if the video content itself, instead of the fingerprint data, is used.
In the present invention, discussions are focused on the handling and processing of digital video data streams. The method can be extended to audio data streams by using variations of the techniques and will not be discussed here.
Video monitoring requires that video content distribution activity be identified by time, location, channel assignment or network origination address. In addition, the monitoring must be identified by its source content. The preferred way to monitor video content is to have a recording of the video content as it's distributed. Such recording typically is in digital formats, stored as data files in a computer system. The recording also has the additional information attached to the recording itself, such as time, location, etc. At a later time, when an operator decides to verify the video content distribution, he or she can simply retrieve that video recording and view it in person.
In this invention, we provide a method for facilitating the archiving and search of video content.
As shown in
Alternatively, the extracted fingerprint data can be stored in a local storage 25 that is not removable but can be transferred via the network interface 23 and the network connection to the data center 4 at non-regular time intervals or at pre-scheduled time. This is shown in
The data organization within the transfer buffer 205 is as shown in
The data packets 300 are transferred to the data center 4 and become part of the fingerprint archive database. The database will be organized by the data packets 300 received, as shown in
The fingerprint archive database can hold such packets 300 from potentially large number of collectors 2 over a long period of time durations, and across many television channels or video source. For example, it may contain data for all of the television channels distributed in an entire country over the last ten years.
The database can be searched according to some specific rules. For example, it is possible to search the archive and extract the entries for a specific location or for specific time duration.
Before further describing how to make use of the fingerprint archive database, we first discuss the basic processing components with the data center 4. This is shown in
The data center 4 typically operates as follows. A user submits a video content clip 14, which is for specific time duration of video content. The video clip 14 preferably is in digital compressed video data format. The converter 21 takes the video clip 14 as input and performs decompression and passes the resulting digitized video frame data to the fingerprint extractor 22, which obtains the fingerprint samples for the video frames transferred from the converter 21. The output of the fingerprint extractor 22 contains the fingerprint samples associated with the video clip 14.
The fingerprint extractor 22 preferably operates as shown in
Illustration below will focus on the internal operations of the fingerprint extractor 22 in some greater detail, see
In
In the preferred embodiment, each video frames 100 are sampled exactly the same way. In other words, image samples from the same positions are sampled for different images, and the same number of samples is taken from different images. In addition, the images are sampled consecutively.
The samples are then organized as part of the continuous streams of image samples and be placed into the transfer buffer 205. The image samples from different frames are organized together into the transfer buffer 205 before it's sent out. Sampling on images may be performed non-consecutively. In other words, the number of samples taken from each image may be different.
Separately, as shown in
Next it is to show how the selected entries from the archive database 400 can be prepared for the matching operation with the video clip 14. This is illustrated in
Fingerprint archive database 400 holds entries for fingerprint streams from many locations, channels and over potentially very long period of time durations. Fingerprint entries can be retrieved from the database according to location, channel, and time. Fingerprint entries selected meet a specific attributes. Further search among this data results in video information meeting the same attributes.
The fingerprint data segments are then copied out of the entries and assembled into a continuous fingerprint data stream. This stream is the restored output of the sub-sampler within fingerprint extractor 22 shown in
The matcher or correlator 600 in
The matcher 600 takes in two fingerprint data sets. The first is the finite duration fingerprint obtained from the input video clip 14. The second is the fingerprint stream reconstructed from the searched fingerprint archive database 400. The matching result between the two, combined with the additional information obtained from the archive entries, such as time, location, channel and content types, are then put together as the search report. The output of the formatter 204 therefore contains information on when and where the original video clip 14 appeared in the video signals captured by the remote collectors 2.
More specifically, as shown in
Once the fingerprint reconstruction is complete, it is aligned and compared against the fingerprint data obtained from the searching video clip (steps 803, 804). If the result is a match, it means that the two fingerprints represent two visually identical pieces of video content. In this case, the additional information obtained form the data entries, such as location, channel, time and any optional information, will be combined with the information on the searching video clip to product a single report message (step 805). If the two fingerprints do not match, then the fingerprint array obtained from (step 802) is advanced by one frame relative to the searching fingerprint (step 806), and the corresponding information obtained from the newly included fingerprint data points will be updated as well (step 807). The process is then repeated at (step 803).
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CN2008/071028 | 5/21/2008 | WO | 00 | 5/30/2008 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2009/140817 | 11/26/2009 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
3919479 | Moon et al. | Nov 1975 | A |
5019899 | Boles et al. | May 1991 | A |
5870754 | Dimitrova et al. | Feb 1999 | A |
5926223 | Hardiman | Jul 1999 | A |
6037986 | Zhang et al. | Mar 2000 | A |
6084539 | Yamada | Jul 2000 | A |
6374260 | Hoffert et al. | Apr 2002 | B1 |
6473529 | Lin | Oct 2002 | B1 |
6834308 | Ikezoye et al. | Dec 2004 | B1 |
7336841 | Neogi | Feb 2008 | B2 |
7523312 | Kalker et al. | Apr 2009 | B2 |
7809154 | Lienhart et al. | Oct 2010 | B2 |
20030126276 | Kime et al. | Jul 2003 | A1 |
20040021669 | Fredlund et al. | Feb 2004 | A1 |
20040240562 | Bargeron et al. | Dec 2004 | A1 |
20050141707 | Haitsma et al. | Jun 2005 | A1 |
20050213826 | Neogi | Sep 2005 | A1 |
20060129822 | Snijder et al. | Jun 2006 | A1 |
20060184961 | Lee et al. | Aug 2006 | A1 |
20070055987 | Lu et al. | Mar 2007 | A1 |
20070071330 | Oostveen et al. | Mar 2007 | A1 |
20070124796 | Wittkotter | May 2007 | A1 |
20070136782 | Ramaswamy et al. | Jun 2007 | A1 |
20070162571 | Gupta et al. | Jul 2007 | A1 |
20070186228 | Ramaswamy et al. | Aug 2007 | A1 |
20070186229 | Conklin et al. | Aug 2007 | A1 |
20080148309 | Wilcox et al. | Jun 2008 | A1 |
20080310731 | Stojancic et al. | Dec 2008 | A1 |
20090063277 | Bernosky et al. | Mar 2009 | A1 |
20090074235 | Lahr et al. | Mar 2009 | A1 |
20090154806 | Chang et al. | Jun 2009 | A1 |
20090213270 | Ismert et al. | Aug 2009 | A1 |
20090324199 | Haitsma et al. | Dec 2009 | A1 |
20100077424 | Ramaswamy et al. | Mar 2010 | A1 |
20100122279 | Zhang | May 2010 | A1 |
20100158488 | Roberts et al. | Jun 2010 | A1 |
20100169911 | Zhang | Jul 2010 | A1 |
20100306791 | Deng | Dec 2010 | A1 |
Number | Date | Country |
---|---|---|
1190218 | Aug 1998 | CN |
2387588 | Jul 2000 | CN |
1262003 | Aug 2000 | CN |
1341240 | Mar 2002 | CN |
1574953 | Feb 2005 | CN |
1628302 | Jun 2005 | CN |
1679051 | Oct 2005 | CN |
1679261 | Oct 2005 | CN |
1719909 | Jan 2006 | CN |
1723458 | Jan 2006 | CN |
1739121 | Feb 2006 | CN |
2914526 | Jun 2007 | CN |
101002472 | Jul 2007 | CN |
101021852 | Aug 2007 | CN |
101047833 | Oct 2007 | CN |
101120594 | Feb 2008 | CN |
0838960 | Apr 1998 | EP |
1482734 | Dec 2004 | EP |
1760693 | Mar 2007 | EP |
2419489 | Apr 2006 | GB |
9274467 | Oct 1997 | JP |
20020001088 | Jan 2002 | KR |
WO 0209447 | Jan 2002 | WO |
WO 02065782 | Aug 2002 | WO |
WO 2006059053 | Jun 2006 | WO |
WO 2007080133 | Jul 2007 | WO |
WO 2007148264 | Dec 2007 | WO |
Entry |
---|
Non-final Office Action dated Oct. 26, 2010, for U.S. Appl. No. 12/085,928. |
Non-final Office Action dated Oct. 26, 2010, for U.S. Appl. No. 12/085,754. |
Final Office Action dated Apr. 12, 2011, for U.S. Appl. No. 12/085,754. |
Non-final Office Action dated Nov. 29, 2010, for U.S. Appl. No. 12/085,830. |
Final Office Action dated Aug. 19, 2011, for U.S. Appl. No. 12/085,830. |
Non-final Office Action dated Nov. 29, 2010, for U.S. Appl. No. 12/085,927. |
Final Office Action dated May 26, 2011, for U.S. Appl. No. 12/085,927. |
Non-final Office Action dated Oct. 13, 2010, for U.S. Appl. No. 12/085,765. |
Final Office Action dated Mar. 31, 2011, for U.S. Appl. No. 12/085,765. |
Non-final Office Action dated Oct. 29, 2010, for U.S. Appl. No. 12/085,835. |
Final Office Action dated Apr. 13, 2011, for U.S. Appl. No. 12/085,835. |
Non-final Office Action dated Dec. 1, 2010, for U.S. Appl. No. 12/085,827. |
Final Office Action dated Apr. 19, 2012, for U.S. Appl. No. 12/085,827. |
Non-final Office Action dated Oct. 13, 2010, for U.S. Appl. No. 12/085,764. |
Final Office Action dated Mar. 24, 2011, for U.S. Appl. No. 12/085,764. |
Cheung et al., “Efficient Video Similarity Measure With Video Signature”, Jan. 2003. |
Oostveen et al., “Feature Extraction and a Database Strategy for Video Fingerprinting”, 2002. |
Non-final Office Action dated Nov. 10, 2010, for U.S. Appl. No. 12/085,834. |
Final Office Action dated Mar. 25, 2011, for U.S. Appl. No. 12/085,834. |
Non-final Office Action dated Dec. 14, 2010, for U.S. Appl. No. 12/085,823. |
Notice of Allowance and Fees Due dated Aug. 19, 2011, for U.S. Appl. No. 12/085,823. |
Non-final Office Action dated Jul. 26, 2012, for U.S. Appl. No. 12/085,752. |
International Preliminary Report on Patentability and Written Opinion dated Dec. 25, 2008, for PCT Application No. PCT/CN2008/071039. |
International Search Report dated Dec. 25, 2008, for PCT/CN2008/071039. |
International Preliminary Report on Patentability and Written Opinion dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071083. |
International Search Report dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071083. |
International Preliminary Report on Patentability and Written Opinion dated Mar. 12, 2009, for PCT Application No. PCT/CN2008/071082. |
International Search Report dated Mar. 12, 2009, for PCT Application No. PCT/CN2008/071082. |
International Preliminary Report on Patentability and Written Opinion dated Sep. 4, 2008, for PCT Application No. PCT/CN2008/071046. |
International Search Report dated Sep. 4, 2008, for PCT Application No. PCT/CN2008/071046. |
International Preliminary Report on Patentability and Written Opinion dated Sep. 4, 2008, for PCT Application No. PCT/CN2008/071047. |
International Search Report dated Sep. 4, 2008, for PCT Application No. PCT/CN2008/071047. |
International Preliminary Report on Patentability and Written Opinion dated Mar. 5, 2009, for PCT Application No. PCT/CN2008/071023. |
International Search Report dated Mar. 5, 2009, for PCT Application No. PCT/CN2008/071023. |
International Preliminary Report on Patentability and Written Opinion dated Mar. 5, 2009, for PCT Application No. PCT/CN2008/071028. |
International Search Report dated Mar. 5, 2009, for PCT Application No. PCT/CN2008/071028. |
International Preliminary Report on Patentability and Written Opinion dated Feb. 19, 2009, for PCT Application No. PCT/CN2008/071029. |
International Search Report dated Feb. 19, 2009, for PCT Application No. PCT/CN2008/071029. |
International Preliminary Report on Patentability and Written Opinion dated Feb. 12, 2009, for PCT Application No. PCT/CN2008/071030. |
International Search Report dated Feb. 12, 2009, for PCT Application No. PCT/CN2008/071030. |
International Preliminary Report on Patentability and Written Opinion dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071038. |
International Search Report dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071038. |
International Preliminary Report on Patentability and Written Opinion dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071033. |
International Search Report dated Feb. 26, 2009, for PCT Application No. PCT/CN2008/071033. |
International Preliminary Report on Patentability and Written Opinion dated Oct. 16, 2008, for PCT Application No. PCT/CN2008/071041. |
International Search Report dated Oct. 16, 2008, for PCT Application No. PCT/CN2008/071041. |
English Abstract of JP 9274467 A dated Oct. 21, 1997. |
English Abstract of CN 101002472 A dated Jul. 18, 2007. |
English Abstract of CN 1739121 A dated Feb. 22, 2006. |
English Abstract of CN 101120594 A dated Feb. 6, 2008. |
English Abstract of CN 1719909 A dated Jan. 11, 2006. |
English Abstract of WO 0209447 A1 dated Jan. 31, 2002. |
English Abstract of CN 101047833 A dated Oct. 3, 2007. |
English Abstract CN 2914526 Y dated Jun. 20, 2007. |
English Abstract CN 1262003 A dated Aug. 2, 2000. |
English Abstract CN 2387588 Y dated Jul. 12, 2000. |
English Abstract CN 1679261 A dated Oct. 5, 2005. |
English Abstract CN 1574953 A dated Feb. 2, 2005. |
English Abstract CN 1628302 A dated Jun. 15, 2005. |
English Abstract CN 1190218 A dated Aug. 12, 1998. |
English Abstract CN 1341240 A dated Mar. 20, 2002. |
English Abstract CN 1723458 A dated Jan. 18, 2006. |
English Abstract CN 101021852 A dated Aug. 22, 2007. |
English Abstract CN 1679051 A dated Oct. 5, 2005. |
English Abstract KR 20020001088 A dated Jan. 9, 2002. |
Number | Date | Country | |
---|---|---|---|
20100169358 A1 | Jul 2010 | US |