The subject matter of this application is also related to that of Ser. No. 09/620,019, 60/257,822, 60/232,163, and 09/404,291.
The present invention relates to computer-based systems, and more particularly relates to systems that identify electronic or physical objects (e.g., audio, printed documents, video, etc.), and trigger corresponding responses.
In application Ser. No. 09/571,422 (now laid-open as PCT publication WO 00/70585), the present assignee described technology that can sense an object identifier from a physical or electronic object, and trigger a corresponding computer response.
In application Ser. Nos. 09/574,726 and 09/476,686, the present assignee described technology that uses a microphone to sense audio sounds, determine an identifier corresponding to the audio, and then trigger a corresponding response.
Although the cited patent applications focused on use of digital watermarks to identify the subject objects/audio, they noted that the same applications and benefits can be provided with other identification technologies.
One such suitable technology—variously known as robust hashing, fingerprinting, etc.—involves generating an identifier from attributes of the content. This identifier can then be looked-up in a database (or other data structure) to determine the song (or other audio track) to which it corresponds.
Various fingerprinting technologies are known. For example, a software program called TRM, from Relatable Software, was written up in the Washington Post as follows:
Tuneprint is another such audio fingerprinting tool. Tuneprint is understood to utilize a model of human hearing used to predict how audio will appear after it's been distorted by the human ear, and the parts of neural processing that are understood. This is some of the same information that led to MP3 encoders achieving exceptional audio compression. Characteristics that uniquely identify the track are then identified by picking out the most important, surprising, or significant features of the sound.
Yet another fingerprinting program is Songprint, available as an open source library from freetantrum.org.
One form of fingerprint may be derived by applying content—in whole or part, and represented in time—or frequency format—to a neural network, such as a Kohonen self-organizing map. For example, a song may be identified by feeding the first 30 seconds of audio, with 20 millisecond fourier transformed windows, into a Kohonen network having 64 outputs. The 64 outputs can, themselves, form the fingerprint, or they can be further processed to yield the fingerprint.
A variety of other fingerprinting tools and techniques are known to artisans in the field. Others are disclosed, e.g., in applications 60/257,822, Ser. No. 09/563,664 (now U.S. Pat. No. 6,505,160), and Ser. No. 09/578,551. See also the chapter on Fingerprinting by John Hyeon Lee, in Information Hiding: Techniques for Steganography and Digital Watermarking edited by Stefan Katzenbeisse and Fabien A. P. Petitcolas, published by Artech House.
One way to generate a fingerprint is to “hash” the audio, to derive a shorter code that is dependent, in a predetermined way, on the audio data. However, slight differences in the audio data (such as sampling rate) can cause two versions of the same song to yield two different hash codes. While this outcome is advantageous in certain outcomes, it is disadvantageous in many others.
Generally preferable are audio fingerprinting techniques that yield the same fingerprints, even if the audio data are slightly different. Thus, a song sampled at a 96K bitrate desirably should yield the same fingerprint as the same song sampled at 128K. Likewise, a song embedded with steganographic watermark data should generally yield the same fingerprint as the same song without embedded watermark data.
One way to do this is to employ a hash function that is insensitive to certain changes in the input data. Thus, two audio tracks that are acoustically similar will hash to the same code, notwithstanding the fact that individual bits are different. A variety of such hashing techniques are known.
Another approach does not rely on “hashing” of the audio data bits. Instead, the audio is decomposed into elements having greater or lesser perceptibility. Audio compression techniques employ such decomposition methods, and discard the elements that are essentially imperceptible. In fingerprinting, these elements can also be disregarded, and the “fingerprint” taken from the acoustically significant portions of the audio (e.g., the most significant coefficients after transformation of the audio into a transform domain, such as DCT).
Some fingerprinting techniques do not rely on the absolute audio data (or transformed data) per se, but rather rely on the changes in such data from sample to sample (or coefficient to coefficient) as an identifying hallmark of the audio.
Some fingerprinting algorithms consider the entire audio track (e.g., 3 minutes). Others work on much shorter windows—a few seconds, or fractions of seconds. The former technique yields a single fingerprint for the track. The latter yields plural fingerprints—one from each excerpt. (The latter fingerprints can be concatenated, or otherwise combined, to yield a master fingerprint for the entire audio track.) For compressed audio, one convenient unit from which excerpts can be formed is the frame or window used in the compression algorithm (e.g., the excerpt can be one frame, five frames, etc.).
One advantage to the excerpt-based techniques is that a song can be correctly identified even if it is truncated. Moreover, the technique is well suited for use with streaming media (in which the entire song data is typically not available all at once as a single file).
In database look-up systems employing fingerprints from short excerpts, a first fingerprint may be found to match 10 songs. To resolve this ambiguity, subsequent excerpt-fingerprints can be checked.
One way of making fingerprints “robust” against variations among similar tracks is to employ probabilistic methods using excerpt-based fingerprints. Consider the following, over-simplified, example:
In this situation, it appears most probable that the fingerprints correspond to song B, since three of the four excerpt-fingerprints support such a conclusion. (Note that one of the excerpts—that which yielded Fingerprint 2—does not match song B at all.)
More sophisticated probabilistic techniques, of course, can be used.
Once a song has been identified in a database, a number of different responses can be triggered. One is to impose a set of usage controls corresponding to terms set by the copyright holder (e.g., play control limitations, record control, fee charges, etc.) Another is to identify metadata related to the song, and provide the metadata to a user (or a link to the metadata). In some such applications, the song is simply identified by title and artist, and this information is returned to the user, e.g., by email, instant messaging, etc. With this information, the user can be given an option to purchase the music in CD or electronic form, purchase related materials (t-shirts, concert tickets), etc. A great variety of other content-triggered actions are disclosed in the cited applications.
One of the advantages of fingerprint-based content identification systems is that they do not require any alteration to the content. Thus, recordings made 50 years ago can be fingerprinted, and identified through such techniques.
Going forward, there are various advantages to encoding the content with the fingerprint. Thus, for example, a fingerprint identifier derived from a song can be stored in a file header of a file containing that song. (MP3 files, MPEG files, and most other common content file formats include header fields in which such information can readily be stored.) The fingerprint can then obtained in two different ways—by reading the header info, and by computation from the audio information. This redundancy offers several advantages. One aids security. If a file has a header-stored fingerprint that does not match a fingerprint derived from the file contents, something is amiss—the file may be destructive (e.g., a bomb or virus), or the file structure may mis-identify the file contents.
In some embodiments, the fingerprint data (or watermark data) stored in the header may be encrypted, and/or authenticated by a digital signature such as a complete hash, or a few check bits or CRC bits. In such cases, the header data can be the primary source of the fingerprint (watermark) information, with the file contents being processed to re-derive the fingerprint (watermark) only if authentication of the fingerprint stored in the header fails. Instead of including the fingerprint in the header, the header can include an electronic address or pointer data indicating another location (e.g., a URL or database record) at which the fingerprint data is stored. Again, this information may be secured using known techniques.
Similarly, the fingerprint can point to a database that contains one or more IDs that are added via a watermark. This is useful when CDs are being converted to MP3 files (i.e. ripped) and the fingerprint is calculated from a hash of the table of contents (TOC) such as done with CDDB.com, or from all of the songs. In this case, the database entry for that fingerprint could include a list of IDs for each song, and these IDs are added via a watermark and/or frame header data. This can also be useful where the content is identified based upon a group of fingerprints from plural excerpts, in which case the database that determines the content also contains an identifier, unrelated to the fingerprint(s) for that piece of content that can be embedded via a watermark.
Instead of, or in addition to, storing a fingerprint in a file header, the fingerprint data may be steganographically encoded into the file contents itself, using known watermarking techniques (e.g., those disclosed in application Ser. No. 09/503,881 (now U.S. Pat. No. 6,614,914) and U.S. Pat. Nos. 6,061,793, 6,005,501 and 5,940,135). For example, the fingerprint ID can be duplicated in the data embedded via a watermark.
In some arrangements, a watermark can convey a fingerprint, and auxiliary data as well. The file header can also convey the fingerprint, and the auxiliary data. And even if the file contents are separated from the header, and the watermark is corrupted or otherwise lost, the fingerprint can still be recovered from the content. In some cases, the lost auxiliary data can alternatively be obtained from information in a database record identified by the fingerprint (e.g., the auxiliary information can be literally stored in the record, or the record can point to another source where the information is stored).
Instead of especially processing a content file for the purpose of encoding fingerprint data, this action can be done automatically each time certain applications process the content for other purposes. For example, a rendering application (such as an MP3 player or MPEG viewer), a compression program, an operating system file management program, or other-purposed software, can calculate the fingerprint from the content, and encode the content with that information (e.g., using header data, or digital watermarking). It does this while the file is being processed for another purpose, e.g., taking advantage of the file's copying into a processing system's RAM memory, from slower storage.
In formats in which content is segregated into portions, such as MP3 frames, a fingerprint can be calculated for, and encoded in association with, each portion. Such fingerprints can later be cross-checked against fingerprint data calculated from the content information, e.g., to confirm delivery of paid-for content. Such fingerprints may be encrypted and locked to the content, as contemplated in application Ser. No. 09/620,019.
In addition, in such frame based systems, the fingerprint data and/or watermark data can be embedded with some or all data throughout each frames. This way a streaming system can use the header to first check the song for identification, and if that identification is absent or not authenticated, the system can check for the watermark and/or calculate the fingerprint. This improves the efficiency and cost of the detecting system.
Before being encrypted and digitally signed, the data in the frame header can be modified by the content, possibly a hash of the content or a few critical bits of content. Thus, the frame header data cannot be transferred between content. When reading the data, it must be modified by the inverse transform of the earlier modification. This system can be applied whether the data is embedded throughout each frame or all in a global file header and is discussed in application Ser. No. 09/404,291 entitled “Method And Apparatus For Robust Embedded Data” by Ken Levy on Sep. 23, 1999. Reading this secure header data is only slightly more complex than without the modification, such that the system is more efficient than always having to calculate the fingerprint and/or detect the watermark.
To provide a comprehensive disclosure without unduly lengthening this specification, applicants incorporate by reference the patents and patent applications cited above. It is applicant's express intention to teach that the methods detailed herein are applicable in connection with the technologies and applications detailed in these cited patents and applications.
Although the foregoing specification has focused on audio applications, it will be recognized that the same principles are likewise applicable with other forms of content, including still imagery, motion pictures, video, etc. Thus, for example, Digimarc MediaBridge linking from objects to corresponding internet resources can be based on identifiers derived from captured image data or the like, rather than from embedded watermarks. As such, the technique is applicable to images and video.
This application is a continuation-in-part of application Ser. No. 09/571,422, filed May 15, 2000 (now U.S Pat. No. 6,947,571). Application Ser. No. 09/571,422 claims priority benefit to each of the following provisional applications: 60/141,468, filed Jun. 29, 1999 ; 60/151,586, filed Aug. 30, 1999; 60/158,015, filed Oct. 6, 1999; 60/163,332, filed Nov. 3, 1999; and 60/164,619, filed Nov. 10, 1999. Application Ser. No. 09/571,422 is also a continuation-in-part of each of the following utility applications: Ser. No. 09/314,648, filed May 19, 1999(now Pat. No. 6,681,028); 09/342,688, filed Jun. 29, 1999 (now Pat. No. 6,650,761); 09/342,689, filed Jun. 29, 1999 (now Pat. No. 6,311,214); 09/342,971, filed Jun. 29, 1999(published as US 20030040957, now abandoned); 09/343,101, filed Jun. 29, 1999(now abandoned); 09/343,104, filed Jun. 29, 1999(now abandoned); 09/531,076, filed Mar. 18, 2000; Ser. No. 09/543,125, filed Apr. 6, 2000(now Pat. No. 7,143,949); 09/547,664, filed Apr. 12, 2000; and Ser. No. 09/552,998, filed Apr. 19, 2000 (now abandoned). This application is also a continuation-in-part of copending application Ser. Nos. 09/574,726,filed May 18, 2000, and 09/476,686, filed Dec. 30, 1999, both of which claim priority to application 60/134,782, filed May 19, 1999. The present application claims priority benefit to the foregoing applications.
Number | Name | Date | Kind |
---|---|---|---|
3919479 | Moon et al. | Nov 1975 | A |
4230990 | Lert, Jr. et al. | Oct 1980 | A |
5486686 | Zdybel, Jr. et al. | Jan 1996 | A |
5572653 | DeTemple et al. | Nov 1996 | A |
5574519 | Manico et al. | Nov 1996 | A |
5574962 | Fardeau et al. | Nov 1996 | A |
5579124 | Aijala et al. | Nov 1996 | A |
5581800 | Fardeau et al. | Dec 1996 | A |
5584070 | Harris et al. | Dec 1996 | A |
5659726 | Sandford, II et al. | Aug 1997 | A |
5748763 | Rhoads | May 1998 | A |
5768426 | Rhoads | Jun 1998 | A |
5774452 | Wolosewicz | Jun 1998 | A |
5774666 | Portuesi | Jun 1998 | A |
5781914 | Stork et al. | Jul 1998 | A |
5838790 | McAuliffe et al. | Nov 1998 | A |
5918223 | Blum et al. | Jun 1999 | A |
5938727 | Ikeda | Aug 1999 | A |
5940135 | Petrovic et al. | Aug 1999 | A |
5982956 | Lahmi | Nov 1999 | A |
5986651 | Reber et al. | Nov 1999 | A |
5995751 | Kosugi et al. | Nov 1999 | A |
6005501 | Wolosewicz | Dec 1999 | A |
6037984 | Isnardi et al. | Mar 2000 | A |
6061793 | Tewfik et al. | May 2000 | A |
6081827 | Reber et al. | Jun 2000 | A |
6121530 | Sonoda | Sep 2000 | A |
6233684 | Stefik et al. | May 2001 | B1 |
6282362 | Murphy et al. | Aug 2001 | B1 |
6282654 | Ikeda et al. | Aug 2001 | B1 |
6311214 | Rhoads | Oct 2001 | B1 |
6351815 | Adams | Feb 2002 | B1 |
6389055 | August et al. | May 2002 | B1 |
6421450 | Nakano | Jul 2002 | B2 |
6425081 | Iwamura | Jul 2002 | B1 |
6437933 | Sugiyama et al. | Aug 2002 | B1 |
6505160 | Levy et al. | Jan 2003 | B1 |
6590998 | Rhoads | Jul 2003 | B2 |
6614914 | Rhoads et al. | Sep 2003 | B1 |
6650761 | Rodriguez et al. | Nov 2003 | B1 |
6681028 | Rodriguez et al. | Jan 2004 | B2 |
6704432 | Eversole et al. | Mar 2004 | B2 |
6807285 | Iwamura | Oct 2004 | B1 |
6834308 | Ikezoye et al. | Dec 2004 | B1 |
6931451 | Logan et al. | Aug 2005 | B1 |
6941275 | Swierczek | Sep 2005 | B1 |
6947571 | Rhoads et al. | Sep 2005 | B1 |
6970886 | Conwell et al. | Nov 2005 | B1 |
6973669 | Daniels | Dec 2005 | B2 |
6987862 | Rhoads | Jan 2006 | B2 |
20010021144 | Oshima et al. | Sep 2001 | A1 |
20010026618 | Van Wie et al. | Oct 2001 | A1 |
20010032312 | Runje et al. | Oct 2001 | A1 |
20010055391 | Jacobs | Dec 2001 | A1 |
20020010826 | Takahashi et al. | Jan 2002 | A1 |
20020016915 | Kumazawa et al. | Feb 2002 | A1 |
20020023020 | Kenyon et al. | Feb 2002 | A1 |
20020030907 | Ikeda et al. | Mar 2002 | A1 |
20020032698 | Cox et al. | Mar 2002 | A1 |
20020036800 | Nozaki et al. | Mar 2002 | A1 |
20020049580 | Kutaragi et al. | Apr 2002 | A1 |
20020059522 | Hirano et al. | May 2002 | A1 |
20020059580 | Kalker et al. | May 2002 | A1 |
20020085737 | Kitamura | Jul 2002 | A1 |
20020107803 | Lisanke et al. | Aug 2002 | A1 |
20020126872 | Brunk et al. | Sep 2002 | A1 |
20030011684 | Narayanaswami et al. | Jan 2003 | A1 |
20030076955 | Alve et al. | Apr 2003 | A1 |
20030167173 | Levy et al. | Sep 2003 | A1 |
20030174861 | Levy et al. | Sep 2003 | A1 |
20050013462 | Rhoads | Jan 2005 | A1 |
Number | Date | Country |
---|---|---|
0967803 | Dec 1999 | EP |
1173001 | Jan 2002 | EP |
WO9626494 | Aug 1996 | WO |
WO0079709 | Dec 2000 | WO |
WO0106703 | Jan 2001 | WO |
WO0115021 | Mar 2001 | WO |
WO0203385 | Jan 2002 | WO |
WO0211123 | Feb 2002 | WO |
WO0227600 | Apr 2002 | WO |
Number | Date | Country | |
---|---|---|---|
20020032864 A1 | Mar 2002 | US |
Number | Date | Country | |
---|---|---|---|
60141468 | Jun 1999 | US | |
60151586 | Aug 1999 | US | |
60158015 | Oct 1999 | US | |
60163332 | Nov 1999 | US | |
60164619 | Nov 1999 | US | |
60134782 | May 1999 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 09571422 | May 2000 | US |
Child | 09858189 | US | |
Parent | 09574726 | May 2000 | US |
Child | 09571422 | US | |
Parent | 09552998 | Apr 2000 | US |
Child | 09574726 | US | |
Parent | 09547664 | Apr 2000 | US |
Child | 09552998 | US | |
Parent | 09543125 | Apr 2000 | US |
Child | 09547664 | US | |
Parent | 09531076 | Mar 2000 | US |
Child | 09543125 | US | |
Parent | 09476686 | Dec 1999 | US |
Child | 09531076 | US | |
Parent | 09342688 | Jun 1999 | US |
Child | 09476686 | US | |
Parent | 09342689 | Jun 1999 | US |
Child | 09342688 | US | |
Parent | 09342971 | Jun 1999 | US |
Child | 09342689 | US | |
Parent | 09343101 | Jun 1999 | US |
Child | 09342971 | US | |
Parent | 09343104 | Jun 1999 | US |
Child | 09343101 | US |