The invention relates to systems and methods for gathering data reflecting receipt of, and/or exposure to, audio data by encoding and obtaining both signature data and additional data and identifying the audio data based on both.
There is considerable interest in identifying and/or measuring the receipt of, and or exposure to, audio data by an audience in order to provide market information to advertisers, media distributors, and the like, to verify airing, to calculate royalties, to detect piracy, and for any other purposes for which an estimation of audience receipt or exposure is desired.
The emergence of multiple, overlapping media distribution pathways, as well as the wide variety of available user systems (e.g. PC's, PDA's, portable CD players, Internet, appliances, TV, radio, etc.) for receiving audio data, has greatly complicated the task of measuring audience receipt of, and exposure to, individual program segments. The development of commercially viable techniques for encoding audio data with program identification data provides a crucial tool for measuring audio data receipt and exposure across multiple media distribution pathways and user systems.
One such technique involves adding an ancillary code to the audio data that uniquely identifies the program signal. Most notable among these techniques is the methodology developed by Arbitron Inc., which is already providing useful audience estimates to numerous media distributors and advertisers.
An alternative technique for identifying program signals is extraction and subsequent pattern matching of “signatures” of the program signals. Such techniques typically involve the use of a reference signature database, which contains a reference signature for each program signal the receipt of which, and exposure to which, is to be measured. Before the program signal is broadcast, these reference signatures are created by measuring the values of certain features of the program signal and forming a feature set or “signature” from these values, commonly termed “signature extraction”, which is then stored in the database. Later, when the program signal is broadcast, signature extraction is again performed, and the signature obtained is compared to the reference signatures in the database until a match is found and the program signal is thereby identified.
However, one disadvantage of using such pattern matching techniques is that, after a signature is extracted from a program signal, the signature must be compared to numerous reference signatures in the database until a match is found. This problem is further exacerbated in systems that do not use a “cue” or “start” code to trigger the extraction of the signature at a particular predetermined point in the program signal, as such systems require the program signal to continually undergo signature extraction, and each of these many successive signatures extracted from a single program signal must be compared to each and every reference signature in the database until a match is found. This, of course, requires a tremendous amount of data processing, which, due to the ever increasing methods and amounts of audio data transmission, is becoming more and more economically impractical.
Accordingly, it is desired to provide techniques for gathering data reflecting receipt of and/or exposure to audio data that require minimal processing and storage resources.
It is also desired to provide such data gathering techniques which are likely to be adaptable to future media distribution paths and user systems.
For this application, the following terms and definitions shall apply, both for the singular and plural forms of nouns and for all verb tenses:
The term “data” as used herein means any indicia, signals, marks, domains, symbols, symbol sets, representations, and any other physical form or forms representing information, whether permanent or temporary, whether visible, audible, acoustic, electric, magnetic, electromagnetic, or otherwise manifested.
The term “audio data” as used herein means any data representing acoustic energy, including, but not limited to, audible sounds, regardless of the presence of any other data, or lack thereof, which accompanies, is appended to, is superimposed on, or is otherwise transmitted or able to be transmitted with the audio data.
The term “network” as used herein means networks of all kinds, including both intra-networks, such as a single-office network of computers, and inter-networks, such as the Internet, and is not limited to any particular such network.
The term “source identification code” as used herein means any data that is indicative of a source of audio data, including, but not limited to, (a) persons or entities that create, produce, distribute, reproduce, communicate, have a possessory interest in, or are otherwise associated with the audio data, or (b) locations, whether physical or virtual, from which data is communicated, either originally or as an intermediary, and whether the audio data is created therein or prior thereto.
The terms “audience” and “audience member” as used herein mean a person or persons, as the case may be, who access media data in any manner, whether alone or in one or more groups, whether in the same or various places, and whether at the same time or at various different times.
The term “processor” as used herein means data processing devices, apparatus, programs, circuits, systems, and subsystems, whether implemented in hardware, software, or both and whether operative to process analog or digital data, or both.
The terms “communicate” and “communicating” as used herein include both conveying data from a source to a destination, as well as delivering data to a communications medium, system or link to be conveyed to a destination. The term “communication” as used herein means the act of communicating or the data communicated, as appropriate.
The terms “coupled”, “coupled to”, and “coupled with” shall each mean a relationship between or among two or more devices, apparatus, files, programs, media, components, networks, systems, subsystems, and/or means, constituting any one or more of (a) a connection, whether direct or through one or more other devices, apparatus, files, programs, media, components, networks, systems, subsystems, or means, (b) a communications relationship, whether direct or through one or more other devices, apparatus, files, programs, media, components, networks, systems, subsystems, or means, or (c) a functional relationship in which the operation of any one or more of the relevant devices, apparatus, files, programs, media, components, networks, systems, subsystems, or means depends, in whole or in part, on the operation of any one or more others thereof.
In accordance with one aspect of the present invention, a method is provided for identifying audio data received at an audience member's location. The method comprises obtaining signature data from the received audio data characterizing the received audio data; obtaining additional data from the received audio data; and producing an identification of the received audio data based both on the signature data and the additional data.
In accordance with another aspect of the present invention, a system is provided for identifying audio data received at an audience member's location. The system comprises a first means to obtain signature data from the received audio data characterizing the received audio data; a second means to obtain additional data from the received audio data; and a third means to produce an identification of the received audio data based both on the signature data and the additional data.
In accordance with a further aspect of the present invention, a method is provided for encoding audio data for gathering data reflecting receipt of and/or exposure to the audio data. The method comprises forming a database having a plurality of reference signature data sets, each of which signature data sets characterizes identified audio data; grouping the reference signatures into a plurality of signature data groups; and encoding audio data to be monitored with data denoting one of the signature data groups.
In accordance with still another aspect of the present invention, a system is provided for encoding audio data for gathering data reflecting receipt of and/or exposure to the audio data. The system comprises a database having a plurality of signature groups, each of which groups has at least one reference signature data set, each of which signature data sets characterizes identified audio data; and an encoder to encode audio data to be monitored with data denoting one of the signature data groups.
The particular audio data to be monitored varies between particular embodiments and can include any audio data which may be reproduced as acoustic energy, the measurement of the receipt of which, or exposure to which, may be desired. In certain advantageous embodiments, the audio data represents commercials having an audio component, monitored, for example, in order to estimate audience exposure to commercials or to verify airing. In other embodiments, the audio data represents other types of programs having an audio component, including, but not limited to, television programs or movies, monitored, for example, in order to estimate audience exposure or verify their broadcast. In yet other embodiments, the audio data represents songs, monitored, for example, in order to calculate royalties or detect piracy. In still other embodiments, the audio data represents streaming media having an audio component, monitored, for example, in order to estimate audience exposure. In yet other embodiments, the audio data represents other types of audio files or audio/video files, monitored, for example, for any of the reasons discussed above.
After the system 30 receives the audio data, in certain embodiments, the system 30 reproduces the audio data as acoustic audio data, and the system 16 further includes a monitoring device 40 that detects this acoustic audio data. In other embodiments, the system 30 communicates the audio data via a connection to monitoring device 40, or through other wireless means, such as RF, optical, magnetic and/or electrical means. While system 30 and monitoring device 40 are shown as separate boxes in
After the audio data is received by the monitoring device 40, which in certain embodiments comprises one or more processors, the monitoring device 40 forms signature data characterizing the audio data. Suitable techniques for extracting signatures from audio data are disclosed in U.S. Pat. No. 5,612,729 to Ellis, et al. and in U.S. Pat. No. 4,739,398 to Thomas, et al., each of which is assigned to the assignee of the present invention and both of which are incorporated herein by reference.
Still other suitable techniques are the subject of U.S. Pat. No. 2,662,168 to Scherbatskoy, U.S. Pat. No. 3,919,479 to Moon, et al., U.S. Pat. No. 4,697,209 to Kiewit, et al., U.S. Pat. No. 4,677,466 to Lert, et al., U.S. Pat. No. 5,512,933 to Wheatley, et al, U.S. Pat. No. 4,955,070 to Welsh, et al., U.S. Pat. No. 4,918,730 to Schulze, U.S. Pat. No. 4,843,562 to Kenyon, et al., U.S. Pat. No. 4,450,531 to Kenyon, et al., U.S. Pat. No. 4,230,990 to Lert, et al., U.S. Pat. No. 5,594,934 to Lu, et al., and PCT publication WO91/11062 to Young, et al., all of which are incorporated herein by reference.
Specific methods for forming signature data include the techniques described below. It is appreciated that this is not an exhaustive list of the techniques that can be used to form signature data characterizing the audio data.
In certain embodiments, the audio signature data is formed by using variations in the received audio data. For example, in some of these embodiments, the signature is formed by forming a signature data set reflecting time-domain variations of the received audio data, which set, in some embodiments, reflects such variations of the received audio data in a plurality of frequency sub-bands of the received audio data. In others of these embodiments, the signature is formed by forming a signature data set reflecting frequency-domain variations of the received audio data.
In certain other embodiments, the audio signature data is formed by using signal-to-noise ratios that are processed for a plurality of predetermined frequency components of the audio data and/or data representing characteristics of the audio data. For example, in some of these embodiments, the signature is formed by forming a signature data set comprising at least some of the signal-to-noise ratios. In others of these embodiments, the signature is formed by combining selected ones of the signal-to-noise ratios. In still others of these embodiments, the signature is formed by forming a signature data set reflecting time-domain variations of the signal-to-noise ratios, which set, in some embodiments, reflects such variations of the signal-to-noise ratios in a plurality of frequency sub-bands of the received audio data, which, in some such embodiments, are substantially single frequency sub-bands. In still others of these embodiments, the signature is formed by forming a signature data set reflecting frequency-domain variations of the signal-to-noise ratios.
In certain other embodiments, the signature data is obtained at least in part from the additional data and/or from an identification code in the audio data, such as a source identification code. In certain of such embodiments, the code comprises a plurality of code components reflecting characteristics of the audio data and the audio data is processed to recover the plurality of code components. Such embodiments are particularly useful where the magnitudes of the code components are selected to achieve masking by predetermined portions of the audio data. Such component magnitudes, therefore, reflect predetermined characteristics of the audio data, so that the component magnitudes may be used to form a signature identifying the audio data.
In some of these embodiments, the signature is formed as a signature data set comprising at least some of the recovered plurality of code components. In others of these embodiments, the signature is formed by combining selected ones of the recovered plurality of code components. In yet other embodiments, the signature can be formed using signal-to-noise ratios processed for the plurality of code components in any of the ways described above. In still further embodiments, the code is used to identify predetermined portions of the audio data, which are then used to produce the signature using any of the techniques described above. It will be appreciated that other methods of forming signatures may be employed.
After the signature data is formed in the monitoring device 40, it is communicated to a reporting system 50, which processes the signature data to produce data representing the identity of the program segment. While monitoring device 40 and reporting system 50 are shown as separate boxes in
In addition to the signature data, additional data is also communicated to the reporting system 50, which uses the additional data, in conjunction with the signature data, to identify the program segment.
As shown in
In certain embodiments, the reporting system 50 has a database 54 containing reference audio signature data of identified audio data, with which the audio signature data formed in the monitoring device 40 is compared in order to identify the received audio data, as will be further explained below. In certain advantageous embodiments, prior to encoding the audio data with the additional data, the reference signatures forming the database 54 are grouped into a plurality of signature groups 82, 84, 86, 88. Accordingly, when the audio data to be monitored is encoded with the additional data, this additional data denotes the signature group in which the reference signature corresponding to the signature that is extracted from the monitored audio data is located. This type of encoded data has certain advantages that may be desired, such as, for example, drastically reducing the maximum number of reference signatures against which signature data extracted from the monitored audio data must be compared in order to ensure that a match occurs.
In some embodiments, the reference signatures may be grouped arbitrarily. In other embodiments, the reference signatures may be grouped according to some attribute of the audio data, such as a characteristic of the audio data itself, such as, for example, its duration, or a characteristic of the content of the program segment, such as, for example the program type (e.g. “commercial”). Similarly, in other embodiments, the reference signatures may be grouped according to the expected uses of the audio data, such as, for example, the ranges of time during which the audio data will be broadcast, such that particular reference signature groups may be compressed during periods when reference to the signatures in those groups is not required, which reduces the amount of storage space needed, or such that this data may be archived and stored at a location remote from the location where signature comparisons are performed, and particular reference signature groups may be retrieved therefrom only when needed, deleted when not needed, and then retrieved again when needed again.
As shown in
In some embodiments, the encoder 18 will encode the audio data with the additional data prior to its communication from the encoder 18 to the source 20. However, as noted above, the audio data may be encoded with the additional data at the source 20, such as, for example, when the reference signatures are not grouped arbitrarily, but instead, are grouped in accordance with a particular attribute of the program segment, such as, for example, by program type (e.g. “commercial”).
The additional data may be added to the audio data using any encoding technique suitable for encoding audio signals that are reproduced as acoustic energy, such as, for example, the techniques disclosed in U.S. Pat. No. 5,764,763 to Jensen, et al., and modifications thereto, which is assigned to the assignee of the present invention and which is incorporated herein by reference. Other appropriate encoding techniques are disclosed in U.S. Pat. No. 5,579,124 to Aijala, et al., U.S. Pat. Nos. 5,574,962, 5,581,800 and 5,787,334 to Fardeau, et al., U.S. Pat. No. 5,450,490 to Jensen, et al., and U.S. patent application Ser. No. 09/318,045, in the names of Neuhauser, et al., each of which is assigned to the assignee of the present application and all of which are incorporated herein by reference.
Still other suitable encoding techniques are the subject of PCT Publication WO 00/04662 to Srinivasan, U.S. Pat. No. 5,319,735 to Preuss, et al., U.S. Pat. No. 6,175,627 to Petrovich, et al., U.S. Pat. No. 5,828,325 to Wolosewicz, et al., U.S. Pat. No. 6,154,484 to Lee, et al., U.S. Pat. No. 5,945,932 to Smith, et al., PCT Publication WO 99/59275 to Lu, et al., PCT Publication WO 98/26529 to Lu, et al., and PCT Publication WO 96/27264 to Lu, et al, all of which are incorporated herein by reference.
In certain advantageous embodiments, the audio signature data is formed from at least a portion of the program segment containing the additional data. This type of signature formation has certain advantages that may be desired, such as, for example, the ability to use the additional data as part of, or as part of the process for forming, the audio signature data, as well as the availability of other information contained in the encoded portion of the program segment for use in creating the signature data.
In another advantageous embodiment, the audio data communicated from the audio source 20 to the system 30 also includes a source identification code. The source identification code may include data identifying any individual source or group of sources of the audio data, which sources may include an original source or any subsequent source in a series of sources, whether the source is located at a remote location, is a storage medium, or is a source that is internal to, or a peripheral of, the system 30. In certain embodiments, the source identification code and the additional data are present simultaneously in the audio data, while in other embodiments they are present in different time segments of the audio data.
As shown in
The audio reproducing system 32 may be any device capable of reproducing audio data from any of the audio sources referenced above at an audience member's location, including, but not limited to, a radio, a television, a stereo system, a home theater system, an audio system in a commercial establishment or public area, a personal computer, a web appliance, a gaming console, a cell phone, a pager, a PDA (Personal Digital Assistant), an MP3 player, any other device for playing digital audio files, or any other device for reproducing prerecorded media.
The system 32 causes the audio data received to be reproduced as acoustic energy. The system 32 typically includes a speaker 70 for reproducing the audio data as acoustic audio data. While the speaker 70 may form an integral part of the system 32, it may also, as shown in
In certain embodiments, the acoustic audio data is received by a transducer, illustrated by input device 43 of monitoring device 42, for producing electrical audio data from the received acoustic audio data. While the input device 43 typically is a microphone that receives the acoustic energy, the input device 43 can be any device capable of detecting energy associated with the speaker 70, such as, for example, a magnetic pickup for sensing magnetic fields, a capacitive pickup for sensing electric fields, or an antenna or optical sensor for electromagnetic energy. In other embodiments, however, the input device 43 comprises an electrical or optical connection with the system 32 for detecting the audio data.
In certain advantageous embodiments, the monitoring device 42 comprising one or more processors, is a portable monitoring device, such as, for example, a portable meter to be carried on the person of an audience member. In these embodiments, the portable device 42 is carried by an audience member in order to detect audio data to which the audience member is exposed. In some of these embodiments, the portable device 42 is later coupled with a docking station 44, which includes or is coupled to a communications device 60, in order to communicate data to, or receive data from, at least one remotely located communications device 62.
The communications device 60 is, or includes, any device capable of performing any necessary transformations of the data to be communicated, and/or communicating/receiving the data to be communicated, to or from at least one remotely located communications device 62 via a communication system, link, or medium. Such a communications device may be, for example, a modem or network card that transforms the data into a format appropriate for communication via a telephone network, a cable television system, the Internet, a WAN, a LAN, or a wireless communications system. In embodiments that communicate the data wirelessly, the communications device 60 includes an appropriate transmitter, such as, for example, a cellular telephone transmitter, a wireless Internet transmission unit, an optical transmitter, an acoustic transmitter, or a satellite communications transmitter.
In certain advantageous embodiments, the reporting system 52 comprises one or more processors and has a database 54 containing reference audio signature data of identified audio data. After audio signature data is formed in the monitoring device 42, it is compared with the reference audio signature data contained in the database 54 in order to identify the received audio data.
There are numerous advantageous and suitable techniques for carrying out a pattern matching process to identify the audio data based on the audio signature data. Some of these techniques are disclosed in U.S. Pat. No. 5,612,729 to Ellis, et al. and in U.S. Pat. No. 4,739,398 to Thomas, et al., each of which is assigned to the assignee of the present invention and both of which are incorporated herein by reference.
Still other suitable techniques are the subject of U.S. Pat. No. 2,662,168 to Scherbatskoy, U.S. Pat. No. 3,919,479 to Moon, et al., U.S. Pat. No. 4,697,209 to Kiewit, et al., U.S. Pat. No. 4,677,466 to Lert, et al., U.S. Pat. No. 5,512,933 to Wheatley, et al, U.S. Pat. No. 4,955,070 to Welsh, et al., U.S. Pat. No. 4,918,730 to Schulze, U.S. Pat. No. 4,843,562 to Kenyon, et al., U.S. Pat. No. 4,450,531 to Kenyon, et al., U.S. Pat. No. 4,230,990 to Lert, et al. U.S. Pat. No. 5,594,934 to Lu et al., and PCT Publication WO91/11062 to Young et al., all of which are incorporated herein by reference.
In certain embodiments, the signature is communicated to a reporting system 52 having a reference signature database 54, and pattern matching is carried out by the reporting system 52 to identify the audio data. In other embodiments, the reference signatures are retrieved from the reference signature database 54 by the monitoring device 42 or the docking station 44, and pattern matching is carried out in the monitoring device 42 or the docking station 44. In the latter embodiments, the reference signatures in the database can be communicated to the monitoring device 42 or the docking station 44 at any time, such as, for example, continuously, periodically, when a monitoring device 42 is coupled to a docking station 44 thereof, when an audience member actively requests such a communication, or prior to initial use of the monitoring device 42 by an audience member.
After the audio signature data is formed and/or after pattern matching has been carried out, the audio signature data, or, if pattern matching has occurred, the identity of the audio data, is stored on a storage device 56 located in the reporting system.
In certain embodiments, the reporting system 52 is a single device containing both a reference signature database 54, a pattern matching subsystem (not shown for purposes of simplicity and clarity) and the storage device. In other embodiments, 56 the reporting system 52 contains only a storage device 56 for storing the audio signature data. Such embodiments have certain advantages that may be desired, such as, for example, limiting the amount of storage space required in the device that performs the pattern matching, which can be achieved, for example, by only retrieving particular groups or subgroups of reference signatures as explained above.
Referring to
In yet other embodiments, as illustrated in
In still further embodiments, the source is another audio reproducing system, as defined below, such that a plurality of audio reproducing systems receive and communicate audio data in succession. Each system in such a series of systems may be coupled either directly or indirectly to the system located before or after it, and such coupling may occur permanently, temporarily, or intermittently, as illustrated stepwise in
Referring to
In certain other embodiments, as illustrated in
In still other embodiments, as illustrated in
In still further embodiments, as illustrated in
Referring to
Although the invention has been described with reference to particular arrangements and embodiments of services, systems, processors, devices, features and the like, these are not intended to exhaust all possible arrangements or embodiments, and indeed many other modifications and variations will be ascertainable to those of skill in the art.
Number | Name | Date | Kind |
---|---|---|---|
2662168 | Scherbatsoy | Dec 1953 | A |
3919479 | Moon et al. | Nov 1975 | A |
4230990 | Lert et al. | Oct 1980 | A |
4450604 | Oowada | May 1984 | A |
4677466 | Lert et al. | Jun 1987 | A |
4697209 | Kiewit et al. | Sep 1987 | A |
4739398 | Thomas et al. | Apr 1988 | A |
4843562 | Kenyon et al. | Jun 1989 | A |
4918730 | Schulze | Apr 1990 | A |
4955070 | Welsh et al. | Sep 1990 | A |
4972471 | Gross et al. | Nov 1990 | A |
5319735 | Preuss et al. | Jun 1994 | A |
5425100 | Thomas et al. | Jun 1995 | A |
5450490 | Jensen et al. | Sep 1995 | A |
5481294 | Thomas et al. | Jan 1996 | A |
5512933 | Wheatley et al. | Apr 1996 | A |
5574962 | Fardeau et al. | Nov 1996 | A |
5579124 | Aijala et al. | Nov 1996 | A |
5581800 | Fardeau et al. | Dec 1996 | A |
5594934 | Lu et al. | Jan 1997 | A |
5612729 | Ellis et al. | Mar 1997 | A |
5764763 | Jensen et al. | Jun 1998 | A |
5787334 | Fardeau et al. | Jul 1998 | A |
5828325 | Wolosewicz et al. | Oct 1998 | A |
5945932 | Smith et al. | Aug 1999 | A |
6035177 | Moses et al. | Mar 2000 | A |
6061793 | Tewfik et al. | May 2000 | A |
6154484 | Lee et al. | Nov 2000 | A |
6175627 | Petrovic et al. | Jan 2001 | B1 |
6208735 | Cox et al. | Mar 2001 | B1 |
6647128 | Rhoads | Nov 2003 | B1 |
6700990 | Rhoads | Mar 2004 | B1 |
6738744 | Kirovski et al. | May 2004 | B2 |
20040022322 | Dye | Feb 2004 | A1 |
Number | Date | Country |
---|---|---|
WO 9111062 | Jul 1991 | WO |
WO 9512278 | May 1995 | WO |
WO 9627264 | Sep 1996 | WO |
WO 9810539 | Mar 1998 | WO |
WO 9826529 | Jun 1998 | WO |
WO 9832251 | Jul 1998 | WO |
WO 9959275 | Nov 1999 | WO |
WO 0004662 | Jan 2000 | WO |
WO 0072309 | Nov 2000 | WO |
Number | Date | Country | |
---|---|---|---|
20040122679 A1 | Jun 2004 | US |