Methods and apparatus for transcoding metadata

Information

  • Patent Grant
  • 9900652
  • Patent Number
    9,900,652
  • Date Filed
    Wednesday, March 22, 2017
    7 years ago
  • Date Issued
    Tuesday, February 20, 2018
    6 years ago
Abstract
Methods and apparatus for transcoding metadata are disclosed. Example apparatus disclosed herein include means for accessing first metadata accompanying media to be presented by a media device, the first metadata in a first format not detectable by a meter collecting audience measurement data associated with the media device. The example apparatus also include means for transcoding the first metadata into second metadata having a second format different from the first format, the second metadata detectable by the meter and supported by an application programming interface (API) provided by the media device and accessible by the meter. The example apparatus further include means for communicating the second metadata in a second transmission for receipt by the meter via the API provided by the media device, the second transmission being separate from a first transmission communicating the media and the first metadata, but not communicating the second metadata, to the media device.
Description
FIELD OF THE DISCLOSURE

This disclosure relates generally to transcoding and, more particularly, to methods and apparatus for transcoding metadata.


BACKGROUND

Through the advancement of technology, media content is finding its way into homes by various non-traditional means. For instance, with the advent of broadband data connections, media content is now being supplied to homes via the Internet and other digital distribution sources as well as the traditional means of television and radio broadcasts. In addition, home networking standards such as HAVi and wired and wireless protocols such as IEEE 1394, IEEE 802.11, ultra wide band (UWB), cellular and pager networks and Bluetooth now allow a variety of different media consumption devices to communicate via a home network. Thus, media content received at a home can be distributed via a home network for display on any of the media consumption devices connected to the home network.


These advancements in home networking and media content delivery mechanisms provide users with enhanced access to media content, but also present new challenges for measuring the consumption of media content. More specifically, conventional media consumption meters are typically configured to measure media signals consumed in a particular format by a specific type of media consumption device. For example, television content consumption is measured using meters adapted to measure signals formatted for television. Likewise, computer content consumption is measured using meters adapted to measure signals formatted for a computer. Although this approach works well for media consumption devices that are not networked together, it does not allow for media consumption measurement in a home networked environment because of the manner in which media content is transmitted via a home network. Specifically, a home network typically comprises a variety of media consumption devices that are networked together and that are each adapted to process media provided in a particular format. To enable communication between the networked devices, home networks include one or more transcoders that transcode or convert media content signals transmitted via the home network between the various signal formats required by the networked devices. For example, a typical home network may be configured to include a television coupled to a computer with a transcoder disposed between the television and the computer. Media content that is received at the television and then selected for viewing at the computer is converted by the transcoder from a television format to a computer format so that the media content received at the television can be processed for display/viewing at the computer.


Unfortunately, the format conversion process performed by home network transcoders to enable communication between dissimilar devices also causes the corruption or loss of valuable audience measurement data. More specifically, data that enables media consumption measurement is embedded into media content by media content providers. The data is then extracted by media content consumption meters for use in determining the identity of the media content as well as other information about the media content which can then be reported for purposes of measuring consumption of that media content. However, home network transcoders are not adapted to convert this embedded data into a format suitable for use by downstream consumption measuring devices and, as a result, the embedded data is either stripped from the media content or corrupted during the transcoding process. Consequently, consumption of the converted media content cannot be accurately measured.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an example home network having a set of transcoders that convert data embedded in media content for use by a set of media consumption metering devices.



FIG. 2 is a block diagram that depicts an example manner in which the transcoder of FIG. 1 may be implemented.



FIG. 3 is a block diagram that depicts an example manner in which one of the media consumption metering devices of FIG. 1 may be implemented.



FIG. 4 is a flow chart that depicts a method for performing metadata transcoding.



FIG. 5A is a diagram that depicts a set of metadata fields suitable for use with media content provided in a broadcast television format.



FIG. 5B is a diagram that depicts a set of metadata fields suitable for use with media content provided in a streaming media format.



FIG. 6 is a flow chart that depicts an example method for creating a watermark that represents a particular media content and for creating correlation information that may be used to correlate the watermark with the particular media content.



FIGS. 7A-7C are flow charts that align to depict an example method for determining the sensing capabilities of a metering device.





DETAILED DESCRIPTION

Referring to FIG. 1, a home network 10 includes a plurality of media consumption devices 12, each representing one of a variety of devices including, for example, a TV, a radio, a personal computer, a personal digital assistant (PDA), a telephone and a digital video disk (DVD) player, a personal video recorder (PVR). Each of the media consumption devices 12 are adapted to receive media content from one or more different media sources collectively represented in FIG. 1 as a media content cloud 14. The content provided by the media content cloud 14 may include, for example, broadcast content, streaming or web content from the Internet, or content supplied by a local device, such as, for example, a DVD player, a video cassette recorder (VCR), a media server or any other local source. The media consumption devices 12 are in communication with each other such that media content transmitted via the home network 10 may be consumed via the media consumption devices 12, may be shared between multiple media consumption devices 12 or may be stored on a media storage database 16, also coupled to the home network 10, for later retrieval and consumption. The home network may be configured in either a wired or wireless manner. In some embodiments, the home network may include a combination of both wired and wireless communication.


To enable audience measurement, a set of content consumption metering devices 18, hereinafter “meters”, are also coupled to the home network 10 and are adapted to measure the content consumed by the media consumption devices 12. Each of the meters 18 may be configured to meter a single one of the media consumption devices 12 or may be configured to meter multiple media consumption devices 12. Additionally, each meter 18 may be adapted to operate independently or each may instead be adapted to operate under the control of a master or central metering unit (not shown). The meters 18 may be coupled to the network 10 wirelessly or in a wired fashion. For example, if the meter 18 is implemented as a portable meter to be carried by a household member for purposes of measuring the consumption habits of that household member, then the meter will likely communicate with the network wirelessly. Such communication may be as limited as receiving codes from a media consumption device or may be as robust as two-way communication between the portable meter and other network devices. Audience measurement data collected by each meter 18 is supplied to a measurement collection unit 20 where the data is either analyzed or stored for later transmission to a remote data collection facility (not shown) for analysis. The measurement collection unit 20 may be disposed within the same residence as the home network 10 or may instead be disposed outside of the residence that houses the home network 10. As can be seen in FIG. 1, the meters 18 may be separate from, or integral with, the media consumption devices 12 being metered.


Referring still to FIG. 1, a set of transcoders 22 are coupled to the home network 10 at various locations and are configured to transcode or convert media content transmitted via the home network 10 between one or more formats thereby causing the media content to be suitable for consumption by one or more of the media consumption devices 12. In an example configuration, the transcoders 22 are adapted to receive media content from one or more of the media consumption devices 12, to transcode the media content and to provide the transcoded media content for consumption to one or more of the media consumption devices 12 according to a method described in greater detail below.


A home network media server 24 coupled to and in communication with the home network media storage database 16 may be adapted to control communication on the home network 10, may be adapted to function as a media content consumption device 12 and may be further adapted to receive content from the media cloud 14. The home network media server 24 may provide content to the various devices coupled to the home network 10 by streaming the content, pushing the content, allowing the content to be downloaded or by any other manner. Additionally, the home network media server 24 may act as a central repository for all of the media content that is consumed in the home network 10. Alternatively, the media content may be received at any of the media consumption devices 12 and then supplied to the home network 10 for consumption/viewing/display at any of the other media consumption devices 12.


As will be appreciated by one having ordinary skill in the art, the home network 10 may be configured in any desired manner and may include any number of network devices. Moreover, the devices coupled to the home network 10 may communicate and interface in any desired manner including, for example, using either a wired or wireless communication protocol such as HAVi, Wi-Fi, BlueTooth, IEEE 1394, DVI, HDMI or any other high speed interface protocol or using an Ethernet protocol.


Referring now to FIG. 2, in an example embodiment, each meter 18 is equipped with a processor 26 which executes a set of instructions stored in a memory 28 to control the operation of the meter 18 and a home network interface 30 that allows communication between the meter 18 and the other devices coupled to the home network 10 and that may be implemented using, for example, a software application program interface (API). As will be appreciated by one having ordinary skill in the art, the home network interface 30 may be implemented using any type of device capable of translating between a communication protocol used by the home network 10 and a communication protocol used by the processor 26 disposed in the meter 18. In addition, each meter 18 includes one or more interfaces 32a, 32b for interfacing with other devices and one or more sensors 32c-32e for sensing media content consumption. The interfaces 32a, 32b may include, for example, an interface 32a adapted to communicate with a people meter that senses the presence of one or more network users/audience members and/or a data collection interface 32b adapted to communicate with the collection measurement unit 20. The sensors 32c-32e supply sensed information to the processor 26 which processes the information and causes the information to be transmitted to the measurement collection unit 20 via the home network 10. The measurement collection unit 20 then transmits the data to a remote data collection facility (not shown) for analysis. In addition, the processor 26 is adapted to respond to queries sent by the transcoder 22 via the home network 10. Alternatively, the measurement collection unit 20 need not be coupled to the home network 10 but may instead be adapted to receive media consumption measurement data directly from the meters 18 via one or more data collection interfaces 32b, disposed in one or more of the meters 18.


The sensors 32c-32e associated with each meter 18 are adapted to sense the type of output signals supplied by a corresponding media consumption device and may be implemented using, for example, a microphone 32c for sensing audio signals, a video sensor 32d for sensing video signals, and/or a digital data sensor 32e for sensing data supplied in a digital bitstream. Due to the plurality of sensor types that may be installed in the meters 18, the meters 18 may be adapted to sense a variety of signals and may be further adapted to recognize and process a variety of codes embedded in such signals. These codes may include video based codes such as closed captioning, automated measurement of lineup (AMOL), interactive television triggers in the vertical blanking interval (VBI) of a traditional NTSC or PAL television signal, imperceptible codes in active analog/digital video, and codes included in the user data of digital video packets, to name a few. In addition, these codes may include inaudible audio codes, auxiliary data codes, digitally compressed audio packets as well as information in digital packets containing program guide information such as PSI, PSIP and A-90 data packets. Such codes may be formatted using existing broadcast standards or may instead be formatted according to standards that are currently being developed or that will be developed in the future such as, for example, the standard recently developed by the BBC and presented to TV-Anytime, the content identifier standard being developed by ATSC called Versioned ISAN (VISAN) and the standard known as AD-ID that was recently defined by the American Association of Advertising Agencies. Instead of, or in addition to, sensing signals containing codes associated with media content, one or more of the sensors may be adapted to receive signals that indicate usage of a computer software application. For example, one or more of the sensors may be adapted to monitor a computer software application that controls access to the database 16 to determine, for example, how often particular media content is being transferred into, or out of, the database 16.


The codes are transmitted as metadata via the media signal. The metadata may include a variety of information associated with the media content, such as, for example, content identification information, source identification information (SID), destination device identification information, distribution channel identification information and/or data and timestamps identifying the creation and/or transmission dates of the media content. Metadata may additionally include signal characteristics such as frequency, format, signal strength, bitrates, frame rates, and sampling frequency, to name a few. The signal format information may be used to transcode the signal from a first format to a second format to enable consumption of the signal at a consumption device that is able to recognize and process the second format.


As will be appreciated by one having ordinary skill in the art, the meters 18 may be implemented using any combination of software and hardware adapted to perform core metering functions such as receiving and processing consumption data and supplying the consumption data to a central data collection facility. As will further be appreciated by one having ordinary skill in the art, existing media consumption metering technology spans from the very simple to the complex. Yet, the present invention is not limited to use with either type of metering technology but instead may be used with a variety of meters 18 having a range of complexity. For example, the meters 18 used with the present invention may be capable of communicating via the home network 10 as well as metering media consumption or the meters 18 may instead be standalone devices that are configured to meter a particular type of media consumed via a particular type of media consumption device 12. In addition, the meters 18 may be implemented as simple, hardware based devices that collect channel tuning information or as more sophisticated, software based devices that perform intelligent data collection and processing functions.


The location of each of the meters 18 within the home network 10 depends on the type of metering to be performed. For example, if the meter 18 is adapted to perform metering by sensing a signal output by one of the media consumption devices 12, such as an audio or video signal, then the meter 18 is typically positioned near that content consumption device 12 so that it may access the signal to be metered. If, instead, the meter 18 is adapted to meter a signal supplied by the transcoder 22, then the meter 18 need not be positioned in close, physical proximity to the media consumption device 12 but must be either proximate to the transcoder 22 or arranged to remotely receive the signal output by the transcoder 22. Of course, in the latter embodiment, the meter 18 detects the flow of signals through the transcoder 22 instead of the consumption of those signals by a media content consumption device 12. Such an arrangement may be especially useful with a meter 18 adapted to detect the flow of signals through the transcoder 22 and having a software application interface (not shown) adapted to query the transcoder 22 about the media content being transcoded.


Referring to FIG. 3, the transcoder 22 of the present invention is adapted to transcode both media content and metadata transmitted with the media content and may be implemented to include a home network interface 30, an input media demultiplexer 34, a processor 36, a memory 38, an output media codec 40, an encode manager 42, and an audio/video watermark codec 44 that is adapted to detect watermarks and to encode new watermarks. The various components included in the transcoder 22 are controlled by the processor 36 which executes a set of software instructions stored in the memory 38 and each transcoder component may be implemented using software, firmware, hardware of any combination thereof. For example, commercially available devices may be used to implement any of the transcoder components provided that the functionality of each device has been modified as needed to operate as described herein. In addition, one or more of the components, such as the encode manager 42, may be implemented as software routines stored in the memory 38 and executed by the processor 36.


Referring also to FIG. 4, the transcoder 22 may be adapted to perform a method 50 that enables the transcoder 22 to transcode media content and accompanying metadata received from a first media consumption device 12 coupled to the home network 10, such as a television, and to supply the transcoded media content and accompanying metadata to a second media consumption device 12, such as a personal computer, also coupled to the home network 10. The method 50 may be performed by the various components of the transcoder 22, as described below, operating under the control of the processor 36 which executes a set of software instructions stored in the memory 38. In an example embodiment, the method 50 may begin when the transcoder 22 receives media content at the home network interface 30 (block 52). The home network interface 30 disposed in the transcoder 22 enables communication between the transcoder 22 and the other devices in the home network 10. As described with respect to the home network interface 30 disposed in the meter 18, the home network interface 30 disposed in the transcoder 22 may be implemented using any type of device capable of translating between the communication protocol used by the home network 10 and the communication protocol used by the processor 36 disposed in the transcoder 22. Also at the block 52, the home network interface 30 stores the media content in the memory 38.


After the media content has been stored in the memory 38, the input media demultiplexer 34 demultiplexes the media content to extract the metadata transmitted therewith (block 54). The media content received at the home network interface 30 need not be stored before being demultiplexed by the demultiplexer 34 but may instead be delivered by the home network interface 30 to the demultiplexer 34 as the media content arrives at the transcoder 22. The demultiplexer 34 may demultiplex the media content as it is received and then cause the demultiplexed content, i.e., the media content and the extracted metadata to be stored in the memory 38. The method 50 may then continue with the encode manager 42 examining the extracted metadata to identify a media format, “Format A”, in which the media content was supplied to the transcoder 22 by the television (block 56). For example, the metadata may indicate that the media content received at the transcoder 22 is formatted as a television broadcast signal such that Format A is a standard broadcast television signal format such as, for example, an NTSC signal. In addition, the encode manager 42 examines the metadata to identify a metadata format, Format A1, in which the metadata extracted from the media content is formatted. In this example embodiment, the transcoder 22 is described as being adapted to examine the incoming media content and metadata to determine the identities of Format A and Format A1. The ability of the transcoder 22 to make such a determination is especially useful for transcoders adapted to receive media content and metadata from a variety of media consumption device types. If, instead, the transcoder 22 is configured within the home network 10 to receive media content and metadata from only a single media consumption device type, then the transcoder 22 may be pre-programmed with the identities of Format A and Format A1 such that the encode manager 42 need not determine the identities of Format A and Format A1 (block 56).


Next, the encode manager determines a media format, “Format B”, associated with the second network device to which the transcoded media content shall be delivered for consumption (block 58). By way of example, the personal computer may have a Windows operating system and a Windows media player such that the media content must be supplied to the personal computer in a streaming media format suitable for display via the Windows media player. Thus, for this example, media Format B is a streaming media signal format. The encode manager 42 may be pre-programmed with the media format used by the personal computer, i.e., Format B, such that the encode manager retrieves this format information from the memory 38. Alternatively, the encode manager 42 may be configured to query the personal computer to obtain media Format B. In addition, the encode manager 42 also identifies a metadata format, Format B1, to which the extracted metadata shall be transcoded before delivery to the personal computer (block 58). Format B1 is the metadata format recognized by the meter 18 configured to measure media consumption at the personal computer. The encode manager 42 may be pre-programmed with the identity of the metadata format, Format B1, or the encode manager 42 may be adapted to obtain the identity of Format B1 from the meter 18 associated with the personal computer by performing one or more of a set of querying methods described in greater detail below.


Once the encode manager 42 has identified the applicable media formats, Formats A and B, and the applicable metadata formats, Format A1 and B1, the method continues with the encode manager 42 instructing the output media codec 40 to transcode the media content from Format A to Format B and the extracted metadata from Format A1 to Format B1 and providing the output media codec 40 with the parameters needed to perform such transcoding functions (block 60). In addition, the encode manager 42 supplies the output media codec 40 with instructions and parameters necessary for combining the transcoded metadata with the transcoded media content to form an output media signal for subsequent delivery to the personal computer. Lastly, the transcoded media signal having the transcoded metadata embedded therein is supplied by the transcoder 22 to the personal computer for consumption thereat (block 62).


As will be appreciated by one having ordinary skill in the art, methods for performing transcoding functions that convert a signal from a first media format to a second media format are well known in the art. For example, the transcoder 22 may be adapted to convert a signal containing media content from a broadcast format to any of a 1) streaming media format, 2) JPEG format, e.g., deriving a still picture from a movie for use in a digital photo frame, 3) MP3 format, e.g., playing a soundtrack to a broadcast movie, and/or 4) a digital video recorder format. Alternatively, the transcoder 22 may be adapted to convert a signal from a digital video recorder format to any of a 1) streaming media format, 2) MP3 format, and/or 3) a JPEG format. By way of further example, the transcoder 22 may be adapted to convert a signal from a streaming media format to either of a 1) JPEG format and/or 2) MP3 format. As will further be appreciated by one having ordinary skill in the art, any of these well-known transcoding techniques may be adapted to enable transcoding of the metadata from a first metadata format to a second metadata format.


As will further be appreciated by one having ordinary skill in the art, the metadata formats may be configured in a variety of different ways and, in some instances, the metadata formats may be at least partially dependent on the media format of the media content associated with the metadata. For example, if the media content associated with the metadata is provided in a broadcast television media format, then the metadata format will likely be formatted to include data fields related to broadcast television. Referring now to FIG. 5A, Format A1, if associated with a digital broadcast television media format, such as, for example, ATSC, may include a set of fields 70 for containing a variety of broadcast television signal information such as a minor channel number 72, a major channel number 74, a transport ID 76, a name of a program transmitted in the media content 78, a program number 80, a program description 82, a source ID 84, a date/time stamp 86, a transport rate 88, a video bitrate 90, an audio bitrate 92, an audio sampling rate 94, and a video frame rate 96. In this example, the minor channel number and major channel number are each 10 bits long, the transport ID is 16 bits long, the program name is represented as a sequence of one to seven 16 bit character codes coded in accordance with the Basic Multilingual Plane (BMP) of Unicode™, as specified in ISO 10646-1, the program number is 8 bits long, the program description is represented as a sequence of one to seven 16-bit character codes coded in accordance with the Basic Multilingual Plane (BMP) of Unicode™, as specified in ISO 10646-1, the source ID is 16 bits long, and the date/time stamp is a 32 bit long number in UTC format.


In contrast, metadata associated with media content that is to be supplied in a streaming media format for consumption/display at a personal computer will likely be formatted as a data header including data fields 98 for containing information about the media content in a digital data format. Thus, for example, Format B1 may include fields for containing information such as the program name 100, the program number 102, the program description 104, a source ID 106, a broadcast date/time stamp 108, a transcoded date/time stamp 110, an author 112, a device ID 114, a home ID 116, a format 118, a video bitrate 120, an audio bitrate 122, an audio sampling rate 124, and a video frame rate 126. In addition, the data fields of Format B1 may be defined to include data in a specific code or a data string having a specific number of bits. For example, the fields of Format B1 may be defined such that the program name 100 is ASCII encoded, the program number 102 is 8 bits long, the program description 104 is ASCII encoded, the source ID 106 is identical in format to the source ID of Format A, the broadcast date/time stamp 108 is the same as in the broadcast metadata, the transcoded date/time stamp 110 is a 32 bit number reflecting the date and time of transcoding by transcoder 104, the author field 112 is an ASCII representation of the user of the transcoder 104, the device ID 114 is an 8 bit ASCII encoded number identifying the transcoder 104, and the home ID 116 is a 32 bit ASCII encoded number assigned to identify the home network 10. In addition to identifying Formats A1 and B1, the encode manager 42 may be configured to generate the data contained in one or more of the fields of Format B1 including, for example, the data associated with the author 112, device ID 114, home ID 116 and format 118 fields.


The metadata format, Format B1, may also be affected by the type of meter 18 that will be used to collect the metadata for purposes of measuring consumption of the media content that is associated with the metadata. For example, if the meter 18 is configured to detect audio codes, then the metadata format, Format B1, will be suitable for insertion into an audio signal and recognition by an audio code meter and/or if the meter is configured to detect video codes, then the metadata format, Format B1, will be suitable for insertion into a video signal and recognition by a video code meter. Likewise, if the meter 18 is configured to detect digital data, then the metadata format, Format B1, will be suitable for insertion into a digital data stream and detection by a digital data sensor.


The transcoding capabilities of the transcoder 22 may also vary depending upon the position of the transcoder 22 within the home network 10. For example, if the transcoder 22 is positioned to receive input from multiple media consumption devices 12 and to provide transcoded media content to multiple media consumption devices 12, then the transcoder 22 will likely be adapted to transcode media content between the multiple formats used by the media consumption devices 12 coupled thereto. Alternatively, if the transcoder 22 is positioned to receive input from a single media consumption device 12 and to supply transcoded media content to a single media consumption device 12, then the transcoder 22 need only be adapted to transcode between the media formats used by the two media consumption devices 12 configured to supply/receive media content to/from the transcoder 22.


Referring now to FIG. 6, depending on the capabilities of the meter 18 associated with the second network device, e.g., the personal computer, and whether one or more conditions are satisfied, as is described in greater detail below, the method 50 may be expanded to include a submethod, identified generally with reference numeral 130, for causing the audio/video watermark codec 44 to insert a watermark into the media content in a manner that causes the inserted watermark to be imperceptible to the human senses so that the inserted watermark does not interfere with the consumption of the media content at the personal computer. In such an arrangement, the encode manager 42 may use the metadata associated with the media content to identify the media content and then cause the audio/video watermark codec 44 to generate a watermark that uniquely represents that media content and to insert the watermark into the transcoded media content (block 132). Next, the transcoder 22 supplies information that correlates the identity of the media content with the unique watermark to the meter 18 which may be adapted to transmit this correlation information directly to a remote data collection facility (not shown) or via the measurement collection unit 20 (see FIG. 1) (block 122). Alternatively, the transcoder 22 may be adapted to supply this correlation information directly to the measurement collection unit 20 via the home network 10. When the media content having the inserted watermark is consumed at the personal computer, and the meter 18 associated with the personal computer extracts the watermark and reports the presence of the watermark to the measurement collection unit 20 or the remote data collection facility, the data collection facility may use the correlation information to determine the identity of the consumed media content based on the reporting of the unique watermark.


Referring again to FIG. 1, it should be understood that not all of the transcoding functions need to be performed by a single network component but may instead be performed by any of the consumption devices 12, or any other network device or combination of network devices, provided that such devices are adapted to perform the transcoding functions described herein. For example, the transcoder 22 may be configured to demultiplex an incoming media content signal to separate the metadata contained in the signal from the programming or other media content as described with respect to FIGS. 3 and 4. Alternatively, a media consumption device 12 may be configured to perform this demultiplexing function such that the metadata extracted thereby is transmitted to the transcoder 22 separately from the media content. Likewise, any of the media consumption devices 12 may be configured to decode or otherwise process the media content signal before the signal is transmitted to the transcoder 22. For example, any of the media consumption devices 12 may be configured to receive and demodulate/downconvert a digitally compressed broadcast signal provided in an ATSC, DVB or MPEG format via an RF transmission. Any of the media consumption devices 12 may further be configured to include a demultiplexer that demultiplexes the demodulated bitstream to obtain a set of elementary components, including an elementary video stream, an elementary audio stream and the metadata transmitted in the broadcast. The elementary video and audio streams may then be decoded by an MPEG/AC3 audio decoder and an MPEG video decoder, respectively, both of which may also be disposed in any of the media consumption devices 12. The decoded video and audio streams may subsequently be transmitted to the transcoder 22 for transcoding in accordance with the blocks 56-62 of the method 50 shown in FIG. 4 described above. After the transcoding has been performed, the transcoder 22 may be adapted to yield a streaming media signal that may be delivered to one of the media consumption devices 12 configured to consume/display streaming media or may instead be stored in the memory 16 by the server 24 for consumption at a later time. In addition, the meter 18 used to meter the media consumption device 12 at which the streaming media is consumed may be configured to receive the transcoded metadata directly from the transcoder 22 or may instead be configured to extract the transcoded metadata embedded in the streaming media signal delivered by the transcoder 22 as it is consumed by the media consumption device 12.


Referring still to FIG. 1, in an embodiment in which the transcoder 22 supplies the transcoded metadata directly to a first meter 18, the transcoder 22 may be disposed in or in communication with a second meter 18 configured to meter consumption at a media consumption device 12. The second meter 18 may be implemented using, for example, a set top box 18 that is configured to meter consumption of television programming via, for example, a television 12. In this embodiment, the set top box 18 collects metadata as the corresponding television programming is consumed at the television 12. The set top box 18 then supplies the collected metadata to the transcoder 22 which transcodes the metadata from the first format, Format B1, to the second format, Format B2, suitable for delivery to and reception at the first meter 18. The transcoder 22 may be integrated with the set top box 18, disposed within the set top box 18, or separate from but in communication with the set top box 18. The first meter 18 may be implemented using, for example, a portable meter 18 designed to collect audio codes or designed to collect data transmitted wirelessly using any wireless protocol. Of course, if the first meter 18 is designed to receive data wirelessly using a wireless protocol such as Bluetooth, then the transcoder 22 will be configured to transcode the metadata to a Bluetooth format and the second meter 18 will be configured to transmit the data in Bluetooth format. The portable meter 18 may be configured to be carried by a household member for purposes of metering that household member's viewing habits. The portable meter 18 may additionally be configured to meter media consumption by that household member that occurs out of the home by detecting audio codes emitted by media consumption devices 12 disposed outside the home and may be configured to detect media consumption within the home by wireless reception of transcoded metadata from the set top box 18 as described above. Alternatively, the portable meter 18 may be designed to detect audio codes emitted by the television 12 as well as metadata transcoded and transmitted by the set top box 12 such that the portable meter 18 receives two sets of data representing the same viewing event while disposed in the home. These two data sets may then be compared for consistency to increase confidence in the accuracy of the data. Of course, the two data sets must be treated such that the sets are only counted as a single viewing event. In another example embodiment, the set top box 18 may configured to supplement the code capturing capabilities of the portable meter 18 by detecting codes that are not detectable by the portable meter 18, by transcoding such codes to a format suitable for detection by the portable meter 18 and by then supplying the transcoded codes to the portable meter 18. In such an embodiment, the portable meter 18 may be designed to detect a first type of code such as an SID, emitted by the television 12, but not a second type of code, such as a time stamp. The set top meter 18 may be designed to detect either or both of the first and second types of code. For example, the set top meter 18 may detect the SID and the time stamp emitted by the television 12 and may transcode the SD and timestamp to a format suitable for reception/processing by the portable meter 18. Thereafter, the set top meter 18 may transmit one or both of the transcoded time stamp and SID to the portable meter 18 via, for example, radio frequency signals, infra-red signals, Wi-Fi signals, audio signals, etc. Of course, the set top and portable meters 18 will have to be equipped to communicate using such signals.


The transcoder 22 may be pre-programmed with the identity of Format A1 and Format B1 such that the transcoder 22 may automatically convert metadata received at the transcoder from Format A1 to Format B1. In fact, such an arrangement may be preferred when the transcoder 22 is adapted to receive media in a single format only and to transmit data in a single format only. Alternatively, referring also to FIG. 7A, as mentioned above, the transcoder 104 may be adapted to perform a variety of methods to query one or more networked meters 18 for information about the detection capabilities of the meters 18 so that the format in which the metadata is to be supplied to the meters 18, i.e., Format B1, can be determined. For example, one such method 140 may comprise a set of submethods, each submethod designed to test for the presence of a different type of sensor. The first such submethod 142 may test for the presence of an audio sensor capable of sensing audio watermarks and may begin when the transcoder 22 queries the meter 18 to determine whether it includes an audio watermark sensor (block 144). If the meter 18 responds to the query in the negative, i.e., the meter 18 does not have an audio watermark sensor, then the transcoder 22 bypasses the remainder of the submethod 142 and instead proceeds to additional submethods for testing whether the meter 18 includes one or more other types of sensors as described in greater detail below with reference to FIGS. 7B and 7C.


If instead, in response to the query performed at the block 64, the meter 18 responds in the positive, i.e., the meter 18 does have an audio watermark sensor, then the submethod 62 continues with the transcoder 22 determining whether an audio watermark has been detected in the media content supplied to the transcoder 22 for transcoding (block 146). If an audio watermark is not detected in the media content, then the transcoder 22 creates a new audio watermark and causes the new audio watermark to be embedded into the media content (block 148). As will be appreciated by one having ordinary skill in the art, the audio/video watermark codec 44 may be configured to detect the presence of an audio watermark supplied in the media content and to create a new audio watermark for insertion in the content. Moreover, the capabilities and functionality of a standard audio/video watermark codec are well known in the art are not described further herein. The transcoder 22 may also cause all or a portion of the submethod 130, described with respect to FIG. 6, to be performed thereby causing correlation information to be generated and transmitted to the measurement collection unit 20 and/or a remote data collection facility where it may be used to correlate the watermark to the program or other content it represents.


If an audio watermark has been detected at the block 146, then the transcoder 22 determines whether the audio watermark, if inserted into the signal to be output by the transcoder 22, will survive the compression performed by the output media codec 40 (block 150). Specifically, the output media codec 40 is adapted to compress the media content signal having the inserted watermark before the signal is transmitted via the home network 10. The output media codec 40 compresses the signal by suppressing one or more of the signal frequencies. However, watermarks are created by modulating a particular set of signal frequencies in a manner such that the modulated frequencies uniquely represent a particular program or other media content. Thus, the compression performed by the output media codec 40, may cause one or more of the frequencies modulated to create the watermark to be suppressed thereby causing the audio watermark to be unrecoverable by the meter 18. The transcoder 22 may be adapted to perform a variety of methods for determining whether the watermark will survive compression by the output media codec 40. For example, the transcoder 22 may cause the output media codec 40 to insert the watermark into the media content and the resulting signal may be processed by the transcoder 22, in much the same way an input signal would be processed, to determine whether the audio watermark is recoverable. In another embodiment, the transcoder 22 may be pre-programmed with information pertaining to signal compression ratios that the watermark will be able to withstand/survive. Specifically, before inserting a particular watermark into a media content signal, the watermark may be tested to determine a range of suitable signal compression ratios, i.e., compression ratios that the watermark will survive. These suitable ratios may then be provided to the manufacturers/developers of the transcoder 22 and used to pre-program the transcoder 22 so that when watermarks are encountered, the transcoder 22 may use the pre-programmed information to compare to the ratio used by the output media codec 40 to determine whether the compression ratio used by the output media codec 40 is suitable inserting and for transmitting the watermark in a recoverable, distortion-free manner. In a still further embodiment, information about suitable compression ratios may be transmitted with the signal containing the watermark and extracted from the signal by the transcoder 22 for use in determining whether the watermark will survive the compression ratio used by the output media codec 40. If the audio watermark will survive, then the sub method 142 causes the output media codec 40 to insert the watermark (block 152) (if it is not already inserted) after which the submethod 142 is complete and the method 140 continues at another submethod described below with respect to FIG. 7B and FIG. 7C.


If, at the block 150, the transcoder 22 determines that the compression ratio used by the output media codec 40 is not suitable, i.e., will cause the watermark to be unrecoverable, then the transcoder 22 may decrease the compression ratio by a desired amount. As will be understood by one having ordinary skill in the art, the compression ratio refers to the ratio of the data in the uncompressed signal to the data in the compressed signal. In addition, the signal compression ratio is inversely related to the output bit rate, i.e., as the compression ratio decreases, the output bit rate increases. Thus, if the compression ratio is too high and would cause the watermark to be unrecoverable, then the transcoder 22 may decrease the compression ratio of the output media codec 40 by increasing the output bit rate of the output media codec 40 by a desired amount such as, for example, 30 Kb/sec (block 154). After increasing the output bit rate, the transcoder 22 determines whether the new, increased output bit rate exceeds the maximum allowable bit rate supported by the home network 10 (block 156). As will be appreciated by one having ordinary skill in the art, if the bit rate is higher than that supported by the home network 10, then the home network 10 may not be able to carry the signal without causing it to be distorted. As a result, the transcoder 22 decreases the output bit rate to a value that is within the bandwidth of the home network 10 (block 158), and the submethod 142 returns to the block 148 at which a new audio watermark is created and inserted into the media content, as described above. As is also described above, when the transcoder 22 causes the output media codec 40 to create a new audio watermark the transcoder 22 may also perform all or a portion of the submethod 130 shown in FIG. 6, for causing correlation information to be created and transmitted to the meter 18 for reporting to the remote data collection facility (not shown).


If, the output bit rate does not exceed the maximum rate supported by the home network 10 (block 156), then the submethod 142 returns to the block 150, and the blocks subsequent thereto, at which the transcoder 22 again tests to determine whether the codec compression ratio is suitable for transmission of the watermark, as described above.


Referring also to FIG. 7B, which includes a continuation of the flow chart of FIG. 7A as indicated by the alignment points A, after performing the submethod 142 for detecting the presence of an audio watermark sensor, a submethod 162 for determining whether the meter has a sensor capable of sensing a video watermark may be performed. The submethod 162 for detecting the presence of a video watermark sensor begins when the transcoder 22 queries the meter 18 to determine whether it includes a video watermark sensor (block 164). If the meter 18 responds to the query in the negative, i.e., the meter 18 does not have a video watermark sensor, then the transcoder 22 proceeds to additional submethods for testing whether the meter includes one or more other types of sensors as described in greater detail below with reference to FIG. 7C. If instead, in response to the query performed at the block 164, the meter 18 responds in the positive, i.e., the meter 18 does have a video watermark sensor, then the submethod 162 continues with the transcoder 22 querying the meter 18 to determine whether a video watermark has been detected in the media content supplied to the transcoder 22 for transcoding (block 166). If a video watermark is not detected in the media content, then the submethod 162 may continue at a block 168 at which the transcoder 22 creates a new video watermark and causes the new video watermark to be inserted into the media content. As discussed above, the capabilities and functionality of a standard audio/video watermark codec are well known in the art and are not described further herein. The transcoder 22 may also cause all or a portion of the submethod 130, described with respect to FIG. 6, to be performed thereby causing correlation information to be generated and transmitted to the measurement collection unit 20 and/or a remote data collection facility (not shown) where it may be used to correlate the watermark to the program or other content it represents.


If a video watermark has been detected at the block 166, then the transcoder 22 determines whether the video watermark, if inserted into the signal to be output by the transcoder 22, will survive the compression performed by the output media codec 40 (block 170). Specifically, the output media codec 40 is adapted to compress the media content signal having the inserted watermark before the signal is transmitted via the home network 10. The output media codec 40 compresses the signal by suppressing one or more of the signal frequencies. However, watermarks are created by modulating a particular set of signal frequencies in a manner such that the modulated frequencies uniquely represent a particular program or other media content. Thus, the compression performed by the output media codec, may cause one or more of the frequencies modulated to create the watermark to be suppressed thereby causing the video watermark to be unrecoverable by the meter 18. The transcoder 22 may be adapted to perform a variety of methods for determining whether the watermark will survive compression by the output media codec 40. For example, the transcoder 22 may cause the output media codec 40 to insert the watermark into the media content and the resulting signal may be processed by the transcoder 22, in much the same way that an input signal is processed, to determine whether the video watermark is recoverable. In another embodiment, the transcoder 22 may be pre-programmed with information pertaining to signal compression ratios that the watermark will be able to withstand/survive. Specifically, before inserting a particular watermark into a media content signal, the watermark may be tested to determine a range of suitable signal compression ratios, i.e., compression ratios that the watermark will survive. These suitable ratios may then be provided to the manufacturers/developers of the transcoder 22 and used to pre-program the transcoder 22 so that when watermarks are encountered, the transcoder 22 may use the pre-programmed information to compare to the compression ratio of the output media codec 40 to determine whether the compression ratio used by the output media codec 40 is suitable for inserting and transmitting the watermark in a recoverable, distortion-free manner. In a still further embodiment, information about suitable compression ratios may be transmitted with the signal containing the watermark and extracted from the signal by the transcoder 22 for use in determining whether the watermark will survive the compression ratio used by the output media codec 40. If the video watermark will survive compression, the submethod 162 causes the output media codec 40 to insert the watermark (block 172) into the media content (if it is not already inserted) after which the submethod 162 is complete and the method 140 continues at another submethod described below with respect to FIG. 7B and FIG. 7C.


If, at the block 170, the transcoder 22 determines that the compression ratio of the output media codec 40 is not suitable, i.e., will cause the watermark to be unrecoverable, then the transcoder 22 may decrease the compression ratio by a desired amount. As will be understood by one having ordinary skill in the art, the compression ratio refers to the ratio of the quantity of data in the uncompressed signal to the quantity of data in the compressed signal. In addition, the signal compression ratio is inversely related to the output bit rate, i.e., as the compression ratio decreases, the output bit rate increases. Thus, if the compression ratio is too high and would cause the watermark to be unrecoverable, then the transcoder 22 may decrease the compression ratio of the output media codec 40 by increasing the output bit rate of the output media codec 40 by a desired amount such as, for example, 30 Kb/sec (block 174). After increasing the output bit rate, the transcoder 22 determines whether the new, increased output bit rate exceeds the maximum allowable bit rate supported by the home network 10 (block 176). As will be appreciated by one having ordinary skill in the art, if the bit rate is higher than that supported by the home network 10, then the home network 10 may not be able to carry the signal without causing it to be distorted. As a result, the transcoder 22 decreases the output bit rate to a value that is within the bandwidth of the home network 10 (block 178), and the submethod 162 returns to the block 168 at which a new video watermark is created for insertion into the media content, as described above. As is also described above, the block 168 at which the transcoder 22 causes the output media codec 40 to create a new video watermark may also include all or a portion of the submethod 130 shown in FIG. 6, for causing correlation information to be created and transmitted to the meter 18 for reporting to the remote data collection facility (not shown).


If, at the block 176, the output bit rate does not exceed the maximum rate supported by the home network 10, then the submethod 162 returns to the block 170, and the blocks subsequent thereto, at which the transcoder 22 again tests to determine whether the codec compression ratio is suitable for transmission of the watermark, as described above.


In addition to testing for the presence of an audio watermark sensor and a video watermark sensor, the transcoder 22 may be adapted to query the 18 meter to determine whether it includes other types of sensors as well, such as digital sensors, database sensors and/or software sensors. Specifically, with reference also to FIG. 7C which includes a continuation of the flow chart of FIGS. 7A and 7B as indicated by the alignment points B, such a submethod 200 may begin with the transcoder 22 querying the meter 18 to determine whether it includes a sensor capable of parsing a digital bitstream to decode metadata embedded in the stream (block 202). If the meter 18 responds that it does include such a digital sensor, then the transcoder 22 causes the output media codec 40 to encode the metadata received with the media content into the desired format (block 204) and to digitally insert the encoded metadata into the bitstream of the transcoded media content to be output by the transcoder (block 206).


If the meter 18 indicates that it does not include a digital sensor (block 202), then the method continues at a submethod 210 at which the transcoder 22 queries the meter 18 to determine whether it includes a database sensor, i.e., a sensor that can identify when media content is being read from a media database (block 212). If such a database sensor is present, then the submethod 200 continues with the transcoder 22 transcoding the metadata, embedding the transcoded metadata in the transcoded media content and then causing the media content to be stored in a media database (block 214).


If, at the block 212, a database meter is not detected, then the method continues at a submethod 216 for determining whether the meter 18 includes a software sensor adapted to extract metadata from a media consumption device using software APIs associated with the consumption device, e.g. DASE/MHP API (block 218). If such a software sensor is detected, then the transcoder 22 causes the output media codec 40 to format the metadata in a manner suitable for extraction by the software sensor and to embed the metadata into the transcoded media signal that is output by the transcoder 22 (block 220).


If a software sensor is not detected, then the transcoder 22 may query the home network 10 for the presence of other meters 18 that are configured to meter the consumption device to which the transcoder 22 supplies transcoded media content (block 222). If another meter 18 is detected, then the transcoder 22 may return to the beginning of the method 140 and cause it to be performed again with respect to the newly detected meter 18. If another meter is not detected, then the transcoder 22 may forego repeating the method 140.


As described, the method 140 for querying a meter 18 to determine the sensing capabilities of the meter 18 actually comprises a set of sub-methods each adapted to query the meter 18 for a specific type of sensor. Although the sub-methods are described as being performed in a specific order, the sub-methods may actually be performed in any desired order. Likewise, the submethods may be performed in parallel instead of serially. In addition, the transcoder 22 need not be configured to perform all of the submethods of FIGS. 7A-7C but may instead be configured to perform any combination of a subset of these sub-methods. Moreover, the transcoder 22 may be configured to perform any number of additional sub-methods as necessary to determine the sensing capabilities of the meter 18 so that the metadata may be formatted accordingly.


The querying methods described above need not be performed at all if the transcoder 22 is pre-programmed with information about the sensing capabilities of the meter 18. In such an embodiment, the meters 18 need not be capable of communicating via the home network 10 and need not even be coupled to the home network 10. Instead, the meters 18 need only be capable of metering consumption at a media consumption device 12.


Depending on whether frequent changes to the configuration of the home network 10 are anticipated, the transcoder 22 may be adapted to perform the method 140 every time new media content is received or only a single time, e.g., upon installing the transcoder 22 in the home network 10. Alternatively, the transcoder 22 may be adapted to query for sensor types only after the home network 10 has been reconfigured. Of course, all or portions of the submethods 142 and 162 for detecting the presence of an audio watermark sensor and a video watermark sensor, respectively, may need to be performed every time media content is supplied since at least portions of the submethods 142 and 162 operate to test for the presence of an audio watermark or a video watermark supplied with the media content.


The submethods of FIGS. 7A, 7B and 7C are described as including blocks at which the transcoder 22 determines whether a particular type of sensor is detected and, if such a sensor type is detected, then causes the metadata to be transcoded in a manner suitable for detection by that sensor type, and then proceeds to perform tests for a different type of sensor. However, the submethods may instead be structured such that the transcoder first detects the presence (or absence) of each type of sensor and then, after each sensor type has been detected, causes the metadata to transcoded in a manner suitable for the detected sensor types. In addition, the metadata may be transcoded into more than one metadata format thereby enabling detection by more than one type of sensor.


As described herein, the media content and its corresponding metadata received at the transcoder 22 are stored in the memory device 38 and accessed by the various transcoder components for purposes of transcoding the media content and metadata. Instead, the transcoder 22 may include a plurality of memory devices arranged as registers associated with the various components of the transcoder 22 between which the data may be transferred. Alternatively, the transcoders 22 may be adapted to process and store the media content and metadata in any desired manner.


The home network 10, although described as being disposed within a home residence, may instead be disposed at any type of location and may be configured to enable communication between network devices located at any number of different locations. For example, the home network 10 may be installed in a place of business or at any public location. Any network that enables communication between multiple media consumption devices is sufficient to qualify as a “home network,” as that term is used herein.


As will be appreciated by one having ordinary skill in the art, if the consumption device 12 is metered using only a single meter 18 and that single meter 18 is limited to signature sensing only, then any metadata extracted from the media content need not be transcoded at the trancoder 22 because signature metering involves capturing signal characteristic information, i.e., signature information, and does not involve the extraction of codes. Thus, the querying method 60 of FIGS. 7A-7C does not illustrate querying for the presence of a signature sensor. Or, as is more often the case, a consumption device 12 may be metered using multiple meters, one of which is capable of sensing signatures. In a system configured in this manner, metadata transcoding will likely be required as signature sensing is more often used as a back up sensing mechanism instead of a primary sensing mechanism. Thus, the transcoders 22 will not typically be configured to forego metadata transcoding functions upon the detection of a meter 18 having signature sensing capabilities.


The transcoder 22 may additionally be adapted to query the media content consumption device 12 that supplies media content to the transcoder 22 for identification purposes such as, for example, device type and/or model information and the transcoder 22 may then transmit this identifying information to the meter 18 configured to measure consumption at the media consumption device to which the transcoder delivers the transcoded media content.


While the present invention has been described with respect to several embodiments, it is to be understood that the invention is not limited to these disclosed embodiments. To the contrary, the invention is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the appended claims. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.

Claims
  • 1. An apparatus to meter a media device, the apparatus comprising: means for accessing first metadata accompanying media to be presented by the media device, the first metadata conveying a first identifier, the first metadata in a first format not detectable by a meter collecting audience measurement data associated with the media device;means for transcoding the first metadata into second metadata that is to convey the first identifier, the second metadata to be associated with the media, the second metadata being transcoded to have a second format that is different from the first format, the second metadata having the second format being detectable by the meter and supported by an application programming interface (API) provided by the media device, the API being accessible by the meter; andmeans for communicating the second metadata having the second format in a second transmission for receipt by the meter via the API provided by the media device, the second transmission being separate from a first transmission communicating the media and the first metadata having the first format, but the first transmission not communicating the second metadata having the second format different from the first format, to the media device.
  • 2. The apparatus of claim 1, wherein the first metadata corresponds to at least one of an audio watermark or a video watermark.
  • 3. The apparatus of claim 2, wherein the second metadata corresponds to digital data formatted according to a data communications protocol.
  • 4. The apparatus of claim 3, wherein the digital data includes a plurality of data fields.
  • 5. The apparatus of claim 4, further including means for inserting first information in a first one of the plurality of data fields, the first information corresponding to second information conveyed by the first metadata.
  • 6. The apparatus of claim 5, wherein the first information corresponds to the first identifier conveyed by the first metadata.
  • 7. The apparatus of claim 1, wherein the first transmission is communicated via a first communication medium, and the second transmission is communicated via a wireless communication medium different from the first communication medium.
  • 8. An apparatus to transcode metadata included in a media signal conveyed via a network from a first media format associated with a first media consumption device to a second media format detectable by a media metering device configured to collect audience measurement data associated with a second media consumption device, the apparatus comprising: means for extracting the metadata from the media signal to form extracted metadata;means for querying the media metering device via the network to determine a capability of the media metering device to monitor a media output signal from the second media consumption device, the second media consumption device communicatively coupled to the network;means for identifying the second media format based on the determined capability of the media metering device, the second media format detectable by the media metering device;means for converting the extracted metadata from the first media format associated with the first media consumption device to the identified second media format detectable by the media metering device to form converted media information, the first media consumption device communicatively coupled to the network; andmeans for sending the converted media information to at least one of the second media consumption device or the media metering device via the network.
  • 9. The apparatus of claim 8, wherein the means for identifying the second media format includes means for identifying at least one of an audio watermark sensor, a video watermark sensor, a digital bitstream sensor, a database sensor, or a software sensor associated with the media metering device.
  • 10. The apparatus of claim 8, wherein the means for converting the extracted metadata from the first media format to the second media format to form the converted media information includes: means for detecting a watermark associated with the media signal;means for identifying a signal compression ratio associated with the watermark; andmeans for modifying the signal compression ratio based on the second media format.
  • 11. The apparatus of claim 10, wherein the means for modifying the signal compression ratio based on the second media format includes: means for comparing an output bit rate associated with the signal compression ratio to a network bit rate associated with the network; andmeans for changing the output bit rate based on the network bit rate.
  • 12. The apparatus of claim 8, wherein the means for converting the extracted metadata from the first media format to the second media format to form the converted media information includes: means for generating a watermark based on the second media format; andmeans for inserting the watermark in the converted media information.
  • 13. The apparatus of claim 8, wherein the means for converting the extracted metadata from the first media format to the second media format to form the converted media information includes: means for encoding the extracted metadata in the second media format; andmeans for digitally inserting encoded metadata into a bitstream associated with the converted media information.
  • 14. An apparatus to meter media presented by a media consumption device, the apparatus comprising: means for detecting first metadata accompanying the media presented by the media consumption device, the first metadata being detected in a media signal output from the media consumption device;means for receiving second metadata from a transcoder, the second metadata associated with the media presented by the media consumption device, the second metadata being received in a second signal different from the media signal output from the media consumption device, the second metadata comprising third metadata that has been transcoded to form the second metadata, wherein the third metadata is not detectable by the apparatus, and the second metadata corresponds to the third metadata after being transcoded into a format capable of being received and processed by the apparatus; andmeans for processing the first metadata and the second metadata to meter the media presented by the media consumption device.
  • 15. The apparatus of claim 14, wherein the first metadata includes an audio code and the second metadata includes data received wirelessly from the transcoder.
  • 16. The apparatus of claim 14, wherein the third metadata includes a timestamp not detectable by the apparatus, and the second metadata includes the timestamp after being transcoded into a format capable of being received wirelessly by the apparatus.
  • 17. The apparatus of claim 14, wherein the means for processing the first metadata and the second metadata includes means for comparing the first metadata and the second metadata for consistency.
  • 18. The apparatus of claim 14, wherein the means for processing the first metadata and the second metadata includes means for supplementing the first metadata with the second metadata.
  • 19. The apparatus of claim 14, wherein the apparatus is portable, and the second metadata is received wirelessly from the transcoder.
  • 20. The apparatus of claim 14, further including means for receiving a query from the transcoder for information regarding a metadata detection capability of the apparatus, the second metadata determined based on a response to the query.
RELATED APPLICATIONS

This patent is a continuation of U.S. patent application Ser. No. 14/089,279, entitled “Methods and Apparatus for Transcoding Metadata” and filed on Nov. 25, 2013, which is a continuation of U.S. patent application Ser. No. 12/890,216, entitled “Methods and Apparatus for Transcoding Metadata” and filed on Sep. 24, 2010, which is a continuation of U.S. patent application Ser. No. 10/540,611, entitled “Methods and Apparatus for Transcoding Metadata” and filed on Jun. 24, 2005, which is a U.S. national stage application of International Patent Application Serial No. PCT/US03/14970, entitled “Methods and Apparatus for Transcoding Metadata” and filed on May 13, 2003, which claims priority from U.S. Provisional Application Ser. No. 60/436,714, entitled “Transcoding of Metadata” and filed on Dec. 27, 2002. U.S. patent application Ser. Nos. 10/540,611, 12/890,216 and 14/089,279, International Patent Application Serial No. PCT/US03/14970 and U.S. Provisional Application Ser. No. 60/436,714 are hereby incorporated by reference in their respective entireties.

US Referenced Citations (429)
Number Name Date Kind
4230990 Lert, Jr. et al. Oct 1980 A
4647974 Butler et al. Mar 1987 A
4677466 Lert, Jr. et al. Jun 1987 A
4697209 Kiewit et al. Sep 1987 A
4745468 Von Kohorn May 1988 A
4876592 Von Kohorn Oct 1989 A
4876736 Kiewit Oct 1989 A
4926255 Von Kohorn May 1990 A
4973952 Malec et al. Nov 1990 A
5003591 Kauffman et al. Mar 1991 A
5019899 Boles et al. May 1991 A
5023929 Call Jun 1991 A
5034807 Von Kohorn Jul 1991 A
5057915 Von Kohorn Oct 1991 A
5081680 Bennett Jan 1992 A
5128752 Von Kohorn Jul 1992 A
5227874 Von Kohorn Jul 1993 A
5249044 Von Kohorn Sep 1993 A
5283734 Von Kohorn Feb 1994 A
5331544 Lu et al. Jul 1994 A
5401946 Weinblatt Mar 1995 A
5425100 Thomas et al. Jun 1995 A
5481294 Thomas et al. Jan 1996 A
5512933 Wheatley et al. Apr 1996 A
5524195 Clanton, III et al. Jun 1996 A
5543856 Rosser et al. Aug 1996 A
5559716 Gaalswyk Sep 1996 A
5574962 Fardeau et al. Nov 1996 A
5579124 Aijala et al. Nov 1996 A
5594934 Lu et al. Jan 1997 A
5629739 Dougherty May 1997 A
5659366 Kerman Aug 1997 A
5666293 Metz et al. Sep 1997 A
5719634 Keely et al. Feb 1998 A
5734413 Lappington et al. Mar 1998 A
5740035 Cohen et al. Apr 1998 A
5815671 Morrison Sep 1998 A
5841978 Rhoads Nov 1998 A
5848155 Cox Dec 1998 A
5850249 Massetti et al. Dec 1998 A
5872588 Aras et al. Feb 1999 A
5880789 Inaba Mar 1999 A
5893067 Bender et al. Apr 1999 A
5910987 Ginter Jun 1999 A
5918223 Blum et al. Jun 1999 A
5930369 Cox et al. Jul 1999 A
5933789 Byun et al. Aug 1999 A
5956716 Kenner et al. Sep 1999 A
5966120 Arazi et al. Oct 1999 A
5974299 Massetti Oct 1999 A
5974396 Anderson et al. Oct 1999 A
5987855 Dey et al. Nov 1999 A
6034722 Viney et al. Mar 2000 A
6049830 Saib Apr 2000 A
6055573 Gardenswartz et al. Apr 2000 A
6061793 Tewfik et al. May 2000 A
6088659 Kelley et al. Jul 2000 A
6108637 Blumenau Aug 2000 A
6115654 Eid et al. Sep 2000 A
6154209 Naughton et al. Nov 2000 A
6208735 Cox et al. Mar 2001 B1
6216129 Eldering Apr 2001 B1
6282573 Darago et al. Aug 2001 B1
6286036 Rhoads Sep 2001 B1
6286140 Ivanyi Sep 2001 B1
6298348 Eldering Oct 2001 B1
6308327 Liu et al. Oct 2001 B1
6310956 Morito Oct 2001 B1
6327619 Blumenau Dec 2001 B1
6331876 Koster et al. Dec 2001 B1
6335736 Wagner et al. Jan 2002 B1
6363159 Rhoads Mar 2002 B1
6363488 Ginter et al. Mar 2002 B1
6377993 Brandt et al. Apr 2002 B1
6389055 August et al. May 2002 B1
6400827 Rhoads Jun 2002 B1
6411725 Rhoads Jun 2002 B1
6421445 Jensen et al. Jul 2002 B1
6442285 Rhoads et al. Aug 2002 B2
6463445 Suzuki Oct 2002 B1
6487564 Asai et al. Nov 2002 B1
6505160 Levy et al. Jan 2003 B1
6512836 Xie et al. Jan 2003 B1
6513014 Walker et al. Jan 2003 B1
6519509 Nierlich et al. Feb 2003 B1
6522771 Rhoads Feb 2003 B2
6539095 Rhoads Mar 2003 B1
6546556 Kataoka et al. Apr 2003 B1
6553178 Abecassis Apr 2003 B2
6642966 Limaye Nov 2003 B1
6647269 Hendrey et al. Nov 2003 B2
6651253 Duclkiewicz et al. Nov 2003 B2
6654480 Rhoads Nov 2003 B2
6665873 Van Gestel et al. Dec 2003 B1
6671732 Weiner Dec 2003 B1
6675383 Wheeler et al. Jan 2004 B1
6683966 Tian Jan 2004 B1
6710815 Billmaier et al. Mar 2004 B1
6714683 Tian et al. Mar 2004 B1
6714684 Yamaguchi et al. Mar 2004 B2
6750985 Rhoads Jun 2004 B2
6766523 Herley Jul 2004 B2
6795972 Rovira Sep 2004 B2
6804379 Rhoads Oct 2004 B2
6829368 Meyer et al. Dec 2004 B2
6853634 Davies et al. Feb 2005 B1
6871323 Wagner et al. Mar 2005 B2
6873688 Aamio Mar 2005 B1
6941275 Swierczek Sep 2005 B1
6956575 Nakazawa et al. Oct 2005 B2
6968315 Nakisa Nov 2005 B1
6968564 Srinivasan Nov 2005 B1
6970886 Conwell et al. Nov 2005 B1
6996213 De Jong Feb 2006 B1
7003731 Rhoads et al. Feb 2006 B1
7007166 Moskowitz et al. Feb 2006 B1
7032178 McKnight et al. Apr 2006 B1
7050603 Rhoads et al. May 2006 B2
7051086 Rhoads et al. May 2006 B2
7058697 Rhoads Jun 2006 B2
7082434 Gosselin Jul 2006 B2
7095871 Jones et al. Aug 2006 B2
7143949 Hannigan Dec 2006 B1
7158943 van der Riet Jan 2007 B2
7171018 Rhoads et al. Jan 2007 B2
7174293 Kenyon et al. Feb 2007 B2
7181042 Tian Feb 2007 B2
7185201 Rhoads et al. Feb 2007 B2
7194752 Kenyon et al. Mar 2007 B1
7197156 Levy Mar 2007 B1
7215280 Percy et al. May 2007 B1
7221405 Basson et al. May 2007 B2
7227972 Brundage et al. Jun 2007 B2
7254249 Rhoads et al. Aug 2007 B2
7273978 Uhle Sep 2007 B2
7317716 Boni et al. Jan 2008 B1
7328153 Wells et al. Feb 2008 B2
7346512 Li-Chun Wang et al. Mar 2008 B2
7356700 Noridomi et al. Apr 2008 B2
7363278 Schmelzer et al. Apr 2008 B2
7369678 Rhoads May 2008 B2
7421723 Harkness et al. Sep 2008 B2
7443292 Jensen et al. Oct 2008 B2
7451317 Oh et al. Nov 2008 B2
7463143 Forr et al. Dec 2008 B2
7519658 Anglin et al. Apr 2009 B1
7592908 Zhang et al. Sep 2009 B2
7623823 Zito et al. Nov 2009 B2
7643090 Ramaswamy et al. Jan 2010 B2
7689823 Shen et al. Mar 2010 B2
7712114 Ramaswamy May 2010 B2
7761465 Nonaka et al. Jul 2010 B1
7827312 Ramaswamy et al. Nov 2010 B2
7962934 Eldering et al. Jun 2011 B1
8065703 Wilson et al. Nov 2011 B2
8103879 Levy et al. Jan 2012 B2
8176322 Lee et al. May 2012 B2
8179475 Sandrew May 2012 B2
8451868 Johnson et al. May 2013 B2
8578272 Pantos et al. Nov 2013 B2
8645373 Knight et al. Feb 2014 B2
8706685 Smith et al. Apr 2014 B1
8732185 Lynn et al. May 2014 B1
8839338 Eyer Sep 2014 B2
8856869 Brinskelle Oct 2014 B1
8990844 Oh et al. Mar 2015 B2
9197421 Besehanic Nov 2015 B2
9209978 Besehanic Dec 2015 B2
9210208 Ramaswamy et al. Dec 2015 B2
9515904 Besehanic et al. Dec 2016 B2
9609034 Ramaswamy et al. Mar 2017 B2
20010031066 Meyer Oct 2001 A1
20010044851 Rothman et al. Nov 2001 A1
20010044899 Levy Nov 2001 A1
20020001395 Davis et al. Jan 2002 A1
20020012443 Rhoads et al. Jan 2002 A1
20020016969 Kimble Feb 2002 A1
20020032734 Rhoads Mar 2002 A1
20020033842 Zetts Mar 2002 A1
20020052885 Levy May 2002 A1
20020053078 Holtz et al. May 2002 A1
20020056094 Dureau May 2002 A1
20020059218 August May 2002 A1
20020062382 Rhoads et al. May 2002 A1
20020069037 Hendrickson et al. Jun 2002 A1
20020083324 Hirai Jun 2002 A1
20020091991 Castro Jul 2002 A1
20020108125 Joao Aug 2002 A1
20020111934 Narayan Aug 2002 A1
20020112002 Abato Aug 2002 A1
20020120925 Logan Aug 2002 A1
20020124246 Kaminsky et al. Sep 2002 A1
20020133705 Tagashira et al. Sep 2002 A1
20020138852 Reynolds et al. Sep 2002 A1
20020144262 Plotnick et al. Oct 2002 A1
20020144273 Reto Oct 2002 A1
20020150247 Linnartz et al. Oct 2002 A1
20020157112 Kuhn Oct 2002 A1
20020162118 Levy et al. Oct 2002 A1
20020174425 Markel et al. Nov 2002 A1
20020194592 Tsuchida et al. Dec 2002 A1
20020197063 Cho Dec 2002 A1
20030021441 Levy et al. Jan 2003 A1
20030039465 Bjorgan et al. Feb 2003 A1
20030055949 Coulombe et al. Mar 2003 A1
20030079131 Reefman Apr 2003 A1
20030088674 Ullman et al. May 2003 A1
20030093810 Taniguchi May 2003 A1
20030105870 Baum Jun 2003 A1
20030108200 Sako Jun 2003 A1
20030115598 Pantoja Jun 2003 A1
20030149890 Shen et al. Aug 2003 A1
20030177488 Smith et al. Sep 2003 A1
20030185232 Moore et al. Oct 2003 A1
20030212710 Guy Nov 2003 A1
20040003394 Ramaswamy Jan 2004 A1
20040009763 Stone et al. Jan 2004 A1
20040019690 Cardno et al. Jan 2004 A1
20040025181 Addington et al. Feb 2004 A1
20040037271 Liscano et al. Feb 2004 A1
20040038692 Muzaffar Feb 2004 A1
20040064319 Neuhauser Apr 2004 A1
20040073916 Petrovic et al. Apr 2004 A1
20040073951 Bac et al. Apr 2004 A1
20040088347 Yeager et al. May 2004 A1
20040088556 Weirauch May 2004 A1
20040125125 Levy Jul 2004 A1
20040128514 Rhoads Jul 2004 A1
20040137929 Jones et al. Jul 2004 A1
20040146161 De Jong Jul 2004 A1
20040153649 Rhoads et al. Aug 2004 A1
20040156489 Vishik et al. Aug 2004 A1
20040199387 Wang et al. Oct 2004 A1
20040254887 Jacoby Dec 2004 A1
20050028189 Heine et al. Feb 2005 A1
20050033758 Baxter Feb 2005 A1
20050058319 Rhoads et al. Mar 2005 A1
20050086682 Burges et al. Apr 2005 A1
20050120113 Bunch et al. Jun 2005 A1
20050144004 Bennett et al. Jun 2005 A1
20050152287 Yokomitsu et al. Jul 2005 A1
20050177738 Van Der Veen et al. Aug 2005 A1
20050188297 Knight et al. Aug 2005 A1
20050192933 Rhoads et al. Sep 2005 A1
20050193425 Still et al. Sep 2005 A1
20050204379 Yamamori Sep 2005 A1
20050216346 Kusumoto et al. Sep 2005 A1
20050234774 Dupree Oct 2005 A1
20060026431 Campello De Souza Feb 2006 A1
20060031297 Zuidema Feb 2006 A1
20060056625 Nakabayashi et al. Mar 2006 A1
20060059277 Zito et al. Mar 2006 A1
20060062426 Levy et al. Mar 2006 A1
20060095401 Krikorian et al. May 2006 A1
20060107195 Ramaswamy et al. May 2006 A1
20060107302 Zdepski May 2006 A1
20060136564 Ambrose Jun 2006 A1
20060161635 Lamkin et al. Jul 2006 A1
20060167747 Goodman et al. Jul 2006 A1
20060168613 Wood et al. Jul 2006 A1
20060195614 Sena et al. Aug 2006 A1
20060195886 Ashley Aug 2006 A1
20060212705 Thommana et al. Sep 2006 A1
20060221173 Duncan Oct 2006 A1
20060224798 Klein et al. Oct 2006 A1
20060242325 Ramaswamy et al. Oct 2006 A1
20070006250 Croy et al. Jan 2007 A1
20070016918 Alcorn et al. Jan 2007 A1
20070055987 Lu et al. Mar 2007 A1
20070074020 Nishimura Mar 2007 A1
20070083611 Farago et al. Apr 2007 A1
20070110089 Essafi et al. May 2007 A1
20070112837 Houh et al. May 2007 A1
20070118375 Kenyon et al. May 2007 A1
20070118873 Houh et al. May 2007 A1
20070124771 Shvadron May 2007 A1
20070127717 Herre et al. Jun 2007 A1
20070129952 Kenyon et al. Jun 2007 A1
20070133223 Fredley et al. Jun 2007 A1
20070136753 Bovenschulte et al. Jun 2007 A1
20070136777 Hasek et al. Jun 2007 A1
20070149114 Danilenko Jun 2007 A1
20070157262 Ramaswamy et al. Jul 2007 A1
20070162927 Ramaswamy et al. Jul 2007 A1
20070186228 Ramaswamy et al. Aug 2007 A1
20070186288 Peterson et al. Aug 2007 A1
20070198738 Angiolillo et al. Aug 2007 A1
20070201835 Rhoads Aug 2007 A1
20070226760 Neuhauser et al. Sep 2007 A1
20070250901 McIntire et al. Oct 2007 A1
20070274523 Rhoads Nov 2007 A1
20070276925 La Joie et al. Nov 2007 A1
20070276926 La Joie et al. Nov 2007 A1
20070288476 Flanagan, III et al. Dec 2007 A1
20070294057 Crystal et al. Dec 2007 A1
20070294132 Zhang et al. Dec 2007 A1
20070294705 Gopalakrishnan et al. Dec 2007 A1
20070294706 Neuhauser et al. Dec 2007 A1
20080027734 Zhao et al. Jan 2008 A1
20080028223 Rhoads Jan 2008 A1
20080040354 Ray et al. Feb 2008 A1
20080046499 Cabrera et al. Feb 2008 A1
20080059160 Saunders et al. Mar 2008 A1
20080065507 Morrison et al. Mar 2008 A1
20080077956 Morrison et al. Mar 2008 A1
20080082510 Wang et al. Apr 2008 A1
20080082922 Biniak et al. Apr 2008 A1
20080083003 Biniak et al. Apr 2008 A1
20080104624 Narasimhan et al. May 2008 A1
20080120661 Ludvig et al. May 2008 A1
20080126420 Wright et al. May 2008 A1
20080133223 Son et al. Jun 2008 A1
20080133504 Messer et al. Jun 2008 A1
20080134232 Rhoads Jun 2008 A1
20080139182 Levy et al. Jun 2008 A1
20080140573 Levy et al. Jun 2008 A1
20080168503 Sparrell Jul 2008 A1
20080184132 Zato Jul 2008 A1
20080200999 Hakansson Aug 2008 A1
20080209491 Hasek Aug 2008 A1
20080219496 Tewfik et al. Sep 2008 A1
20080219637 Sandrew Sep 2008 A1
20080235077 Harkness et al. Sep 2008 A1
20080235209 Rathod et al. Sep 2008 A1
20080240490 Finkelstein et al. Oct 2008 A1
20080249961 Harkness et al. Oct 2008 A1
20080263579 Mears et al. Oct 2008 A1
20080294487 Nasser Nov 2008 A1
20080310629 Van Der Veen et al. Dec 2008 A1
20090007169 Headley et al. Jan 2009 A1
20090015599 Bennett et al. Jan 2009 A1
20090070408 White Mar 2009 A1
20090083417 Hughes et al. Mar 2009 A1
20090086812 Ducharme Apr 2009 A1
20090103887 Choi et al. Apr 2009 A1
20090119723 Tinsman May 2009 A1
20090129588 Takakusu et al. May 2009 A1
20090133093 Hodge May 2009 A1
20090150553 Collart et al. Jun 2009 A1
20090157731 Zigler et al. Jun 2009 A1
20090158318 Levy Jun 2009 A1
20090164378 West et al. Jun 2009 A1
20090164564 Willis Jun 2009 A1
20090210892 Ramaswamy et al. Aug 2009 A1
20090228492 Valdez et al. Sep 2009 A1
20090248886 Tan et al. Oct 2009 A1
20090259325 Topchy et al. Oct 2009 A1
20090265214 Jobs et al. Oct 2009 A1
20090276313 Wilhelm Nov 2009 A1
20090305680 Swift et al. Dec 2009 A1
20090307061 Monighetti et al. Dec 2009 A1
20090307084 Monighetti et al. Dec 2009 A1
20100008586 Meyer et al. Jan 2010 A1
20100009722 Levy et al. Jan 2010 A1
20100023405 Liu Jan 2010 A1
20100083299 Nelson et al. Apr 2010 A1
20100088583 Schachter Apr 2010 A1
20100094897 Sumrall et al. Apr 2010 A1
20100121936 Liu et al. May 2010 A1
20100135638 Mio Jun 2010 A1
20100169904 Nielsen et al. Jul 2010 A1
20100172540 Davis et al. Jul 2010 A1
20100174774 Kern et al. Jul 2010 A1
20100226526 Modro et al. Sep 2010 A1
20100241963 Kulis et al. Sep 2010 A1
20100246955 Wright et al. Sep 2010 A1
20100262711 Bouazizi Oct 2010 A1
20100280641 Harkness et al. Nov 2010 A1
20100306257 Levy Dec 2010 A1
20100318600 Furbeck Dec 2010 A1
20110016231 Ramaswamy et al. Jan 2011 A1
20110030031 Lussier et al. Feb 2011 A1
20110055314 Rosenstein et al. Mar 2011 A1
20110066437 Luff Mar 2011 A1
20110078721 Wang et al. Mar 2011 A1
20110088053 Lee Apr 2011 A1
20110123062 Hilu May 2011 A1
20110145246 Prager et al. Jun 2011 A1
20110145581 Malhotra et al. Jun 2011 A1
20110154185 Kern et al. Jun 2011 A1
20110157475 Wright et al. Jun 2011 A1
20110173200 Yang et al. Jul 2011 A1
20110196921 Sylthe Aug 2011 A1
20110252118 Pantos et al. Oct 2011 A1
20110320287 Holt et al. Dec 2011 A1
20110321003 Doig et al. Dec 2011 A1
20120023516 Wolinsky et al. Jan 2012 A1
20120036350 Kuno et al. Feb 2012 A1
20120045054 Main et al. Feb 2012 A1
20120096546 Dilley et al. Apr 2012 A1
20120124605 Praden May 2012 A1
20120137015 Sun May 2012 A1
20120209949 Deliyannis et al. Aug 2012 A1
20120239809 Mazumdar et al. Sep 2012 A1
20120265735 McMillan et al. Oct 2012 A1
20120284804 Lindquist et al. Nov 2012 A1
20120308071 Ramsdell et al. Dec 2012 A1
20120311126 Jadallah et al. Dec 2012 A1
20130007298 Ramaswamy et al. Jan 2013 A1
20130007794 Besehanic et al. Jan 2013 A1
20130054972 Thorwirth Feb 2013 A1
20130061275 Seo et al. Mar 2013 A1
20130073960 Eppolito et al. Mar 2013 A1
20130097285 van Zwol et al. Apr 2013 A1
20130124747 Harrang et al. May 2013 A1
20130166868 Jarnikov et al. Jun 2013 A1
20130202150 Sinha et al. Aug 2013 A1
20130205319 Sinha et al. Aug 2013 A1
20130226942 Denoual et al. Aug 2013 A1
20130231931 Kulis et al. Sep 2013 A1
20130247078 Nikankin et al. Sep 2013 A1
20130268623 Besehanic et al. Oct 2013 A1
20130268630 Besehanic et al. Oct 2013 A1
20130290508 Besehanic et al. Oct 2013 A1
20130291001 Besehanic et al. Oct 2013 A1
20130297410 Oh et al. Nov 2013 A1
20130297737 Wajs et al. Nov 2013 A1
20130311776 Besehanic Nov 2013 A1
20130311780 Besehanic Nov 2013 A1
20140082220 Ramaswamy et al. Mar 2014 A1
20140105392 Robert et al. Apr 2014 A1
20140229629 Besehanic Aug 2014 A1
20140229970 Besehanic Aug 2014 A1
20140244828 Besehanic Aug 2014 A1
20140298365 Matsubara et al. Oct 2014 A1
20140301386 Harrenstien et al. Oct 2014 A1
20150222939 Gallant et al. Aug 2015 A1
20160043916 Ramaswamy et al. Feb 2016 A1
20160127466 Albrecht et al. May 2016 A1
Foreign Referenced Citations (79)
Number Date Country
8976601 Feb 2002 AU
9298201 Apr 2002 AU
2003230993 Nov 2003 AU
2012272872 Aug 2015 AU
112901 Jun 2003 BR
309598 Feb 2005 BR
2483104 Apr 2003 CA
1457600 Nov 2003 CN
1592906 Mar 2005 CN
1647160 Jul 2005 CN
101115124 Jan 2008 CN
769749 Apr 1997 EP
1176826 Jan 2002 EP
1349370 Oct 2003 EP
1406403 Apr 2004 EP
1307833 Jun 2006 EP
1745464 Oct 2007 EP
1853026 Nov 2007 EP
1704695 Feb 2008 EP
1504445 Aug 2008 EP
2002247610 Aug 2002 JP
2003524199 Aug 2003 JP
2004320752 Nov 2004 JP
95027349 Oct 1995 WO
97002672 Jan 1997 WO
0004662 Jan 2000 WO
0019699 Apr 2000 WO
01019088 Mar 2001 WO
01024027 Apr 2001 WO
01031497 May 2001 WO
01040963 Jun 2001 WO
01046782 Jun 2001 WO
2001046782 Jun 2001 WO
01053922 Jul 2001 WO
01075743 Oct 2001 WO
01091109 Nov 2001 WO
02005517 Jan 2002 WO
02011123 Feb 2002 WO
02015081 Feb 2002 WO
02017591 Feb 2002 WO
02019625 Mar 2002 WO
02027600 Apr 2002 WO
02037381 May 2002 WO
02045034 Jun 2002 WO
02061652 Aug 2002 WO
02065305 Aug 2002 WO
02065318 Aug 2002 WO
02069121 Sep 2002 WO
03009277 Jan 2003 WO
03091990 Nov 2003 WO
03094499 Nov 2003 WO
03096337 Nov 2003 WO
2004010352 Jan 2004 WO
2004040416 May 2004 WO
2004040475 May 2004 WO
2004061699 Jul 2004 WO
2005025217 Mar 2005 WO
2005064885 Jul 2005 WO
2005101243 Oct 2005 WO
2005111998 Nov 2005 WO
2006012241 Feb 2006 WO
2006025797 Mar 2006 WO
2006203639 Sep 2006 WO
2007056531 May 2007 WO
2007056532 May 2007 WO
2008042953 Apr 2008 WO
2008044664 Apr 2008 WO
2008045950 Apr 2008 WO
2008110002 Sep 2008 WO
2008110790 Sep 2008 WO
2009011206 Jan 2009 WO
2009061651 May 2009 WO
2009064561 May 2009 WO
2010095320 Aug 2010 WO
2010127268 Nov 2010 WO
2012177866 Dec 2012 WO
2012177870 Dec 2012 WO
2012177872 Dec 2012 WO
2012177874 Dec 2012 WO
Non-Patent Literature Citations (190)
Entry
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International Application Serial No. PCT/US2010/033201, dated Oct. 1, 2010 (16 pages).
Patent Cooperation Treaty, “International Preliminary Examination Report,” issued in connection with International Application Serial No. PCT/US2003/031180, dated Aug. 17, 2004 (4 pages).
Patent Cooperation Treaty, “International Preliminary Report on Patentability,” issued in connection with International Application Serial No. PCT/US2005/026426, dated Feb. 1, 2007 (9 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International Application Serial No. PCT/US2005/026426, dated Aug. 18, 2006 (10 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International Application Serial No. PCT/US2009/061479, dated May 26, 2010 (15 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International Application Serial No. PCT/US2009/061750, dated Mar. 3, 2010 (10 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International Application Serial No. PCT/US2009/061827, dated Mar. 15, 2010 (12 pages).
Patent Cooperation Treaty, “International Search Report,” issued in connection with International Application Serial No. PCT/US2003/031180, dated Jun. 8, 2004 (5 pages).
Shazam, “Shazam and VidZone Digital Media announce UK1s first fixed price mobile download service for music videos,” http://www.shazam.com/music/web/newsdetail.html?nid=NEWS136, Feb. 11, 2008 (1 page).
Shazam, “Shazam launches new music application for Facebook fans,” http://www.shazam.com/music/web/newsdetail.html?nid=NEWS135, Feb. 18, 2008 (1 page).
Shazam, “Shazam turns up the volume on mobile music,” http://www.shazam.com/music/web/newsdetail.html?nid=NEWS137, Nov. 28, 2007 (1 page).
Shazam, Company summary outline and list of products, undated (1 page).
Stross, “Apple Wouldn't Rist Its Cool Over a Gimmick, Would It?,” The New York Times, Nov. 14, 2009 (3 pages).
Stultz, “Handheld Captioning at Disney World Theme Parks,” article retrieved on Mar. 8, 2013, http://goflorida.about.com/od/disneyworld/a/wdw—captioning.htm, (2 pages).
Sullivan, “Google Cozies Up to SMBs for Digital Content” MediaPost News, Mar. 19, 2009 (2 pages).
USPTO, “Advisory Action,” issued in connection with U.S. Appl. No. 11/618,245, dated Sep. 30, 2009 (3 pages).
USPTO, “Final Office Action,” issued in connection with U.S. Appl. No. 10/530,233, dated Mar. 18, 2010 (9 pages).
USPTO, “Final Office Action,” issued in connection with U.S. Appl. No. 11/618,245, dated Jul. 21, 2009 (26 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/530,233, dated Sep. 16, 2009 (8 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 11/618,245, dated Feb. 5, 2009 (35 pages).
Wactlar et al., “Digital Video Archives: Managing Through Metadata” Building a National Strategy for Digital Preservation: Issues in Digital Media Archiving, http://www.informedia.cs.cmu.edu/documents/Wactlar-CLIR-final.pdf, Apr. 2002 (14 pages).
Wang, “An Industrial-Strength Audio Search Algorithm,” Shazam Entertainment Ltd., in Proceedings of the Fourth International Conference on Music Information Retrieval, Baltimore, Oct. 26-30, 2003 (7 pages).
CIPO, “Office Action,” issued in connection with Canadian Application Serial No. 2,574,998, dated Aug. 26, 2008 (4 pages).
CIPO, “Office Action,” issued in connection with Canadian Application Serial No. 2,574,998, dated Mar. 23, 2009 (5 pages).
CIPO, “Office Action,” issued in connection with Canadian Application Serial No. 2,574,998, dated Nov. 13, 2009 (10 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion ”, issued in connection with International application No. PCT/US2012/043539, dated Jan. 17, 2013, (9 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion ”, issued in connection with International application No. PCT/US2012/043544, dated Jan. 31, 2013, (15 pages).
Patent Cooperation Treaty, “International Search Report and Written Opinion ”, issued in connection with International application No. PCT/US2012/043546, dated Dec. 10, 2012, (6 pages).
U.S. Appl. No. 61/499,520, filed Jun. 21, 2011, (51 pages).
U.S. Appl. No. 61/568,631, filed Dec. 8, 2011, (80 pages).
USPTO, “Final Office Action,” issued in connection with U.S. Appl. No. 11/618,245, dated Oct. 26, 2011 (38 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 11/618,245, dated Apr. 28, 2011 (48 pages).
USPTO, “Requirement for Restriction,” issued in connection with U.S. Appl. No. 10/530,233, dated Jun. 10, 2009 (20 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 13/341,661, mailed Mar. 19, 2013 (10 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 13/181,147, dated Nov. 21, 2012 (30 pages).
USPTO, “Restriction,” issued in connection with U.S. Appl. No. 13/181,147, dated Aug. 10, 2012 (4 pages).
PCT Application No. PCT/US12/43544, filed Jun. 21, 2012 (72 pages).
PCT Application No. PCT/US12/43539, filed Jun. 21, 2012 (50 pages).
PCT Application No. PCT/US12/43546, filed Jun. 21, 2012, (61 pages).
PCT Application No. PCT/US12/43535, filed Jun. 21, 2012, (83 pages).
U.S. Appl. No. 13/767,548, filed Feb. 14, 2013, (68 pages).
U.S. Appl. No. 13/778,108, filed Feb. 26, 2013, (66 pages).
U.S. Appl. No. 13/793,959, filed Mar. 11, 2013, (68 pages).
U.S. Appl. No. 13/793,974, filed Mar. 11, 2013, (58 pages).
U.S. Appl. No. 13/793,991, filed Mar. 11, 2013, (47 pages).
United States Patent and Trademark Office, “Office Action,” issued in connection with U.S. Appl. No. 13/472,170 dated Nov. 8, 2013, 17 pages.
United States Patent and Trademark Office, “Examiner's Answer to Appeal Brief”, issued in connection with U.S. Appl. No. 13/181,147, dated Dec. 22, 2014 (34 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/767,548, dated Feb. 3, 2015 (10 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,959 dated Jan. 30, 2015 (11 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/793,983, dated Jan. 9, 2015 (12 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,974, dated Feb. 18, 2015 (12 pages).
Mexican Patent Office, “Office Action”, issued in connection with Mexican Patent Application No. MX/a/2014/000280, dated Jan. 21, 2015 (5 pages, English translation included).
United States Patent and Trademark Office, “Corrected Notice of Allowance”, issued in connection with U.S. Appl. No. 13/472,170, dated Feb. 12, 2015 (5 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/472,170, dated Jun. 18, 2014 (18 pages).
U.S. Appl. No. 13/472,170, filed May 15, 2012, (50 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/472,170, dated Dec. 5, 2014 (12 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/443,596, dated Feb. 26, 2015 (25 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Feb. 27, 2015 (21 pages).
“What is Jacked?,” http://www.jacked.com/, retrieved on Dec. 3, 2009 (1 page).
Anderson, “Google to compete with Nielsen for TV-ratings info?,” Ars Technica, Jun. 19, 2006 (2 pages).
Boehret, “Yahoo Widgets Lend Brains to Boob Tube,” The Wall Street Journal, Mar. 25, 2009 (3 pages).
Claburn, “Google Researchers Propose TV Monitoring,” Information Week, Jun. 7, 2006 (3 pages).
Evain, “TV-Anytime Metadata—A Preliminary Specification on Schedule!,” EBU Technical Review, Sep. 2000 [http://www.ebu.ch/en/technical/trev/trev—284-contents.html, retrieved on Jul. 20, 2006] (14 pages).
Fink et al., “Social- and Interactive-Television Applications Based on Real-Time Ambient-Audio Identification,” EuroITV, 2006 (10 pages).
Heuer et al., “Adaptive Multimedia Messaging based on MPEG-7—The M3-Box,”, Proc. Second Int'l Symposium on Mobile Multimedia Systems Applications, Nov. 9-10, 2000 (8 pages).
Heussner, “Sound-Triggered Smart Phone Ads Seek You Out,” Adweek.com, http://www.adweek.com/news/advertising-branding/sound-triggered-smartphone-ads-seek-you-out-136901, Dec. 7, 2011 (1 page).
Hopper, “EBU Project Group P/META Metadata Exchange Standards,” EBU Technical Review, Sep. 2000 [http://www.ebu.ch/en/technical/trev/trev—284-contents.html, retrieved on Jul. 20, 2006] (24 pages).
Kane, “Entrepreneur Plans On-Demand Videogame Service,” The Wall Street Journal, Mar. 24, 2009 (2 pages).
Laven,“EBU Technical Review (Editorial),” No. 284, Sep. 2000 [http://www.ebu.ch/en/technical/trev/trev—284-contents.html, retrieved on Jul. 20, 2006] (3 pages).
Mulder, “The Integration of Metadata From Production to Consumer,” EBU Technical Review, Sep. 2000 [http://www.ebu.ch/en/technical/trev/trev—284-contents.html, retrieved on Jul. 20, 2006] (5 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Feb. 27, 2015 (18 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/341,646, dated Mar. 3, 2015 (7 pages).
Japanese Patent Office, “Notice of Reasons for Rejection”, issued in connection with Japanese Patent Application No. P2014-517158, dated Mar. 3, 2015 (7 pages).
Mexican Patent Office, “Notice of Allowance”, issued in connection with Mexican Patent Application No. MX/a/2014/000281, dated Feb. 25, 2015 (1 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/793,983, dated Mar. 16, 2015 (9 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/472,170, dated Mar. 26, 2015 (10 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/341,661, dated Mar. 26, 2015 (13 pages).
European Patent Office, “European Search Report”, issued in connection with European Patent Application No. 12803215.8, dated Apr. 20, 2015 (9 pages).
Canadian Patent Office, “Office Action”, issued in connection with Canadian Patent Application No. 2,840,092, dated Apr. 20, 2015 (4 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/455,961, dated May 20, 2015 (14 pages).
State Intellectual Property Office, “Notice of Allowance” issued in connection with Application No. 201210105474.3 , May 25, 2015, 5 pages.
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,974, dated May 28, 2015 (13 pages).
Canadian Intellectual Property Office, “Office Action”, issued in connection with Canadian Patent Application No. 2,840,094, dated May 19, 2015 (4 pages).
European Patent Office, “European Search Report”, issued in connection with European Patent Application No. 12802805.7, dated May 27, 2015 (8 pages).
European Patent Office, “European Search Report”, issued in connection with European Patent Application No. 12802746.3, dated May 27, 2015 (9 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/341,646, dated Jun. 19, 2015 (9 pages).
Mexican Patent Office, “Notice of Allowance”, issued in connection with Mexican Patent Application No. MX/a/2014/000280, dated Jun. 12, 2015 (1 page).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/472,170, dated Jul. 7, 2015 (9 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/793,983, dated Jul. 7, 2015 (10 pages).
European Patent Office, “European Search Report” issued in connection with European Patent Application No. 12802202.7 dated May 28, 2015 (7 pages).
IP Australia, “Notice of Acceptance”, issued in connection with Australian Patent Application No. 2012272868, dated Jul. 22, 2015 (2 pages).
CIPO, “Office Action,” issued in connection with Canadian Application Serial No. 2,773,567, on Mar. 27, 2015, 6 pages
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/767,548 dated Aug. 11, 2015 (13 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/778,108 dated Aug. 13, 2015 (17 pages).
IP Australia, “Notice of Acceptance”, issued in connection with Australian Patent Application No. 2012272872 dated Aug. 6, 2015 (2 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,959, dated Sep. 11, 2015 (12 pages).
IP Australia, “Notice of Acceptance”, issued in connection with Australian Patent Application No. 2012272874, dated Sep. 11, 2015 (2 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/455,961, dated Sep. 24, 2015 (16 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,974, dated Sep. 24, 2015 (14 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/443,596, dated Oct. 20, 2015 (23 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Oct. 22, 2015 (20 pages).
State Intellectual Property Office of China, “Office Action”, issued in connection with Chinese Patent Application No. 201280032737.0, dated Nov. 10, 2015 (5 pages).
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 12/890,216 dated Aug. 6, 2013, 14 pages.
Patent Cooperation Treaty, “International Preliminary Examination Report,” issued in connection with International Application Serial No. PCT/US2003/014970, dated Aug. 21, 2004 (6 pages).
Patent Cooperation Treaty, “International Search Report,” issued in connection with International Application Serial No. PCT/US2003/14970, dated Feb. 10, 2004 (1 page).
USPTO, “Advisory Action,” issued in connection with U.S. Appl. No. 10/540,611, dated Jan. 22, 2010 (4 pages).
USPTO, “Final Office Action,” issued in connection with U.S. Appl. No. 10/540,611, dated Sep. 29, 2009 (20 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/540,611, dated Mar. 4, 2009 (19 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 10/540,611, dated Sep. 15, 2008 (14 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 12/890,216, dated Apr. 2, 2012 (21 pages).
USPTO, “Notice of Allowance,” issued in connection with U.S. Appl. No. 10/540,611, dated Jun. 22, 2010 (7 pages).
Van Beek et al., “Metadata-Driven Multimedia Access,” IEEE Signal Processing Magazine, vol. 20, No. 2, Institute of Electric and Electronic Engineers, Inc., New York, New York, USA, Mar. 2003 (13 pages).
Vetro et al., “Video Transcoding Architectures and Techniques: An Overview,” IEEE Signal Processing Magazine, vol. 20, No. 2, Institute of Electric and Electronic Engineers, Inc., New York, New York, USA, Mar. 2003 (12 pages).
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 14/725,877, dated May 5, 2017 (10 pages).
United States Patent and Trademark Office, “Examiner's Answer to Appeal Brief,” issued in connection with U.S. Appl. No. 13/778,108, dated May 5, 2017 (24 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 15/331,568, dated Aug. 1, 2017 (9 pages).
USPTO, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 12/890,216, dated Nov. 29, 2012 (29 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/089,279, dated Mar. 28, 2014 (13 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 14/089,279, dated Nov. 21, 2014 (22 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/089,279, dated Apr. 23, 2015 (23 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 14/089,279, dated Nov. 17, 2015 (25 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/181,147, dated Dec. 3, 2015 (10 pages).
IP Australia, “Notice of Grant,” issued in connection with Application No. 2012272872, Dec. 3, 2015, 1 page.
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/767,548, dated Dec. 4, 2015 (9 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/778,108, dated Jan. 8, 2016 (13 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/793,959, dated Jan. 26, 2016 (9 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/341,661, dated Feb. 1, 2016 (8 pages).
Japanese Patent Office, “Notice of Reasons for Rejection”, issued in connection with Japanese Patent Application No. P2014-517158, dated Jan. 26, 2016 (4 pages).
United States Patent and Trademark Office, “Notice of Allowance” issued in connection with U.S. Appl. No. 13/181,147, dated Feb. 18, 2016 (8 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/922,918, dated Feb. 23, 2016 (17 pages).
Canadian Intellectual Property Office, “Office Action”, issued in connection with Canadian Patent Application No. 2,773,567, dated Mar. 9, 2016 (4 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/455,961, dated Mar. 23, 2016 (17 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/443,596, dated Apr. 6, 2016 (25 pages).
European Patent Office, “Communication Pursuant to Article 94(3) EPS,” issued in connection with application No. 12002599.4 on Mar. 4, 2016, (4 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Apr. 8, 2016 (23 pages).
IP Australia, “Notice of Acceptance”, issued in connection with Australian Patent Application No. 2013204488, dated Apr. 26, 2016 (3 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2012272876, dated Apr. 26, 2016 (3 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/089,279, dated May 5, 2016 (29 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,974, dated May 23, 2016 (14 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/778,108, dated May 23, 2016 (13 pages).
The State Intellectual Property Office of China, “Office Action”, issued in connection with Chinese Patent Application No. 201280032740.2, dated May 31, 2016 (22 pages).
IP Australia, “Notice of Acceptance”, issued in connection with Australian Patent Application No. 2012272876, dated Jun. 6, 2016 (2 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S Appl. No. 13/341,661, dated Aug. 19, 2016 (9 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/922,918, dated Sep. 9, 2016 (16 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2015252031, dated Oct. 21, 2016 (3 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 14/089,279, dated Nov. 14, 2016 (13 pages).
The State Intellectual Property Office of China, “First Office Action”, issued in connection with Chinese Patent Application No. 201280032738.5, dated Dec. 16, 2016 (13 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 14/725,877, dated Jan. 26, 2017 (14 pages).
United States Patent and Trademark Office, “Non-Final Office Action,” issued in connection with U.S. Appl. No. 15/181,399, dated Aug. 9, 2016, 44 pages.
United States Patent and Trademark Office, “Corrected Notice of Allowance,” issued in connection with U.S. Appl. No. 15/181,399, dated Feb. 16, 2017, 6 pages.
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 15/181,399, dated Jan. 27, 2017, 9 pages.
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 15/331,568, dated Feb. 24, 2017 (9 pages).
European Patent Office, “Examination Report”, issued in connection with European Patent Application No. 12802746.3, dated Feb. 23, 2017 (6 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 14/922,918, dated Apr. 6, 2017 (18 pages).
Canadian Intellectual Property Office, “Office Action,” issued in connection with application No. CA 2,773,567, dated Mar. 6, 2014, 2 pages.
State Intellectual Property Office of China, “First Office Action,” issued in connection with application No. CN 201210105474.3, dated Feb. 8, 2014, 15 pages.
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/341,646, dated Jun. 5, 2014 (17 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/181,147, dated Mar. 10, 2014 (48 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/443,596, dated Apr. 9, 2014 (20 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Apr. 11, 2014 (15 pages).
International Bureau, “International Preliminary Report on Patentability” issued in connection with International Application No. PCT/US2012/043544, dated Jan. 9, 2014 (9 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2012272868, dated Jun. 27, 2014 (3 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2012272874, dated Jun. 27, 2014 (3 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2012272872, dated Jun. 24, 2014 (4 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/341,661 dated Jul. 8, 2014 (8 pages).
United States Patent and Trademark Office, “Final Office Action”, issued in connection with U.S. Appl. No. 13/793,983, dated Jun. 6, 2014 (13 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2012272876, dated Sep. 18, 2014 (4 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2013203778, dated Aug. 21, 2014 (5 pages).
IP Australia, “Examination Report”, issued in connection with Australian Patent Application No. 2013204488, dated Aug. 12, 2014 (5 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/443,596, dated Sep. 25, 2014 (22 pages).
United States Patent and Trademark Office, “Notice of Allowance”, issued in connection with U.S. Appl. No. 13/341,646, dated Nov. 3, 2014 (10 pages).
U.S. Appl. No. 13/341,646, filed Dec. 30, 2011, (78 pages).
U.S. Appl. No. 13/341,661, filed Dec. 30, 2011, (79 pages).
U.S. Appl. No. 13/181,147, filed Jul. 12, 2011, (41 pages).
United States Patent and Trademark Office, “Office Action,” issued in connection with U.S. Appl. No. 13/793,983 dated Nov. 8, 2013, 13 pages.
Patent Cooperation Treaty, “International Search Report and Written Opinion,” issued in connection with International application No. PCT/ US2012/043535, dated Feb. 21, 2013, 15 pages.
United States Patent and Trademark Office, “Notice of Allowance,” issued in connection with U.S. Appl. No. 13/341,661 dated Sep. 23, 2013, 10 pages.
United States Patent and Trademark Office, “Office Action,” issued in connection with U.S. Appl. No. 13/181,147 dated Aug. 15, 2013, 46 pages.
United States Patent and Trademark Office, “Office Action,” issued in connection with U.S. Appl. No. 13/443,596 dated Nov. 21, 2013, 25 pages.
United States Patent and Trademark Office, “Office Action,” issued in connection with U.S. Appl. No. 13/793,991 dated Dec. 6, 2013, 21 pages.
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/455,961, dated Dec. 5, 2014 (13 pages).
United States Patent and Trademark Office, “Non-Final Office Action”, issued in connection with U.S. Appl. No. 13/793,991, dated Nov. 10, 2014 (19 pages).
U.S. Appl. No. 13/455,961, filed Apr. 25, 2012, (61 pages).
“Video: timed text tracks”, Windows Internet Explorer, Microsoft, 2012, http://msdn.microsoft.com/en-us/library/ie/hh673566(v=vs.85).aspx, (6 pages).
Apple Inc. “HTTP Live Streaming Overview”, Apr. 1, 2011, (36 pages).
Apple Inc. “Timed Metadata for HTTP Live Streaming”, Apr. 28, 2011, (12 pages).
R. Pantos, Ed., & W. May, Apple Inc. “HTTP Live Streaming: draft-pantos-http-live-streaming-07”, Sep. 2011, (33 pages).
Eric Winkelman, “Timed Text Tracks and TV Services”, Aug. 15, 2011, (5 pages).
U.S. Appl. No. 13/443,596, filed Apr. 10, 2012, (50 pages).
U.S. Appl. No. 13/793,983, filed Mar. 11, 2013, (68 pages).
Related Publications (1)
Number Date Country
20170195716 A1 Jul 2017 US
Provisional Applications (1)
Number Date Country
60436714 Dec 2002 US
Continuations (3)
Number Date Country
Parent 14089279 Nov 2013 US
Child 15466547 US
Parent 12890216 Sep 2010 US
Child 14089279 US
Parent 10540611 US
Child 12890216 US