Method and apparatus for identifying media content presented on a media playing device

Information

  • Patent Grant
  • 10194187
  • Patent Number
    10,194,187
  • Date Filed
    Friday, January 21, 2011
    13 years ago
  • Date Issued
    Tuesday, January 29, 2019
    5 years ago
Abstract
A media device generates a digital fingerprint of perceptual features of a segment of a work. The segment is less than an entirety of the work and includes at least one of audio or video media content. The media device sends a query to a lookup server for data associated with the work, the query including said digital fingerprint. The media device receives the data associated with said work from said lookup server in response to said query, wherein the data was identified based on a comparison of the digital fingerprint to a collection of digital fingerprints of known works. The media device then processes the received data.
Description
BACKGROUND OF THE INVENTION
1. Field of the Invention

This invention pertains generally to media content identification methods. More particularly, the invention is an apparatus and method for identifying media content presented on a media playing device and presenting media-content related information and/or action to the user of the media playing device.


2. The Prior Art

The use of a data processing means, such as a computer or personal digital assistant (PDA), for playing works is well known. For purposes of the present discussion, a work is anything that is fixed in a tangible medium. Some examples of works include, but are not limited to, audio renderings, video renderings, images, video/audio renderings, and software. An example of an audio rendering is a song or other audio track. Examples of video renderings include animation or video sequence. Examples of an image include photographs and paintings. Examples of audio/video renderings include movies, television shows, and cartoons. Examples of software include word processing programs and video games.


Some examples of playing a work include using a personal computer (PC) to play songs from audio compact disks (CD), normally with the use of a CD-ROM drive, a sound card, and speakers. Other types of media, including various formats of audio and video, are also commonly played using a PC.


With the increasing popularity of the global information network commonly known as the “Internet”, various audio and video formats have been introduced to provide live (as well as archived) broadcasts of works over the Internet. These broadcasts may be viewed by users connected to the Internet with the use of a PC and the proper client application. For example, Real Networks™ provides Real Player™ (a client application) for playing streaming audio and video works (in Real™ format) which is broadcast over the Internet. Various servers (also connected to the Internet) carry out the operation of making such works available and streaming data for the appropriate works to users upon request. In this way, Real™ media content of the work may be played by a user using Real Player™ on the user's PC. Like other media client applications, Real Player™ plays audio content of a work via the user's sound card and speakers and video content of a work via the user's video card and monitor (or other viewing device).


Internet radio broadcasts (or webcasts) are also known in the art. In general, Internet radio broadcasts are provided over the Internet by one or more server computers equipped to provide streaming audio works, much like traditional AM or FM radio broadcast. A user that would like to listen to an Internet radio broadcast would use a client application (such as Real Player™, Microsoft™ media player, or Apple QuickTime™ viewer, for example) and direct the client application to the appropriate server computer. The server computer then transmits the media content of a work to the client application via the Internet. The client application receives the media content of the work transmitted from the server computer and plays the content of the work using the user's sound card and speaker.


Internet video broadcasts (another form of webcast) are carried out in a similar fashion to that of Internet radio broadcasts, except video broadcasts typically contain both audio and video content of a work. As such, the audio component of the work is played to the user via a sound card and speakers, while the video component of the work is displayed using a video card and a monitor.


In many cases, webcasts will not provide content identifying information, such as song title, artist name or album name, as part of a broadcast work. Other information, such as where the content of the work can be purchased in the form of a CD, DVD (digital video disk) or VHS (video home system) tape, for example, are also typically not included. The primary reason a webcast provider (station) fails to provide such content-related information of the work is that the station does not have real-time access to accurate playlists. For example, it is common for a station disk jockey to make real-time adjustments to the playlist order and timing. Thus current radio station systems do not transmit or attach identification data of a work with the broadcast.


In addition to webcasts, many Internet sites provide archived or pre-recorded works for download and playback on a user's PC. Examples of such content of works include “way”, “mp3”, “mov”, and “midi” files, among others. Once these files have been downloaded from the Internet site to the user's PC, the user is able to play back the audio or video content of a work using an appropriate client application. Many of these media files, like webcasts, also fail to provide identifying data or other information for a work.


Additionally, PCs often have the ability to play other forms of media content for other works via PC sound card and speakers and/or PC video card and monitor. For example, tuner expansion cards (tuners) which may be inserted into (or interfaced with) a user's PC are presently available (some computers already include such tuners). In either case, “television” (TV) tuners allow a user of the PC to view conventional television, cable or satellite transmissions on the PC monitor and hear the audio via the PC speaker. “Radio” tuners allow a user to hear conventional (FM or AM) radio broadcast using the PC speaker. Conventional TV and radio broadcast also fail to provide rich content-related information of a work to the user.


Other media playing devices such as televisions, radio/stereo systems, and portable audio devices also typically do not provide content-related information of a work on demand for the user. For example, in automobiles, current car stereo systems fail to provide a system and method for identifying a song being played on the stereo. Typically a user would either rely on the disk jockey to identify the artist and song title or call the radio station to ascertain the information.


Users may also read and/or store media content of works from storage mediums such as compact disk and magnetic tapes. Sometimes these copies of works do not include metadata or the metadata for the work may be incorrect. Furthermore, there may be no proper way to identify the work.


Accordingly, there is a need for an apparatus and method which provides real-time media content-related information of a work and context specific action choices to users viewing and/or listening to media content of the work over media playing devices. The present invention satisfies these needs, as well as others, and generally overcomes the deficiencies found in the background art.


BRIEF DESCRIPTION OF THE INVENTION

The present invention is a system and method for identifying a work from media content presented over a media playing device, such as a computer. The system generates a media sample or analytical representation from the media content of the work, such as audio and/or video, played on the media player. The media sample or representation is compared to a database of samples of media content or representations of known works to query and ascertain content-related information related to the work. This media content-related information of the work is then displayed on the media player.


The invention further relates to machine readable media which are stored embodiments of the present invention. It is contemplated that any media suitable for retrieving instructions is within the scope of the present invention. By way of example, such media may take the form of magnetic, optical, or semiconductor media. The invention also relates to data structures that contain embodiments of the present invention, and to the transmission of data structures containing embodiments of the present invention.


By way of example only, and not of limitation, the media player may comprise any data processing means or computer executing the present invention including devices carrying out the invention via an embedded system. For example, the media player may comprise cellular or mobile phones, portable media players, and fixed media players including analog broadcast receivers (such as car stereo, home stereos, and televisions, for example).


In a first embodiment, the database of media content samples of known works resides on a “lookup server” which carries out the database query described above. In general, the lookup server is operatively coupled for communication to one or more media players (client) via the Internet as is known in the art. Under this arrangement, media samples of works are first communicated from the client to the lookup server before the media samples are compared to the database. Then after the database query is carried out on the lookup server, the content-related information of the work is transmitted to the client for presentation thereon.


According to this embodiment, the system of the present invention generally includes one or more client media players, each operatively coupled for communication to a lookup server. As noted above, the lookup server is generally connected to the client media players via an Internet connection, although any suitable network (including wired or wireless) connection may also be used.


For example, the client media player may comprise a computer. The client computer typically plays audio files via a sound card and speakers. Video files are played via a video card and a monitor. The client computer has executing thereon a “client engine” software application comprising a sampling unit, a media intercept unit (intercept unit), and a user-interface.


By way of example, and not of limitation, the intercept unit carries out the operation of monitoring the client computer for media content of a work presented thereon. In general, the intercept unit carries out the operation of monitoring audio signals transmitted to the sound card and/or video signals transmitted to the video card. The intercept unit may also monitor the operations of other hardware suitable for communicating audio and/or video signals, such as USB (universal serial bus) and IEEE (Institute of Electrical and Electronics Engineers) standard 1394 connections, for example. Preferably, the intercept unit maintains a FIFO (first in first out) buffer of fixed size (150 seconds, for example) containing media content played on the client.


The sampling unit carries out the operation of creating a media sample from the media content of a work played on the client computer in order to represent and identify the work. In general, the sampling unit creates a sample of the work from the FIFO buffer maintained by the intercept unit. The sampling unit may create media samples from the media content of a work according to a predetermined or user-defined interval or upon the request of the user of the client computer. A user request may be received via commands issued using the user-interface. In general, the media sample created by the sampling unit comprises a “digital fingerprint” or “digital signature” using conventional digital signal processing known in the art. For example, a sound file may be sampled according to its acoustic/perceptual features over time. It is noted that “media sample” is defined herein to include actual recordings of the media content and/or analytical representations of the media content of the work. One example of such a digital signal processing technique for creating a media sample is described in U.S. Pat. No. 5,918,223 issued to Blum which is hereby incorporated by reference as if set forth in its entirety. Other examples of methods for creating media samples are given in PCT application number WO200444820 having inventors Jin S. Seo, Jaap A. Haitsma and Antonius A. C. M. Kalker and assigned to Koninklijke Philips Electronics and PCT application number WO03091990 having inventors Daniel Culbert and Avery Li-Chun Wang and assigned to Shazam Entertainment, Ltd. The media sample is then transmitted to the lookup server for further processing, as described in further detail below.


The user interface carries out the operation of receiving commands from a user of the client computer, and displaying content-related information to the user. As noted above, a user may issue a request for content-related information via the user-interface. This request is communicated to the sampling unit for further processing. As described above, in response to this request, the sampling unit creates a sample of the media content being played on the client computer and transmits the sample to the lookup server for further processing. In response, the lookup server provides the information related to the media sample to the client computer. This content-related information is received by the user interface which then displays the received information to the user of the client computer.


The content-related information for the work returned from the lookup server may also be used for a plurality of purposes including, for example, generating a log of the user activity, providing an option to purchase media to the user, and displaying the content-related information on the media playing device, among others.


The lookup server has executing thereon a “server engine” software application comprising a lookup unit, and a log unit. The lookup unit is further coupled to a media database, and the log unit is coupled to a log database. While the present invention describes the lookup server as a single machine, a group or cluster of server machines may be used to carry out the tasks of the lookup server described herein to thereby balance the load between several machines as is known in the art.


The lookup unit carries out the operation of receiving media samples of the work from client computers and performing database queries on the media database to ascertain content information related to the media sample provided. This content-related information of the work may include such information as song title, artist, and album name, for example. The content-related information of the work may also include product fulfillment information, such as how and where to purchase media containing the work, advertising banners, and/or promotional offers, for example. This content-related information of the work is transmitted back to the client computer for further processing thereon.


The log unit caries out the operation of tracking media requests made by the users of the client computers to the lookup server. The log unit maintains a plurality of information related to the requests made by the user including, for example, media type, genre or category. This user information is maintained in the log database.


The source of the media content of the work played on the client computer includes conventional media sources such as Internet sources or webcasts, including streaming media and archived media. The media content source may also be audio CDs, DVD, or other formats suitable for presentation on media playing devices, such as the client computer.


In a second embodiment of the present invention, the database of sampled media content resides within the client computer. The database resides on conventional storage medium, such as a computer memory, a hard drive, CD-ROM, or other appropriate storage medium. Under this arrangement, the database query is carried out “locally” on the computer playing the media content. It will be readily apparent to those skilled in the art that various other topological arrangements of the system elements (including the location of the database) may be used with the invention without departing from the scope and spirit of the invention.


An object of the invention is to provide an apparatus and method for identifying media content presented over a media playing device which overcomes the deficiencies of the prior art.


Another object of the invention is to provide an apparatus and method for identifying media content presented over a media playing device which does not require the media content provider to provide content-related information.


Further objects and advantages of the invention will be brought out in the following portions of the specification, wherein the detailed description is for the purpose of fully disclosing the preferred embodiment of the invention without placing limitations thereon.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will be more fully understood by reference to the following drawings, which are for illustrative purposes only.



FIG. 1 is a functional block diagram of a media content identifying system in accordance with the invention.



FIG. 2 is a functional block diagram of a media content identifying system having a plurality of client nodes in accordance with the invention.



FIG. 3 is a flow chart showing generally the processes associated with the client engine in accordance with the invention.



FIG. 4 is flow chart showing generally the processes associated with the media intercept unit in accordance with the invention.



FIG. 5 is flow chart showing generally the processes associated with the server engine in accordance with the invention.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Persons of ordinary skill in the art will realize that the following description of the present invention is illustrative only and not in any way limiting. Other embodiments of the invention will readily suggest themselves to such skilled persons having the benefit of this disclosure.


Referring more specifically to the drawings, for illustrative purposes the present invention is embodied in the apparatus shown FIG. 1 and FIG. 2 and the method outlined in FIG. 3 through FIG. 5. It will be appreciated that the apparatus may vary as to configuration and as to details of the parts, and that the method may vary as to details and the order of the acts, without departing from the basic concepts as disclosed herein. The invention is disclosed generally in terms of an apparatus and method for identification of a work on a personal computer, although numerous other uses for the invention will suggest themselves to persons of ordinary skill in the art.


Referring first to FIG. 1, there is generally shown a functional block diagram of a media content identifying system 10 in accordance with the invention. The system 10 comprises a lookup server 12 and at least one client media player 14. The lookup server 12 can be any standard data processing means or computer, including a minicomputer, a microcomputer, a UNIX® machine, a mainframe machine, a personal computer (PC) such as INTEL® based processing computer or clone thereof, an APPLE® computer or clone thereof or, a SUN® workstation, or other appropriate computer. Lookup server 12 generally includes conventional computer components (not shown), such as a motherboard, central processing unit (CPU), random access memory (RAM), hard disk drive, display adapter, other storage media such as diskette drive, CD-ROM, flash-ROM, tape drive, PCMCIA cards and/or other removable media, a monitor, keyboard, mouse and/or other user interface means, a modem, and/or other conventional input/output devices. Lookup server 12 also includes a Network Interface 17 for communication with other computers using an appropriate network protocol.


Lookup server 12 has loaded in its RAM a conventional server operating system (not shown) such as UNIX®, WINDOWS® NT, NOVELL®, SOLARIS®, or other server operating system. Lookup server 12 also has loaded in its RAM server engine software 16, which is discussed in more detail below.


Client media player 14, may comprise a standard computer such as a minicomputer, a microcomputer, a UNIX® machine, mainframe machine, personal computer (PC) such as INTEL®, APPLE®, or SUN® based processing computer or clone thereof, or other appropriate computer. As such, client media player 14 is normally embodied in a conventional desktop or “tower” machine, but can alternatively be embodied in a portable or “laptop” computer, a handheld personal digital assistant (PDA), a cellular phone capable of playing media content, a dumb terminal playing media content, a specialized device such as Tivo® player, or an internet terminal playing media content such as WEBTV®, among others. As described above, client media player 14 may also comprise other media playing devices such as a portable stereo system, fixed stereo systems, and televisions suitable for use with embedded systems carrying out the operations of the client engine as described further below.


Client media player 14 also includes typical computer components (not shown), such as a motherboard, central processing unit (CPU), random access memory (RAM), hard disk drive, other storage media such as diskette drive, CD-ROM, flash-ROM, tape drive, PCMCIA cards and/or other removable media, keyboard, mouse and/or other user interface means, a modem, and/or other conventional input/output devices.


Client media player 14 also has loaded in RAM an operating system (not shown) such as UNIX®, WINDOWS® XP or the like. Client media player 14 further has loaded in RAM a media client application program 18 such as Real Player™, Windows™ Media Player, Apple Quicktime™ Player, or other appropriate media client application for playing audio and/or video via client media player 14. Client media player 14 also has loaded in its RAM client engine software 30, which is discussed in more detail below.


Client media player 14 further includes a conventional sound card 20 connected to speakers 22. As is known in the art, the media client application 18 will generally play audio signals through the sound card device 20 and speakers 22. For example, audio streams, audio files, audio CDs, and other audio sources are communicated from the client application 18 to the sound card 20 for output. The sound card 18 then produces the audio signal via speakers 22.


Client media player 14 also includes a conventional video card 24 connected to a display means 26. As is known in the art, the media client 18 will generally play video content through the video card 24, which then produces an appropriate video signal suitable for display on the display means 26. The display means 26 may be a standard monitor, LCD, or other appropriate display apparatus. Client media player 14 also includes a Network Interface 28 for communication with other computers using an appropriate network protocol. The network interface 28 may comprise a network interface card (NIC), a modem, ISDN (Integrated Services Digital Network) adapter, ADSL (Asynchronous digital subscriber line) adapter or other suitable network interface.


Client media player 14 is networked for communication with lookup server 12. Typically, client media player 14 is operatively coupled to communicate with lookup server 12 via an Internet connection 32 through a phone connection using a modem and telephone line (not shown), in a standard fashion. The user of client media player 14 will typically dial the user's Internet service provider (ISP) (not shown) through the modem and phone line to establish a connection between the client media player 14 and the ISP, thereby establishing a connection between the client media player 14 and the Internet 32. Generally, lookup server 12 is also connected to the Internet 32 via a second ISP (not shown), typically by using a fast data connection such as T1, T3, multiple T1, multiple T3, or other conventional data connection means (not shown). Since computers connected to the Internet are themselves connected to each other, the Internet 32 establishes a network communication between client media player 14 and server 12. One skilled in the art will recognize that a client may be connected to any type of network including but not limited to a wireless network or a cellular network. Generally, client media player 14 and lookup server 12 communicate using the IP (internet protocol). However, other protocols for communication may also be utilized, including PPTP, NetBEUI over TCP/IP, and other appropriate network protocols.


Client media player 14 and server 12 can alternatively connect to the Internet 32 using a network means, wireless means, satellite means, cable means, infrared means or other means for establishing a connection to the Internet, as alternatives to telephone line connection. Alternative means for networking client media player 14 and server 12 may also be utilized, such as a direct point to point connection using modems, a local area network (LAN), a wide area network (WAN), wireless connection, satellite connection, direct port to port connection utilizing infrared, serial, parallel, USB, FireWire/IEEE-1394, ISDN, DSL and other means known in the art.


In general, client engine 30 is a software application executing within the client media player 14. However, client engine 30 may also be an embedded system carrying out the functions described herein and suitable for use with various media player platforms. Client engine 30 comprises a sampling unit 34, an intercept unit 36, and a user interface 38.


The intercept unit 36 carries out the operation of monitoring the client media player 14 for media content of a work presented thereon, typically by media client application 18, but may be from any media source. In operation, the intercept unit 36 monitors audio signals transmitted from the media client 18 to the sound card 20 and/or video signals transmitted from the media client 18 to the video card 24. The intercept unit 36 may also monitor the operations of other hardware (not shown) suitable for communicating audio and/or video signals, such as USB, IEEE standard 1394 connections, and analog input devices (such as microphones). For example, in a media player comprising a mobile stereo system, the intercept unit 36 intercepts the audio signal played on thereon.


Preferably, the intercept unit 36 maintains a FIFO (first in first out) buffer 40 of fixed size (150 seconds, for example) for storing media content of a work played on the client media player 14, and from which media samples of the work are created by sampling unit 34. The operation of the intercept unit is further described below in conjunction with FIG. 4.


The sampling unit 34 carries out the operation of creating a media sample from the media content of a work played on the client media player 14. In operation, the sampling unit 34 creates a media sample from the FIFO buffer 40 maintained by the intercept unit 36. The sampling unit 34 may create media samples according to a predetermined or user-defined interval or upon the request of user of the client media player 14. For example, user requests may be received via commands issued using the user-interface 38.


As noted above, the media sample created by the sampling unit 34 typically comprises a “digital fingerprint” using conventional digital signal processing known in the art. For example, a sound file may be sampled according to its acoustic/perceptual features over time. One example of such a digital signal processing technique is described in U.S. Pat. No. 5,918,223 issued to Blum which is hereby incorporated by reference as if set forth in its entirety. Other examples are described in PCT application number WO200444820 having inventors Jin S. Seo, Jaap A. Haitsma and Antonius A. C. M. Kalker and assigned to Koninklijke Philips Electronics and PCT application number WO03091990 having inventors Daniel Culbert and Avery Li-Chun Wang and assigned to Shazam Entertainment, Ltd. The media sample generated by sampling unit 34 is then transmitted to the lookup server 12 for further processing, as described in further detail below.


The user interface 38 carries out the operation of receiving commands from a user of the client media player 14, and displaying content-related information to the user. The user interface 38 may be any conventional interfaces including, for example, a graphical user interface (GUI), a command line user interface (CLUI), a voice recognition interface, a touch screen interface, a momentary contact switch on a stereo system, or other appropriate user interface.


In operation, a user may issue a request for content-related information via the user-interface 38. This request is communicated to the sampling unit 34 for further processing. As described above, in response to such a request, the sampling unit 34 creates a media sample from the media content of a work being played on the client media player 14 and transmits the sample to the lookup server 12 for further processing. In response, the lookup server 12 provides the information related to the work, if available, to the client media player. This content-related information of the work is received by the user interface 38 which then displays the received information to the user of the client media player 14.


The server engine 16 is a software application executing within the lookup server 12. Server engine 16 comprises a lookup unit 42 and a log unit 44. Lookup unit 42 is operatively coupled for communication with a media database 46. Log unit 44 is operatively coupled for communication with a log database 48.


The lookup unit 42 carries out the operation of receiving media samples from the client media player 14 and comparing the media samples to a collection of samples of the media content of known works (reference samples). The collection of reference samples of known works will typically reside in an in-memory structure (designated 47). Upon initialization of the lookup unit 42, the collection of reference samples of known works from the media database 46 is stored in the in-memory structure 47.


The lookup unit 42 sequentially compares each reference sample of known works in structure 47 to the media sample provided by the media player 14. The lookup unit will cut a media sample of the received work into a series of overlapping frames, each frame the exact size of the particular reference sample under examination. The lookup unit 42 will then compare the reference sample to every possible “frame” of information within the media samples of the known works and compute a “distance” between each frame and the reference sample. This process repeats for each reference sample. The reference sample which has the smallest distance to any frame in the sample is considered for a match. If this distance is below a predefined threshold, a match is considered to be found.


When a match is determined, the information related to the matching record of the known work is returned to the client media player for presentation thereon. This content-related information of a work may include such information as song title, artist, and album name, for example. The content-related information may also include product fulfillment information, such as how and where to purchase media containing the media sample, advertising banners, and/or promotional offers, for example. This content-related information of the work is transmitted back to the client media player 14 for further processing thereon.


The log unit 44 caries out the operation of tracking media requests made by the users of the client media player 14 to the lookup server 12. The log unit 44 maintains a plurality of information related to the requests made by the user including, for example, media type, genre or category. This user information is maintained in the log database 48


Media database 46 and log database 48 comprise conventional relational database storage facilities. Thus, media database 46 may comprise one or more tables (not shown) for storing data associated with the media content samples of known works and related content-specific information (song name, title, album name, etc.) of the known works. In general, another computer (not shown) may be connected to the lookup server 12 for entering the content-specific media information into the media database 46. Log database 46 may comprise one or more tables (not shown) for storing data associated with users and the user's related media content requests.


Referring now to FIG. 2, there is generally shown a functional block diagram of a media content identifying system 50 having a plurality of client nodes in accordance with the invention.


System 50 comprises a lookup server 54 operatively coupled for communication with a plurality of client media players 52a through 52n via Internet connection 32. System 50 operates in substantially the same manner as system 10 described above. That is, lookup server 54 operates in the same manner as lookup server 12, and clients 52a through 52n operate in the same manner as client media player 14. However, in system 50, lookup server 12 handles a plurality of media sample identification requests from the client 52a through 52n.


As depicted in. FIG. 2, clients 52a through 52n communicate with lookup server 54 using the HTTP (hypertext transfer protocol) over IP. Although http is used in the example, one skilled in the art will recognize that any suitable protocol may be used. More particularly, the media sample signal generated by each client 52a through 52n is wrapped using a transport protocol before being transmitted over HTTP to lookup server 54 for processing. When lookup server 54 receives the HTTP transmission from the client media player, a transport protocol converts (“unwraps”) the signal for processing therein by the server engine. Likewise, before transmitting content-related information retrieved from the media database to the appropriate client media player, the information is first wrapped using the transport protocol. The content-related information may be communicated using “XML” tags, or other appropriate programming, scripting, or markup language. The appropriate client media player upon receiving the transmission from the lookup server 54 unwraps the signal for processing therein by the client engine.


The method and operation of the invention will be more fully understood by reference to the flow charts of FIG. 3 through FIG. 5. The order of acts as shown in FIG. 3 through FIG. 5 and described below are only exemplary, and should not be considered limiting.


Referring now to FIG. 3, as well as FIG. 1 and FIG. 2, there is shown the processes associated with the client engine 30 in accordance with the invention. As described above, the client engine 30 is a software application or embedded system operating within a client media player for identifying media content played via the client media player.


At process 100, the client engine 30 is initiated. The client engine 30 may be initiated by the user of client media player 14 or may initiate automatically upon the recognition of media content being played via the sound card 20 and/or the video card 24. Box 110 is then carried out.


At box 110, the intercept unit 36 is initiated. As described above, the intercept unit 36 carries out the operation of monitoring media content of a work played on client media player 14 (typically via sound card 20 and/or video card 24). The processes of the intercept unit 36 are described more fully below in conjunction with FIG. 4. After the intercept unit 36 is initiated, box 120 is then carried out.


At box 120, a user request for media information is received via user interface 38. This request is communicated from the user interface 38 to the sampling unit 34 for further processing. Box 130 is then carried out.


At box 130, the sampling unit 34 creates a media sample from the media data of a work contained in the FIFO buffer 40. The media sample created by the sampling unit 34 comprises a “digital fingerprint” using conventional digital signal processing known in the art. As noted above, a sound file may be sampled according to its acoustic/perceptual features over time. Box 14 is then carried out.


At box 140, the sampling unit 34 transmits the media sample created from box 130 to the lookup server 12 for content identification and content-related information. In the illustrative system depicted in FIG. 1 and FIG. 2, the media sample is first wrapped in a transport protocol and then communicated over IP (HTTP) via network interface 28. Box 150 is then carried out.


At box 150, the client media player 14 receives from the lookup server 12 the content-related information of the work requested in box 140. The process for generating the content-related information by the lookup server 12 is described in further detail below in conjunction with FIG. 5. This content data is first received via the network interface 28, unwrapped using the appropriate transport protocol and then communicated to the client engine 30. In the client engine 30, the user interface 38 receives the content-related information, and parses the data according to the appropriate format (XML tags, for example) transmitted by the lookup server 12. Box 160 is then carried out.


At box 160, the user interface 38 presents the content-related information of the work to the user via video card 24 and display 26, or other display device such as an LCD screen, or standard broadcast television. As described above, this content-related information of the work may include such information as song title, artist, and album name, for example. The content-related information of the work may also include product fulfillment information, such as how and where to purchase media containing the media sample, advertising banners, and/or promotional offers, for example.


Referring now to FIG. 4, there is generally shown the processes associated with the media intercept unit 36 in accordance with the invention. In general, the media intercept unit 36 maintains a FIFO buffer of predetermined size (150 seconds, for example) of media content played via client media player 14, 52a through 52n.


At process 200, the intercept unit 36 is initiated. This is carried out from box 110 of FIG. 3, during the start up of the client engine software 30. Box 210 is then carried out.


At box 210, the intercept unit 36 monitors the media hardware devices of the client media player 14. As described above, the media hardware devices may comprise a sound card 20 and/or a video card 24. Other hardware devices suitable for playing media content are also monitored by the intercept unit 36. Diamond 220 is then carried out.


At diamond 220, the intercept unit 36 determines whether any media hardware devices are playing media content of a work. If the intercept unit 36 determines that media content is currently being played, diamond 230 is carried out. Otherwise box 210 is repeated.


At diamond 230, the intercept unit 36 determines whether the FIFO buffer 40 is currently full. If the buffer is determined to be full, box 240 is carried out. Otherwise, box 250 is carried out.


At box 240, the intercept unit 36 has determined that all the buffer 40 is full and deletes the older sample in the buffer. As noted above, the FIFO buffer 40 is commonly configured with a predetermined size. For example, if the sampling rate is 22 KHz (HiloHertz), the oldest 1/22000th of a second sample would be deleted, and the newest 1/22,000th of a second sample would be added to the buffer. Using this method, the intercept unit 36 is able to maintain the most recent one hundred fifty (150) seconds of media content of a work played via client media player 14. Box 250 is then carried out.


At box 250, the intercept unit 36 stores the media content of the work currently being played on client media player 14 into the FIFO buffer 40. Box 210 is then repeated to continue monitoring media hardware devices.


Referring now to FIG. 5, there is generally shown the processes associated with the server engine 16 and the lookup server 12 in accordance with the invention. It is noted that the lookup server 12 is structured and configured to handle a plurality of requests from a plurality of client media players as depicted in FIG. 2 above.


Prior to box 300 as described below, the in-memory structure 47 is populated with reference media samples of known works. Upon initialization of the lookup unit 42, the collection of reference samples from the media database 46 is stored in the in-memory structure 47.


At box 300, the lookup server 12 receives a media sample request from one of the client media players 14, 52a through 52n. As described above, such requests are generally communicated over an Internet connection 32, although any suitable network connection may also be used. The request is first received via network interface 17 via the IP (HTTP) protocol or other suitable protocol. An appropriate transport protocol unwraps the message and communicates the request to the server engine 16 for processing. Box 310 is then carried out.


At box 310, the lookup unit 42 sequentially compares each reference sample of the known works in structure 47 to the media sample of the work provided by the media player 14, 52a through 52n, as described above. The lookup unit will cut the media sample into a series of overlapping frames, each frame the exact size of the particular reference sample under examination. The lookup unit 42 will then compare the reference sample to every possible “frame” of information within the media sample and compute a “distance” between each frame and the reference sample. This process repeats for each reference sample. Diamond 320 is then carried out.


At diamond 320, the lookup unit 42 determines whether a match was obtained. As noted above, the reference sample of the known work which has the smallest distance to any frame in the sample is considered for a match. If the distance is below a predefined threshold, a match is considered to be found. If a match is determined, box 330 is then carried out. Otherwise box 350 is carried out.


At box 330, the media sample from box 300 matches a corresponding record in the media database 46, and the lookup unit 42 retrieves the content-related information of the known work associated with the matching record in the media database 46. As noted above, this content-related information may include such information as song title, artist, album name, for example as well as other information such as product fulfillment information such as a sponsor or an advertiser. Box 340 is then carried out.


At box 340, the content-related information of the known work obtained in box 330 is transmitted to the client media player submitting the media sample request in box 300. In the preferred embodiment, this transmission is carried out over the Internet using the IP (HTTP) protocol or other suitable protocol. Box 370 is then carried out to generate a log.


At box 350, the lookup unit 42 has determined that the media sample of the work from box 300 does not match any reference sample in the in-memory structure 47. Box 360 is then carried out.


At box 360, a notice indicating that the requested information is not available is transmitted to the requesting client media player of box 300. Box 370 is then carried out.


At box 370 the log process is initiated. This log process involves tracking the user and related-media information associated with the request of box 300. The log unit 44 maintains a plurality of information related to the requests made by the user including, for example, media type, genre or category. This user information is maintained in the log database 48 and may be used for various marketing strategies, including analyzing user behavior, for example.


Accordingly, it will be seen that this invention provides an apparatus and method which provides real-time media content-related information to users viewing media content over a personal computer, or other data processing means. Although the description above contains many specificities, these should not be construed as limiting the scope of the invention but as merely providing an illustration of the presently preferred embodiment of the invention. Thus the scope of this invention should be determined by the appended claims and their legal equivalents.

Claims
  • 1. A method comprising: receiving, by a media player, streamed or broadcast audio or video media content of a work, wherein the media player comprises an analog broadcast receiver, and wherein at least a portion of the audio or video media content of the work is received via the analog broadcast receiver;playing, by the media player, the streamed or broadcast audio or video media content of the work;determining, by an intercept unit of an embedded system of the media player, that the audio or video media content of the work is currently being played on the media player;intercepting, by the intercept unit, audio media signals corresponding to the currently playing audio or video media content of the work;writing, by the intercept unit, the audio media signals to a first in first out (FIFO) buffer;receiving, via a user interface of the media player, a user request for information related to the work currently being played by the media player;responsive to receiving the user request, generating, by a sampling unit of the embedded system of the media player and using the audio media signals, a media sample based on a segment of the media content that is less than an entirety of the work, the media sample comprising a digital fingerprint of the segment of the media content that is based on acoustical features of the segment;sending, by the media player, a query to a lookup server for data associated with the work, the query comprising the media sample;receiving, by the media player, the data associated with said work in response to said query, the data having been identified by the lookup server based on separate comparisons of a plurality of portions of the media sample to a collection of reference samples of known works, each of the plurality of portions of the media sample having a size that is smaller than the media sample; anddisplaying the data via the user interface while the work is being played.
  • 2. The method of claim 1, further comprising: buffering the segment of the media content by the media player while the segment is played on the media player.
  • 3. The method of claim 1, further comprising: generating a log of user activity based on the received data; andoffering for purchase, by the media player, at least one of a product or a service related to said work based on said log.
  • 4. The method of claim 1, wherein said media player is at least one of a portable media player or a mobile phone.
  • 5. The method of claim 1, wherein said work includes at least one of a song, a musical audio track or a nonmusical audio track.
  • 6. A non-transitory machine-readable medium including instructions that, when executed by a media player, cause the media player to: receive, by the media player, streamed or broadcast audio or video media content of a work, wherein the media player comprises an analog broadcast receiver, and wherein at least a portion of the audio or video media content of the work is received via the analog broadcast receiver;play, by the media player, the streamed or broadcast audio or video media content of the work;determine, by an intercept unit of an embedded system of the media player, that the audio or video media content of the work is currently being played on the media player;intercept, by the intercept unit, audio media signals corresponding to the currently playing audio or video media content of the work;write, by the intercept unit, the audio media signals to a first in first out (FIFO) buffer;receive, via a user interface of the media player, a user request for information related to the work currently being played on the media player;responsive to receiving the user request, generate, by a sampling unit of the embedded system of the media player and using the audio media signals, a media sample based on a segment of the media content that is less than an entirety of the work, the media sample comprising a digital fingerprint of the segment of the media content that is based on acoustical features of the segment;send a query to a lookup server for data associated with the work, the query comprising the media sample;receive the data associated with said work from said lookup server in response to said query, the data having been identified by the lookup server based on separate comparisons of a plurality of portions of the media sample to a collection of reference samples of known works, each of the plurality of portions of the media sample having a size that is smaller than the media sample; anddisplay the data via the user interface while the work is being played.
  • 7. The non-transitory machine-readable medium of claim 6, the at least one processing unit further to: generate a log of user activity based on the received data; andoffer for purchase at least one of a product or a service related to said work based on said log.
  • 8. The non-transitory machine-readable medium of claim 6, wherein said work includes at least one of a song, a musical audio track or a nonmusical audio track.
  • 9. A media player comprising: a receiver to receive streamed or broadcast audio media content of a work, wherein the receiver comprises an analog broadcast receiver, wherein at least a portion of the audio media content of the work is received via the analog broadcast receiver, and wherein the media player is to play the streamed or broadcast audio media content of the work on a speaker;a speaker;a display;a user interface;a first in first out (FIFO) buffer; andan embedded system comprising an intercept unit and a sampling unit; wherein the intercept unit of the embedded system is to: determine that the audio media content of the work is being played;intercept audio media signals corresponding to the audio media content of the work;and write the audio media signals to the FIFO buffer;wherein responsive to the media player receiving, via the user interface, a user request for information related to the work played by the media player, the sampling unit is to use the audio media signals to generate a media sample based on a segment of the media content that is less than an entirety of the work, the media sample comprising a digital fingerprint of the segment of the media content that is based on acoustical features of the segment; andwherein the media player is further to: send a query to a lookup server for data associated with the work, the query including said digital fingerprint;receive the data associated with said work from said lookup server in response to said query, the data having been identified by the lookup server based on separate comparisons of distinct portions of the digital fingerprint to a collection of digital fingerprints of known works; andoutput, via the user interface, the data on the display while the work is played.
  • 10. The media player of claim 9, the at least one processing unit further to: generate a log of user activity based on the received data; andoffer for purchase at least one of a product or a service related to said work based on said log.
  • 11. The media player of claim 9, wherein said media player is at least one of a portable media player or a mobile phone.
  • 12. The media player of claim 9, wherein said work includes at least one of a song, a musical audio track or a nonmusical audio track.
  • 13. The method of claim 1, wherein the media player is selected from a group consisting of a car stereo, a home stereo, and a television.
  • 14. The non-transitory machine-readable medium of claim 6, wherein the media player is selected from a group consisting of a car stereo, a home stereo, and a television.
  • 15. The media player of claim 9, wherein the media player is selected from a group consisting of a car stereo, a home stereo, and a television.
RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 12/251,404, filed on Oct. 14, 2008 now U.S. Pat. No. 7,917,645, which is a continuation of U.S. patent application Ser. No. 10/955,841, filed Sep. 29, 2004, now U.S. Pat. No. 7,500,007, which is a continuation of U.S. patent application Ser. No. 09/511,632, filed on Feb. 17, 2000, now U.S. Pat. No. 6,834,308. U.S. patent application Ser. No. 12/251,404, U.S. Pat. No. 7,500,007, and U.S. Pat. No. 6,834,308 are hereby incorporated by reference.

US Referenced Citations (324)
Number Name Date Kind
2539767 Anderson Jan 1951 A
3919479 Moon et al. Nov 1975 A
4230990 Lert, Jr. et al. Oct 1980 A
4449249 Price May 1984 A
4450531 Kenyon May 1984 A
4454594 Hefron et al. Jun 1984 A
4623837 Efron Nov 1986 A
4677455 Okajima Jun 1987 A
4677466 Lert, Jr. et al. Jun 1987 A
4739398 Thomas et al. Apr 1988 A
4843562 Kenyon et al. Jun 1989 A
4918730 Schulze Apr 1990 A
5210820 Kenyon May 1993 A
5283819 Glick Feb 1994 A
5327521 Savic et al. Jul 1994 A
5437050 Lamb et al. Jul 1995 A
5442645 Ugon Aug 1995 A
5504518 Ellis Apr 1996 A
5581658 O'Hagan et al. Dec 1996 A
5588119 Vincent Dec 1996 A
5612974 Astrachan Mar 1997 A
5613004 Cooperman et al. Mar 1997 A
5638443 Stefik Jun 1997 A
5692213 Goldberg et al. Nov 1997 A
5701452 Siefert Dec 1997 A
5710916 Barbara et al. Jan 1998 A
5758257 Herz et al. Feb 1998 A
5724605 Wissner Mar 1998 A
5732193 Aberson Mar 1998 A
5790691 Narayanswamy et al. Aug 1998 A
5850388 Anderson Dec 1998 A
5862260 Rhoads Jan 1999 A
5918223 Blum Jun 1999 A
5924071 Morgan et al. Jul 1999 A
5930369 Cox et al. Jul 1999 A
5930749 Maes Jul 1999 A
5943422 Van Wie et al. Aug 1999 A
5949885 Leighton Sep 1999 A
5959659 Dokic Sep 1999 A
5983176 Hoffert et al. Nov 1999 A
6006183 Lai et al. Dec 1999 A
6006256 Zdepski et al. Dec 1999 A
6011758 Dockes Jan 2000 A
6012051 Sammon, Jr. Jan 2000 A
6026411 Delp Feb 2000 A
6026439 Chowdhury Feb 2000 A
6044402 Jacobson Mar 2000 A
6067369 Kamei May 2000 A
6067517 Bahl May 2000 A
6088455 Logan et al. Jul 2000 A
6092040 Voran Jul 2000 A
6096961 Bruti Aug 2000 A
6118450 Proehl et al. Sep 2000 A
6192340 Abecassis Feb 2001 B1
6195693 Berry Feb 2001 B1
6243615 Neway Jun 2001 B1
6243725 Hempleman et al. Jun 2001 B1
6253193 Ginter Jun 2001 B1
6253337 Maloney et al. Jun 2001 B1
6279010 Anderson Aug 2001 B1
6279124 Brouwer et al. Aug 2001 B1
6285596 Miura et al. Sep 2001 B1
6330593 Roberts Dec 2001 B1
6345256 Milsted et al. Feb 2002 B1
6345274 Zhu Feb 2002 B1
6360265 Falck et al. Mar 2002 B1
6363381 Lee Mar 2002 B1
6370513 Kolawa Apr 2002 B1
6374260 Hoffert et al. Apr 2002 B1
6385596 Wiser May 2002 B1
6418421 Hurtado et al. Jul 2002 B1
6422061 Sunshine Jul 2002 B1
6425081 Iwamura Jul 2002 B1
6434520 Kanevsky Aug 2002 B1
6438556 Malik et al. Aug 2002 B1
6449226 Kumagai Sep 2002 B1
6452874 Otsuka et al. Sep 2002 B1
6453252 Laroche Sep 2002 B1
6460050 Pace et al. Oct 2002 B1
6463508 Wolf et al. Oct 2002 B1
6477704 Cremia Nov 2002 B1
6484182 Dunphy Nov 2002 B1
6487641 Cusson et al. Nov 2002 B1
6490279 Chen et al. Dec 2002 B1
6496802 van Zoest et al. Dec 2002 B1
6526411 Ward Feb 2003 B1
6542869 Foote Apr 2003 B1
6550001 Corwin et al. Apr 2003 B1
6550011 Sims, III Apr 2003 B1
6570991 Scheirer May 2003 B1
6591245 Klug Jul 2003 B1
6609093 Gopinath et al. Aug 2003 B1
6609105 van Zoest et al. Aug 2003 B2
6628737 Timus Sep 2003 B1
6636965 Beyda Oct 2003 B1
6654757 Stern Nov 2003 B1
6675174 Bolle et al. Jan 2004 B1
6690835 Brockmeyer Feb 2004 B1
6714921 Stefik Mar 2004 B2
6732180 Hale May 2004 B1
6735699 Sasaki et al. May 2004 B1
6763069 Divakaran Jul 2004 B1
6771885 Agnihotri et al. Aug 2004 B1
6788800 Carr Sep 2004 B1
6834308 Ikezoye Dec 2004 B1
6868440 Gupta et al. Mar 2005 B1
6947909 Hoke, Jr. Sep 2005 B1
6961858 Fransdonk Nov 2005 B2
6968337 Wold Nov 2005 B2
6990453 Wang Jan 2006 B2
7020635 Hamilton Mar 2006 B2
7043536 Philyaw May 2006 B1
7058223 Cox et al. Jun 2006 B2
7080049 Truitt Jul 2006 B2
7090128 Farley Aug 2006 B2
7107462 Fransdonk Sep 2006 B2
7150045 Koelle Dec 2006 B2
7181398 Thong et al. Feb 2007 B2
7194752 Kenyon Mar 2007 B1
7228293 DeTreville Jun 2007 B2
7263205 Lev Aug 2007 B2
7266645 Garg et al. Sep 2007 B2
7269556 Kiss et al. Sep 2007 B2
7281272 Rubin et al. Oct 2007 B1
7289643 Brunk Oct 2007 B2
7334720 Hulst Feb 2008 B2
7349552 Levy et al. Mar 2008 B2
7363278 Schmelzer et al. Apr 2008 B2
7424747 DeTreville Sep 2008 B2
7426750 Cooper et al. Sep 2008 B2
7443797 Cheung et al. Oct 2008 B2
7474759 Sternberg Jan 2009 B2
7500007 Ikezoye Mar 2009 B2
7529659 Wold May 2009 B2
7546120 Ulvenes et al. Jun 2009 B1
7562012 Wold Jul 2009 B1
7565327 Schmelzer Jul 2009 B2
7587502 Crawford Sep 2009 B2
7593576 Meyer et al. Sep 2009 B2
7613686 Rui Nov 2009 B2
7630562 Gong Dec 2009 B2
7653210 Rhoads Jan 2010 B2
7701941 O'Callaghan et al. Apr 2010 B2
7707088 Schmelzer et al. Apr 2010 B2
7711586 Aggarwal May 2010 B2
7711652 Schmelzer et al. May 2010 B2
7770013 Rhoads Aug 2010 B2
7797249 Schmelzer et al. Sep 2010 B2
7853664 Wang Dec 2010 B1
7877438 Schrempp et al. Jan 2011 B2
7917645 Ikezoye Mar 2011 B2
7942317 Racz May 2011 B2
8006314 Wold Aug 2011 B2
8033458 Hulst Oct 2011 B2
8061598 Racz Nov 2011 B2
8082150 Wold Dec 2011 B2
8086445 Wold Dec 2011 B2
8112818 Wold Feb 2012 B2
8122339 Bastos dos Santos Feb 2012 B2
8130746 Schrempp et al. Mar 2012 B2
8199651 Schrempp et al. Jun 2012 B1
8316238 Mergen Nov 2012 B2
8332326 Schrempp et al. Dec 2012 B2
8458156 Sharifi Jun 2013 B1
8472669 Sharma Jun 2013 B2
8484691 Schmelzer et al. Jul 2013 B2
8639178 Anniballi Jan 2014 B2
8645279 Schmelzer Feb 2014 B2
8732858 Wold May 2014 B2
8775317 Schmelzer Jul 2014 B2
8843952 Pora Sep 2014 B2
8862795 Castro Scorsi Oct 2014 B2
8886635 Cho Nov 2014 B2
8972481 Schrempp et al. Mar 2015 B2
9049468 Ikezoye Jun 2015 B2
9081778 Garside et al. Jul 2015 B2
9268921 Wold Feb 2016 B2
9374183 Anniballi Jun 2016 B2
9411050 Weber Aug 2016 B1
9461759 Anniballi Oct 2016 B2
9589141 Schmelzer Mar 2017 B2
9608824 Garside et al. Mar 2017 B2
9785757 Wold Oct 2017 B2
9949044 Choi, II Apr 2018 B1
20010013061 DeMartin Aug 2001 A1
20010027493 Wallace Oct 2001 A1
20010027522 Saito Oct 2001 A1
20010034219 Hewitt et al. Oct 2001 A1
20010037304 Paiz Nov 2001 A1
20010041989 Vilcauskas et al. Nov 2001 A1
20010051996 Cooper et al. Dec 2001 A1
20010056430 Yankowki Dec 2001 A1
20020002899 Gjerdingen Jan 2002 A1
20020019858 Kaiser Feb 2002 A1
20020023020 Kenyon Feb 2002 A1
20020023220 Kaplan Feb 2002 A1
20020037083 Weare Mar 2002 A1
20020042754 Del Beccaro et al. Apr 2002 A1
20020049760 Scott Apr 2002 A1
20020064149 Elliott et al. May 2002 A1
20020069098 Schmidt Jun 2002 A1
20020072325 Hayter Jun 2002 A1
20020073316 Collins Jun 2002 A1
20020076051 Nii Jun 2002 A1
20020082999 Lee Jun 2002 A1
20020083060 Wang Jun 2002 A1
20020087885 Peled et al. Jul 2002 A1
20020120577 Hans et al. Aug 2002 A1
20020123990 Abe et al. Sep 2002 A1
20020129140 Peled et al. Sep 2002 A1
20020133494 Goedken Sep 2002 A1
20020133499 Ward Sep 2002 A1
20020141384 Liu et al. Oct 2002 A1
20020152262 Arkin et al. Oct 2002 A1
20020156737 Kahn Oct 2002 A1
20020157005 Brunk Oct 2002 A1
20020158737 Yokoyama Oct 2002 A1
20020178410 Haitsma Nov 2002 A1
20020184517 Tadayon Dec 2002 A1
20020186887 Rhoads Dec 2002 A1
20020198789 Waldman Dec 2002 A1
20030014530 Bodin Jan 2003 A1
20030018709 Schrempp et al. Jan 2003 A1
20030023852 Wold Jan 2003 A1
20030033321 Schrempp et al. Feb 2003 A1
20030037010 Schmelzer et al. Feb 2003 A1
20030061352 Bohrer Mar 2003 A1
20030051100 Patel May 2003 A1
20030095660 Lee et al. May 2003 A1
20030105739 Essafi Jun 2003 A1
20030135623 Schrempp et al. Jul 2003 A1
20030191719 Ginter et al. Oct 2003 A1
20030191764 Richards Oct 2003 A1
20030195852 Campbell et al. Oct 2003 A1
20030223554 Zhang Dec 2003 A1
20040008864 Watson et al. Jan 2004 A1
20040010495 Kramer et al. Jan 2004 A1
20040028281 Cheng Feb 2004 A1
20040053654 Kokumai et al. Mar 2004 A1
20040073513 Stefik et al. Apr 2004 A1
20040089142 Georges et al. May 2004 A1
20040133797 Arnold Jul 2004 A1
20040148191 Hoke, Jr. Jul 2004 A1
20040163106 Schrempp et al. Aug 2004 A1
20040167858 Erickson Aug 2004 A1
20040199387 Wang Oct 2004 A1
20040201784 Dagtas et al. Oct 2004 A9
20050021783 Ishii Jan 2005 A1
20050038819 Hicken Feb 2005 A1
20050039000 Erickson Feb 2005 A1
20050044189 Ikezoye et al. Feb 2005 A1
20050097059 Shuster May 2005 A1
20050141707 Haitsma Jun 2005 A1
20050154678 Schmelzer Jul 2005 A1
20050154680 Schmelzer Jul 2005 A1
20050154681 Schmelzer Jul 2005 A1
20050216433 Bland et al. Sep 2005 A1
20050267945 Cohen et al. Dec 2005 A1
20050289065 Weare Dec 2005 A1
20060034177 Schrempp Feb 2006 A1
20060062426 Levy et al. Mar 2006 A1
20060118619 Hulst Jun 2006 A1
20070033409 Brunk Feb 2007 A1
20070074147 Wold Mar 2007 A1
20070078769 Way et al. Apr 2007 A1
20070124756 Covell May 2007 A1
20070186229 Conklin et al. Aug 2007 A1
20070226365 Hildreth et al. Sep 2007 A1
20070271248 Albernoz Nov 2007 A1
20080008173 Kanevsky et al. Jan 2008 A1
20080019371 Anschutz et al. Jan 2008 A1
20080133415 Ginter et al. Jun 2008 A1
20080141379 Schmelzer Jun 2008 A1
20080154730 Schmelzer Jun 2008 A1
20080155116 Schmelzer Jun 2008 A1
20080250080 Arrasvuori et al. Oct 2008 A1
20080263579 Mears Oct 2008 A1
20080320545 Schwartz Dec 2008 A1
20090019468 Ganesan Jan 2009 A1
20090030651 Wold Jan 2009 A1
20090031326 Wold Jan 2009 A1
20090043870 Ikezoye et al. Feb 2009 A1
20090063974 Bull Mar 2009 A1
20090131152 Busse May 2009 A1
20090132391 Jacobs May 2009 A1
20090192640 Wold Jul 2009 A1
20090240361 Wold et al. Sep 2009 A1
20090306966 Hejna, Jr. Dec 2009 A1
20090328236 Schmelzer Dec 2009 A1
20100042843 Brunk Feb 2010 A1
20100061560 Cronin Mar 2010 A1
20100104259 Shakya et al. Apr 2010 A1
20100281042 Windes et al. Nov 2010 A1
20100290667 Lienhart et al. Nov 2010 A1
20100290867 Leinhart et al. Nov 2010 A1
20110069230 Polumbus Mar 2011 A1
20110078719 Kenyon Mar 2011 A1
20110114723 Hulst May 2011 A1
20110119149 Ikezoye May 2011 A1
20120124679 Wold May 2012 A1
20120223135 Meek Sep 2012 A1
20120318071 Biehl et al. Dec 2012 A1
20130011008 Ikezoye Jan 2013 A1
20130159021 Felsher Jun 2013 A1
20130276138 Schmelzer Oct 2013 A1
20130318071 Cho Nov 2013 A1
20130318114 Emerson, III Nov 2013 A1
20140056433 Emerson, III Feb 2014 A1
20140089307 Garside et al. Mar 2014 A1
20140115716 Schmelzer Apr 2014 A1
20140215643 Wold Jul 2014 A1
20140335834 Emerson, III Nov 2014 A1
20140336798 Emerson, III Nov 2014 A1
20140336799 Emerson, III Nov 2014 A1
20140357234 Sullivan Dec 2014 A1
20150078575 Selig Mar 2015 A1
20150154273 Schrempp et al. Jun 2015 A1
20150234814 Ikezoye Aug 2015 A1
20150270976 Garside et al. Sep 2015 A1
20160132664 Wold May 2016 A1
20160203824 Park Jul 2016 A1
20170193104 Garside et al. Jul 2017 A1
20170366847 Covell Dec 2017 A1
20180032705 Wold Feb 2018 A1
Foreign Referenced Citations (40)
Number Date Country
0349106 Jan 1990 EP
0402210 Jun 1990 EP
0459046 Dec 1991 EP
0517405 May 1992 EP
0689316 Dec 1995 EP
0731446 Sep 1996 EP
0859503 Aug 1998 EP
0459046 Apr 1999 EP
01449103 Aug 2004 EP
01485815 Dec 2004 EP
1492767 Sep 2005 EP
01593018 Nov 2005 EP
1354276 Dec 2007 EP
1485815 Oct 2009 EP
177191 Feb 2012 EP
1449103 Mar 2012 EP
2464049 Dec 2012 GB
WO 9636163 Nov 1996 WO
WO 9820672 May 1998 WO
WO 0005650 Feb 2000 WO
WO 0039954 Jul 2000 WO
WO 0063800 Oct 2000 WO
WO 0123981 Apr 2001 WO
WO 2001062004 Aug 2001 WO
WO 0203203 Jan 2002 WO
WO 0215035 Feb 2002 WO
WO 2002027600 Apr 2002 WO
WO 02037316 May 2002 WO
02086803 Oct 2002 WO
WO 02082271 Oct 2002 WO
WO 03007235 Jan 2003 WO
WO 03009149 Jan 2003 WO
WO 03036496 May 2003 WO
WO 03067459 Aug 2003 WO
WO 03091990 Nov 2003 WO
WO 04044820 May 2004 WO
WO 2003007235 Jul 2004 WO
WO 04070558 Aug 2004 WO
WO 06015168 Feb 2006 WO
WO 09017710 Feb 2009 WO
Non-Patent Literature Citations (212)
Entry
Patel et al (Video classification using speaker identification) (Year: 1997).
Baum, L., et al., “A Maximization Technique Occurring in the Statistical Analysis of Probabilistic Functions of Markov Chains,” The Annals of Mathematical Statistics, vol. 41, No. 1, pp. 164-171, 1970.
Beritelli, F., et al., “Multilayer Chaotic Encryption for Secure Communications in packet switching Networks,” IEEE, vol. 2Aug. 2000, pp. 1575-1582.
Blum, T., Keislar, D., Wheaton, J., and Wold, E., “Audio Databases with Content-Based Retrieval,” Proceedings of the 1995 International Joint Conference on Artificial Intelligence (IJCAI) Workshop on Intelligent Multimedia Information Retrieval, 1995.
Breslin, Pat, et al., Relatable Website, “Emusic uses Relatable's open source audio recognition solution, TRM, to signature its music catabblog for MusicBrainz database,” http://www.relatable.com/news/pressrelease/001017.release.html, Oct. 17, 2000.
Business Wire, “Cisco and Fox Host Groundbreaking Screening of Titan A.E.; Animated Epic Will Be First Film Ever to be Digitaly Transmitted Over the Internet Monday,” Jun. 5, 2000, 08:14 EDT.
Business Wire, “IBM: IBM Announces New Descrambler Software; First to Offer Software to Work With Digital Video Chips,” Jun. 5, 25, 1997, 07:49.
Chen, et al., Yong-Cong, A Secure and Robust Digital Watermaking Technique by the Blcok Cipher RC6 and Secure Hash Algorithm, Deparment of Computer Science, National Tsing Hua University, 2001.
Cosi, P., De Poli, G., Prandoni, P., “Timbre Characterization with Mel-Cepstrum and Neural Nets,” Proceedings of the 1994 International Computer Music Conference, pp. 42-45, San Francisco, No date.
Dempster, A.P., et al., “Maximum Likelihood from Incomplete Data via the EM Algorithm” Journal of the Royal Statistical Society, Series B (Methodological), vol. 39, Issue 1, pp. 31-38, 1977.
Feiten, B. and Gunzel, S., “Automatic Indexing of a Sound Database Using Self-Organizing Neural Nets,” Computer Music Journal, 18:3, pp. 53-65, Fall 1994.
Fischer, S., Lienhart, R., and Effelsberg, W., “Automatic Recognition of Film Genres,” Reihe Informatik, Jun. 1995, Universitat Mannheim, Praktische Informatik IV, L15, 16, D-68131 Mannheim.
Foote, J., “A Similarity Measure for Automatic Audio Classification,” Institute of Systems Science, National University of Singapore, 1977, Singapore.
Gasaway Laura, Close of Century Sees New Copyright Amendments, Mar. 200, Information Outlook, 4, 3, 42.
Gonzalez, R. and Melih, K., “Content Based Retrieval of Audio,” The Institute for Telecommunication Research, University of Wollongong, Australia, No date.
Haitsma, J., et al., “Robust Audio Hashing for Content Identification”, CBMI 2001, Second International Workshop on Content Based Multimedia and Indexing, Brescia, Italy, Sep. 19-21, 2001.
Harris, Lesley Ellen, “To register or not,” Mar. 2006, Information Outlook, 10, 3, 32(s).
Kanth, K.V. et al. “Dimensionality Reduction or Similarity Searching in Databases,” Computer Vision and Image understanding, vol. 75, Nos. 1/2 Jul./Aug. 1999, pp. 59-72, Academic Press. Santa Barbara, CA, USA.
Keislar, D., Blum, T., Wheaton, J., and Wold, E., “Audio Analysis for Content-Based Retrieval” Proceedings of the 1995 International Computer Music Conference.
Ohtsuki, K., et al. , “Topic extraction based on continuous speech recognition in broadcast-news speech,” Proceedings IEEE Workshop on Automated Speech Recognition and Understanding, 1997, pp. 527-534, N.Y., N.Y., USA.
Packethound Tech Specs, www.palisdesys.com/products/packethount/tck specs/prodPhtechspecs.shtml, 2002.
“How does PacketHound work?”, www.palisdesys.com/products/packethound/how_does_it_work/prod_Pghhow.shtml 2002.
Pankanti, Sharath, “Verification Watermarks on Fingerprint Recognition and Retrieval,” Part of IS&T/SPIE Conference on Security and Watermarking of Multimedia Contents, San Jose, CA Jan. 1999, SPIE vol. 3657, pp. 66-78.
Pellom, B. et al., “Fast Likelihood Computation Techniques in Nearest-Neighbor search for Continuous Speech Recognition.”, IEEE Signal Processing Letters, vol. 8, pp. 221-224 Aug. 2001.
Reynolds, D., et al. , “Robust Text-Independent Speaker Identification Using Gaussian Mixture Speaker Models”, IEEE Transactions on Speech and Audio Processing, vol. 3, No. 1, pp. 72-83 Jan. 1995.
Scheirer, E., Slaney, M., “Construction and Evaluation of a Robust Multifeature Speech/Music Discriminator,” pp. 1-4, Proceedings of ICASSP-97, Apr. 2-24, Munich, Germany.
Scheirer, E.D., “Tempo and Beat Analysis of Acoustic Musical Signals,” Machine Listening Group, E15-401D MIT Media Laboratory, pp. 1-21, Aug. 8, 1997, Cambridge, MA.
Smith, Alan J., “Cache Memories,” Computer Surveys, Sep. 1982, University of California, Berkeley, California, vol. 14, No. 3, pp. 1-61.
Vertegaal, R. and Bonis, E., “ISEE: An Intuitive Sound Editing Environment,” Computer Music Journal, 18:2, pp. 21-22, Summer 1994.
Wang, Yao, et al., “Multimedia Content Analysis,” IEEE Signal Processing Magazine, pp. 12-36, Nov. 2000, IEEE Service Center, Piscataway, N.J., USA.
Wold, Erling, et al., “Content Based Classification, Search and Retrieval of Audio,” IEEE Multimedia, vol. 3, No. 3, pp. 27-36, 1996 IEEE Service Center, Piscataway, N.J., USA.
Zawodny, Jeremy, D., “A C Program to Compute CDDB discids on Linus and FreeBSD,” [internet] http://jeremy.zawodny.com/c/discid-linux-1.3tar.gz, 1 page, Apr. 14, 2001, retrieved Jul. 17, 2007.
European Patent Application No. 02752347.1, Supplementary European Search Report dated May 8, 2006, 4 pages.
European Patent Application No. 02756525.8, Supplementary European Search Report dated Jun. 28, 2006, 4 pages.
European Patent Application No. 02782170, Supplementary European Search Report dated Feb. 7, 2007, 4 pages.
European Patent Application No. 02725522.3, Supplementary European Search Report dated May 12, 2006, 2 Pages.
European Patent Application No. 04706547.9 European Search Report dated Feb. 25, 2010, 3 Pages.
European Patent Application No. 05778109.8 European Search Report dated Sep. 10, 2010, 7 Pages.
PCT Search Report PCT/US01/50295, International Search Report dated May 14, 2003, 5 Pages.
PCT Search Report PCT/US02/10615, International Search Report dated Aug. 7, 2002, 5 Pages.
PCT Search Report PCT/US02/33186, International Search Report dated Dec. 16, 2002, 6 Pages.
PCT Search Report PCT/US04/02748, International Search Report and Written Opinion dated Aug. 20, 2007, 8 Pages.
PCT Search Report PCT/US05/26887, International Search Report dated May 3, 2006, 3 Pages.
PCT Search Report PCT/US08/09127, International Search Report dated Oct. 30, 2008, 8 Pages.
Audible Magic Office Action for U.S. Appl. No. 09/511,632 dated Dec. 4, 2002.
Audible Magic Office Action for U.S. Appl. No. 09/511,632 dated May 13, 2003.
Audible Magic Office Action for U.S. Appl. No. 09/511,632 dated Aug. 27, 2003.
Audible Magic Office Action for U.S. Appl. No. 09/511,632 dated Feb. 5, 2004.
Audible Magic Notice of Allowance for U.S. Appl. No. 09/511,632 dated Aug. 10, 2004.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/955,841 dated Sep. 25, 2006.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/955,841 dated Mar. 23, 2007.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/955,841 dated Sep. 11, 2007.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/955,841 dated Feb. 25, 2008.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/251,404 dated May 14, 2010.
Audible Magic Office Action for U.S. Appl. No. 08/897,662 dated Aug. 13, 1998.
Audible Magic Notice of Allowance for U.S. Appl. No. 08/897,662 dated Jan. 29, 1999.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated May 5, 2004.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Nov. 12, 2004.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated May 9, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Nov. 1, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Jun. 23, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Nov. 7, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Mar. 29, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Sep. 17, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated May 29, 2008.
Audible Magic Office Action for U.S. Appl. No. 09/706,227 dated Jan. 9, 2009.
Audible Magic Office Action for U.S. Appl. No. 10/192,783 dated Dec. 13, 2004.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/192,783 dated Jun. 7, 2005.
Audible Magic Office Action for U.S. Appl. No. 11/239,543 dated Apr. 23, 2008.
Audible Magic Notice of Allowance for U.S. Appl. No. 11/239,543 dated Nov. 6, 2008.
Audible Magic Notice of Allowance for U.S. Appl. No. 11/239,543 dated Feb. 25, 2009.
Audible Magic Office Action for U.S. Appl. No. 12/410,445 dated Aug. 10, 2010.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/410,445 dated Oct. 20, 2010.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Nov. 17, 2004.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated May 16, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Sep. 29, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Jun. 23, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Aug. 8, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Jan. 25, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/910,680 dated Dec. 5, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Apr. 6, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Oct. 6, 2005.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Apr. 7, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Oct. 6, 2006.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Mar. 7, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Aug. 20, 2007.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Jan. 7, 2008.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Jun. 27, 2008.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Dec. 22, 2008.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Jul. 20, 2009.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Dec. 21, 2009.
Audible Magic Office Action for U.S. Appl. No. 09/999,763 dated Jun. 23, 2010.
Audible Magic Notice of Allowance for U.S. Appl. No. 09/999,763 dated Sep. 16, 2010.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated May 3, 2005.
Audible Magic Notice of Allowance for U.S. Appl. No. 10/072,238 dated May 3, 2005.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Oct. 25, 2005.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Apr. 25, 2006.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Sep. 19, 2007.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Apr. 7, 2008.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Oct. 1, 2008.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Jan. 9, 2009.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Mar. 31, 2009.
Audible Magic Office Action for U.S. Appl. No. 10/072,238 dated Aug. 6, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Dec. 13, 2004.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Apr. 8, 2005.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Oct. 7, 2005.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Apr. 20, 2006.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Jul. 31, 2006.
Audible Magic Office Action for U.S. Appl. No. 11/116,710 dated Jan. 16, 2007.
Audible Magic Notice of Allowance for U.S. Appl. No. 11/116,710 dated Nov. 19, 2007.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/042,023 dated Dec. 29, 2008.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/042,023 dated Jun. 25, 2009.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/042,023 dated Mar. 8, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/048,307 dated Aug. 22, 2007.
Audible Magic Office Action for U.S. Appl. No. 11/048,307 dated May 16, 2008.
Audible Magic Notice of Allowance for U.S. Appl. No. 11/048,307 dated May 29, 2009.
Audible Magic Office Action for U.S. Appl. No. 12/488,504 dated Nov. 10, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/048,308 dated Feb. 25, 2008.
Audible Magic Office Action for U.S. Appl. No. 11/048,308 dated Mar. 5, 2009.
Audible Magic Notice of Allowance for U.S. Appl. No. 11/048,308 dated Aug. 7, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Apr. 18, 2007.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Oct. 11, 2007.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Jan. 14, 2008.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Jul. 9, 2008.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Jan. 7, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Jul. 6, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Dec. 28, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/048,338 dated Jun. 24, 2010.
Audible Magic Office Action for U.S. Appl. No. 12/035,599 dated Nov. 17, 2008.
Audible Magic Office Action for U.S. Appl. No. 12/035,599 dated May 29, 2009.
Audible Magic Office Action for U.S. Appl. No. 12/035,599 dated Nov. 24, 2009.
Audible Magic Office Action for U.S. Appl. No. 12/035,599 dated Jun. 9, 2010.
Audible Magic Office Action for U.S. Appl. No. 12/035,609 dated Dec. 29, 2008.
Audible Magic Office Action for U.S. Appl. No. 12/035,609 dated Jun. 24, 2009.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/035,609 dated Dec. 11, 2009.
Audible Magic Notice of Allowance for U.S. Appl. No. 12/277,291 dated May 12, 2010.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated May 24, 2006.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Nov. 2, 2006.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Apr. 11, 2007.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Nov. 1, 2007.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated May 9, 2008.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Jan. 6, 2009.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Jun. 15, 2009.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Jan. 21, 2010.
Audible Magic Office Action for U.S. Appl. No. 10/356,318 dated Jan. 7, 2011.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated Jul. 17, 2008.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated Jan. 9, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated Apr. 28, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated Nov. 19, 2009.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated May 25, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/191,493 dated Oct. 4, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/829,662 dated Oct. 8, 2010.
Audible Magic Office Action for U.S. Appl. No. 11/923,491 dated Nov. 12, 2010.
Albert, Douglas, et al., “Combatting Software Piracy by encryption and key management,” IEEE Computer Society, vol. 17. No. 4, Apr. 1984, 6 pages.
Audible Magic Corporation, “Audio Identification Technology Provides the Cornerstone for Online Distribution,” 2000, http://www.audiblemagic.com/documents/Technology_Summary.pdf.
USPTO Office Action for U.S. Appl. No. 12/482,313 dated Feb. 4, 2011.
USPTO Notice of Allowance for U.S. Appl. No. 12/482,313 dated Aug. 23, 2011.
USPTO Notice of Allowance for U.S. Appl. No. 12/410,445 dated Aug. 12, 2011.
USPTO Office Action for U.S. Appl. No. 09/910,680 dated May 21, 2013.
USPTO Office Action for U.S. Appl. No. 12/488,504 dated Apr. 26, 2013.
USPTO Office Action for U.S. Appl. No. 12/035,599 dated Jun. 9, 2010.
USPTO Office Action for U.S. Appl. No. 12/035,599 dated Aug. 7, 2012.
USPTO Notice of Allowance for U.S. Appl. No. 12/035,599 dated Mar. 11, 2013.
USPTO Office Action for U.S. Appl. No. 10/356,318 dated Jun. 17, 2011.
USPTO Notice of Allowance for U.S. Appl. No. 10/356,318 dated Oct. 16, 2012.
USPTO Notice of Allowance for U.S. Appl. No. 11/191,493 dated Feb. 17, 2011.
USPTO Notice of Allowance for U.S. Appl. No. 11/923,491 dated Sep. 29, 2011.
USPTO Office Action for U.S. Appl. No. 13/355,424 dated Jan. 18, 2013.
USPTO Office Action for U.S. Appl. No. 13/355,424 dated May 24, 2013.
USPTO Office Action for U.S. Appl. No. 12/405,174 dated Mar. 2, 2011.
USPTO Office Action for U.S. Appl. No. 12/405,174 dated Sep. 9, 2011.
USPTO Office Action for U.S. Appl. No. 13/011,776 dated Jul. 10, 2013.
USPTO Office Action for U.S. Appl. No. 12/488,504 dated May 23, 2013.
USPTO Notice of Allowance for U.S. Appl. No. 12/488,504 dated Sep. 13, 2013.
USPTO Notice of Allowance for U.S. Appl. No. 11/829,662 dated Apr. 11, 2011.
USPTO Office Action for U.S. Appl. No. 13/915,511 dated Aug. 23, 2013.
Lin, et al., “Generating Robust Digital Signature for Image/Video Authentication,” Proc. Multimedia and Security workshop at ACM Multimedia'98, Sep. 1, 1998, pp. 49-54.
USPTO Office Action for U.S. Appl. No. 13/011,776 dated Feb. 5, 2013.
USPTO Office Action for U.S. Appl. No. 13/011,776 dated Jul. 19, 2013.
USPTO Office Action for U.S. Appl. No. 13/011,776 dated Dec. 13, 2013.
USPTO Office Action for U.S. Appl. No. 09/910,680 dated Dec. 31, 2013.
USPTO Notice of Allowance for U.S. Appl. No. 13/915,511 dated Feb. 18, 2014.
USPTO Notice of Allowance for U.S. Appl. No. 13/355,424 dated Jan. 3, 2014.
USPTO Notice of Allowance for U.S. Appl. No. 12/405,174 dated Jan. 4, 2012.
International Search Report & Written Opinion dated Apr. 15, 2014, for PCT/US2013/061750.
USPTO Office Action for U.S. Appl. No. 13/620,144 dated May 29, 2014.
USPTO Notice of Allowance for U.S. Appl. No. 09/910,680 dated Nov. 5, 2014.
USPTO Office Action for U.S. Appl. No. 13/626,709 dated Oct. 12, 2014.
USPTO Office Action for U.S. Appl. No. 14/245,630 dated Jun. 18, 2015.
USPTO Notice of Allowance for U.S. Appl. No. 14/245,630 dated Sep. 24, 2015.
USPTO Office Action for U.S. Appl. No. 14/139,643 dated Dec. 2, 2014.
Kunio Kashino et al, “Robust Search Methods for Music Signals Based on Simple Representation”, 2007 IEEE International Conference on Acoustics, Speech, and Signal Processing Apr. 15-20, 2007 Honolulu, HI, USA, Piscataway, NJ, USA, (Oct. 23, 2007), doi:10.1109/ICASSP.2007.367346, ISBN 978-1-4244-0727-9, pp. IV-1421, XP055296602 [X] 1-15.
Jaap Haitsma et al, “A Highly Robust Audio Fingerprinting System With an Efficient Search Strategy”, Journal of New Music Research, Lisse, doi:10.1076/jnmr.32.2.211.16746, (Oct. 17, 2002), pp. 211-221, URL: http://www.cs.northwestern.edu/—pardo/courses/casa2009/casa_papers_2009/papers/audio fingerprint-haitsma.pdf, (Mar. 18, 2016), XP055259740 [X] 1-15.
Cong Jin et al, “Applications of digital fingerprinting and digital watermarking for E-commerce security mechanism”, Audio, Language and Image Processing, 2008. ICALIP 2008. International Conference on, IEEE, Piscataway, NJ, USA, (Jul. 7, 2008), ISBN 978-1-4244-1723-0, pp. 536-540, XP031298539 [A] 1-15.
USPTO Non-Final Office Action for U.S. Appl. No. 14/139,643 dated Jun. 7, 2016.
USPTO Non-Final Office Action for U.S. Appl. No. 14/733,578 dated Mar. 23, 2016.
USPTO Notice of Allowance for U.S. Appl. No. 14/733,578 dated Nov. 8, 2016.
USPTO Notice of Allowance for U.S. Appl. No. 14/139,643 dated Oct. 21, 2016.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Feb. 6, 2013.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Jul. 10, 2013.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Dec. 10, 2013.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Jun. 5, 2014.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Jan. 28, 2015.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Aug. 17, 2015.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Mar. 16, 2016.
USPTO Office Action for U.S. Appl. No. 14/610,983 dated Nov. 7, 2016.
White, R., “How Computers Work,” Oct. 15, 2003, Que Publishing, 7th Ed., 23 pages.
European Patent Application No. 13840744.0 extended European Search Report dated Sep. 5, 2016, 9 pages.
USPTO Non-Final Office Action for U.S. Appl. No. 14/610,983 dated May 2, 2017.
USPTO Final Office Action for U.S. Appl. No. 14/610,983 dated Nov. 1, 2017.
USPTO Non-Final Office Action for U.S. Appl. No. 14/699,850 dated May 3, 2017.
USPTO Final Office Action for U.S. Appl. No. 14/699,850 dated Nov. 15, 2017.
Related Publications (1)
Number Date Country
20110119149 A1 May 2011 US
Continuations (3)
Number Date Country
Parent 12251404 Oct 2008 US
Child 13011776 US
Parent 10955841 Sep 2004 US
Child 12251404 US
Parent 09511632 Feb 2000 US
Child 10955841 US