This invention generally relates to the matching of unknown media data, such as video or audio segments, against a massive database of reference media files.
Systems for automatic content recognition (ACR) of audio or video media are well known to persons skilled in the art. However, such ACR systems pose many technical challenges, including managing potentially very large databases of encoded audio or video information as well as managing large indices needed for addressing information in said databases.
Also well known to those skilled in the art, is that large database indices such as may be used by this invention, can be generated using certain hashing functions. Another method of addressing a database might be by applying binary tree structures also known as b-trees. Both methods are commonly utilized in data management systems.
Whatever the method employed to index a large database, said index is often too large to reside in its entirety in the main memory of a computer server as used in a typical ACR system. When said database cannot fit entirely in the memory of a computer system, it is typically stored on magnetic disk storage and parts of said database are then read into memory in blocks corresponding to the index value providing the address. Said means of recalling partial database information is also known to one skilled in the art as “paging” which is a process common to many different computer software systems.
The present invention is an extension of the invention referenced above and is a system and method for matching unknown digital media such as television programming to a database of known media using a signal processing means employing a modified path pursuit algorithm, as described in the first invention.
Another novel aspect of the system and method as disclosed herein is its distance associative hash indexing means which can be subdivided into a plurality of independently addressable segments where each of said segments can address a portion of the database each of which can reside in its entirety in the main memory of a server means. The resulting cluster of servers of the indexing means each hosts a sector of the index addressing associated data of a larger database of searchable audio or video information. This indexing means of the invention results in a significant improvement in the speed and accuracy of the ACR system so enabled as to identify unknown media even when the television display is showing content where a user is changing channels, rewinding, fast-forwarding or even pausing video from a digital video recorder.
In some embodiments, an exemplary method related to addressing a media database using distance associative hashing may include receiving one or more indications of a sample of a video segment; determining, for at least one patch of the sample of a video segment including at least one or more pixels of the at least one patch, an algorithmically-derived value of the one or more pixels of each patch; subtracting a median point value established for each patch from the mean value for each patch; transforming the values resulting from the subtraction using a function pre-derived to distribute the values evenly; constructing a hash value from the transformed values; referencing a number of most significant bits of the constructed hash value to determine a database sector; and storing at least the hash value on the determined database sector.
In some embodiments, at least one of the receiving, determining, subtracting, transforming, constructing, referencing, or storing of the foregoing exemplary method is at least partially implemented using one or more processing devices. In some embodiments of the foregoing exemplary method, receiving one or more indications of a sample of a video segment may include receiving one or more indications of at least one of a frame or a still image. In some embodiments of the foregoing exemplary method, receiving one or more indications of a sample of a video segment may include receiving one or more indications of a sample of a video segment, the one or more indications of a sample of a video segment associated with at least one indication of a channel, at least one indication of a video segment, and at least one indication of a timecode offset from the beginning of the video segment.
In some embodiments of the foregoing exemplary method, determining, for at least one patch of the sample of a video segment including at least one or more pixels of the at least one patch, an algorithmically-derived value of the one or more pixels of each patch includes at least determining, for at least one patch of the sample of a video segment including at least one or more pixels of the at least one patch, a mean value of the one or more pixels of each patch. In some embodiments of the foregoing exemplary method, subtracting a median point value established for each patch from the mean value for each patch may include subtracting a median point value established for each patch from the mean value for each patch, the median point value established for each patch having been previously determined utilizing data from each patch for a plurality of channels over at least one period of time.
In some embodiments of the foregoing exemplary method, transforming the values resulting from the subtraction using a function pre-derived to distribute the values evenly may include forming a variable matrix including at least the values resulting from the subtraction; obtaining a static matrix which, when crossed with the variable matrix, will more evenly distribute the transformed values; and computing a dot product of the variable matrix and the static matrix, the dot product including at least the more evenly-distributed transformed values. In some embodiments of the foregoing exemplary method, obtaining a static matrix which, when crossed with the variable matrix, will more evenly distribute the transformed values may include determining, using locality-sensitive hashing at least partially based on one or more previously obtained hash values, a static matrix which, when crossed with a variable matrix, will more evenly distribute the transformed values of the variable matrix.
In some embodiments of the foregoing exemplary method, constructing a hash value from the transformed values may include constructing a hash value from the transformed values, including at least reducing the fidelity of the transformed values via reducing each transformed value to a binary representation. In some embodiments of the foregoing exemplary method, reducing the fidelity of the transformed values via reducing each transformed value to a binary representation may include determining for each transformed value whether the transformed value is a positive number and, if the transformed value is a positive number, assigning a one to the hash value and otherwise assigning a zero to the hash value.
In some embodiments of the foregoing exemplary method, referencing a number of most significant bits of the constructed hash value to determine a database sector may include referencing a number of most significant bits of the constructed hash value to determine a database server, wherein the number of most significant bits is pre-determined to address a plurality of database servers, wherein a number of database servers associated with the number of most significant bits is established to enable at least one index associated with a database sector to reside entirely in memory of a corresponding database server. In some embodiments of the foregoing exemplary method, storing at least the hash value on the determined database sector may include storing at least the hash value on the determined database sector, including at least storing at least one indication of a channel, at least one indication of a video segment, and at least one indication of a timecode offset from the beginning of the video segment at a database location at least partially based on the hash value.
In one or more alternative embodiments of the foregoing exemplary method, related systems include but are not limited to circuitry and/or programming for effecting the herein-referenced method embodiments; the circuitry and/or programming can be virtually any combination of hardware, software, and/or firmware configured to effect the herein-referenced method aspects depending upon the design choices of the system designer.
In a different embodiment, an exemplary method related to addressing a media database using distance associative hashing may include receiving a cue, the cue constructed via one or more operations associated with a media storage operation; referencing a number of most significant bits of the received cue to determine a database sector; and returning at least one indication of at least one candidate from the database sector based at least partially on the received cue.
In some embodiments of the foregoing exemplary method, receiving a cue, the cue constructed via one or more operations associated with a media storage operation may include receiving a cue associated with a sample of a video buffer of a client system, including at least receiving one or more indications related to an epoch time associated with the sample of the video buffer of the client system. In some embodiments of the foregoing exemplary method, receiving a cue, the cue constructed via one or more operations associated with a media storage operation may include receiving a cue, the cue associated with a sample of a video buffer of a client system, the cue at least partially determined by hashing at least some values associated with the video buffer.
In some embodiments of the foregoing exemplary method, receiving a cue, the cue associated with a sample of a video buffer of a client system, the cue at least partially determined by hashing at least some values associated with the video buffer may include receiving a cue, the cue associated with a sample of a video buffer of a client system, the cue at least partially determined by hashing at least some values associated with the video buffer, the hashing based at least partially one or more of at least one operand or at least one algorithm also utilized in an associated media storage operation. In some embodiments of the foregoing exemplary method, receiving a cue, the cue constructed via one or more operations associated with a media storage operation may include receiving a cue, the cue determined via one or more operations including at least receiving one or more indications of at least one content of a video buffer of a client system; determining, for at least one patch of the at least one content of the video buffer including at least one or more pixels of the at least one patch, an algorithmically-derived value of the one or more pixels of each patch; subtracting a median point value from the mean value for each patch; transforming the values resulting from the subtraction; constructing a hash value from the transformed values; and associating the cue at least partially with the constructed hash value, wherein at least one of the determining, subtracting, transforming, or constructing operations utilize one or more of at least one operand or at least one algorithm also utilized in an associated media storage operation.
In some embodiments of the foregoing exemplary method, returning at least one indication of at least one candidate from the database sector based at least partially on the received cue may include returning at least one indication of at least one candidate from the database sector based at least partially on a probabilistic point location in equal balls (“PPLEB”) algorithm as a function of the received cue. In some embodiments of the foregoing exemplary method, returning at least one indication of at least one candidate from the database sector based at least partially on the received cue may include returning at least one indication of at least one candidate from the database sector based at least partially on the received cue, the at least one candidate being within a predetermined inverse percentage distribution radius of the received cue.
In one or more alternative embodiments of the foregoing exemplary method, related systems include but are not limited to circuitry and/or programming for effecting the herein-referenced method embodiments; the circuitry and/or programming can be virtually any combination of hardware, software, and/or firmware configured to effect the herein-referenced method aspects depending upon the design choices of the system designer.
In a different embodiment, an exemplary method related to addressing a media database using distance associative hashing may include receiving at least one indication of at least one candidate and at least one indication of at least one cue; adding a token to a bin associated with at least one received candidate; and determining whether a number of tokens in a bin exceeds a value associated with a probability that a client system is displaying a particular video segment associated with at least one cue and, if the number of tokens in a bin exceeds a value associated with a probability that a client system is display a particular video segment associated with at least one cue, returning at least some data associated with the particular video segment based at least partially on the bin.
In some embodiments of the foregoing exemplary method, adding a token to a bin associated with at least one received candidate may include adding a token to a time bin associated with at least one received candidate. In some embodiments of the foregoing exemplary method, adding a token to a bin associated with at least one received candidate may include determining a relative time, including at least subtracting a candidate time associated with the at least one candidate from an arbitrary time associated with the at least one cue; and adding a token to a time bin associated with the candidate based at least partially on the determined relative time. In some embodiments of the foregoing exemplary method, the method may include removing one or more tokens from a time bin based at least partially on a time period elapsing.
In one or more alternative embodiments of the foregoing exemplary method, related systems include but are not limited to circuitry and/or programming for effecting the herein-referenced method embodiments; the circuitry and/or programming can be virtually any combination of hardware, software, and/or firmware configured to effect the herein-referenced method aspects depending upon the design choices of the system designer.
In a different embodiment, an exemplary system related to addressing a media database using distance associative hashing may include, but is not limited to, one or more computing devices; and one or more instructions that, when executed on at least some of the one or more computing devices, cause at least some of the one or more computing devices to at least receive at least one stream of rasterized video; create at least one hash value associated with at least one sample of at least one received rasterized video stream; determine at least one database sector for storing a created at least one hash value; and store a created at least one hash value on at least one determined database sector.
In a different embodiment, an exemplary system related to addressing a media database using distance associative hashing may include, but is not limited to, one or more computing devices; and one or more instructions that, when executed on at least some of the one or more computing devices, cause at least some of the one or more computing devices to at least receive one or more indications associated with at least one video buffer of at least one client system; determine a cue based at least partially on the at least one video buffer and at least one epoch time associated with the at least one video buffer, wherein one or more of at least one operand or at least one function associated with determining the cue is also utilized in an associated media storage operation; reference a number of most significant bits of a determined cue to determine a database sector; and return at least one indication of at least one candidate from a determined database sector based at least partially on a determined cue.
In a different embodiment, an exemplary system related to addressing a media database using distance associative hashing may include, but is not limited to, one or more computing devices; and one or more instructions that, when executed on at least some of the one or more computing devices, cause at least some of the one or more computing devices to at least receive at least one indication of at least one candidate and at least one indication of at least one cue; add a token to a bin associated with at least one received candidate; and determine whether a number of tokens in a bin exceeds a value associated with a probability that a client system is receiving a particular video segment associated with at least one received cue and, if the number of tokens in a bin exceeds a value associated with a probability that a client system is receiving a particular video segment associated with at least one received cue, returning at least some data associated with the particular video segment based at least partially on the bin.
In addition to the foregoing, various other methods, systems and/or program product embodiments are set forth and described in the teachings such as the text (e.g., claims, drawings and/or the detailed description) and/or drawings of the present disclosure.
The foregoing is a summary and thus contains, by necessity, simplifications, generalizations and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is NOT intended to be in any way limiting. Other aspects, embodiments, features and advantages of the device and/or processes and/or other subject matter described herein will become apparent in the teachings set forth herein.
Certain embodiments of the present invention are described in detail below with reference to the following drawings:
The first invention which relates to this invention is a system and method of matching unknown video to a database of known video using a novel signal processing means employing a modified path pursuit algorithm, among other means, as described in the aforementioned publication.
A novel means of the new invention is its Distance Associated Hashing with its attendant provision of utilizing a sectored-index database access. Said indexing means provides a highly computationally-efficient means for matching an unknown media segment to a reference database of known media, such as audio or video content.
This indexing means of the invention results in a significant improvement in the speed and accuracy of the ACR system so enabled as to track the identity of media even when the television display is showing content where a user is changing channels, rewinding, fast-forwarding or even pausing video from a digital video recorder.
Both the building, updating, and the subsequent accessing of the media matching database will describe a system capable of generating and addressing a sectored database such that the database sectors can each reside in the main memory of a respective multiplicity of server means without resorting to a paging means within each of the respective server means. This collective means of addressing a sectored database through locality sensitive hashing provides a significant improvement in efficiency of operation.
The construction of a sectored video matching database begins with the process as illustrated in
The system of the invention first builds and then continuously updates a reference database of television programming fingerprints described in the original application as cues or cue values. For purposes of building said reference database of video cues, the invention performs the acquisition of one or more patches of video 102 which are read from the video frame buffer 103. Said video patches can be any arbitrary shape or pattern but for the purposes of this example shall be 10 pixels horizontally by 10 pixels vertically. Also for the sake of this example, assume that there are 25 pixel patch positions within the video frame buffer that are evenly distributed within the boundaries of said buffer, though they do not have to be evenly distributed. Each pixel shall consist of a red, a green and a blue value, 104, typically represented by an eight bit binary value for each color for a total of 24 bits or three bytes per patch location.
This composite data structure is populated with the average pixel values from a number of pixel patch positions from the video buffer. A pixel patch is defined as a typically square pixel area of the video buffer with dimensions of perhaps ten pixels by ten rows of pixels 304. The number of pixel patch positions might typically be between ten and fifty locations within the video buffer.
The average pixel values 305 are assembled with a time code 306 referencing the “epoch time” from the processor means of the television system. Epoch time is defined as the time in fractions of a second that have elapsed since midnight, Jan. 1, 1970 which is an accepted convention in computing systems, particularly with Unix (or Linux)-based systems.
In addition, metadata may be included and together a data structure 106 is defined called a tagged fingerprint, “cue”, or a “point”, as taught in the original patent application. Such metadata attributes might be derived from closed-captioning data from the currently displaying video program or it could be keywords extracted by means of a speech recognition system operating within the processor means of the television system which converts audio from the respective television program into text information. Said textual information may then be searched for relevant key words or sent in its entirety as part of the cue data structure to the central server means for further processing.
The cue records 201 are passed in
The ten by ten pixel patch 302 shown in this particular example would have one hundred pixels and is mathematically averaged resulting in a mean pixel value 305 for red, green and blue values, respectively. Alternatively, any averaging function can be used in place of a simple mean.
A plurality of such pixel patches are extracted from the video frame. If, by way of example, 25 such pixel patches are extracted from the video frame, then the result will be a point representing a position in a 75-dimension space. The skilled person will know that such a large search space could require extensive computing resources to later locate, even approximately, said value in combination with the other 74 values representing one frame of video.
It is an advantage of system and method of distance associated hashing described herein to reduce the computational load and improve accuracy of matching unknown video frames to a known video frame database.
The creation of the hash value involves several steps beginning with computing an algorithmically-derived value of each point as shown in
The resulting hash value points to more or less evenly distributed values across the data storage area. The hash value 203 can be further divided,
The division point of the hash value that defines the individual sector address space is calculated such that the data of the database sector's index fits within the memory confines of the processor systems of said memory sector. Otherwise, said database would be subject to paging which would diminish the effectiveness of this process.
To contrast the system and method taught by the present invention with that which is known to those skilled in the relevant prior art,
Such accessing means using paging are common but considerably reduce the efficiency of a computer system. In fact, such an approach could not be uses with ACR systems searching large media database as the read/write speed of magnetic hard drives is insufficient to keep up with the task. Many different algorithmic approaches have been developed over the years to address this issue of splitting the search into smaller parts and allocating smaller searches to multiple computer server systems.
A notable example might be the considerable Google search engine. The skilled person knows this system to be one of the largest computer systems built to date. The speed and accuracy of the Google search process is remarkable. However, the Google search means is considerably different and not at all applicable to matching unknown media to a database of known media even if the two databases were the same very large size. This is because the Google search means employs the map-reduce algorithm which is designed for searching large databases of essentially unassociated data. While an advance over paging system, map-reduce is a computationally-intense process which also requires significant data communications bandwidth between the participating computer systems. In contrast, this invention is efficient in the use of processing and communications resources.
In this invention, the distance associative hashing function provides a means to address a database in sectors such that the data of said addressing means fits in the main memory of an individual server means of group of servers. Said grouping is accomplished by grouping the data related by distance in a multi-dimensional array into the same sector using the distance associative hashing step as a means to achieve said grouping. The sector identification for addressing a data element is calculated from the hash index generated from said process by extracting a subset of the total bits of said hash function and using said subset to address the desired sector in which to store data in the reference database.
In this manner, the hash index subset is the address of the sector that contains the distance associated hash values, called buckets in the first invention. The remainder of the hash address is then used to address a bucket of the sector for storing the new data. Alternatively, the sector address can be found by means of re-hashing the first hash value.
This system and method of database addressing by means of multiple hash-indexing steps produces a highly efficient database accessing scheme with significant performance benefits and increased efficiency over traditional methods of database access as described above.
The distance associative hashing provides a means to address a very complex (multi-dimensional) database quickly by finding data that is not an exact match but rather is within a predetermined radius (distance associative) of the value sought. Importantly, sometimes this addressing means will result in no match at all. Where a business-oriented database cannot tolerate inaccuracy, a media matching system can readily tolerate missed matches and will simply continue the matching process upon the arrival of the next data received and taught in the first patent. Data arrival from the unknown source that is to be determined by the ACR system is periodic, of course, but can be commanded by the system of the invention to arrive at differing intervals based on the requirement for accuracy or by requirements imposed by the state of the system such as when the system might be nearing overload and the sending clients are then commanded to send a lower sample rate. A typical data reception rate might be 1/10 second intervals, for example.
For the reference media database, the group of pixel values are derived from every frame of video from every video source that is to be part of the reference database. The group of pixel values and are then appended with the broadcast time of the video program as well as with certain metadata, which is information about the program such as the content identification (ID), title of the program, actors name, time of airing, short synopsis, etc. Said metadata is generally acquired from commercial electronic program guide sources.
Said array of processed pixel values with the addition of the timecode plus the metadata are then stored in the reference database and the address of said stored data is then added to the hash index at the respective hash value and sector ID value. In addition, a second database index is built and maintained by using the content ID from the metadata as another means for addressing the reference database.
The process of building and continuously updating the database is continuous and the number of days of data maintained by the database is based on the needs of the user but for example might range from one day to one month.
The process of identifying an unknown video segment from data received from a multiplicity of client devices begins with a procedure similar to that used above for building the reference database. In
Said system of the invention then calculates a distance associative hash index of the collected mean values identical to the content ingest function described above. The resulting sector identification (ID) value is extracting as a subset of the total bits of the hash index also identically to the ingest system described above. The remainder of the hash index is used to address the desired sector in which to search for all candidate (potential) matches belonging to the same bucket as the unknown data point.
Optionally, if a good guess of a match (a successful match) was available from the process above, the system of the invention will also collect candidates from the database responsible for said sector belonging to the potential content ID, using the content ID index, created during the ingest process as described above, to address reference cues around time radius r′ of the timestamp (of the successfully matched candidate). Duplicate candidates are next removed as well as candidates that are too far from the unknown point by radius r, as taught in the first patent.
In order to test for a match of an unknown video segment against a reference database of known video data, assume the list of candidates from the previous step where each candidate (i.e. each possible match) has associated with it the following data items: content ID, media time, inverse percentage distribution radius which is calculated as the distance from the current unknown point (from unknown video stream) where 100% represents the exact value of the unknown point and 0% is a value beyond the radius r (distribution) from the unknown point.
Each matching candidate 501 is assigned a data structure 502 in the memory of the matching system of the invention. The data structure consists of, among other things, arbitrary time bins grouped by some arbitrary amount (e.g. approximately one second). For the sake of example, assume said data structure consists of one hundred bins representing ten seconds of video cue points. The bins are generally not equally spaced in time.
For each candidate found in the reference (match) database: first, a relative time is calculated by subtracting candidate time from the arbitrary time of the unknown video. Candidate time is the time of broadcast of each video cue associated with the candidate during the reference program airing.
The arbitrary time of the unknown video came from the internally generated epoch time of the television monitor from the application of the invention operating in the memory of said television or in a set-top box attached to said television and sent by said application to the central server means of the invention with the sampled video cue points. Epoch time is well known to the skilled person and is typically employed in computer systems. Said time is calculated as the current number of units of time since Jan. 1, 1970.
If, for example, the time difference between the arbitrary time from the television (in the home) and the true media time is 100 seconds, then the relative time of the actually matching candidates should be close to that value. Likewise, candidates that are not a good match are not likely to have relative times close to the 100 seconds of this example.
In the candidate data structure, when a cue point of the unknown video matches a reference cue point, the system of the invention adds a token to the respective bin of the candidate data structure. Said system then repeats the process for the next candidate as described in the previous paragraph.
Another, and important, step for the scoring of results is to apply time discounting to all bins. This is a relatively simple process that decrements the value in all bins by a small amount for each cycle of time. The skilled person would recognize this as a “leaky bucket” method of scoring. By definition, bins that are no longer being filled by means of matching cue points will ultimately decrement to zero over a number of cycles of said process. Also, bins that are filled slowly by random noise in the system will likewise be decremented. Hence, time discounting ultimate clears bins that are filled by false-positive matches and random noise. The skilled person would also clearly see that without said time discount binning, all bins would eventually fill to capacity and no results could be obtained from the process.
Said time discounting also decrements to zero any bins with levels, such as 503, that are above the matching threshold 504 when the video stream from the client television monitor is in any way changed by any of the following: changing channels, rewind, fast forward, pausing video, etc.
If any bin of the candidate data structure is above a certain threshold 504, such as bin 503, then the content is declared a match. Further means to qualify a match might include testing for contiguous matches of the candidate segment for greater that a determined number of seconds (e.g. three seconds).
When partition a dataset in this manner, the computed median is not necessarily in the middle of the dataset nor does it need to be. The desired result is to spread out the data so that when said data is partitioned and assigned to individual servers, the system accessing said servers more uniformly. In contrast, the unoptimized data of
After a start operation, the operational flow 1000 moves to operation 1002. Operation 1002 depicts receiving one or more indications of a sample of a video segment. For example, as shown in and/or described with respect to
Then, operation 1004 depicts determining, for at least one patch of the sample of a video segment including at least one or more pixels of the at least one patch, an algorithmically-derived value of the one or more pixels of each patch. For example, as shown in and/or described with respect to
Then, operation 1006 depicts subtracting a median point value established for each patch from the mean value for each patch. For example, as shown in and/or described with respect to
Then, operation 1008 depicts transforming the values resulting from the subtraction using a function pre-derived to distribute the values evenly. For example, as shown in and/or described with respect to
Then, operation 1010 depicts constructing a hash value from the transformed values. For example, as shown in and/or described with respect to
Then, operation 1012 depicts referencing a number of most significant bits of the constructed hash value to determine a database sector. For example, as shown in and/or described with respect to
Then, operation 1014 depicts storing at least the hash value on the determined database sector. For example, as shown in and/or described with respect to
Operation 1102 illustrates at least one of the receiving 1002, determining 1004, subtracting 1006, transforming 1008, constructing 1010, referencing 1012, or storing 1014 operations being at least partially implemented using one or more processing devices. In some instances, one of the foregoing operations may be at least partially implemented using one or more computer processors. Other processing devices may include Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), digital signal processors (DSPs), or any other circuitry configured to effect the result of at least one of the foregoing operations.
Operation 1202 illustrates receiving one or more indications of at least one of a frame or a still image. For example, as shown in and/or described with respect to
Further, operation 1204 illustrates receiving one or more indications of a sample of a video segment, the one or more indications of a sample of a video segment associated with at least one indication of a channel, at least one indication of a video segment, and at least one indication of a timecode offset from the beginning of the video segment. For example, as shown in and/or described with respect to
Operation 1302 illustrates determining, for at least one patch of the sample of a video segment including at least one or more pixels of the at least one patch, a mean value of the one or more pixels of each patch. For example, as shown in and/or described with respect to
Operation 1402 illustrates subtracting a median point value established for each patch from the mean value for each patch, the median point value established for each patch having been previously determined utilizing data from each patch for a plurality of channels over at least one period of time. For example, as shown in and/or described with respect to
Operation 1502 illustrates forming a variable matrix including at least the values resulting from the subtraction. For example, as shown in and/or described with respect to
Operation 1504 illustrates obtaining a static matrix which, when crossed with the variable matrix, will more evenly distribute the transformed values. For example, as shown in and/or described with respect to
Operation 1506 illustrates computing a dot product of the variable matrix and the static matrix, the dot product including at least the more evenly-distributed transformed values. For example, as shown in and/or described with respect to
Operation 1602 illustrates determining, using locality-sensitive hashing at least partially based on one or more previously obtained hash values, a static matrix which, when crossed with a variable matrix, will more evenly distribute the transformed values of the variable matrix. For example, as shown in and/or described with respect to
Operation 1702 illustrates constructing a hash value from the transformed values, including at least reducing the fidelity of the transformed values via reducing each transformed value to a binary representation. For example, as shown in and/or described with respect to
Operation 1702 may include operation 1704. Operation 1704 illustrates determining for each transformed value whether the transformed value is a positive number and, if the transformed value is a positive number, assigning a one to the hash value and otherwise assigning a zero to the hash value. For example, as shown in and/or described with respect to
Operation 1802 illustrates referencing a number of most significant bits of the constructed hash value to determine a database server, wherein the number of most significant bits is pre-determined to address a plurality of database servers, wherein a number of database servers associated with the number of most significant bits is established to enable at least one index associated with a database sector to reside entirely in memory of a corresponding database server. For example, as shown in and/or described with respect to
Operation 1902 illustrates storing at least the hash value on the determined database sector, including at least storing at least one indication of a channel, at least one indication of a video segment, and at least one indication of a timecode offset from the beginning of the video segment at a database location at least partially based on the hash value. For example, as shown in and/or described with respect to
After a start operation, the operational flow 2000 moves to operation 2002. Operation 2002 depicts receiving a cue, the cue constructed via one or more operations associated with a media storage operation. For example, as shown in and/or described with respect to
Then, operation 2004 depicts referencing a number of most significant bits of the received cue to determine a database sector. For example, as shown in and/or described with respect to
Then, operation 2006 depicts returning at least one indication of at least one candidate from the database sector based at least partially on the received cue. For example, as shown in and/or described with respect to
Operation 2102 illustrates receiving a cue associated with a sample of a video buffer of a client system, including at least receiving one or more indications related to an epoch time associated with the sample of the video buffer of the client system. For example, as shown in and/or described with respect to
Operation 2104 illustrates receiving a cue, the cue associated with a sample of a video buffer of a client system, the cue at least partially determined by hashing at least some values associated with the video buffer. For example, as shown in and/or described with respect to
Operation 2106 illustrates receiving a cue, the cue associated with a sample of a video buffer of a client system, the cue at least partially determined by hashing at least some values associated with the video buffer, the hashing based at least partially one or more of at least one operand or at least one algorithm also utilized in an associated media storage operation. For example, as shown in and/or described with respect to
Operation 2202 illustrates receiving one or more indications of at least one content of a video buffer of a client system. For example, as shown in and/or described with respect to
Operation 2204 illustrates determining, for at least one patch of the at least one content of the video buffer including at least one or more pixels of the at least one patch, an algorithmically-derived value of the one or more pixels of each patch. For example, as shown in and/or described with respect to
Operation 2206 illustrates subtracting a median point value from the mean value for each patch. For example, as shown in and/or described with respect to
Operation 2208 illustrates transforming the values resulting from the subtraction. For example, as shown in and/or described with respect to
Operation 2210 illustrates constructing a hash value from the transformed values. For example, as shown in and/or described with respect to
Operation 2212 illustrates associating the cue at least partially with the constructed hash value. For example, as shown in and/or described with respect to
Operation 2214 illustrates at least one of the determining 2204, subtracting 2206, transforming 2208, or constructing 2210 operations utilize one or more of at least one operand or at least one algorithm also utilized in an associated media storage operation. For example, as shown in and/or described with respect to
Operation 2302 illustrates returning at least one indication of at least one candidate from the database sector based at least partially on a probabilistic point location in equal balls (“PPLEB”) algorithm as a function of the received cue. For example, as shown in and/or described with respect to
Operation 2304 illustrates returning at least one indication of at least one candidate from the database sector based at least partially on the received cue, the at least one candidate being within a predetermined inverse percentage distribution radius of the received cue. For example, as shown in and/or described with respect to
After a start operation, the operational flow 2400 moves to operation 2402. Operation 2402 depicts receiving at least one indication of at least one candidate and at least one indication of at least one cue. For example, as shown in and/or described with respect to
Then, operation 2404 depicts adding a token to a bin associated with at least one received candidate. For example, as shown in and/or described with respect to
Then, operation 2406 depicts determining whether a number of tokens in a bin exceeds a value associated with a probability that a client system is displaying a particular video segment associated with at least one cue and, if the number of tokens in a bin exceeds a value associated with a probability that a client system is display a particular video segment associated with at least one cue, returning at least some data associated with the particular video segment based at least partially on the bin. For example, as shown in and/or described with respect to
Operation 2502 illustrates adding a token to a time bin associated with at least one received candidate. For example, as shown in and/or described with respect to
Operation 2602 illustrates determining a relative time, including at least subtracting a candidate time associated with the at least one candidate from an arbitrary time associated with the at least one cue. For example, as shown in and/or described with respect to
Operation 2604 illustrates adding a token to a time bin associated with the candidate based at least partially on the determined relative time. For example, as shown in and/or described with respect to
Operation 2606 illustrates removing one or more tokens from a time bin based at least partially on a time period elapsing. For example, as shown in and/or described with respect to
In varying embodiments, pixel locations may relate to one or many colors and/or color spaces/models (e.g. red, blue, green; red, blue, green, and yellow; cyan, magenta, yellow, and black; a single pixel value uniquely identifying a color e.g. a 24-bit value associated with a pixel location; hue, saturation, brightness; etc.). Differing numbers of pixels in a patch may be used, and the patch does not have to be a square patch. Further, resolution of the video buffer of the client system may vary. Resolutions and/or color densities at the client system and the ingest system may vary. The system may be operable with various raster resolutions, including but not limited to 1920 by 1080, 3840 by 2160, 1440×1080, 1366×768, or other resolutions. It is expected that over the next two decades, increases in pixel resolution of common programming, televisions, and/or client systems will occur; the same basic operations may be utilized although pixel patch number, size, sampling rate, or other aspects may vary. Further, an up-conversion, down-conversion, or other transformation operation associated with resolution and/or color density may occur and/or be interposed between other operations described herein.
Certain aspects of the present invention include process steps and instructions described herein in the form of an algorithm. It should be noted that the process steps and instructions of the present invention could be embodied in software, firmware or hardware, and when embodied in software, could be downloaded to reside on and be operated from different platforms used by real-time network operating systems.
The present invention also relates to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application specific integrated circuits (ASICs), or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
Furthermore, computers or computing means referred to in the specification may include a single processor or may employ multiple-processor designs for increased computing capability.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may also be used with programs in accordance with the teachings herein, or it may prove convenient to construct more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description above. In addition, the present invention is not described with reference to any particular programming language or operating systems. It is appreciated that a variety of programming languages and operating systems may be used to implement the teachings of the present invention as described herein.
The system and methods, flow diagrams, and structure block diagrams described in this specification may be implemented in computer processing systems including program code comprising program instructions that are executable by a computer processing system. Other implementations may also be used. Additionally, the flow diagrams and structure block diagrams herein described describe particular methods and/or corresponding acts in support of steps and corresponding functions in support of disclosed structural means, may also be utilized to implement corresponding software structures and algorithms, and equivalents thereof.
Embodiments of the subject matter described in this specification can be implemented as one or more computer program products, i.e., one or more modules of computer program instructions encoded on a tangible program carrier for execution by, or to control the operation of, data processing apparatus. The computer readable medium can be a machine readable storage device, a machine readable storage substrate, a memory device, or a combination of one or more of them.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a suitable communication network.
The processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
The essential elements of a computer are a processor for performing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Processors suitable for the execution of a computer program include, by way of example only and without limitation, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read only memory or a random access memory or both.
To provide for interaction with a user or manager of the system described herein, embodiments of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well. For example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
Embodiments of the subject matter described in this specification can be implemented in a computing system that includes back end component(s) including one or more data servers, or that includes one or more middleware components such as application servers, or that includes a front end component such as a client computer having a graphical user interface or a Web browser through which a user or administrator can interact with some implementations of the subject matter described is this specification, or any combination of one or more such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication, such as a communication network. The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client server relationship to each other.
While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or of what may be claimed, but rather as descriptions of features that may be specific to particular embodiments of particular inventions. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment.
Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the embodiments described above should not be understood as requiring such separation in all embodiments, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
This written description sets forth the best mode of the invention and provides examples to describe the invention and to enable a person of ordinary skill in the art to make and use the invention. This written description does not limit the invention to the precise terms set forth. Thus, while the invention has been described in detail with reference to the examples set forth above, those of ordinary skill in the art may effect alterations, modifications and variations to the examples without departing from the scope of the invention.
This application is a continuation of U.S. patent application Ser. No. 15/240,815, filed on Aug. 18, 2016, which is a continuation of U.S. patent application Ser. No. 14/678,856, filed on Apr. 3, 2015, now U.S. Pat. No. 9,449,090, which is a continuation-in-part of Ser. No. 12/788,721, filed on May 27, 2010, now U.S. Pat. No. 8,595,781, and a continuation-in-part of U.S. patent application Ser. No. 12/788,748, filed on May 27, 2010, now U.S. Pat. No. 8,769,584, and a continuation-in-part of U.S. Pat. No. 14,089,003, filed on Nov. 25, 2013, now U.S. Pat. No. 8,898,714, and a continuation-in-part of U.S. patent application Ser. No. 14/217,039, filed on Mar. 17, 2014, now U.S. Pat. No. 9,055,335, and a continuation-in-part of U.S. patent application Ser. No. 14/217,075, filed on Mar. 17, 2014, now U.S. Pat. No. 9,055,309, and a continuation-in-part of U.S. patent application Ser. No. 14/217,094, filed Mar. 17, 2014, now U.S. Pat. No. 8,930,980, and a continuation-in-part of U.S. patent application Ser. No. 14/217,375, filed on Mar. 17, 2014, now U.S. Pat. No. 9,094,714, and a continuation-in-part of U.S. patent application Ser. No. 14/217,425, filed on Mar. 17, 2014, now U.S. Pat. No. 9,071,868, and a continuation-in-part of U.S. patent application Ser. No. 14/217,435, filed on Mar. 17, 2014, now U.S. Pat. No. 9,094,715, and a continuation-in-part of U.S. patent application Ser. No. 14/551,933, filed on Nov. 24, 2014, now U.S. Pat. No. 9,906,834. U.S. patent application Ser. No. 12/788,721 claims the benefit of and priority to U.S. Provisional Application No. 61/182,334 filed on May 29, 2009 and U.S. Provisional Application No. 61/290,714 filed on Dec. 29, 2009. U.S. patent application Ser. No. 14/217,039 claims the benefit of and priority to U.S. Provisional Application No. 61/791,578 filed on Mar. 15, 2013. All of these applications are hereby incorporated by reference in their entireties for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
4677466 | Lert, Jr. et al. | Jun 1987 | A |
4697209 | Kiewit et al. | Sep 1987 | A |
4739398 | Thomas et al. | Apr 1988 | A |
5019899 | Boles et al. | May 1991 | A |
5193001 | Kerdranvrat | Mar 1993 | A |
5210820 | Kenyon | May 1993 | A |
5319453 | Copriviza et al. | Jun 1994 | A |
5436653 | Ellis et al. | Jul 1995 | A |
5473607 | Hausman | Dec 1995 | A |
5481294 | Thomas et al. | Jan 1996 | A |
5557334 | Legate | Sep 1996 | A |
5572246 | Ellis et al. | Nov 1996 | A |
5812286 | Li | Sep 1998 | A |
5826165 | Echeita et al. | Oct 1998 | A |
5918223 | Blum et al. | Jun 1999 | A |
6008802 | Goldschmidt et al. | Dec 1999 | A |
6025837 | Matthews, III et al. | Feb 2000 | A |
6035177 | Moses et al. | Mar 2000 | A |
6064764 | Bhaskaran et al. | May 2000 | A |
6381362 | Deshpande et al. | Apr 2002 | B1 |
6415438 | Blackketter et al. | Jul 2002 | B1 |
6463585 | Hendricks et al. | Oct 2002 | B1 |
6469749 | Dimitrova | Oct 2002 | B1 |
6577346 | Perlman | Jun 2003 | B1 |
6577405 | Kranz et al. | Jun 2003 | B2 |
6628801 | Powell et al. | Sep 2003 | B2 |
6647548 | Lu et al. | Nov 2003 | B1 |
6675174 | Bolle et al. | Jan 2004 | B1 |
6771316 | Iggulden | Aug 2004 | B1 |
6804659 | Graham et al. | Oct 2004 | B1 |
6978470 | Swix et al. | Dec 2005 | B2 |
6990453 | Wang et al. | Jan 2006 | B2 |
7028327 | Dougherty et al. | Apr 2006 | B1 |
7039930 | Goodman et al. | May 2006 | B1 |
7050068 | Bastos et al. | May 2006 | B1 |
7051351 | Goldman et al. | May 2006 | B2 |
7064796 | Roy et al. | Jun 2006 | B2 |
7089575 | Agnihotri et al. | Aug 2006 | B2 |
7098959 | Mishima et al. | Aug 2006 | B2 |
7136875 | Anderson et al. | Nov 2006 | B2 |
7178106 | Lamkin et al. | Feb 2007 | B2 |
7210157 | Devara | Apr 2007 | B2 |
7243239 | Kirovski | Jul 2007 | B2 |
7325013 | Caruso | Jan 2008 | B2 |
7346512 | Wang et al. | Mar 2008 | B2 |
7356830 | Dimitrova | Apr 2008 | B1 |
7421723 | Harkness et al. | Sep 2008 | B2 |
7545984 | Kiel et al. | Jun 2009 | B1 |
7590998 | Hanley | Sep 2009 | B2 |
7623823 | Zito et al. | Nov 2009 | B2 |
7787696 | Wilhelm et al. | Aug 2010 | B2 |
7793318 | Deng | Sep 2010 | B2 |
7933451 | Kloer | Apr 2011 | B2 |
8001571 | Schwartz et al. | Aug 2011 | B1 |
8094872 | Yagnik et al. | Jan 2012 | B1 |
8171004 | Kaminski, Jr. et al. | May 2012 | B1 |
8171030 | Peira et al. | May 2012 | B2 |
8175413 | Ioffe et al. | May 2012 | B1 |
8189945 | Stojancic et al. | May 2012 | B2 |
8195689 | Ramanathan et al. | Jun 2012 | B2 |
8229227 | Stojancic et al. | Jul 2012 | B2 |
8335786 | Peira et al. | Dec 2012 | B2 |
8364703 | Ramanathan et al. | Jan 2013 | B2 |
8385644 | Stojancic et al. | Feb 2013 | B2 |
8392789 | Biscondi et al. | Mar 2013 | B2 |
8494234 | Djordjevic et al. | Jul 2013 | B1 |
8522283 | Laligand et al. | Aug 2013 | B2 |
8595781 | Neumeier et al. | Nov 2013 | B2 |
8619877 | McDowell | Dec 2013 | B2 |
8625902 | Baheti et al. | Jan 2014 | B2 |
8769854 | Battaglia | Jul 2014 | B1 |
8776105 | Sinha et al. | Jul 2014 | B2 |
8832723 | Sinha et al. | Sep 2014 | B2 |
8856817 | Sinha et al. | Oct 2014 | B2 |
8893167 | Sinha et al. | Nov 2014 | B2 |
8893168 | Sinha et al. | Nov 2014 | B2 |
8898714 | Neumeier et al. | Nov 2014 | B2 |
8918804 | Sinha et al. | Dec 2014 | B2 |
8918832 | Sinha et al. | Dec 2014 | B2 |
8930980 | Neumeier et al. | Jan 2015 | B2 |
8959202 | Haitsma et al. | Feb 2015 | B2 |
9055309 | Neumeier et al. | Jun 2015 | B2 |
9055335 | Neumeier | Jun 2015 | B2 |
9071868 | Neumeier et al. | Jun 2015 | B2 |
9083710 | Yadav | Jul 2015 | B1 |
9094714 | Neumeier et al. | Jul 2015 | B2 |
9094715 | Neumeier et al. | Jul 2015 | B2 |
9262671 | Unzueta | Feb 2016 | B2 |
9368021 | Touloumtzis | Jun 2016 | B2 |
9432616 | Hurd | Aug 2016 | B1 |
9449090 | Neumeier et al. | Sep 2016 | B2 |
9465867 | Hoarty | Oct 2016 | B2 |
9838753 | Neumeier et al. | Dec 2017 | B2 |
9955192 | Neumeier et al. | Apr 2018 | B2 |
10080062 | Neumeier et al. | Sep 2018 | B2 |
10116972 | Neumeier et al. | Oct 2018 | B2 |
10169455 | Neumeier et al. | Jan 2019 | B2 |
10185768 | Neumeier et al. | Jan 2019 | B2 |
10192138 | Neumeier et al. | Jan 2019 | B2 |
20010039658 | Walton | Nov 2001 | A1 |
20010044992 | Jahrling | Nov 2001 | A1 |
20020026635 | Wheeler et al. | Feb 2002 | A1 |
20020054069 | Britt et al. | May 2002 | A1 |
20020054695 | Bjorn et al. | May 2002 | A1 |
20020056088 | Silva, Jr. et al. | May 2002 | A1 |
20020059633 | Harkness et al. | May 2002 | A1 |
20020078144 | Lamkin et al. | Jun 2002 | A1 |
20020083060 | Wang et al. | Jun 2002 | A1 |
20020100041 | Rosenberg et al. | Jul 2002 | A1 |
20020105907 | Bruekers et al. | Aug 2002 | A1 |
20020120925 | Logan | Aug 2002 | A1 |
20020122042 | Bates | Sep 2002 | A1 |
20020162117 | Pearson et al. | Oct 2002 | A1 |
20020162118 | Levy et al. | Oct 2002 | A1 |
20030026422 | Gerheim et al. | Feb 2003 | A1 |
20030086341 | Wells | May 2003 | A1 |
20030105794 | Jasinschi | Jun 2003 | A1 |
20030118243 | Sezer | Jun 2003 | A1 |
20030121037 | Swix et al. | Jun 2003 | A1 |
20030121046 | Roy et al. | Jun 2003 | A1 |
20030147561 | Faibish | Aug 2003 | A1 |
20030182291 | Kurupati | Sep 2003 | A1 |
20030188321 | Shoff et al. | Oct 2003 | A1 |
20040045020 | Witt et al. | Mar 2004 | A1 |
20040059708 | Dean et al. | Mar 2004 | A1 |
20040183825 | Stauder et al. | Sep 2004 | A1 |
20040216171 | Barone et al. | Oct 2004 | A1 |
20040221237 | Foote et al. | Nov 2004 | A1 |
20040226035 | Hauser | Nov 2004 | A1 |
20040240562 | Bargeron et al. | Dec 2004 | A1 |
20050015795 | Iggulden | Jan 2005 | A1 |
20050015796 | Bruckner et al. | Jan 2005 | A1 |
20050027766 | Ben | Feb 2005 | A1 |
20050066352 | Herley | Mar 2005 | A1 |
20050120372 | Itakura | Jun 2005 | A1 |
20050172312 | Lienhart et al. | Aug 2005 | A1 |
20050207416 | Rajkotia | Sep 2005 | A1 |
20050209065 | Schlosser et al. | Sep 2005 | A1 |
20050235318 | Grauch et al. | Oct 2005 | A1 |
20060029286 | Lim et al. | Feb 2006 | A1 |
20060029368 | Harville | Feb 2006 | A1 |
20060031914 | Dakss et al. | Feb 2006 | A1 |
20060116989 | Bellamkonda | Jun 2006 | A1 |
20060120624 | Jojic | Jun 2006 | A1 |
20060133647 | Werner et al. | Jun 2006 | A1 |
20060153296 | Deng | Jul 2006 | A1 |
20060155952 | Haas | Jul 2006 | A1 |
20060173831 | Basso et al. | Aug 2006 | A1 |
20060187358 | Lienhart et al. | Aug 2006 | A1 |
20060193506 | Dorphan et al. | Aug 2006 | A1 |
20060195857 | Wheeler et al. | Aug 2006 | A1 |
20060195860 | Eldering et al. | Aug 2006 | A1 |
20060245724 | Hwang et al. | Nov 2006 | A1 |
20060245725 | Lim | Nov 2006 | A1 |
20060253330 | Maggio et al. | Nov 2006 | A1 |
20060277047 | DeBusk et al. | Dec 2006 | A1 |
20060294561 | Grannan et al. | Dec 2006 | A1 |
20070033608 | Eigeldinger | Feb 2007 | A1 |
20070050832 | Wright et al. | Mar 2007 | A1 |
20070061724 | Slothouber et al. | Mar 2007 | A1 |
20070061831 | Savoor et al. | Mar 2007 | A1 |
20070083901 | Bond | Apr 2007 | A1 |
20070094696 | Sakai | Apr 2007 | A1 |
20070109449 | Cheung | May 2007 | A1 |
20070113263 | Chatani | May 2007 | A1 |
20070139563 | Zhong | Jun 2007 | A1 |
20070143796 | Malik | Jun 2007 | A1 |
20070168409 | Cheung | Jul 2007 | A1 |
20070180459 | Smithpeters et al. | Aug 2007 | A1 |
20070192782 | Ramaswamy | Aug 2007 | A1 |
20070250901 | McIntire et al. | Oct 2007 | A1 |
20070261070 | Brown et al. | Nov 2007 | A1 |
20070261075 | Glasberg | Nov 2007 | A1 |
20070271300 | Ramaswamy | Nov 2007 | A1 |
20070274537 | Srinivasan | Nov 2007 | A1 |
20070300280 | Turner et al. | Dec 2007 | A1 |
20080005086 | Moore | Jan 2008 | A1 |
20080044102 | Ekin | Feb 2008 | A1 |
20080046945 | Hanley | Feb 2008 | A1 |
20080089551 | Heather et al. | Apr 2008 | A1 |
20080138030 | Bryan et al. | Jun 2008 | A1 |
20080155588 | Roberts et al. | Jun 2008 | A1 |
20080155627 | O'Connor et al. | Jun 2008 | A1 |
20080172690 | Kanojia et al. | Jul 2008 | A1 |
20080208891 | Wang et al. | Aug 2008 | A1 |
20080240562 | Fukuda et al. | Oct 2008 | A1 |
20080263620 | Berkvens et al. | Oct 2008 | A1 |
20080276266 | Huchital et al. | Nov 2008 | A1 |
20080310731 | Stojancic et al. | Dec 2008 | A1 |
20080313140 | Pereira et al. | Dec 2008 | A1 |
20090007195 | Beyabani | Jan 2009 | A1 |
20090024923 | Hartwig et al. | Jan 2009 | A1 |
20090028517 | Shen et al. | Jan 2009 | A1 |
20090052784 | Covell et al. | Feb 2009 | A1 |
20090087027 | Eaton et al. | Apr 2009 | A1 |
20090088878 | Otsuka et al. | Apr 2009 | A1 |
20090100361 | Abello et al. | Apr 2009 | A1 |
20090116702 | Conradt | May 2009 | A1 |
20090131861 | Braig et al. | May 2009 | A1 |
20090172728 | Shkedi et al. | Jul 2009 | A1 |
20090172746 | Aldrey et al. | Jul 2009 | A1 |
20090175538 | Bronstein et al. | Jul 2009 | A1 |
20090213270 | Ismert et al. | Aug 2009 | A1 |
20090235312 | Morad et al. | Sep 2009 | A1 |
20100010648 | Bull et al. | Jan 2010 | A1 |
20100042790 | Mondal | Feb 2010 | A1 |
20100083299 | Nelson | Apr 2010 | A1 |
20100115543 | Falcon | May 2010 | A1 |
20100125870 | Ukawa et al. | May 2010 | A1 |
20100166257 | Wredenhagen | Jul 2010 | A1 |
20100199295 | Katpelly et al. | Aug 2010 | A1 |
20100235486 | White et al. | Sep 2010 | A1 |
20100253838 | Garg et al. | Oct 2010 | A1 |
20100269138 | Krikorian et al. | Oct 2010 | A1 |
20100306805 | Neumeier | Dec 2010 | A1 |
20100306808 | Neumeier et al. | Dec 2010 | A1 |
20110015996 | Kassoway et al. | Jan 2011 | A1 |
20110026761 | Radhakrishnan et al. | Feb 2011 | A1 |
20110041154 | Olson | Feb 2011 | A1 |
20110055552 | Francis et al. | Mar 2011 | A1 |
20110096955 | Voloshynovskiy et al. | Apr 2011 | A1 |
20110251987 | Buchheit | Apr 2011 | A1 |
20110247042 | Mallinson | Oct 2011 | A1 |
20110289099 | Quan | Nov 2011 | A1 |
20110299770 | Vaddadi et al. | Dec 2011 | A1 |
20110307683 | Spackman | Dec 2011 | A1 |
20120017240 | Shkedi | Jan 2012 | A1 |
20120054143 | Doig et al. | Mar 2012 | A1 |
20120076357 | Yamamoto et al. | Mar 2012 | A1 |
20120095958 | Pereira et al. | Apr 2012 | A1 |
20120117584 | Gordon | May 2012 | A1 |
20120158511 | Lucero et al. | Jun 2012 | A1 |
20120174155 | Mowrey et al. | Jul 2012 | A1 |
20120177249 | Levy et al. | Jul 2012 | A1 |
20120185566 | Nagasaka | Jul 2012 | A1 |
20120272259 | Cortes | Oct 2012 | A1 |
20120294586 | Weaver et al. | Nov 2012 | A1 |
20120317240 | Wang | Dec 2012 | A1 |
20120324494 | Burger et al. | Dec 2012 | A1 |
20120327450 | Sagan | Dec 2012 | A1 |
20130007191 | Klappert et al. | Jan 2013 | A1 |
20130042262 | Riethmueller | Feb 2013 | A1 |
20130050564 | Adams et al. | Feb 2013 | A1 |
20130054356 | Richman et al. | Feb 2013 | A1 |
20130067523 | Etsuko et al. | Mar 2013 | A1 |
20130070847 | Iwamoto et al. | Mar 2013 | A1 |
20130108173 | Lienhart et al. | May 2013 | A1 |
20130139209 | Urrabazo et al. | May 2013 | A1 |
20130198768 | Kitazato | Aug 2013 | A1 |
20130202150 | Sinha et al. | Aug 2013 | A1 |
20130209065 | Yeung | Aug 2013 | A1 |
20130212609 | Sinha et al. | Aug 2013 | A1 |
20130290502 | Bilobrov | Oct 2013 | A1 |
20130297727 | Levy | Nov 2013 | A1 |
20130318096 | Cheung | Nov 2013 | A1 |
20140016696 | Nelson | Jan 2014 | A1 |
20140052737 | Ramanathan et al. | Feb 2014 | A1 |
20140082663 | Neumeier et al. | Mar 2014 | A1 |
20140088742 | Srinivasan | Mar 2014 | A1 |
20140123165 | Mukherjee et al. | May 2014 | A1 |
20140130092 | Kunisetty | May 2014 | A1 |
20140153571 | Neugebauer | Jun 2014 | A1 |
20140188487 | Perez Gonzalez | Jul 2014 | A1 |
20140193027 | Scherf | Jul 2014 | A1 |
20140195548 | Harron | Jul 2014 | A1 |
20140201769 | Neumeier et al. | Jul 2014 | A1 |
20140201772 | Neumeier et al. | Jul 2014 | A1 |
20140201787 | Neumeier et al. | Jul 2014 | A1 |
20140219554 | Yamaguchi et al. | Aug 2014 | A1 |
20140237576 | Zhang | Aug 2014 | A1 |
20140258375 | Munoz | Sep 2014 | A1 |
20140270489 | Jaewhan et al. | Sep 2014 | A1 |
20140270504 | Baum et al. | Sep 2014 | A1 |
20140270505 | McCarthy | Sep 2014 | A1 |
20140282671 | McMillan | Sep 2014 | A1 |
20140293794 | Zhong et al. | Oct 2014 | A1 |
20140344880 | Geller et al. | Nov 2014 | A1 |
20150026728 | Carter et al. | Jan 2015 | A1 |
20150040074 | Hofmann et al. | Feb 2015 | A1 |
20150100979 | Moskowitz et al. | Apr 2015 | A1 |
20150112988 | Pereira et al. | Apr 2015 | A1 |
20150120839 | Kannan et al. | Apr 2015 | A1 |
20150121409 | Zhang et al. | Apr 2015 | A1 |
20150128161 | Conrad et al. | May 2015 | A1 |
20150163545 | Freed et al. | Jun 2015 | A1 |
20150181311 | Navin et al. | Jun 2015 | A1 |
20150205727 | Kimmel | Jul 2015 | A1 |
20150256891 | Kim et al. | Sep 2015 | A1 |
20150302890 | Ergen et al. | Oct 2015 | A1 |
20150382075 | Neumeier et al. | Dec 2015 | A1 |
20160112319 | Hasani | Apr 2016 | A1 |
20160154880 | Hoarty | Jun 2016 | A1 |
20160227261 | Neumeier et al. | Aug 2016 | A1 |
20160227291 | Shaw et al. | Aug 2016 | A1 |
20160286244 | Chang et al. | Sep 2016 | A1 |
20160307043 | Neumeier | Oct 2016 | A1 |
20160314794 | Leitman et al. | Oct 2016 | A1 |
20160353172 | Miller et al. | Dec 2016 | A1 |
20160359791 | Zhang et al. | Dec 2016 | A1 |
20170017645 | Neumeier et al. | Jan 2017 | A1 |
20170017651 | Neumeier et al. | Jan 2017 | A1 |
20170017652 | Neumeier et al. | Jan 2017 | A1 |
20170019716 | Neumeier et al. | Jan 2017 | A1 |
20170019719 | Neumeier et al. | Jan 2017 | A1 |
20170026671 | Neumeier et al. | Jan 2017 | A1 |
20170031573 | Kaneko | Feb 2017 | A1 |
20170032033 | Neumeier et al. | Feb 2017 | A1 |
20170032034 | Neumeier et al. | Feb 2017 | A1 |
20170090814 | Yeung | Mar 2017 | A1 |
20170134770 | Neumeier et al. | May 2017 | A9 |
20170186042 | Wong et al. | Jun 2017 | A1 |
20170311014 | Fleischman | Oct 2017 | A1 |
20170353776 | Holden et al. | Dec 2017 | A1 |
Number | Date | Country |
---|---|---|
2501316 | Sep 2005 | CA |
1557096 | Dec 2004 | CN |
101162470 | Apr 2008 | CN |
1681304 | Jul 2010 | CN |
102377960 | Mar 2012 | CN |
101681373 | Sep 2012 | CN |
248 533 | Aug 1994 | EP |
1578126 | Sep 2005 | EP |
1 760 693 | Mar 2007 | EP |
1504445 | Aug 2008 | EP |
2 084 624 | Aug 2009 | EP |
2 352 289 | Aug 2011 | EP |
2 541 963 | Jan 2013 | EP |
2 685 450 | Jan 2014 | EP |
2457694 | Aug 2009 | GB |
0144992 | Jun 2001 | WO |
2005101998 | Nov 2005 | WO |
2007114796 | Oct 2007 | WO |
2008065340 | Jun 2008 | WO |
2009131861 | Oct 2009 | WO |
2009150425 | Dec 2009 | WO |
2010135082 | Nov 2010 | WO |
2011090540 | Jul 2011 | WO |
2012057724 | May 2012 | WO |
2012108975 | Aug 2012 | WO |
2012170451 | Dec 2012 | WO |
2014142758 | Sep 2014 | WO |
2014145929 | Sep 2014 | WO |
2015100372 | Jul 2015 | WO |
WO-2016033676 | Mar 2016 | WO |
2016123495 | Aug 2016 | WO |
2016168556 | Oct 2016 | WO |
2017011758 | Jan 2017 | WO |
2017011792 | Jan 2017 | WO |
Entry |
---|
International Search Report and Written Opinion dated Mar. 31, 2015 for PCT Application No. PCT/US2014/072255, 8 pages. |
International Search Report and Written Opinion dated Apr. 26, 2016 for PCT Application No. PCT/US2016/015681, 13 pages. |
“How to: Watch from the beginning |About DISH” (Dec. 31, 2014) XP055265764, retrieved on Apr. 15, 2016 from URL:http://about.dish.com/blog/hopper/how-watch-beginning 2 pages. |
International Search Report and Written Opinion dated Jun. 24, 2016 for PCT Application No. PCT/US2016/027691, 13 pages. |
Gionis et al., “Similarity Search in High Dimension via Hashing”, Proceedings of the 25th VLDB Conference, 1999, 12 pages. |
Huang , “Bounded Coordinate System Indexing for Real-time Video Clip Search”, Retrieved from the Internet:URL:http://staff.itee.uq.edu.aujzxf/papers/TOIS.pdf, Jan. 1, 2009, 32 pages. |
Kim et al., “Edge-Based Spatial Descriptor Using Color Vector Angle for Effective Image Retrieval”, Modeling Decisions for Artificial Intelligence; [Lecture Notes in Computer Science; Lecture Notes in Artificial Intelligence, Jul. 1, 2005, pp. 365-375. |
Liu et al., “Near-duplicate video retrieval”, ACM Computing Surveys, vol. 45, No. 4, Aug. 30, 2013, pp. 1-23. |
International Search Report and Written Opinion dated Oct. 12, 2016 for PCT Application No. PCT/US2016/042522,13 pages. |
International Search Report and Written Opinion dated Oct. 11, 2016 for PCT Application No. PCT/US2016/042621, 13 pages. |
International Search Report and Written Opinion dated Oct. 20, 2016 for PCT Application No. PCT/US2016/042611, 12 pages. |
Scouarnec et al., “Cache policies for cloud-based systems:To keep or not to keep”, 2014 IEEE 7th International Conference on Cloud Computing, IEEE XP032696624, Jun. 27, 2014, pp. 1-8. |
International Search Report and Written Opinion dated Oct. 25, 2016 for PCT Application No. PCT/US2016/042564, 14 pages. |
Anonymous; “Cache (computing)” Wikipedia, the free encyclopedia, URL:http://en.wikipedia.org/w/index.phpti tle=Cache(computing)&oldid=474222804, Jan. 31, 2012; 6 pages. |
International Search Report and Written Opinion dated Oct. 24, 2016 for PCT Application No. PCT/US2016/042557, 11 pages. |
Anil K. Jain, “Image Coding Via a Nearest Neighbors Image Model” IEEE Transactions on Communications, vol. Com-23, No. 3, Mar. 1975, pp. 318-331. |
Lee et al., “Fast Video Search Algorithm for Large Video Database Using Adjacent Pixel Intensity Difference Quantization Histogram Feature” International Journal of Computer Science and Network Security, vol. 9, No. 9, Sep. 2009, pp. 214-220. |
Li et al., A Confidence Based Recognition System for TV Commercial Extraction, Conferences in Research and Practice in Information Technology vol. 75, 2008. |
International Search Report and Written Opinion dated Jul. 27, 2011 for PCT Application No. PCT/US2010/057153, 8 pages. |
International Search Report and Written Opinion dated Aug. 31, 2011 for PCT Application No. PCT/US2010/057155, 8 pages. |
International Search Report and Written Opinion dated Aug. 26, 2014 for PCT Application No. PCT/US2014/030782; 11 pages. |
International Search Report and Written Opinion dated Jul. 21, 2014 for PCT Application No. PCT/US2014/030795; 10 pages. |
International Search Report and Written Opinion, dated Jul. 25, 2014 for PCT Application No. PCT/US2014/030805, 10 pages. |
Extended European Search Report dated Mar. 7, 2013 for European Application No. 12178359.1, 8 pages. |
Extended European Search Report dated Oct. 11, 2013 for European Application No. 10844152.8, 19 pages. |
Kabal (P.), Ramachandran (R.P.): The computation of line spectral frequencies using Chebyshev polynomials, IEEE Trans. on ASSP, vol. 34, No. 6, pp. 1419-1426, 1986. |
Itakura (F.): Line spectral representation of linear predictive coefficients of speech signals, J. Acoust. Soc. Amer., vol. 57, Supplement No. 1, S35, 1975, 3 pages. |
Bistritz (Y.), Pellerm (S.): Immittance Spectral Pairs (ISP) for speech encoding, Proc. ICASSP'93, pp. 11-9 to 11-12. |
International Search Report and Written Opinion dated Mar. 8, 2016 for PCT Application No. PCT/ US2015/062945; 9 pages. |
Extended European Search Report dated Dec. 21, 2016 for European Application No. 14763506.4, 11 pages. |
Extended European Search Report dated Nov. 23, 2016 for European Application No. 14764182.3, 10 pages. |
Extended European Search Report dated Jan. 24, 2017 for European Application No. 14762850.7, 12 pages. |
Extended European Search Report dated Jun. 16, 2017, for European Patent Application No. 14873564.0, 8 pages. |
U.S. Appl. No. 14/551,933, “Final Office Action”, dated May 23, 2016, 19 pages. |
U.S. Appl. No. 14/551,933 , “Non-Final Office Action”, dated Oct. 17, 2016, 15 pages. |
U.S. Appl. No. 14/551,933 , “Non-Final Office Action”, dated Dec. 31, 2015, 24 pages. |
U.S. Appl. No. 14/551,933 , “Notice of Allowance”, dated Mar. 21, 2017, 8 pages. |
U.S. Appl. No. 14/217,039 , “Non-Final Office Action”, dated May 23, 2014, 27 pages. |
U.S. Appl. No. 14/217,039 , “Final Office Action”, dated Nov. 7, 2014, 15 pages. |
U.S. Appl. No. 14/217,039 , “Notice of Allowance”, dated Jan. 29, 2015, 8 pages. |
U.S. Appl. No. 14/678,856, “Non-Final Office Action”, dated Dec. 1, 2015, 28 pages. |
U.S. Appl. No. 14/678,856, “Notice of Allowance”, dated May 20, 2016, 9 pages. |
U.S. Appl. No. 14/217,075, “Non-Final Office Action”, dated Jul. 16, 2014, 39 pages. |
U.S. Appl. No. 14/217,075, “Notice of Allowance”, dated Feb. 20, 2015, 51 pages. |
U.S. Appl. No. 14/217,094, “Notice of Allowance”, dated Sep. 4, 2014, 30 pages. |
U.S. Appl. No. 14/217,375, “Non-Final Office Action”, dated Apr. 1, 2015, 39 pages. |
U.S. Appl. No. 14/217,375, “Notice of Allowance”, dated Apr. 1, 2015, 31 pages. |
U.S. Appl. No. 14/217,425, “Non-Final Office Action”, dated Apr. 7, 2015, 12 pages. |
U.S. Appl. No. 14/217,425, “Notice of Allowance”, dated May 20, 2015, 15 pages. |
U.S. Appl. No. 14/217,435, “Non-Final Office Action”, dated Nov. 24, 2014, 9 pages. |
U.S. Appl. No. 14/217,435, “Notice of Allowance”, dated Jun. 5, 2015, 9 pages. |
U.S. Appl. No. 15/011,099 , “First Action Interview Office Action Summary”, dated May 9, 2017, 6 pages. |
U.S. Appl. No. 15/011,099 , “First Action Interview Pilot Program Pre-Interview Communication”, dated Feb. 28, 2017, 5 pages. |
U.S. Appl. No. 12/788,721 , “Non-Final Office Action”, dated Mar. 28, 2012, 15 Pages. |
U.S. Appl. No. 12/788,721 , “Final Office Action”, dated Aug. 15, 2012, 22 Pages. |
U.S. Appl. No. 12/788,721 , “Notice of Allowance”, dated Aug. 15, 2013, 16 Pages. |
U.S. Appl. No. 14/763,158 , “Non-Final Office Action”, dated Jun. 27, 2016, 16 Pages. |
U.S. Appl. No. 14/763,158 , “Final Office Action”, dated Sep. 7, 2016, 12 Pages. |
U.S. Appl. No. 14/763,158 , “Notice of Allowance”, dated Mar. 17, 2016, 8 Pages. |
U.S. Appl. No. 14/807,849 , “Non-Final Office Action”, dated Nov. 25, 2015, 12 Pages. |
U.S. Appl. No. 14/807,849 , “Final Office Action”, dated Apr. 19, 2016, 13 pages. |
U.S. Appl. No. 14/807,849 , “Non-Final Office Action”, dated Feb. 28, 2017, 10 Pages. |
U.S. Appl. No. 14/089,003 , “Notice of Allowance”, dated Jul. 30, 2014, 24 Pages. |
U.S. Appl. No. 12/788,748 , “Non-Final Office Action”, dated Jan. 10, 2013, 10 Pages. |
U.S. Appl. No. 12/788,748 , “Final Office Action”, dated Nov. 21, 2013, 13 Pages. |
U.S. Appl. No. 12/788,748 , “Notice of Allowance”, dated Mar. 6, 2014, 7 Pages. |
U.S. Appl. No. 14/953,994 , “Non-Final Office Action”, dated Mar. 3, 2016, 34 Pages. |
U.S. Appl. No. 14/953,994 , “Final Office Action”, dated Jun. 1, 2016, 36 Pages. |
U.S. Appl. No. 14/953,994 , “Notice of Allowance”, dated Aug. 31, 2016, 15 Pages. |
U.S. Appl. No. 14/807,849 , “Final Office Action”, dated Jun. 22, 2017, 10 pages. |
U.S. Appl. No. 15/011,099 , “Final Office Action”, dated Jul. 24, 2017, 22 pages. |
U.S. Appl. No. 15/240,801 , “Non-Final Office Action”, dated Aug. 11, 2017, 18 pages. |
U.S. Appl. No. 15/240,815 , “Non-Final Office Action”, dated Aug. 23, 2017, 15 pages. |
U.S. Appl. No. 15/211,345 , “First Action Interview Pilot Program Pre-Interview Communication”, dated Sep. 19, 2017, 8 pages. |
U.S. Appl. No. 14/807,849 , “Notice of Allowance”, dated Nov. 30, 2017, 9 Pages. |
U.S. Appl. No. 15/240,801 , “Final Office Action”, dated Dec. 22, 2017, 24 pages. |
U.S. Appl. No. 15/011,099, “Non-Final Office Action”, dated Jan. 22, 2018, 23 pages. |
U.S. Appl. No. 15/240,815 , “Final Office Action”, dated Mar. 2, 2018, 14 pages. |
U.S. Appl. No. 15/211,345 , “Final Office Action”, dated Mar. 2, 2018, 14 pages. |
Extended European Search Report dated Mar. 22, 2018 for European Application No. 15865033.3, 10 pages. |
U.S. Appl. No. 15/099,842 , “Final Office Action”, dated Apr. 2, 2018, 8 pages. |
U.S. Appl. No. 15/210,730, “Notice of Allowance”, dated May 23, 2018, 10 pages. |
U.S. Appl. No. 15/796,706, “Non-Final Office Action”, dated Jun. 26, 2018, 17 pages. |
U.S. Appl. No. 15/011,099, “Notice of Allowance”, dated Jun. 28, 2018, 12 pages. |
U.S. Appl. No. 15/796,698, “Non-Final Office Action”, dated Jul. 5, 2018, 15 pages. |
U.S. Appl. No. 15/240,801, “Notice of Allowance”, dated Aug. 30, 2018, 9 pages. |
U.S. Appl. No. 15/211,345, “Non-Final Office Action”, dated Sep. 4, 2018, 13 pages. |
U.S. Appl. No. 15/099,842, “Notice of Allowance”, dated Sep. 7, 2018, 10 pages. |
U.S. Appl. No. 15/240,815, “Notice of Allowance”, dated Sep. 12, 2018, 9 pages. |
U.S. Appl. No. 15/290,848 , “First Action Interview Office Action Summary”, Nov. 2, 2018, 5 pages. |
U.S. Appl. No. 15/796,692, “Notice of Allowance”, dated Dec. 5, 2018, 8 pages. |
U.S. Appl. No. 15/796,698, “Notice of Allowance”, dated Dec. 20, 2018, 8 pages. |
U.S. Appl. No. 15/796,706, “Notice of Allowance”, dated Jan. 11, 2019, 9 pages. |
U.S. Appl. No. 15/211,508, “Non-Final Office Action”, dated Jan. 10, 2019, 20 pages. |
U.S. Appl. No. 15/211,492, “Non-Final Office Action”, dated Jan. 11, 2019, 19 pages. |
U.S. Appl. No. 16/141,598, “First Action Interview Office Action Summary”, dated Jan. 11, 2019, 8 pages. |
Number | Date | Country | |
---|---|---|---|
20190179851 A1 | Jun 2019 | US |
Number | Date | Country | |
---|---|---|---|
61182334 | May 2009 | US | |
61290714 | Dec 2009 | US | |
61791578 | Mar 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15240815 | Aug 2016 | US |
Child | 16214875 | US | |
Parent | 14678856 | Apr 2015 | US |
Child | 15240815 | US | |
Parent | 14089003 | Nov 2013 | US |
Child | 14551933 | US | |
Parent | 12788721 | May 2010 | US |
Child | 14089003 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12788721 | May 2010 | US |
Child | 14678856 | US | |
Parent | 12788748 | May 2010 | US |
Child | 12788721 | US | |
Parent | 14089003 | Nov 2013 | US |
Child | 12788748 | US | |
Parent | 14217039 | Mar 2014 | US |
Child | 14089003 | US | |
Parent | 14217075 | Mar 2014 | US |
Child | 14217039 | US | |
Parent | 14217094 | Mar 2014 | US |
Child | 14217075 | US | |
Parent | 14217375 | Mar 2014 | US |
Child | 14217094 | US | |
Parent | 14217425 | Mar 2014 | US |
Child | 14217375 | US | |
Parent | 14217435 | Mar 2014 | US |
Child | 14217425 | US | |
Parent | 14551933 | Nov 2014 | US |
Child | 14217435 | US |