The field of the present invention relates to targeted television ad replacement using a so-called “smart TV.” In particular, systems and methods are disclosed herein for terminating or altering targeted television ad replacement in response to a user-control action.
A goal of modern television advertising is targeted selection of advertisements for individual households or even for specific viewers in those households. Numerous techniques and methodologies are available for (i) collecting user profile information from one or more sources (online or offline), (ii) using that profile information to select one or more targeted advertisements, (iii) using the profile information to select one or more programs or channels for presenting the targeted advertisements, and (iv) correlating subsequent viewer actions after presenting the targeted advertisements. Some of these techniques and methodologies are described in:
U.S. Pat. No. 7,861,260 entitled “Targeted television advertisements based on online behavior” issued Dec. 28, 2010 to Shkedi;
U.S. Pat. No. 8,566,164 entitled “Targeted online advertisements based on viewing or interacting with television advertisements” issued Oct. 22, 2013 in the names of Shkedi et al.;
U.S. Pat. No. 9,083,853 entitled “Targeted television advertisements associated with online users' preferred television programs or channels” issued Jul. 14, 2015 in the name of Shkedi;
U.S. Pat. No. 8,051,444 entitled “Targeted television advertisements selected on the basis of an online user profile and presented with television programs or channels related to that profile” issued Nov. 1, 2011 to Shkedi; and
U.S. Pat. No. 7,890,609 entitled “Requesting offline profile data for online use in a privacy-sensitive manner” issued Feb. 15, 2011 to Shkedi.
Typically a targeted television advertisement is inserted into the stream of presented television content in place of a non-targeted advertisement. A “targeted advertisement” refers to an advertisement shown to users based on taking into account user-specific profile information. A “non-targeted advertisement” refers to an advertisement shown to a large group of users without use of user-specific profile information; it may be “targeted” in a different sense, such as selected based on broad demographic characteristics, such as the demographic characteristics of users who watch a particular television program. A variety of technological approaches are employed to achieve proper insertion of targeted television advertisements into the stream of presented television content.
In some cases, capabilities of a so-called “smart television” (i.e., smart TV 100 shown in
As also seen in
An example of a conventional method for inserting a targeted television advertisement into stream 200, shown in
In all cases discussed herein, data streams like first data 103 can comprise a so-called “fingerprint” of one or more portions of the primary television content (e.g., a digital video fingerprint generated in any suitable way), a compressed digital encoding of one or more portions of the primary television content, a raw, uncompressed encoding of one or more portions of the primary television content, or other suitable representative data that enables smart TV 100 to identify the corresponding television content in television signal feed 200 (i.e., replaced 203 or preceding 202 content). If a fingerprint of the primary television content is employed, it can be generated according to any one of several known techniques or protocols. One example of a fingerprint of a segment of television content can rely on sampling within every video frame of that segment of the television content. Smart TV 100, or a central server in possession of the segment samples, can compare the samples of the segment with a sample taken by smart TV 100 from television signal feed 200. Based on that comparison, smart TV 100 or the central server can recognize a specific frame within the segment. Likewise, the second data can comprise any suitable compressed or uncompressed encoding of the secondary television content (i.e., replacement content 205). Different or parallel compression protocols can be used for first 103 and second 105 data. If a fingerprint of the primary television content is employed for first data 103, it can be generated according to any one of several known techniques or protocols (one example can include taking a sample of audio, video, or both from TV feed 200 and comparing it with an equivalent sample of known content). A given fingerprint of television content can be generated based on both video and audio portions of that content, on the video portion only, or on the audio portion only.
In some other examples (in which fingerprints are not employed or employed partially), smart TV 100 receives via Internet 140 first 103 or second 105 data in, e.g., MPEG4 or other compressed format. The compressed format saves bandwidth so as to enable the parallel transmission of a multitude of different targeted ads to a multitude of corresponding households watching television simultaneously. In one example, entire advertisements (not just portions; the preceding, replaced, and targeted advertisements) are sent to smart TV 100, which then uncompresses the three advertisements. The uncompressed preceding ad 202 is compared with uncompressed television signal feed 200 transmitted from STB 130, e.g., via HDMI. At the end of preceding ad 202, uncompressed, targeted, replacement advertisement 205 is inserted by smart TV 100 in place of replaced advertisement 203. While presenting targeted, replacement advertisement 205, smart TV 100 monitors television signal feed 200 for changes in replaced advertisement 203 by comparing feed 200 with replaced advertisement 203 received over the Internet 140. Alternatively, instead of comparing uncompressed television signal feed 200 and received advertisements, smart TV 100 instead can compress television signal feed 200 and do the comparisons described above using compressed television signal feed 200 and received advertisements.
The example methods and systems described above create a technical problem relating to viewer control. Typically, television signal source device 130 mediates viewer control of the presentation of the primary television content. Various user-control actions can be executed through television signal source device 130, whether by direct manipulation (e.g., pushing buttons on a STB), or by use of a remote control or auxiliary device 150 (e.g., a tablet or smartphone software application controlling the STB). Examples of user-control actions executed through television signal source device 130 can include a channel change or a display change.
A channel change denotes switching among multiple different streams of distinct broadcast, multicast, unicast, or locally stored television content that can be live, prerecorded, time-shifted, or on-demand. A display change can include: (i) a so-called “trick mode” (e.g.: fast forward at various speeds such as 4×, 8×, 32×, or other; rewind or fast reverse at various speeds such as 4×, 8×, 32×, or other; pause or resume; or skipping forward or backward by various specified time intervals, numbers of frames, or scenes); (ii) loop, repeat, or replay; (iii) slow motion or stop action; (iv) changing the zoom, stretch, or aspect ratio; (v) changing viewing angle or other 3D viewing parameter(s); (vi) changing between, e.g., “standard,” “movie,” “sports,” “game,” or other display presets; or (vii) changing backlight, contrast, brightness, color, tint sharpness, color temperature, or other picture quality parameters.
Such user-control actions, when executed through television signal source device 130, affect television signal feed 200, which in turn affects the presentation by smart TV 100 of the primary television content. In that capacity, smart TV 100 acts as a passive presentation device, becoming “active” (i.e., directly affecting the presentation of television content) only when it presents a replacement advertisement or other replacement television content 205. Note that some display changes can also be executed using smart TV 100 (e.g., items (iv)-(vii) listed above, or even items (i)-(iii) if smart TV 100 has sufficient memory 120 or buffering capabilities), even when presenting television content provided by television signal source device 130 via television signal feed 200. However, only those display changes executed through television signal source device 130 are considered for purposes of the present disclosure.
Because user-control actions, when executed through television signal source device 130, affect only television signal feed 200 from television signal source device 130, such user-control actions ordinarily would have no discernible effect on any replacement television content 205 being presented on smart TV display 101. In other words, in conventional techniques, if a viewer attempts to execute a control action during presentation of replacement content 205, the control action would appear disabled. The control action indeed would affect television signal feed 200 (fast forward, pause, etc.), but display 101 would continue to show replacement content 205, undisturbed by the control action. This state of affairs (i.e., seemingly ineffectual user-control actions) would persist until replacement content 205 finished showing, at which time presentation would revert to the primary television content (presumably modified, unbeknownst to the user, in compliance with the intervening user-control actions, or perhaps just the last user-control action).
It would be desirable to provide systems and methods that mitigate the problem described above.
A method is performed using computer 160 associated with television display 101 and connected to computer network 140 and to television signal source device 130 (collectively, a so-called smart TV 100). Smart TV 100 receives from television signal source device 130 television signal feed 200 having encoded thereon primary television content, and presents the primary television content via television display 101. Computer 160 is programmed to utilize data received via computer network 140, including (i) first data 103 comprising electronic indicia of a to-be-replaced portion 203 of the primary television content and (ii) second data 105 comprising electronic indicia of secondary television content 205. Computer 160 is programmed to present automatically via television display 101, using second data 105, at least a portion of secondary television content 205 in place of the to-be-replaced television content 203, see
The method comprises: (a) during presentation of secondary television content 205, automatically monitoring, with computer 160, television signal feed 200 and comparing it with the first data 103; and (b) in response to detecting any difference, between television signal feed 200 and first data 103, indicative of a user-control action with respect to television signal source device 130, automatically altering, with computer 160, the presentation of secondary television content 205 in accordance with the user-control action.
Objects and advantages pertaining to presenting replacement television content using a smart TV may become apparent upon referring to the embodiments disclosed in the following written description or outlined in the appended claims.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
In one or more of the illustrative embodiments disclosed herein, various examples of methods for inserting a targeted television advertisement into a stream of primary television content can be performed as follows using various examples of systems that include computer network connection 145 and computer 160 of smart TV 100. Parts of those methods proceed substantially as described above.
In a first group of one or more embodiments, using computer 160 and first data 103, smart TV 100 automatically compares, as in 310 of
If replacement television content 205 ends before replaced television content 203, an end portion of replaced television content 203 might be presented after completion of replacement content 205. That may result in a television presentation that may be jarring, confusing, or annoying to viewers due to the abrupt and seemingly mysterious appearance of a fragment of replaced content 203. In a second group of one or more embodiments, if replacement content 205 reaches its end before the end of replaced content 203, smart TV 100 can present a black screen (or a blank screen of a different color, a user-selected color controlled by a user setting in the smart TV, or any other suitable filler material) to fill the gap between the end of replacement content 205 and the end of replaced content 203. A black screen can be used in countries where a black screen identifies the end of one segment of television content and the beginning of a succeeding segment of unrelated television content (e.g., a transition from a television program to a television advertisement, from one television advertisement to another, or from a television advertisement to a television program). In countries where a different type of screen is used to identify the end of one segment of content and the beginning of the next, that corresponding kind of screen will be inserted as filler content by smart TV 100 to fill the gap from the end of replacement content 205 to the end of replaced content 203. In one example implementation of a black screen, smart TV 100 inserts the black frames. When smart TV 100, using computer 160 and first data 103, detects the end of replaced content 203 in television signal feed 200 while presenting such filler content, it can then revert presentation to the primary television content encoded in television signal feed 200, providing the viewer with a smooth transition from replacement content 205 to succeeding content 204.
If, on the other hand, smart TV 100 does detect a discrepancy, in 320 of
However, as with the case where replacement content 205 ends before replaced content 203 ends, this situation also results in presentation of television content that may appear jarring, confusing, or annoying to viewers due to the sudden switch, upon attempting to execute a user-control action, between different television advertisements. An exception is the case wherein the user-control action is a channel change; in that instance the abrupt appearance of different television content would be expected.
A fourth group of one or more embodiments help eliminate the presentation problems described in the previous paragraphs. In response to detecting, at 320 of
If the user-control action is a channel change, then smart TV 100 can simply terminate presentation of replacement television content 205, see 360 of
If smart TV 100 determines that the user-control action is a display change such as a trick mode, then, using computer 160 and second data 105, smart TV 100 can alter presentation of replacement content 205, see 350 of
In a fifth group of one or more embodiments, after replacement content 205 has been presented and its display completed, smart TV 100 can again insert replacement content 205 if, for example, a rewind or fast reverse user-control action is executed beginning from a time after the conclusion of replacement content 205. To accomplish this, the sequence of steps described above can be executed again, with smart TV 100 detecting succeeding content 204, presenting replacement content 205 (possibly altered depending on user-control actions), monitoring television signal feed 200 during a second presentation of replacement content 205, and terminating or altering the repeated presentation of replacement content 205 in response to a user-control action. A purpose of this technique is to allow the user to review replacement content 205 after the succeeding television show (for example) resumes. Without this technique, if the user rewinds from a point after replacement content 205 has been completed, the user would see replaced content 203 instead of replacement content 205, and the user would not be able to find replacement content 205. To enable such “rewinding” of replacement content 205, smart TV 100 receives electronic indicia of additional data 104 that represents a portion of the primary television content that immediately follows replaced content 203 (i.e., succeeding content 204) or creates such additional data by itself, for example by fingerprinting succeeding content 204 or simply recording it. Smart TV 100 automatically monitors the rewound television signal feed 200 it receives from television signal source 130 and compares it with additional data 104 to detect or otherwise determine the beginning of succeeding content 204.
In some instances, the same replacement content 205 can be inserted each time replaced content 203 (or preceding content 202 or succeeding content 204) is detected by smart TV 100. In other instances, different replacement content can be inserted in place of replaced content 203; the choice of replacement content is part of the targeting process when the replacement content is a targeted advertisement. Any suitable sequence of replacement content can be employed for insertion in place of a given portion of replaced content 203 upon repeated encounters with that given portion. In one example, the same advertisement can be presented each time the given portion of replaced content 203 is encountered (forward or backward). In another example, one advertisement can be presented a number of times, followed by another advertisement presented a number of times, and so on. In a third example, number of different advertisements can be rotated through one at a time at each encounter with the replaced content. In a fourth example, each replacement advertisement is presented during only one encounter with the given replacement content; at each subsequent encounter a new advertisement is presented that has not been presented before.
In a sixth group of one or more embodiments, if the television content that includes the preceding and succeeding content 202, 204 is replayed at a later time (e.g., stored on a DVR or retrieved from an “on-demand” system), the same replacement content 205 can be presented each time the primary content is replayed, or different replacement content can be selected than would be selected for the initial or earlier presentations of the content. Subsequent replays can include still other replacement content. Any suitable variation, selection, or rotation of replacement content, including those described in the preceding paragraph, can be employed.
The systems and methods disclosed herein can be implemented as general or special purpose computers or servers or other programmable hardware devices programmed through software, or as hardware or equipment “programmed” through hard wiring, or a combination of the two. A “computer” (e.g., as in a “smart TV”) can comprise a single processor, processor core, or machine or can comprise multiple interacting processors, processor cores, or machines (located at a single location or at multiple locations remote from one another). A computer memory or computer-readable medium can be encoded with a computer program, so that execution of that program by one or more computers causes the one or more computers to perform one or more of the methods disclosed herein. Suitable media can include temporary or permanent storage or replaceable media, such as network-based or Internet-based or otherwise distributed storage of software modules that operate together, RAM, ROM, CD ROM, CD-R, CD-R/W, DVD ROM, DVD±R, DVD±R/W, hard drives, thumb drives, flash memory, optical media, magnetic media, semiconductor media, or any future storage alternatives. Such encoded media can be preinstalled already encoded in a smart TV during its manufacture, can be encoded after installation into a smart TV during its manufacture, or can be encoded with “client” or “application” software copied or downloaded into the smart TV after its manufacture. For example, client software can be downloaded to the smart TV as part of occasional or periodic software updates received by the smart TV, typically via its network connection.
In addition to the preceding, the following examples fall within the scope of the present disclosure or appended claims:
A method performed using a computer associated with a television display and connected to a computer network and to a television signal source device, wherein: a television signal feed having encoded thereon primary television content is received from the television signal source device and the primary television content is presented via the television display; the computer is programmed to utilize data received via the computer network, including (i) first data comprising electronic indicia of a to-be-replaced portion of the primary television content and (ii) second data comprising electronic indicia of secondary television content; the computer is programmed to present automatically via the television display, using the second data, at least a portion of the secondary television content in place of the to-be-replaced television content; and the method comprises: (a) during presentation of the secondary television content, automatically monitoring, with the computer, the television signal feed and comparing it with the first data; and (b) in response to detecting any difference, between the television signal feed and the first data, indicative of a user-control action with respect to the television signal source device, automatically altering, with the computer, the presentation of the secondary television content in accordance with the user-control action.
The method of Example 1 further comprising reverting to presentation of the primary television content after presenting the secondary television content.
The method of any preceding Example wherein (i) the user-control action indicated by the detected difference between the television signal feed and the first data is a channel change, and (ii) altering the presentation of the secondary television content comprises terminating presentation of the secondary television content and resuming presentation of the primary television content.
The method of any preceding Example wherein (i) the user-control action indicated by the detected difference between the television signal feed and the first data is a display change, and (ii) altering the presentation of the secondary television content comprises altering presentation of the secondary television content in a manner analogous to the user-control action indicated by the detected difference between the television signal feed and the first data.
The method of any preceding Example wherein (i) the primary television content comprises one or more television programs and one or more associated television advertisements, and (ii) the secondary television content comprises one or more replacement television advertisements.
The method of any preceding Example further comprising, after presenting the secondary television content, automatically presenting filler television content via the television display in place of an end portion of the to-be-replaced television content.
The method of an preceding Example further comprising, after presenting the secondary television content and after a user-control action of rewinding or fast-reversing the primary television content to a point within or before the to-be-replaced television content, again automatically presenting at least a portion of the secondary television content in place of the to-be-replaced television content.
The method of any preceding Example further comprising: (c) after presenting the secondary television content and during later time-shifted or on-demand presentation of the primary television content via the television display, automatically presenting different secondary television content in place of the to-be-replaced television content; (d) during presentation of the different secondary television content, automatically monitoring, with the computer, the television signal feed and comparing it with the first data; and (e) in response to detecting any difference, between the television signal feed and the first data, indicative of a user-control action with respect to the television signal source device, automatically altering, with the computer, the presentation of the different secondary television content in accordance with the user-control action.
A smart TV comprising: a connection to a television signal source device arranged so as to receive from the television signal source a television signal feed having encoded thereon primary television content; a television display structured and connected to present the primary television content; a computer; and a connection to a computer network, wherein the computer is programed to perform the method of any preceding Example.
An article comprising a computer-readable medium in a smart TV or in a storage device controlled by a server connected to a computer network, which medium encodes computer-readable instructions that, when applied to a computer in a smart TV, instruct the computer to perform the method of any preceding Example.
It is intended that equivalents of the disclosed illustrative embodiments and methods shall fall within the scope of the present disclosure or appended claims. It is intended that the disclosed illustrative embodiments and methods, and equivalents thereof, may be modified while remaining within the scope of the present disclosure or appended claims.
In the foregoing Detailed Description, various features may be grouped together in several illustrative embodiments for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that any claimed embodiment requires more features than are expressly recited in the corresponding claim. Rather, as the appended claims reflect, inventive subject matter may lie in less than all features of a single disclosed embodiment. Thus, the appended claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate disclosed embodiment. However, the present disclosure shall also be construed as implicitly disclosing any embodiment having any suitable set of one or more disclosed or claimed features (i.e., sets of features that are not incompatible or mutually exclusive) that appear in the present disclosure or the appended claims, including those sets that may not be explicitly disclosed herein. In particular, any embodiment that falls within any one of the six groups of embodiments disclosed above, and any embodiment that falls within any combination of two or more of those six groups of embodiments, shall be regarded as falling within the scope of the present disclosure or appended claims. Further, any embodiment explicitly or implicitly disclosed herein can be implemented using any suitable set of one or more known or future developed features, including but not limited to those disclosed in the Background. It should be further noted that the scope of the appended claims do not necessarily encompass the whole of the subject matter disclosed herein.
For purposes of the present disclosure and appended claims, the conjunction “or” is to be construed inclusively (e.g., “a dog or a cat” would be interpreted as “a dog, or a cat, or both”; e.g., “a dog, a cat, or a mouse” would be interpreted as “a dog, or a cat, or a mouse, or any two, or all three”), unless: (i) it is explicitly stated otherwise, e.g., by use of “either . . . or,” “only one of,” or similar language; or (ii) two or more of the listed alternatives are mutually exclusive within the particular context, in which case “or” would encompass only those combinations involving non-mutually-exclusive alternatives. For purposes of the present disclosure or appended claims, the words “comprising,” “including,” “having,” and variants thereof, wherever they appear, shall be construed as open ended terminology, with the same meaning as if the phrase “at least” were appended after each instance thereof.
In the appended claims, if the provisions of 35 USC § 112 ¶ 6 are desired to be invoked in an apparatus claim, then the word “means” will appear in that apparatus claim. If those provisions are desired to be invoked in a method claim, the words “a step for” will appear in that method claim. Conversely, if the words “means” or “a step for” do not appear in a claim, then the provisions of 35 USC § 112 ¶ 6 are not intended to be invoked for that claim.
If any one or more disclosures are incorporated herein by reference and such incorporated disclosures conflict in part or whole with, or differ in scope from, the present disclosure, then to the extent of conflict, broader disclosure, or broader definition of terms, the present disclosure controls. If such incorporated disclosures conflict in part or whole with one another, then to the extent of conflict, the later-dated disclosure controls.
The Abstract is provided as required as an aid to those searching for specific subject matter within the patent literature. However, the Abstract is not intended to imply that any elements, features, or limitations recited therein are necessarily encompassed by any particular claim. The scope of subject matter encompassed by each claim shall be determined by the recitation of only that claim.
This application is a continuation of application Ser. No. 15/429,601, filed Feb. 10, 2017, now U.S. Pat. No. 10,009,636, which is a continuation of application Ser. No. 14/763,963, filed Jul. 28, 2015, now U.S. Pat. No. 9,571,865, which is a U.S. National Stage Entry filed under 35 U.S.C. § 371 of application PCT/US2014/013924, filed Jan. 30, 2014, which claims the benefit of U.S. application Ser. No. 13/754,639, filed Jan. 30, 2013, now U.S. Pat. No. 8,713,600.
Number | Name | Date | Kind |
---|---|---|---|
6160989 | Hendricks et al. | Dec 2000 | A |
6216129 | Eldering | Apr 2001 | B1 |
6366298 | Haitsuka et al. | Apr 2002 | B1 |
6446261 | Rosser | Sep 2002 | B1 |
6698020 | Zigmond et al. | Feb 2004 | B1 |
6832207 | Shkedi | Dec 2004 | B1 |
6845396 | Kanojia | Jan 2005 | B1 |
6925440 | Shkedi | Aug 2005 | B1 |
6973436 | Shkedi | Dec 2005 | B1 |
7020888 | Reynolds et al. | Mar 2006 | B2 |
7043746 | Ma | May 2006 | B2 |
7072853 | Shkedi | Jul 2006 | B2 |
7092926 | Cerrato | Aug 2006 | B2 |
7103904 | Blackketter et al. | Sep 2006 | B1 |
7152237 | Flickinger et al. | Dec 2006 | B2 |
7260823 | Schlack et al. | Aug 2007 | B2 |
7428493 | Shkedi | Sep 2008 | B2 |
7454364 | Shkedi | Nov 2008 | B2 |
7747745 | Shkedi | Jun 2010 | B2 |
7818763 | Sie et al. | Oct 2010 | B2 |
7822637 | Shkedi | Oct 2010 | B2 |
7822639 | Shkedi | Oct 2010 | B2 |
7861260 | Shkedi | Dec 2010 | B2 |
7890609 | Shkedi | Feb 2011 | B2 |
7899915 | Reisman | Mar 2011 | B2 |
7941528 | Hicks et al. | May 2011 | B2 |
7979307 | Shkedi | Jul 2011 | B2 |
7984466 | Eldering et al. | Jul 2011 | B2 |
7987491 | Reisman | Jul 2011 | B2 |
8051444 | Shkedi | Nov 2011 | B2 |
8065696 | Scott, III et al. | Nov 2011 | B2 |
8079052 | Chen et al. | Dec 2011 | B2 |
8108895 | Anderson et al. | Jan 2012 | B2 |
8166501 | Shikuma | Apr 2012 | B2 |
8200822 | Shkedi | Jun 2012 | B1 |
8204783 | Shkedi | Jun 2012 | B2 |
8204965 | Shkedi | Jun 2012 | B2 |
8239264 | Shkedi | Aug 2012 | B2 |
8244574 | Shkedi | Aug 2012 | B2 |
8244582 | Shkedi | Aug 2012 | B2 |
8244583 | Shkedi | Aug 2012 | B2 |
8244586 | Shkedi | Aug 2012 | B2 |
8280758 | Shkedi | Oct 2012 | B2 |
8281336 | Shkedi | Oct 2012 | B2 |
8307390 | Holden | Nov 2012 | B2 |
8341247 | Shkedi | Dec 2012 | B2 |
8375409 | Chang et al. | Feb 2013 | B2 |
8494904 | Shkedi | Jul 2013 | B2 |
8566164 | Shkedi | Oct 2013 | B2 |
8589210 | Shkedi | Nov 2013 | B2 |
8595069 | Shkedi | Nov 2013 | B2 |
8600815 | Shkedi | Dec 2013 | B2 |
8607267 | Shkedi | Dec 2013 | B2 |
8671139 | Shkedi | Mar 2014 | B2 |
8677398 | Shkedi | Mar 2014 | B2 |
8683502 | Shkedi | Mar 2014 | B2 |
8695032 | Shkedi | Apr 2014 | B2 |
8775249 | Shkedi | Jul 2014 | B2 |
8959146 | Shkedi | Feb 2015 | B2 |
8997138 | Shkedi | Mar 2015 | B2 |
9071886 | Shkedi | Jun 2015 | B2 |
9078035 | Shkedi | Jul 2015 | B2 |
9083853 | Shkedi | Jul 2015 | B2 |
9131282 | Shkedi | Sep 2015 | B2 |
9208514 | Shkedi | Dec 2015 | B2 |
9226019 | Shkedi | Dec 2015 | B2 |
9271024 | Shkedi | Feb 2016 | B2 |
9351053 | Shkedi | May 2016 | B2 |
9369779 | Shkedi | Jun 2016 | B2 |
9508089 | Shkedi | Nov 2016 | B2 |
9591380 | Shkedi et al. | Mar 2017 | B2 |
9756372 | Shkedi | Sep 2017 | B2 |
9800917 | Shkedi | Oct 2017 | B2 |
9813778 | Shkedi | Nov 2017 | B2 |
9830615 | Shkedi | Nov 2017 | B2 |
20020013943 | Haberman et al. | Jan 2002 | A1 |
20020019769 | Barrits et al. | Feb 2002 | A1 |
20020059602 | Macrae et al. | May 2002 | A1 |
20020078444 | Krewin et al. | Jun 2002 | A1 |
20020082910 | Kontogouris | Jun 2002 | A1 |
20020120929 | Schwalb et al. | Aug 2002 | A1 |
20020120933 | Knudson et al. | Aug 2002 | A1 |
20020123928 | Eldering et al. | Sep 2002 | A1 |
20020124249 | Shintani | Sep 2002 | A1 |
20020124253 | Eyer et al. | Sep 2002 | A1 |
20020129368 | Schlack et al. | Sep 2002 | A1 |
20020144262 | Plotnick | Oct 2002 | A1 |
20020178443 | Ishii | Nov 2002 | A1 |
20020191950 | Wang | Dec 2002 | A1 |
20030051242 | Donnelly | Mar 2003 | A1 |
20030066078 | Bjorgan et al. | Apr 2003 | A1 |
20030093792 | Labeeb | May 2003 | A1 |
20030126597 | Darby et al. | Jul 2003 | A1 |
20030135853 | Goldman et al. | Jul 2003 | A1 |
20030145323 | Hendricks et al. | Jul 2003 | A1 |
20030149975 | Eldering et al. | Aug 2003 | A1 |
20030154128 | Liga et al. | Aug 2003 | A1 |
20030188317 | Liew et al. | Oct 2003 | A1 |
20030208756 | Macrae et al. | Nov 2003 | A1 |
20030208758 | Schein et al. | Nov 2003 | A1 |
20030226141 | Krasnow et al. | Dec 2003 | A1 |
20030235390 | Nygren | Dec 2003 | A1 |
20040003406 | Billmaier | Jan 2004 | A1 |
20040005143 | Tsuru et al. | Jan 2004 | A1 |
20040034874 | Hord et al. | Feb 2004 | A1 |
20040103429 | Carlucci | May 2004 | A1 |
20040107437 | Reichardt et al. | Jun 2004 | A1 |
20040117827 | Karaoguz et al. | Jun 2004 | A1 |
20040158858 | Paxton et al. | Aug 2004 | A1 |
20040163101 | Swix et al. | Aug 2004 | A1 |
20040172650 | Hawkins et al. | Sep 2004 | A1 |
20040228605 | Quan et al. | Nov 2004 | A1 |
20050034171 | Benya | Feb 2005 | A1 |
20050076359 | Pierson et al. | Apr 2005 | A1 |
20050086112 | Shkedi | Apr 2005 | A1 |
20050108776 | Carver et al. | May 2005 | A1 |
20050207731 | Unger | Sep 2005 | A1 |
20060013556 | Poslinksi | Jan 2006 | A1 |
20060070095 | Newton et al. | Mar 2006 | A1 |
20060218602 | Sherer et al. | Sep 2006 | A1 |
20060222322 | Levitan | Oct 2006 | A1 |
20060235756 | Pellgrino | Oct 2006 | A1 |
20060277569 | Smith | Dec 2006 | A1 |
20070204310 | Hua et al. | Aug 2007 | A1 |
20080127246 | Sylvain | May 2008 | A1 |
20080256572 | Chen | Oct 2008 | A1 |
20090037949 | Birch | Feb 2009 | A1 |
20090049468 | Shkedi | Feb 2009 | A1 |
20090100460 | Hicks, III et al. | Apr 2009 | A1 |
20090106792 | Kan et al. | Apr 2009 | A1 |
20090172723 | Shkedi et al. | Jul 2009 | A1 |
20090217324 | Massimi | Aug 2009 | A1 |
20090290852 | Wright | Nov 2009 | A1 |
20100030639 | Feng | Feb 2010 | A1 |
20100043022 | Kaftan | Feb 2010 | A1 |
20100058380 | Yu et al. | Mar 2010 | A1 |
20100058382 | Yu et al. | Mar 2010 | A1 |
20100269128 | Gordon | Oct 2010 | A1 |
20110321083 | Rouse et al. | Dec 2011 | A1 |
20120023522 | Anderson et al. | Jan 2012 | A1 |
20120143660 | Jiwang et al. | Jun 2012 | A1 |
20120159549 | Douillet et al. | Jun 2012 | A1 |
20120314140 | Wiser et al. | Dec 2012 | A1 |
20130024888 | Sivertsen | Jan 2013 | A1 |
Number | Date | Country |
---|---|---|
2008-048230 | Aug 2008 | JP |
10-2006-0025219 | Mar 2006 | KR |
WO2007047310 | Apr 2007 | WO |
WO2009157903 | Dec 2009 | WO |
Entry |
---|
Roettgers, Janko; Your next TV set may show you advertising spots you actually like; GigaOM; http://gigaom.com; Dec. 14, 2012. |
Constine, Josh; Gracenote's Ad Replacement System That Personalizes TV Commercials Will Start Trials in 2013; http://techcrunch.com; Dec. 26, 2012. |
Freidman, Wayne; Gracenote, Invidi Deal Ups Real-Time Ad for TV; MediaDailyNews; http://www.mediapost.com; Jan. 16, 2013. |
Mandese, Joe; Why Invidi's Deal With Gracenote Will Accelerate Real-Time TV Targeting; RTBlog; http://www.mediapost.com; Jan. 17, 2013. |
“Venture Market Summary”; VentureWire Alert; Mar. 29, 2005. |
“TiVo Launches New Interactive Advertising Technology”; press release from www.tivo.com; Jul. 18, 2005. |
“TiVo Announces First Advertising Search Product for Television”; press release from www.tivo.com; Nov. 28, 2005. |
“TiVo Begins Rollout of Online Services Now Accessible Directly on the TV”; press release from www.tivo.com; Dec. 1, 2005. |
“TiVo Launches Telecision's New Advertising Search Product”; press release from www.tivo.com; May 8, 2006. |
“TiVo to Insert Ads at End of Programs”; The Wall Street Journal Online; Nov. 28, 2006. |
Haughey, M.; “TiVo to add banner ads to service when fast forwarding”; www.PVRblog.com; Nov. 16, 2004. |
Piccalo, G.; “TiVo Will No Longer Skip Past Advertisers; The tool that lets viewers control the TV will soon spoint ‘billboards’ and track viewing habits”; Los Angeles Times; Nov. 17, 2004. |
Rojas, P.; “TiVo planning banner ads for when you fast-forward”; www.engadget.com; Nov. 17, 2004. |
Haughey, M.; “Icon ads over commercials at TiVo”; www.PRVblog.com; Mar. 29, 2005. |
“Targeted Visitor Ad”; web page at http://adservices.zango.com/Advertise/SearchSolutions.aspx; 2007. |
“Zango to Pay $3 Million, Settles FTC Charges”; archived web page at www.marketingvox.com; Nov. 3, 2006. |
Nelson, M.G.; “Microsoft Launches Clickable Video Ads for Kohl's”; web page at http://clickz.com/3625272; Mar. 19, 2007. |
Joe, R.; “EveryZing Makes Everything Searchable”; web page at http://www.speechtechmag.com/Articles/ReadArticle.aspx?ArticleID=39598; Sep. 11, 2007. |
Luening, E.; “Free ISP NetZero beefs up ad services”; web page at http://www.news.com/Free-ISP-NewZero-beefs-up-ad-services/2100-1023_3-241067.html; Jan. 2, 2002. |
Blinkx TV White Paper 1.0; Blinkx Inc.; Jan. 2005. |
Mandese, J.; “AOL Unveils Video ‘Ticker’ Ads, New Alternative to Pre-Roll”; from www.mediapost.com; Nov. 19, 2007. |
International Search Report and Written Opinion of the International Searching Authority (KR); PCT/US2014/013924; dated May 9, 2014. |
Number | Date | Country | |
---|---|---|---|
20190069005 A1 | Feb 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15429601 | Feb 2017 | US |
Child | 16016448 | US | |
Parent | 14763963 | US | |
Child | 15429601 | US | |
Parent | 13754639 | Jan 2013 | US |
Child | 14763963 | US |