Advertising impression determination

Information

  • Patent Grant
  • 11436630
  • Patent Number
    11,436,630
  • Date Filed
    Tuesday, September 22, 2020
    3 years ago
  • Date Issued
    Tuesday, September 6, 2022
    a year ago
Abstract
Systems and methods for verifying an advertisement impression in a digital environment are provided. In some aspects, methods of the subject technology include operations for defining a portion of the digital environment as an impression area, wherein the impression area is associated with a tagged advertisement area, providing a stream of an advertisement to the tagged advertisement area, and updating advertising impression information stored in memory regarding the advertisement, wherein an advertising impression is based on the identification of the character within the impression area and the availability of an unobstructed line-of-sight between the character and the tagged advertisement area. In some aspects, computer readable media are also provided.
Description
BACKGROUND OF THE INVENTION
Field of the Invention

The present invention generally relates to targeted advertising, such as an advertising system and method for dynamically displaying advertisements in the context of video games (i.e., in-game advertising). More specifically, the present invention provides for the determination and tracking of advertising impressions in response to users interacting with video games having in-game advertising functionality.


Description of the Related Art

One of the many ways the advertising industry governs the success of advertising campaigns is through impressions. Impressions refer to the exposure a user has had to an ad or ad campaign. Impressions are typically indexed in to the number of times a potential consumer views a particular advertisement. For example, a print advertisement located in a kiosk in a shopping center might be viewed by 1,000 shoppers over the course of an afternoon. It could be said that the particular advertisement enjoyed 1,000 impressions as each shopper walked past the kiosk and viewed the goods or services advertised therein.


High-traffic areas offer the opportunity for additional impressions. For example, an advertising kiosk located near the entrance of a popular store in a shopping center might enjoy 10,000 impressions due to high shopper traffic whereas an advertising kiosk located near an unsuccessful store (e.g., a store going out of business) may enjoy significantly less advertising impressions. As advertisers seek to have their goods and services viewed by as many persons as possible, there is obviously a demand for advertisement placement in high traffic areas.


The same theory applies to other advertising media. For example, newspapers and magazines with high circulation enjoy increased advertising revenue because those newspapers and magazines offer the possibility for additional impressions whereas an unpopular or unsuccessful newspaper or magazine as do those publications circulated in small towns or with niche (i.e., limited) readership. An advertisement on a billboard in Times Square in New York City will similarly offer more impressions (and demand higher revenue) than a billboard located adjacent a service road in rural Nebraska.


High traffic areas or high impression opportunities thus become a valuable asset in the advertising community. Assigning value to those assets offers a challenge as it is difficult to accurately measure how many impressions a particular advertisement or advertising opportunity might offer.


For example, television relies on the Nielsen TV ratings system whereby an estimate of the number of people watching any particular television program at any particular time is provided. Based on statistical information provided by these ratings, a determination of which programs are the most watched or the most popular can be made. These programs often demand higher advertising fees as the increased popularity offers the opportunity for additional impressions. For example, the Super Bowl is one of the most watched television events; Super Bowl XVI between the San Francisco 49ers and the Cincinnati Bengals in 1982 saw almost 50% of the televisions in the United States ‘tuned-in’ to the game. As such, advertising ‘spots’ for the Super Bowl commonly demand exorbitant costs; a 30-second spot for Super Bowl XXXIV between the St. Louis Rams and the Tennessee Titans in 2000 demanded close to $2M. A similar supply-and-demand theory applies to regularly scheduled programs (e.g., sitcoms). Popular television shows will demand more advertising dollars due to the increased opportunity for advertising impressions whereas less popular shows will demand considerably less.


Other methodologies exist for measuring advertising impressions. For example, and as previously noted, television and magazines traditionally look to circulation to determine advertising rates whereas physical real-estate (e.g., bill boards) will look at a number of factors including location, size of the bill board and general traffic in the area. Radio and the audio media have similar ratings services, for example, those offered by Arbitron Inc., in conjunction with comScore Media Metrix.


Certain advertising models have significant shortcomings. For example, pop-up Internet advertisements that appear in a Web browser are generally viewed as annoying and while usually gaining the attention of the individual ‘surfing’ the web, also garner their disdain for the interruption of their browsing session often leading to immediately closure of the pop-up window or, as is often the case today, the use of a pop-up blocker whereby pop-up advertisements are prevented from popping-up altogether.


Other technological innovations continue to offer additional advertising challenges. For example, digital video recorders (DVR) like those offered by TiVo® provide the ability to ‘skip’ over advertisements by fast forwarding through the advertisement. Due to the digital nature of the television program stored on a TiVo® DVR, ‘skipping’ over advertisements is simple and does not involve the jerkiness, fast-forward/back-up that accompanied VHS tape recorders and video tapes.


The advertising industry is, therefore, increasingly faced with the inability to target its advertisements to individuals due to the decrease in readership in print media, unrefined advertising methodologies on the Internet and the inability to keep audiences ‘captive’ whereby there is a certain degree of assurance that a consumer views a particular advertisement (e.g., users can now ‘skip’ over ads while remaining in front of their television during a television program).


The video game industry is quickly becoming one of the last bastions of captive audience advertising. That is, the player of a video game often offers their undivided attention to the video game environment so that they may remain aware of actions taking place in the game (e.g., being attacked by an enemy, discovering a cache of weapons or treasure trove, identifying a ‘lane’ through which to navigate a running back in a football game). Video games, therefore, offer the opportunity for placing ads before a captive and extremely attentive audience.


There have been—and continue to be—numerous cases wherein actual advertisements of advertisers are deployed and displayed within a video game environment. A classic example is in a driving game, wherein advertisements are pasted onto billboards around a driving course as illustrated in U.S. Pat. Nos. 5,946,664 and 6,539,544, the disclosures of which are incorporated herein by reference. With such in-game advertising, the software publishing company that creates the video game identifies an advertiser, creates texture data based on ad copy provided by the advertiser and places this texture data representative of an advertisement in the video game environment (i.e., posting the advertisement on the billboard).


Online and networked gaming is increasing in popularity throughout the world. With this increase in popularity, there is an expectation that gaming networks assemble standards and evolve into an advertising channel such as television and radio. As a part of this increase and evolution, there is a need for a framework and system for advertisers and media providers to manage and track advertising in video games and other digital environments.


SUMMARY OF THE INVENTION

The present invention may be configured to provide a system and method for deploying and tracking advertisements across a video game network.


The present invention may be configured to provide a method for determining an impression area in a video game environment relative an advertisement in the environment.


The present invention may also be configured to provide a method for identifying an obstruction in the impression area and redefining the impression area based on the presence of the obstruction.


The present invention may also be configured to provide a method for verifying an advertising impression in a video game environment when a video game character is present in an impression area with an unobstructed view of a related advertisement.


The present invention may also be configured to further provide a method for determining the time a video game character is present in an impression area with an unobstructed view of an advertisement.


The present invention may also be configured to provide a system for determining when an advertising impression has been made in a video game environment.


The present invention may also be configured to provide a system for processing a payment based on the presence of a video game character in an impression area in a video game environment.


The present invention may also be configured to allow advertisers to identify popular games and/or effective advertisements to allow for the establishment of proper pricing models, receive feedback on their products, market to various segments and deploy custom programming relating to advertising campaigns in an adaptable in-game advertising network.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates one specific exemplary in-game advertising system as may be utilized in an embodiment of the present invention;



FIG. 2 illustrates one exemplary flow of communication between members of the in-game advertising system as illustrated in FIG. 1;



FIGS. 3A and 3B illustrate the exemplary tagging of assets in a video game environment;



FIG. 4 illustrates exemplary advertising structure information as may be used in an embodiment of the present invention;



FIG. 5 illustrates an exemplary end-user client device as may be used in an embodiment of the in-game advertising system of FIG. 1;



FIG. 6 illustrates an exemplary method of in-game advertising upon activation of game program comprising tagged assets for advertising;



FIG. 7 illustrates an exemplary advertisement output method when game data is provided to a video game client device;



FIG. 8 illustrates line-of-sight obstacles as may be found in a video game environment, in one embodiment in accordance with the present invention;



FIG. 9 illustrates a positional relationship between a game character and an advertisement, in one embodiment in accordance with the present invention;



FIG. 10 illustrates a second positional relationship between a game character and an advertisement, in one embodiment in accordance with the present invention;



FIGS. 11A and 11B illustrate two positional relationships between a game character and an advertisement relative obstacles in an impression area, in one embodiment in accordance with the present invention;



FIG. 12 illustrates an exemplary line-of-sight determination methodology to be used in determining the occurrence of an advertising impression, in one embodiment in accordance with the present invention;



FIGS. 13A-13B illustrates the correlation between an impression counter and a positional relationship of a game character and an advertisement, in one embodiment in accordance with the present invention; and



FIG. 14 illustrates an alternative correlation between an impression counter and a positional relationship of a game character and an advertisement, in one embodiment in accordance with the present invention.





DETAILED DESCRIPTION OF AN EXEMPLARY EMBODIMENT


FIG. 1 illustrates an exemplary in-game advertising system 100 as may be utilized in an embodiment of the present invention. The in-game advertising system 100 of FIG. 1 comprises a network 110. Network 110 may be, for example, a cable television network, a broadband wireless network or on optical fiber network. The present invention does not impose any limitation with regard to the particular type of communication medium(s), whether the network is homogeneous (e.g., end-to-end wireless) or whether the network is proprietary, open or a combination of the two. Network 110 only need provide the means to communicate amongst the various servers and/or terminals coupled to the network 110 and make up the advertising system 100 of an embodiment of the present invention. Network 110 may be a communications network, a data network or a combination of the two.


As shown in FIG. 1, communicatively coupled to the network 110 is a content server 120 as offered by a content provider 120; an advertising server 130 as offered by an advertising agency, the server comprising or coupled to an advertising database 140, the advertising database 140 comprising advertisement information data. The exemplary system 100 illustrated in FIG. 1 further comprises content authors 1501 . . . 150N; advertisers 1601 . . . 160N; and end-user client devices 1701 . . . 170N. Some embodiments of the present system may further comprise a payment processing center 180 and an advertising content creator 190.


The content server 120 may distribute digital content. Content may be requested from networked devices operating in a gaming network. In one embodiment, the content is requested by end-user client devices 1701 . . . 170N. The content distributed by content server 120 may comprise video game content (e.g., actual video games, or portions thereof, accessed by end-user client devices 1701 . . . 170N) as well as other forms of digital media (e.g., music and video). The content server 120 may further provide for the storage of digital content. The content server 120 may store such content locally (e.g., as part of a storage area network) or at a location physically remote from the content server 120 but otherwise communicatively coupled to the server 120 thereby allowing for retrieval and transmission of the content to end-user client devices 1701 . . . 170N. Content served by the content server 120 may be served as the result of a push or pull operation.


The advertising server 130, as previously noted, may be managed by an advertising agency providing for the distribution of advertising content to larger audiences (e.g., end-users). The advertising server 130 may serve audio, video, audio/video and still image content. Content served by the advertising server 130 may be served as the result of a push or pull transaction. Advertising database is a storage mechanism for advertising content such as the aforementioned video and audio content. While advertising images are the most prevalent type of advertising content, advertising content may further comprise element types such as programs, objects, state data, control data, textures, bitmap images, compressed images, sequencing data, authentication data, public key and private key. Advertising database 140 may be integrated with advertising server 130 or may be physically remote from the advertising server but otherwise providing a communicate coupling allowing for the retrieval of content from the database 140 for subsequent transmission to end-user client devices 1701 . . . 170N Content authors 1501 . . . 150N are those entities that develop content for distribution to end-users, for example, video games. Content authors 1501 . . . 150N may also develop audio, video and/or audio/video content. Content developed by content authors 1501 . . . 150N may be generated in any form of media. For example, content may be developed in an optical disk format or in non-volatile memory such as a flash card. Content may also be provided in a pure data format to be transmitted and hosted by another party. For example, content author 1501 . . . 150N may develop a video game but never commercially distribute the content in a physical form of media. Instead, the content may be FTP′d or otherwise transmitted to content server 120 and stored in an appropriate storage means for subsequent delivery to end-user client devices 1701 . . . 170N.


Advertiser 1601 . . . 160N is any entity seeking to place an advertisement in the digital content created by content author 1501 . . . 150N. Advertiser may be from any field of endeavor and need not necessarily be in the entertainment or video game industry.


End-user client devices 1701 . . . 170N are those devices allowing an end-user to access digital content. For example, in the case of a video game, the appropriate end-user client device 1701 . . . 170N may be a home entertainment video game system such as a PlayStation3 from Sony Computer Entertainment Inc. In the instance of digital content being, for example, an on-demand movie or other video program, the end-user client device 1701 . . . 170N may be a set-top cable box. End-user client device 1701 . . . 170N may, in other instances, be a portable device that may be temporarily coupled to a more permanent device (e.g., a desktop computer) to allow for the transfer or updating of digital content via a USB cable as would be the case in, for example, a portable music device such as an MP3 player.


Optional payment processing center 180 allows for the execution of various payment and/or monetary transfer transactions. These payments may be achieved, for example, through direct deposit, automatic funds or wire transfers as is appropriate and/or available. Payment processing center 180 may, for example, be a bank offering these services. In another example, payment processing center 180 may be an on-line escrow agent communicatively coupled to a variety of banks wherein the escrow agent instructs and/or receives notice of various monetary transactions on behalf of various entities in the exemplary in-game advertising system 100 (e.g., advertisers 1601 . . . 160N and content providers authors 1501 . . . 150N).


Advertising content creator 190 is an entity that authors and/or develops advertisements on behalf of advertisers 1601 . . . 160N for placement into digital content. In some instances, advertising content creator 190 may only digitally author content. For example, certain advertising copy (be it audio, video, print or any combination of the three) may have already been created in a non-digital format. In those instances, advertising content creator 190 would manipulate (e.g., digitize) the advertising copy so that it may be placed into the greater context of digital content that is offered by the content server 120. In other instances, advertising content creator may take a script for an advertisement and create the same (e.g., film video, record audio and then combine the two with various special effects). Advertising content creator 190 may also utilize program objects and program scripts including commands related to special effects, program elements, control signals, messaging and various protocols. In still other instances, advertising content creator may develop advertisement campaigns from scratch (e.g., the advertising concept for a campaign) and subsequently create the ad content to correspond to that campaign.



FIG. 2 illustrates an exemplary flow of communication 200 between members of an exemplary in-game advertising system 100. The content author 1501 . . . 150N provides the advertising server 130 with advertisement structure information in step 205. The content and makeup of advertisement structure information is discussed in the context of FIG. 4. The advertisement structure information is registered in the advertising database 140 of the advertisement server 130. The operator of the advertisement server 130 informs the advertiser 1601 . . . 160N of advertisement information such as title, contents, etc., of a newly-registered title in informing step 210. Informing of the advertiser 1601 . . . 160N may occur by traditional mail, electronic mail, listservs, SMS, instant messenger, chat or any other available communication medium.


The advertiser 1601 . . . 160N can access the advertisement server 130 and can view the advertisement information in viewing step 215 and further apply for an advertisement buy from, for example, a web-browser screen in application step 220. Once the advertisers 1601 . . . 160N have been established, advertiser specified information such as advertiser name, time slot, and time period of an advertisement are provided to an appropriate content author 1501 . . . 150N from the advertisement server 130 in notification step 225. Notification may occur by traditional mail, electronic mail, listservs, SMS, instant messenger, chat or any other available communication medium.


Advertiser specified information and advertisement structure information are also supplied to the advertisement content creator 190 via the advertisement server 130 in ordering step 230. The advertisement content creator 190 creates advertisement content (e.g., the advertisement) based on the advertiser specified information and advertisement structure information. The completed advertisement information such as bitmap data or other graphic, audio and/or video data is delivered by the advertisement content creator 190 to the advertisement server 130 in delivery step 235.


Notification of the receipt of the completed advertisement is communicated by the advertisement server 130 to the advertiser 1601 . . . 160N in completion step 240 by traditional mail, electronic mail, listservs, SMS, instant messenger, chat or any other available communication medium.


The advertiser 1601 . . . 160N can view the completed advertisement information on the advertisement server 130 in viewing/approval step 245. If the advertiser 1601 . . . 160N approves of the completed advertisement content (e.g., by pressing an ‘OK’ button in a web-interface), the advertisement content is confirmed and an itinerary is by the advertisement server 130 to the content author 250 by traditional mail, electronic mail, listservs, SMS, instant messenger, chat or any other available communication medium in delivery detail confirmation step 250. The itinerary delivered in step 250 may comprise information related to the advertiser, time slot, period, advertising fees and so forth.


In registration step 255, the content provider 120 correlates certain advertisement information and advertisement content with digital contents to be delivered. That is, the content provider 120 recognizes that particular advertisements are to be delivered with particular portions of digital content and so forth. This correlation of information may comprise authoring new derivative files reflecting both advertisement information and digital content/advertising programs), the embedding of metadata in the digital contents or the implementation of object oriented programming wherein certain data files (e.g., digital contents/advertising programs) call upon other distinct files (e.g., advertising information). The metadata may also comprise information as it pertains to advertising information such as how long a game character must be present within an impression area defined within the video game. The metadata may further provide information defining the parameters of the impression area and certain quality factors as are discussed herein. Tracking parameters and feedback information and/or instructions may further be imbedded in the metadata of the advertisement. Such information may also be contemporaneously downloaded with the advertising information as a separate file whereby the advertising information calls upon certain information related to impressions, reporting and so forth.


A user accesses and/or requests digital content (e.g., a driving simulation video game) using end-user client devices 1701 . . . 170N in content application step 260. As a result of the application for content, the user may start to download the content in download step 265. Alternatively, if the user already has a particular portion of the digital content, this step may involve presenting the user with an update as to that content. This step may also comprise unlocking digital content that is already in possession of the user. Step 265 may also involve authenticating removable media, accommodating registration with a game network or a game ‘lobby’ or ‘waiting room’ and so forth


In some embodiments of the present invention, the user may access digital content using permanent physical media (e.g., an optical disc). The physical media may have embodied thereon instruction for accessing the present in-game advertising system 100 as it pertains specifically to advertising content versus actual video game data. In additional embodiments of the present invention, the user may access a combination of advertising content and actual game data via the exemplary in-game system 100 (e.g., new advertising content and new game levels published after the initial release of the game on physical media). Such embodiments are discussed in further detail below.


During the download of content and/or advertisements in step 265, the content provider 120 notifies the advertisement server 130 of the download request as it relates to particular advertising material in step 270, such correlations between content and advertising having previously been made in registration step 255. The advertisement server 130 then transmits the necessary advertisement data corresponding to the user download to the content provider 120 in step 275. If necessary, the advertising data provided to content provider 120 can be updated over time relative the content being downloaded (e.g., new ad copy).


After downloading the digital contents (including advertisement information or content in ad information delivery step 275), the user (through end-user client device 170) renders the advertisement information within the game contents, the advertisement information having been provided via advertisement server 130. As noted above, some embodiments of the present invention may access solely advertising information or a combination of new game content and advertising information rather than an entire game.


The state of the advertisement, such as the number of distributions or impressions made, may be provided to the advertisement server 130 and, if necessary or desired, to the advertiser 1601 . . . 160N in advertisement status notification step 280 so that certain determinations made be made, for example, the success of an ad campaign with regard to the number of impressions made.


As a result of the notification in step 270, the advertisement server 130 can track the advertisements that have been or are being downloaded to an end-user client device 170. Utilizing certain ad impression and tracking methodologies as discussed herein, the advertisement server 130 can receive feedback in connection with advertisement impressions. Information concerning impressions or other advertisement feedback may be generated at the end-user client device 170, which has been configured with the necessary software to either directly or indirectly implement impression tracking.


Direct impression tracking may be based on software configured at the end-user client device 170 and that operates in conjunction with a game kernel and is further configured to participate in network communications such that textures and objects or indexes to textures and objects related to an advertising campaign may be received. The tracking software may directly monitor the angle and position of various advertising asserts with respect to changing camera perspectives presented to the user who controls the camera perspective utilizing a game controller. Indirect impression tracking may occur through a server or a session master client in a peer-to-peer network participating, facilitating, arbitrating or interrogating functions associated with the campaign program (e.g., extraction of data necessary to yield the determination of an ad impression). Hence, ad impression determinations may occur at, for example, ad server 130 or advertiser 160 in response to information generated or signals sent from the end-user client device 170.


Ad impression data may be batched or transmitted over the network at periodic intervals. Transmission of impression data may occur in accordance with a schedule or in conjunction with other processes or transmissions used to facilitate game play. Impression data may also be streamed or pulled during an inquiry received over the network. Any network element of the advertising system 100 may facilitate or influence the transmission of impression data.


Advertisement impressions may be calculated in various ways. For example, an advertisement located in a virtual kiosk in a virtual shopping center might be viewed by 1,000 gamers over the course of an afternoon. It could be said that the particular advertisement enjoyed 1,000 impressions as each gamer walked their gaming character past the kiosk and viewed the goods or services advertised therein. Impressions may also be calculated through a time threshold index. For example, an impression may be earned, triggered, counted or computed after a user has been exposed to the advertisement for a particular period of time. In this example, an impression may occur after 30 seconds of exposure by the user to an advertisement. The impression may also be tracked and computed based on one or more user's continuous or distributed exposures to the advertisement on the virtual kiosk or as part of an overall ad campaign.


The advertisement content receiving, impression tracking and impression data feedback transmission systems of the present invention may reside in a single software element or in multiple software elements. Software elements may be distributed in whole, or part, on one or more processors or across a local or wide area network.


Impression tracking software may be provided as a result of downloading a necessary software module during download step 265 or the software having been installed directly on physical media (e.g. an optical disk) read by the end-user client device 170 or, alternatively, installed directly in the end-user client device 170. Tracking software or various components of the software may also be installed in the various other components of the advertising system 100 dependent upon the particular configuration of an embodiment.


Similar or identical advertisement state information may be provided to content author 1501 . . . 150N. This notification is made so that the advertiser 1601 . . . 160N may be properly invoiced by the contents author 1501 . . . 150N in accordance with any number of payment plans as are discussed herein. The advertisement server 130 may further provide this information to payment processing center 180 to allow for automatic billing and payment in step 285. These payments may be achieved, for example, through direct deposit, automatic funds or wire transfers or any other money transfer methodology as is appropriate and/or available.


The advertising system 100 and methodology of the present invention and as described in an exemplary embodiment through FIGS. 1 and 2 may be implemented over various communication and data networks.


It should be noted that in some embodiments of the present invention, certain elements of the in-game advertising system 100 may be combined or removed from the system 100 entirely without compromising the operations of the system 100. For example, an embodiment of the in-game advertising system 100 as described herein may function without the need for a payment processing center 180 as proper remuneration of parties in the system 100 may have been established before hand or subject to analysis of certain information after advertisement delivery. Similarly, the ad server 130 and related database 140 may be operated in conjunction with the advertisement content creator 190 or with content provider 120. Various approval and notification steps may also be omitted in the course of FIG. 2 so long as ad buys remain reserved, as is discussed in the context of tagging in FIG. 3, and the proper advertisements delivered relative those tagged assets. Additionally, exchanges of data or various notifications described above may be of a dual-direction or exchanged between various components of system 100. For example, various notifications may be delivered to the ad server 130 in addition by the ad server 130. In that regard, the present invention is not meant to be limited by the exemplary data flows as depicted in FIG. 2 of the present invention.


While FIGS. 1 and 2 illustrate an in-game advertising system 100 wherein content and advertising were both provided to an end-user client device 170 over network 110, in some embodiments, a user may play game as embodied on a physical media such as an optical disk as may be utilized in, for example, a PlayStation3 gaming console or for use in a desktop computer (e.g., to be inserted into a CD-ROM drive). In these instances, various libraries related to the game (e.g., character generation, user interfaces, recognition of user controls and so forth) may be authored and stamped directly onto the physical media instead of accessed or downloaded via content provider 120.


In such an embodiment, advertising content may be embodied on the physical media as well. As has been previously noted, however, such advertising schemes may be ineffective if the popularity of a game turns out to be overrated (wherein an ad buy was likely overpriced) or underrated (wherein an ad buy was likely under priced). Similarly, the relevance of certain advertisements may expire over the course of time (e.g., an advertised event occurs, the advertiser stops selling the product or goes out of business or the advertisements are mock advertisements pertaining to a fictitious product but remain relevant in the context of the game despite the passage of time).


In such physical media, a software client may be embodied in the physical media, the client comprising the operating routines, resources, instructions and so forth that allow an end-user client device 170 reading the optical media or other physical media to access the in-game advertising system 100 like that described in the present invention. Although the user may not necessarily be receiving video game content (e.g., the user does not download or directly access code and other information related to the actual game), the user may still receive advertising content as the client pertains to the instructions and operations necessary to access in-game advertising system 100 and for advertising content to be provided to the system 100.


Through the provision of such an advertising client on physical media, it becomes possible for a variety of parties that develop games that operate on a particular end-user client device (e.g., the PS3 game console) to interact with the in-game advertising system 100. Access to the advertising client code may be subject to a fee charged by the in-game advertising system 100 operator, the costs of which may be recouped by the third-party game developer who passes those costs onto advertisers 1601 . . . 160N that might wish to place content in a particular video game as the popularity and advertising value of that game is assessed.


In order to enjoy the advertising opportunities offered by such a system like that described in the present invention, certain objects in a video game may be ‘tagged’ as subject to advertising. For example, and as shown in FIG. 3A, the face of a billboard 300 may be ‘tagged’ 310 to indicate that an advertisement may be embedded in that location. In this way, the in-game advertising system 100, via the advertising client embodied in video game software, may identify these tagged assets and provide advertising content that may be imposed through texturing or other graphic rendering techniques that make it appear as if the advertisement was designed specifically for the billboard as shown FIG. 3B (320). Such advertising content may be pushed or pulled to the video game environment at an end-user client device 170 via the advertising server 130.


Alternatively, advertising content may be loaded into a game during development and specific advertising campaigns may be purchased after the release of the game. In such an embodiment, advertisement purchases trigger signaling event that index specific ads embedded in a game and unlock the advertisements. The ads are then associated with one or more tagged advertising assets. Preloaded advertisements may be replaceable in whole or in part over the network by a server or via a peer-to-peer arrangement. The replacement of advertisements may occur based upon a user profile, user interaction with an advertisement or ad campaign, geographic location of the user or control signals, messages or communications in connection with the advertisement.


Tags reflect not only the space where an advertisement may be placed but may also reflect information such as size limitations, coloring and shading requirements, pointers to variables that track state and impression data, functions and programs associated with the advertisement, hyperlinks and mini-games associated with the advertisement, user-profile filters and, in some embodiments, even advertising relevance. It should be noted that the said functions and programs associated with the advertisement may access variables that track state and impression data. Tags may be numbered to reflect individual assets wherein advertisements may be imposed or grouped to reflect that one particular advertiser 1601 . . . 160N should have one or more of their advertisements placed in these tagged groups (e.g., all billboards on a city block). The tagging of assets and rules related to the tagging of assets may be imposed by content author 150. Rules embedded in the tagged assets (e.g., ad size) may be recognized by advertisement server 130 to ensure that the proper advertising content is delivered to these tagged areas when called upon in the in-game advertising network 100.


Impressions of or exposure to advertising asset tags are capable of being tracked independently or as a group. Additionally, impressions of or exposure to advertising asset tags are capable of being aggregated against a particular end-user device, versus a particular game or across a network in general. For example, an advertising server 130 may receive ad impression information relating to impressions of specifically tagged assets (e.g., billboard A, billboard B, billboard C, etc.) or assets as they apply to a particular advertiser (e.g., Coke, Pepsi or Sprite) at an end-user device or a plurality of devices. Impressions and exposure may also be aggregated as they pertain to a particular game whereby the number of impressions generated for a particular advertiser in a particular title are determined regardless of the particular asset on which the advertisements were placed. Similar aggregation of impression data can occur across an entire network regardless of the particular game title whereby the total impressions for a particular advertiser are determined against all end-user client devices communicating with the advertising network/system 100. Other aggregation parameters may be utilized at the particular needs of an advertiser.


Tagging is not limited to ‘print-like’ advertisements (e.g., a billboard). Tagging can also be related to other visual formats such as audio and video. For example, a television in a video game may be tagged as to reflect that the user tuning the television to a particular channel will cause a full motion video advertisement to be streamed. Various other advertisements could be streamed or rendered on additional channels subject to the whim of the game designer and the extent of tagging of assets for advertisement introduction.


Similarly, audio may be tagged for advertising purposes. For example, if a user plays a video game with a radio (e.g., driving an automobile), the game designed can create different channels whereby actual music from actual artists is played interspersed with various advertisement that one might hear over the radio. Similarly, the actual music a user listens to may be Top 40 hits or other popular music rather than a one-time generated, static soundtrack. In that way, the user may play the game today or five years from now and be able to listen to not only relevant advertisement but relevant music that is current and popular the day the user plays the game. Similar attributes may be reserved for providing real-time television programs and the like (e.g., short films, movie previews and so forth).


As a result of tagging and the delivery of relevant advertisements into the user's game environment, tracking of advertisements may take place. That is, through in-game advertising system 100, it may be determined exactly how many times a particular advertisement was introduced to an end-user client device 170 subject to any variety of conditions (e.g., nationwide ad buys over the gaming network or geographic or targeted advertisements). Additionally, and as described in FIGS. 10 and 12, for example, it may be determined how many actual impressions of that advertisement occur.


It should be noted that while most networks and computing devices can provide nearly instant rendering of dynamic advertising information, to identify a particular portion of an environment where such dynamic content may be rendered (e.g., identifying a tag), sending relevant information to the advertising server 130 and retrieving the relevant advertising information may take several seconds. If a user has a slow or congested communications network or a computing device with slower processing power, rendering of that dynamic information may take even longer. If extended delays in rendering information result or, worse, stagnation of game play to wait for the rendering of the dynamic advertising information, user's may lose interest in the video game or seek to deactivate the dynamic advertising aspects of the game.


As such, it is necessary for video games to identify references points in the video game environment (e.g., physical points in the video game or subject to the accomplishment of certain tasks or reaching a certain level) to determine when the in-game advertising system 100 should begin to be accessed to acquire the necessary advertising information. For example, while a user may not have yet reached a billboard tagged to render advertising information, the user may have surpassed a reference point earlier in the game such the content begins to load in the background to provide for instant rendering when the user finally does reach the billboard. An example of such dynamic loading methodology is described in U.S. Pat. No. 6,764,403, which is incorporated herein by reference.



FIG. 4 illustrates exemplary advertisement structure information 400 as may be utilized in an embodiment of the present invention. As illustrated in FIG. 4, advertisement structure information 400 comprises a header 410 identifying the content author 150, movie content data 420, which may be a few frames to the entire production, game program data 430, object data 440 and multiple sets of texture data 450-460.


Certain embodiments of advertisement structure information 400 may be abbreviated, that is, it comprises demonstration or short previews of larger portions of content (namely movie content 420 and game program data 430). In other embodiments, only game program data 430 may be present, for example, for game downloads in the context of the present in-game advertising system 100. In even further embodiments, for example, wherein physical media comprises game program information, only header 410 and certain ad information may be present.


Object data 440 comprises coordinate values of information of objects in a game environment, those objects made up of polygon apex data or the like. Texture data 440-450 comprises pattern data of the object data converted from three-dimensional data to two-dimensional data through various conversion techniques known in the art. For example, object data 440 may related to a race car, texture data 450 and 460 would relate to color patterns and logo advertisements on the race car.


In an embodiment of the present invention utilizing advertisement structure information 400, advertisement information AD1 . . . AD4 is embedded in the structure information 400. AD1 may comprise information such as a code indicating that advertisements may be inserted, the nature of the advertisement to be inserted, or information pertaining to frames, resolution and so forth. This data may be linked to the advertisement server 130 whereby advertisements are inserted into a game environment. Advertisement information AD1 . . . AD4 may also be embedded in physical media should it be necessary to download content as is described in certain embodiments of the present invention.



FIG. 5 is a block diagram illustrating an exemplary end-user client device 500 (1701 . . . 170N). End-user client device 1701 . . . 170N comprises a user terminal 510, monitor 520, main data storage 530, sub-data storage 540, controller 550, and network interface 560 that coupled the device 500 to network 110. In an entertainment system represented by a home computer, the main data storage 530 may comprise a hard disk drive or a high-speed optical disk drive or the like. Sub-data storage 540 in such an embodiment may comprise flash memory. Game data G1 . . . GN comprise advertisement information AD1 . . . AD4 (as described in FIG. 4) and is stored in the main data storage 530.



FIG. 6 illustrates a method 600 of in-game advertising upon activation of game program (e.g., a downloaded program or one embodied in a physical media) comprising tagged assets for advertising.


In optional step 610, a determination may be made whether a game program comprising tagged assets has been activated. Once the digital contents have been downloaded or accessed on physical media via end-user client devices 1701 . . . 170N, an advertisement delivery program on advertisement server 130 may be contacted to request advertising content. If an end-user client device 170 operating a game program with tagged assets has not been activated, the server 130 can await a request or indication delivery of content is or will be necessary. If such a program has been activated, the advertising program may access the advertising server 130 in step 620 and, in step 630, make notification of the title of the digital contents activated on the end-user client device 170 user address (e.g., a network address). Depending on the nature of the advertising content to be delivered (e.g., targeted advertising as discussed herein), the advertisement server 130 reads out the advertisement data in advertisement database 140 and transfers this content to the user's address. In step 640, the end-user client device 170 receives the advertising data under control of the advertising program, records this in the main data storage 530 (FIG. 5) in step 650, and in step 660, generates a table of information indicating addresses in the main data storage for advertisement data, and the position of the advertisement target (i.e., the tagged asset).


Once the game starts or action in the game continues in step 670, a determination is made in step 680 with regard to whether or not the tagged asset has been reached in the game environment; that is, has the user reached the position for inserting advertisement information. If the determination is YES, in step 690 the corresponding advertisement data is positioned at the corresponding position in the memory (i.e., the tagged asset). In some embodiments, it may not be necessary for advertisements to be inserted during game play as tags may be associated with advertising content upon game commencement, upon a level change or in response to a control signal relating to an in-game advertising event.


As has been previously noted, advertising information may be dynamically loaded prior to the content being needed. Larger advertising data—for example, full motion video or audio—may be stored in main memory 530 or a graphics engine buffer (not shown) before action in the game commences. Other embodiments may place object data and/or texture data in the main memory 530 either immediately before the game action starts or before the data is used.


In FIG. 7, an exemplary method 700 for the operation of an advertisement server 130 wherein advertising data is requested by the advertising program is illustrated. In step 710, a determination is made regarding whether or not there is access from the end-user client devices 1701 . . . 170N and in the event this determination yields YES, the title of the digital contents and the user address is received from the terminals end-user client devices 1701 . . . 170N in step 720, the advertisement database 140 is referred to in step 730 wherein advertisement data corresponding to the title of targeted advertising is read out in step 740. The advertisement data is transmitted to the end-user client devices 1701 . . . 170N indicated by the user address received in step 750 and a transmission record for the advertising data is created and/or updated in step 760. This record reflects the number of times an advertisement is pushed to a user and/or, depending on feedback from the game environment, the number of impressions of that advertisement that occur whereby calculations as to payment or future valuation may be made.


The presently described in-game advertising system 100 may also be utilized to provide for the targeting of advertisements. Providing information over a communications network requires proper addressing of that information to an end-user. For example, a network address (e.g., an Internet Protocol address) may be static and assigned to a particular user. Identifying the actual user assigned to this address may be achieved through the network service provider (e.g., an ISP) that is assigned the network address and aware of the address of that user. Alternatively, a user may register with a content provider (e.g., an on-line gaming network), which may require providing specific information (e.g., name, e-mail, billing address and so forth).


In the case of acquiring end-user information from a network provider (e.g., through a commercial information sharing agreement), the acquired information may reflect billing information (i.e., certain geographic information). Similarly, registering with the content provider may reflect certain geographic information of the user (e.g., billing information). As a result of this geographic information, an advertiser may target geographic or region-specific advertisements.


For example, an end-user that resides in Boston may have little interest in receiving information concerning New York Yankee season ticket sales. Similarly, there would likely be little value to advertise a regional product or service such as a restaurant in New York to someone who resides in San Francisco where that product or service is not offered. National advertising campaigns concerning a regional product or service would likely be ineffective relative a return on the advertising investment and may annoy the user receiving those advertisements, because the user may have no interest/access to the product or service being advertised. In contrast, a user in San Francisco might have interest in receiving advertisements related to San Francisco Giants season ticket sales or a concert in the area; that is, geographically relevant advertisements.


By acquiring geographic information of a user (either through direct registration or a service provider), advertisements can be targeted so that the appropriate advertisement is directed to the user. In this way, advertising dollars are ensured a greater return on investment. For example, products localized to Boston are advertised to persons living in the Boston area and products specific to San Francisco are advertised to persons residing in the San Francisco area.


Geographic information may also be inferred from other available information. For example, an IP address may identify a particular region of a country through geo-location. While geo-location via an IP address is not as accurate as explicit registration with a service provider, it provides a greater degree of accuracy than would blind advertisement campaigns. Thus, even dynamic IP addresses that are not consistently associated with any particular user (but instead a service provider who may recycle the address amongst a group of users) may have some advertising value due to geo-location techniques. Geographically-specific advertisements may be provided to the geo-located user, although there remains the possibility that such advertisements may be less accurately targeted than an advertisement with a specific geographic affiliation.


Advertisements may also be more accurately tracked with regard to actual impressions thereby allowing for more accurate determinations of advertising campaign value or proper remuneration to a provider of the advertisement relative those impressions. For example, an advertisement may have 1,000,000 impressions over a 2 week period. While this number may be impressive in a vacuum, when it is learned that 75% of those impressions occurred in a geographic region where the product or service is unavailable, the number of impressions becomes much less valuable. Many of the impressions were wasted on portions of the consuming public that will not or, perhaps even worse, cannot purchase the service or product. Thus, an advertiser can purchase a particular number of impressions with the caveat that those impressions be within a particular geographic region to count against a total overall ad buy.


Direct targeting of users may also take place using variations of the aforementioned identification methodologies. For example, in the registration scenario, a user may provide certain ‘likes’ or ‘dislikes’ in a user profile generated during the registration process. A user may indicate favorite sports teams, favorite hobbies, and the like. As a result of the user profile reflecting that a user is a Boston Red Sox fan, the user may be presented with certain advertisements that relate to World Series Memorabilia from the Red Sox 2004 World Series victory, and not a compilation of the New York Yankees World Series victories. Similarly, a user that identifies an affiliation with the San Francisco 49ers may receive 49er related advertisements instead of advertisements related to the Oakland Raiders. Alternative or more generic profile factors may also be implemented and/or utilized as are available and/or relevant to a particular advertiser.


This type of targeted advertising may be extremely useful when a product or service is available nationwide but has limited popularity or sales in particular regions. For example, a product may be available over the Internet (e.g., through Amazon.com) but also available at a number of brick and mortar stores in one particular region of the country (e.g., the West Coast). A user on the East Coast might purchase these products if he was aware of particular sale opportunities or new product releases. If that user does not live on the West Coast where an advertising campaign is in effect, however, they may never receive advertisements related to that product as advertising dollars have been allocated to a in the locale where brick and mortar stores are located. If the East Coast user indicated an affinity for a particular product in a profile, advertisements can be presented to this user via the in-game advertising system 100 even though the user lives in a region where product sales are otherwise low and advertising (in traditional media forms) is low or entirely non-existent. Through such targeted advertising, not only are impressions generated amongst able buyers, but also amongst willing and highly interested buyers making each impression all the more valuable.


Certain learning intelligence may also be implemented to aid in the direct or geographic targeting advertising process. For example, a game user may participate in an on-line baseball league. Registration for that league may be limited solely to a user name and billing information. If the user resides in Southern California, it would be (as a broad-based assumption) unlikely for this game user to be a fan of the Florida Marlins and (as another assumption) probably a fan of the Los Angeles Dodgers or the Anaheim Angels. Such assumptions may prove to be false.


But if the same user, via the on-line baseball league, continually selects the Florida Marlins as his team of choice, the in-game advertising system 100 may recognize the repetitive behavior (e.g., the selection of a particular team, or a particular character in a game). Based on the repetitive behavior of the user, an assumptive profile of a user may be generated.


Further, if the user plays the networked/on-line baseball league fifteen times and elects to play with the Marlins fourteen of those times, it would be an intelligent assumption that the user is a Marlins fan even though the user lives in Southern California. As a result, certain advertisements in the game environment may be directed toward fan merchandise for the Florida Marlins, instead of for the Dodgers or a random advertisement.


Such targeted advertising is not limited to favorites or affiliations of the user. Direct targeting may also utilize demographics such as gender, age, and the nature of the game itself. Gender may be specifically identified or presumed based on the content of a video game. Age may be based on a specific identification or a presumption related to the maturity of a particular game. The nature of the game itself may indicate demographic information of the user or relevant advertising content. For example, a sports game may generate sports advertisement whereas role playing games may generate advertisement specific to the nature of the game such as combat or fantasy. Various combinations or subsets of targeted advertising may also be utilized (e.g., age and gender relative a particular genre of video game).


These intelligent determinations or analyses based on various demographics may take place at the advertisement server 130 via an appropriate software module providing for such deductive or intelligent determinations.


It should be noted that the present disclosure describes numerous inventive components that may operate individually or with other inventive components outlined herein. One such inventive component—tracking what advertisements the user sees—involves monitoring the view perspective of the user (e.g., the point-of-view of the game character or of the actual user via a game camera) and calculating when the user has experienced an ad impression. One embodiment of this method further allows for object occlusion detection. Such impression information may then be returned to an ad server 130 or other component of the system 100 as described in FIG. 2.



FIG. 8 illustrates line-of-sight obstacles as may be found in a game environment 800. FIG. 8 illustrates a game character 810 (as reflected by the camera icon), an advertisement 820, one or more obstacles 830-880 and line-of-sight 890 relative the game character 810 and its current viewing orientation to the advertisement 820. It should be noted that any references to a game character in the present disclosure is also inclusive of a game camera as to include various points-of-view of the game character of the video game as well as that of the actual user. In one embodiment, the point-of-view is that of the actual user (player) of the video game such that the impression of an advertisement or other information is actually that of the user as would be most like an advertising impression in the real-world. Obstacles 830-880 may be representative of any object in the gaming environment 800. Walls, corners, pillars, objects (e.g., boxes, signs, rows of lockers, doors, etc.), other game characters, opaque glass or even other advertisements (that is, not advertisement 820) may be representative of the obstacles 830-880.


In an ideal advertising environment, game character 810 and advertisement 820 would be separated by an unbroken line-of-sight. That is, obstacles 830-880 would not break line-of-sight 890 whereby game character 810 (and its controlling user through, for example, a first-person game view) would have a full and uninterrupted view of and exposure to advertisement 820. Such an uninterrupted view of and exposure to advertisement 820 is desirous in that it provides for an advertising impression most like that as would be encountered in the real-world (e.g., reading a newspaper advertisement, viewing a billboard or attentively viewing a television commercial). That is, persons in the real-world are generally able to view an advertisement (or at least position themselves) such that other objects in the environment do not obscure a view of that advertisement.


But as is shown in FIG. 8, line-of-sight 890 is interrupted at multiple points by obstacles 830, 840 and 850. Obstacle 830 may be a stack of boxes whereas obstacle 840 may be a soda machine while obstacle 850 may be a corner in a hallway. The exact nature of obstacles 830, 840 and 850 is irrelevant except for the fact that they are interrupting line-of-sight 890 between game character 810 and advertisement 820. Such a scenario as illustrated in FIG. 8 is common in complex game environments such as those offered by role-playing or first-person adventure games, where users navigate through the gaming environment 800 and the objects that exist therein.


The interruption of the line-of-sight 890 as caused by obstacles 830, 840 and 850 may partially (or wholly) prevent the character 810 from viewing the advertisement 820. Depending on the exact angle of obstacles 830, 840 and 850, the character 810 may be able to see certain portions of advertisement 820, but those portions may be minimal compared to the greater portion of the advertisement 820 obscured by obstacles 830, 840 and 850. In some cases, an advertiser may have paid significant sums of money for the placement of advertisement 820 in game environment 800. However, the advertisement 820 may never be viewed as was intended by the advertiser (e.g., a full-frontal observation of the advertisement 820 for a given period of time in order to allow the game user controlling game character 810 to review and comprehend the advertisement 820). The advertiser may, therefore, have expended certain sums of money with absolutely no end benefit as the user of the game (via character 810) did not view the advertisement 820. This lack of an advertisement impression results even though character 810 is actually standing directly in front of advertisement 820 and has their line-of-sight 890 oriented in the same direction.



FIG. 9 illustrates a positional relationship between a game character 920 and an advertisement 910 in a game environment 900, in one embodiment in accordance with the present invention. In FIG. 9, advertisement 910 is located at a predetermined position in the game environment 900. This positioning may be along a wall 950 in the game environment 900. For example, advertisement 910 may be a poster advertising an upcoming concert, advertisement 910 having been pushed into the game environment 900 by various components of the in-game advertising system 100 (FIG. 1) (e.g., via asset tagging, whereby the relevance or ‘freshness’ of the advertisement 910 can be continually updated after the initial release of the game). In the present example, the advertisement 910 (poster) may be positioned on the wall 950 of, for example, a record store in the game environment 900.


The game character 920 may enter this particular portion of the game environment 900 (the record store) through, for example, an entryway 960. As can be seen in FIG. 9, the game character 920, upon initial entry into the game environment 900, may be standing directly adjacent to and oriented toward the advertisement 910. Due to the positioning of the game character 920 relative the advertisement 910, however, the game character may not be able to actually see or be exposed to the advertisement 910 regardless of the orientation of the game character 920 in the general direction of the poster (advertisement 910).


The game character's 920 inability to view the advertisement 910 in FIG. 9 is a result of the game character 920 being outside an impression area 930 defined, in part, by a combination of a first angle θi and a second angle θr relative to a surface vector 940. The first angle θi is the angle measured from a ray 970 to the surface normal 990 (θ) wherein normal incidence is an angle of zero. The second angle θr is the angle measured from a ray 980 to the surface normal 990. The second angle θr, at least with regard to isotropic surface, is identical to the first angle θi (i.e., θir). In FIG. 9, the first angle θi and the second angle θr each have an angle of about 30θ relative the surface normal 990.


The surface vector 940 comprising a unit length (e.g., a distance from the advertisement) further defines the impression area 930 for a predetermined distance from the surface of the advertisement 910. The surface vector 940 relative the advertisement 910 is defined, for example, as being 20 feet. Absent any obstructions in the impression area 930, if the game character 920 is within 20 feet of the advertisement 910 and within the angles defined by first ray 970 and second ray 980 (i.e., +/−30° relative the surface normal 990), then the game character 920 is within the impression area 930.


A user controlling the game character 920 within the impression area and facing the advertisement 910 will be able to view the advertisement 910. That is, an impression will be established for the advertisement 910 as would normally occur in the real world (e.g., while the user is standing in front of a billboard). Alternatively, if the game character 920 is not within the impression area 930 as defined by first ray 970, second ray 980 and surface vector 940, then no impression is generated.



FIG. 10 illustrates a second positional relationship between a game character 1020 and an advertisement 1010 in a game environment 1000, in one embodiment in accordance with the present invention. The game environment 1000 of FIG. 10 is similar to that illustrated in FIG. 9 with the exception of the position of game character 1020. In FIG. 10, game character 1020 is positioned directly in front of and exposed to the advertisement 1010.


An impression area 1030 is defined in a manner similar to that of FIG. 9. That is, a first ray 1070 and a second ray 1080 relative surface normal 1090 in conjunction with surface vector/distance 1040. Because game character 1030 is located within the impression area 1030 of the advertisement 1010, an advertising impression is generated.



FIG. 11A illustrates a positional relationship between a game character 1120 and an advertisement 1130 relative an obstacle 1140 in an impression area 1110 in a game environment 1100, in one embodiment in accordance with the present invention. The impression area 1110 of the game environment 1100 is defined in a manner similar to that of FIGS. 9 and 10 (e.g., a first and second ray relative a surface normal with a predetermined distance relative the advertisement 1130). Game character 1120 is positioned within the impression area 1110 and the orientation of the line-of-sight of the game character 1120 is toward the advertisement 1130. The line-of-sight from the game character 1120 to the advertisement 1130 is, however, at least partially obscured by the obstacle 1140 within an obstructed area 1145.


With the increased graphic complexity of many video games, placement of objects about a game environment increasingly provides a challenge to creating advertising impressions. For example, in FIG. 11A, due to the position of the game character 1120 relative the obstacle 1140, the game character 1120 cannot view the advertisement 1130. Therefore, no advertising impression is made to the user controlling the game character 1120.


Determining whether a game character falls within an obstructed area can be accomplished by using line-of-sight determination. In FIG. 12, obstructed line-of-sight 1220 is shown between game character 1200 and advertisement 1210, in one embodiment in accordance with the present invention. FIG. 12 also shows unobstructed line-of-sight 1250 between game character 1200 at a different position relative advertisement 1210. The determination of obstruction may be established by testing line-of-sight 1220 (or 1250) between game character 1200 (1200′) and advertisement 1210 that passes through a center of obstruction probe 1225.


In some embodiments, obstruction probe 1225 is a spherical object with a predetermined radius r. Obstruction probe 1225 travels along the line of sight 1220 (or 1250) between game character 1200 (1200′) and advertisement 1210. If obstruction probe 1225 does not collide with any obstacles, then the line-of-sight between game character 1200 (1200′) and advertisement 1210 is unobstructed. If game character 1200 is located in an impression area and oriented toward the advertisement 1210, an impression of the advertisement 1210 is generated. Unobstructed line-of-sight 1250 illustrates the absence of object obstruction between game character 1200′ and advertisement 1210, which allows for an advertising impression.


Alternatively, line-of-sight 1220 is obstructed as a result of one or more objects 1230a. . . 1230c, preventing an advertising impression. If the obstruction probe 1225, while traveling along (obstructed) line-of-sight 1220 intersects one or more polygonal sides 1240a . . . 1240c of one or more objects 1230a . . . 1230, where each of one or more objects 1230a . . . 1230c is typically constructed from multiple polygonal sides 1240a . . . 1240c, then an unobstructed view of the advertisement 1210 relative the game character 1200 is not possible and no advertising impression is generated notwithstanding the presence of the game character 1200 in an impression area. Such a scenario—an obstructed line-of-sight and absence of an advertising impression despite being in an impression area—is illustrated in the aforementioned FIG. 11A.


In some embodiments of the present invention, partial viewing of and exposure to an advertisement may be sufficient to establish an advertising impression. For example, certain trademarks or logos have established a certain degree of notoriety within the purchasing public. For these famous or easily recognizable trademarks or logos, viewing even a portion of the trademark or logo may be sufficient to establish an advertising impression. Similar ‘partial viewing impressions’ may be acceptable with regard to slogans, celebrities, famous spokespersons, and so forth. In these instances, even though the obstruction probe 1225 may intersect with an object, if the intersection involves only a small percentage of the probe 1225, then a partial impression may be generated. If the object obscures the advertisement in its entirety—100% of the probe 1225 intersects with the object—then no impression is generated.


The radius r of the obstruction probe 1225 may be reduced whereby a collision with a polygonal sides 1240a . . . 1240c of one or more objects 1230a . . . 1230c may be avoided thus allowing for an unobstructed line-of-sight and, subject to presence in an impression area, establishing an advertising impression. In that regard, the radius r of obstruction probe 1225 may be relative to an advertisement to be viewed. Information relative the setting of radius r may be part of advertising data pushed to a video game environment by the advertising server 130.


In some embodiments of the present invention, especially those involving third-person points-of-view, it may be possible to overcome obstructed lines-of-sight in an effort to create an unobstructed line-of-sight. For example, in an in-game advertising system 100 (FIG. 1) where payment of advertisement space is made in advance, an advertiser may seek to have their advertisement viewed at any cost. In these instances, a camera tracking the game character may be adjusted to provide an unobstructed line-of-sight to allow for viewing of the advertisement and creating an ad impression. Exemplary systems and methods for providing an unobstructed view of an advertisement or other target are disclosed in U.S. patent application Ser. No. 10/268,495, which has previously been incorporated by reference.



FIG. 11B illustrates a positional relationship between a game character 1170 and an advertisement 1180 relative an obstacle 1190 in an impression area 1160 of a game environment 1150. The impression area 1160 of the game environment 1150 and relative the advertisement 1180 is defined in a manner similar to that of FIGS. 9 and 10 (e.g., a first and second angle relative a surface normal combined with a surface vector). Game character 1170 is positioned within the impression area 1160 but is not in an obstructed area 1195 as would be generated by obstacle 1190 like that described in FIG. 11A. A determination of the lack of an obstruction area 1195 or the game character 1170 not being positioned in an obstruction area 1195 is determined in a manner similar to that as described in FIG. 12. In that regard, an obstruction probe traverses a line-of-sight between the game character 1170 and the advertisement 1180 and does not intersect with a polygonal side of an object. The lack of intersection thus indicates the presence of an unobstructed line-of-sight and the establishment of an advertisement impression resulting from game character's 1170 exposure to the advertisement.



FIG. 13A illustrates the correlation between an impression counter 1360 and the position of a game character 1320 relative an advertisement 1310 in a game environment 1300. Impression counter 1360 measures the time period that the game character 1320 is positioned in impression area 1330 with an unobstructed line-of-sight 1350 relative the advertisement 1310. As can be seen in FIG. 13A, game character 1320 has an unobstructed line-of-sight 1350 of and exposure to advertisement 1310 notwithstanding the presence of object 1340 and obstructed area 1345.


It should be noted that impression counter 1360 is not necessarily a stopwatch or other timing device as depicted in FIG. 13A. Impression counter 1360, in an embodiment of the present invention, is any timing mechanism governed by hardware or software relative an end-user client device 170 (e.g., a gaming console) that may measure the time period that the game character 1320 is positioned in the impression area 1330 with an unobstructed line-of-sight 1350 relative the advertisement 1310 and may utilize various units and/or measurement schemes. In one embodiment, a temporal period is used as a unit of measure (e.g., seconds, fractions of seconds and so forth).



FIG. 13B further illustrates the correlation between an impression counter 1360 and the position of a game character 1320 relative an advertisement 1310 in a game environment 1300. In FIG. 13B, however, the game character 1320 has moved behind the object 1340 and into the obstructed area 1345. Notwithstanding the fact that the game character 1320 remains within the impression area 1330, the game character 1320 has an obstructed line-of-sight 1370 relative advertisement 1310. When the line-of-sight of the game character 1320 relative the advertisement 1310 becomes obstructed (as is reflected in FIG. 13B), the impression counter 1360 terminates. That is, an impression of the advertisement 1310 is no longer being generated as the advertisement 1310 has been obstructed from the game character's 1320 view.


By measuring the length of exposure to an advertisement via impression counter 1360, an advertiser can determine the value of an ad impression or whether an impression has actually been made if the existence of an impression is tied to the duration of presence in the ad impression area (e.g., the time of exposure to the advertisement). For example, a pricing model may be established wherein an advertiser is charged based on the duration of the advertisement impression. The duration of the advertisement impression is reflected by the impression counter 1360. In another pricing model, an advertiser may pay a fee for a certain number of advertisement impressions. An ad impression may be defined as unobstructed exposure to an advertisement for a certain period of time. For example, and as evidenced in FIG. 13B, the game character 1320 is in the defined impression area 1330 but is unable to actually view the advertisement 1310. As such, the advertiser should not have the presence of the game character 1320 in the impression area 1330 with no view of the actual advertisement 1310 constitute an advertisement impression.


The various ad impression determinations may be implemented utilizing software downloads as discussed in the context of FIG. 2 or through software installed on physical media (e.g. a software client on an optical disk) or may be pre-installed in a gaming device. Various modules that interact with the ad impression determination software (e.g., receiving ad impression information for reporting to an advertiser) may be further located in various other aspects of system 100 (e.g., at ad server 130).


As noted, various pricing models may be based upon the existence of advertising impressions or the quality thereof. For example, an advertiser may be satisfied knowing that their advertising content has made it into a video game. Another advertiser may be more demanding and require information related to actual impressions. Using the methodology described in FIG. 12 with regard to determining the existence of unobstructed lines-of-sight it may be determined whether the user actually viewed the advertisement.


Even more specific, it may be determined how long the user viewed the advertisement. For example, if a user is merely scanning around the room for an exit or a particular object, their line-of-sight may intersect with the advertisement but the scanning of the room was too quick to allow for any meaningful consideration or understanding of what the advertisement portrayed. In this scenario, a timer may be implemented as was described in FIGS. 13A and 13B.


On an even more detailed level, it may be possible to determine the quality of the impression. For example, a user may view an advertisement as a result of being in an impression area. That user may, however, be on the very far edge of the impression area and have slight difficulty viewing the advertisement. This might be the case if a user is utilizing a later model television or computer monitor or is utilizing a computing device that has lower graphics processing power. Notwithstanding graphics output considerations, it is possible to further delineate the impression area into quality impression areas whereby the advertisement is viewed in every instance but better or worse depending on the exact placement of the game character when viewing the advertisement.


A game character may be face-to-face with an advertisement. The character, while clearly within the impression area, may be so close to the advertisement that he cannot fully view the advertisement or the copy that he can view is blurred because of the close proximity of the character relative the advertisement in the gaming environment. Similarly, a user may be too far away to fully appreciate the advertisement. Through delineating quality impressions, advertisers can appreciate a minimal impression (e.g., up close or almost too far away) but also have certain assurances with regard to quality impressions as may be subject to the particular whims of the advertiser.


Various pricing schemes may be based upon these various levels or quality of impression whereby a general impression is charged at one rate while a higher quality impression is charged at a different rate. Similarly, the length of time a user is in an impression area can be correlated to a pricing model. For example, if a user is in an impression area for 2-seconds, an impression may have been made but possibly a minimal one due to the complexity of the advertisement. If the user is in the impression area for 10-seconds, a greater impression has been made and has greater value to the advertiser. Limits may be imposed on such an impression counter such that an advertiser is not charged for a 30-minute impression when a user happens to position his game character in front of an advertisement and then leave to attend to another task for half-an-hour. Notwithstanding the presence in the impression area for that period of time, a thirty-minute impression has not truly been made as the user of the game (the controller of the character) has not been subjected to that advertising copy.


Traditional economic aspects of supply and demand may also come into play with various pricing models. For example, if a game is released with great fanfare and is a ‘must buy for the holiday season,’ ad buys in the game may be more expensive. If the game layer proves to be unpopular for a variety of reasons, the pricing levels may decrease to reflect the demand of the game. These determinations as to supply and demand may be made, in part, based on the location and intrinsic value definition of specific tags, the demand for a tag as driven by the number of times tagged assets in a video game are identified during average, peak and off-peak game play thereby resulting in various requests to the advertising server 130, which may be counted as have been previously described in the context of FIG. 2.


Similarly, certain video games may have indicia identifying a distinct owner as a result of a user profile or information embedded on the game or subject to, for example, a network address. An advertiser may also determine that while 1,000 impressions may have occurred for their advertisement on a particular day that almost half of those were related to a small group of users who continually entered a gaming environment where the advertisement was rendered time-and-again versus 1,000 impressions distributed more equally amongst 800 different, unique game players. The impressions in the latter example are more valuable then the repeated impressions amongst a small group of users in the first example.


Certain embodiments of the exemplary in-game advertising system 100 described in the present invention may also allow for certain interactions with the products advertised in a video game. For example, a game player might approach a vending machine whereby a variety of beverages are available and effectively advertised via their labeling, trademarks or other visual indicia. A user might select a particular beverage for his character to enjoy in the course of the video game by pressing a button on his keypad that corresponds to purchasing a particular beverage from the vending machine. These ‘virtual purchases’ may be conveyed to advertisers in that the advertisement has not only made an advertising impression in that a user has seen the product or related advertising but taken some sort of positive interaction with the product (e.g., purchasing the product in the video game).


Similar game metrics may be implemented with regard to negative connotations. For example, if four beverages are available, the user's selection of one beverage may reflect negatively as to the other three. By further example, a user may be presented with a vending machine for two competing beverages; if the user takes some action relative one of the vending machines (e.g., destroying it with a weapon), that act too may reflect negatively relative advertising metric information.


Through tracking user interactions with advertisements in a video game environment, the video game effectively becomes a user feedback service similar to an advertising focus group. Feedback may also be registered through explicit interactions wherein a user may expressly provide their opinion of a product or service relative the game environment. For example, the user may be prompted as to their opinion of a particular product; the user may then press ‘up’ for a positive reflection or ‘down’ for a negative reflection.


Interactions via, for example, a microphone input are also possible wherein the user provides feedback in connection with the advertisement. Such feedback may be transmitted to an agent on the other end of the communication channel or passed through speech recognition software wherein certain keywords as they relate to a product are recognized and categorized.


To address the feedback features, the impression tracking system may include functions or may interact with functions capable of soliciting or recording user reaction to an advertising campaign. For example, an advertiser may deploy an advertising campaign defined by advertising content that is loaded into a tag with program or pointer to program(s). Such programs may signal the user to perform actions. Other programs may monitor user reaction in and about the advertisement or in response thereto. For example, one such program that may be invoked when accessing a loaded advertisement tag includes a reference to a speech input requirements and definitions.


Advertising beyond the traditional flat, print advertisement may also be implemented utilizing the presently described in-game advertising system 100. For example, in addition to billboards or single page ad copy, rotating billboards may be utilized wherein triangular panels in the billboard rotate relative to one another thereby effectively providing three-billboards-in-one. The billboard panels then rotate every few seconds to reflect a new advertisement on each panel as occurs in the real-world. In this way, a single game asset can be tagged for multiple advertisement introductions.


Other ads in a game environment may be movable. For example, advertisements may be located on the sides of buses that traverse city streets or a series of flyers that might blow down an alley. Television ads with full motion video and audio ads as might be emitted over a radio or a telephone in a gaming environment are also envisioned as being implemented in the present invention.


Just as certain advertisements have higher demand in the real-world (e.g., high traffic areas), certain advertisements in a video game environment may enjoy higher pricing as a result of high traffic areas. For example, advertisements that appear in the beginning of a video game or a level wherein every user will view the advertisement inherently have more value than an advertisement located in a ‘secret Easter Egg’ level or extremely difficult level that many users may never reach.


Other embodiments of the present invention may include rewards based on user interaction with particular advertisements. For example, if a user provides actual feedback in a video game environment, the user may receive merchandise, points or coupon rewards from the producer of the product as an appreciation for their opinion. Such information concerning where to send a reward may be expressly provided during a feedback session or as a result of an association with a user profile. This latter case would be valuable wherein points or rewards are offered for less explicit interactions (e.g., not in response to an advertiser/feedback query) such as casual interactions or favorable behavior relative an advertised product and points or rewards accumulate over time.


Advertisements, especially those ads that are audible in nature or are full motion video, may be subject to real-time limitations. For example, a user in a video game may be changing the channels of a television in the video game environment. If the user only watches two seconds of the advertisement, an impression may or may not be generated. Such limitations in the case of real-time advertising may be subject not only to an impression area but also an impression time and even an impression time relative particular portions of the advertisement.


For example, an advertisement may be thirty-seconds in length but the first five-seconds do not indicate the nature of the product and the last five-seconds concern legal boilerplate required by the particular advertisement. If an impression time is identified as five-seconds, watching the first or last five-seconds of this particular advertisement would technically constitute an impression notwithstanding the fact that the user knows nothing more about the product after those five-seconds than he did prior. In these cases, limitations as to impressions of particular portions of an advertisement may be implemented. For example, for an advertiser to consider there to have been an impression, the user must not only view five-seconds of the advertisement but those five-seconds must be within the middle 20-seconds of the 30-second advertisement.


Video or audio ads may also be subject to start-stop loops. That is, the advertisement starts when the user accesses the advertisement (e.g., tunes to a radio station playing the advertisement) and then stops when the user leaves the advertisement (e.g., changes the radio to another station). If the user then changes back to the original station with the advertisement, the ad may commence where it left off as if no time has passed. Such a methodology better ensures an impression but does so at the risk compromising reality (i.e., real-time passage of time is not in effect). The tag object may track the state of the advertisement impression, such as the index into the location in a video file to start the next sequence for the one or more tags associated with the advertising video loop.


Video games, radios and televisions that offer the user the ability to change channels may be associated with features to track multiple advertisement impressions and campaigns. When a user changes a channel or directs a virtual character in the game environment to change a channel, new advertisements may be provided. Such advertisement changes may be transitioned with white noise or a familiar blur associated with changing a channel according to the nature of the device. Radio or television devices may be configured with channels that access both traditional programming, advertisement content or other content. Other content may include chat wherein the device facilitates communications. Other content may also include other information in connection with the game. Generally, the mixing of advertisement and other content in such devices may have the benefit of catalyzing user exposure to advertisements since the use of the device and changing of the channels may be necessary to facilitate game play.


Other advertisements may be rendered or emitted in true real-time. For example, if a television advertisement in a video game is two minutes in length and the user changes the channel in the video game after thirty-seconds of viewing the advertisement but comes back to the same channel thirty-seconds later, the advertisement will now be at the 60 second point and not the 30 second point as in a start-stop embodiment.


While real-time advertisements may be more realistic, ensuring an impression becomes more difficult relative the portion of the advertisement the user viewed as has been previously noted. Certain impression, especially in the real-time video and audio sense, may be subject to ongoing impression limitations. For example, an impression may constitute viewing 30 seconds of a one minute advertisement. The user may, at one point in the game, view a first 10-seconds of the advertisement, view a second 10-seconds at a different point in the game and view yet another 10-seconds at another point in the game. In this instance, the user—albeit piecemeal—may have viewed enough of the ad over the course of time to constitute an impression.


Other advertisements may limit an impression opportunity to consecutive time or such piecemeal viewing/listening but within an overall time frame. For example, viewing the advertisement in 10-second snippets may suffice as an advertisement but they must occur within 15 minutes of one another. Other advertisements may require the thirty-seconds to occur consecutively or an impression has not been established.


Some of these real-time/consecutive impression implications addressed above are reflected in FIG. 14 of the present application. FIG. 14 illustrates the positional relationship of an in-motion game character 1420 relative an advertisement 1410 in an impression area 1430 in a game environment 1400. In FIG. 14, the game character 1420 is positioned in the impression area 1430 of advertisement 1410. The impression area 1430 is also populated with obstacles 1440A . . . 1440D. As the game character 1420 traverses the game environment 1400, the line-of-sight of the game character changes from an obstructed line-of-sight (as would occur behind obstacles 1440A . . . 1440D) and an unobstructed line-of-sight 1450A . . . 1450D. An impression counter (not shown) would move between an on-and-off state as the line-of-sight alternates between obstructed and unobstructed (1450A . . . 1450D) lines-of sight.


For example, as the game character 1420 moves past object 1440A, the impression counter would begin to measure the existence of an ad impression as provided by unobstructed line-of-sight 1450A. As the game character passes behind object 1440a, the impression counter would stop measuring the existence of an advertisement impression as a result of now obstructed line-of-sight. Once the game character 1420 emerges from behind object 1440a, an unobstructed line-of-sight (1450B) once again exists and the impression counter again would begin to measure the existence of an advertisement impression from the stop point of the previous impression. The measurement of an advertisement impression would continue in a similar fashion as the game character 1420 passes in between remaining objects 1440B. . . 1440D.


In the present embodiment, as the impression counter starts-and-stops, any one segment of time correlating to an advertisement impression may not constitute a single advertisement impression. The ongoing exposure to the advertisement 1410, albeit in an interrupted fashion, may over the course of time constitute an ad impression. For example, by the impression counter reaching a certain time period (e.g., from start point to a point three seconds in time later), this time period may (as a whole) constitute an ad impression. Such a measurement methodology would be desirous in instances where a game character passes by, for example, a number of pillars; a rod iron fence, a series of windows, or a crowded room.


The above description is illustrative and not restrictive. Many variations of the invention will become apparent to those of skill in the art upon review of this disclosure. The scope of the invention should, therefore, be determined not with reference to the above description, but instead should be determined with reference to the appended claims along with their full scope of equivalents.


For example, the embedded advertising ‘tags’ as described in the present invention may be further applied to digital video and audio signals—television and audio broadcasts, for example—as well as movies filmed in a digital format whereby advertisements or other content may be inserted into previously generated audio and/or video content. On-line media such as on-line magazines, newspapers and blogs may also benefit from the implementation of tagging advertising assets (e.g., particular column inches or steaming news broadcasts) as present-day advertising methodologies such as pop-ads become less effective and/or less popular. Advertising content may be offered by network and/or content providers (e.g., cable providers) whereby advertising content is offered on-demand.


Additionally, the various impression area and occlusion concepts disclosed herein may be applied to audio advertisements or other audible emissions. For example, a radio or other audio emitting object may be defined, in part, by an impression area. Such an impression area would be determined in a manner similar to an impression area as it concerns a visual advertisement. An impression area in the context of audio would be representative of where an audio advertisement or other audio emission may be heard by the character in a game environment as the volume of the audio emission decreases as the character moves further away from the advertisement in three-dimensional space or if the character is located behind an object in which case the occlusion determination concepts become applicable (e.g., does a wall separate the character and the audio signal). The quality of audio impressions may also be determined in a manner similar to quality determinations with visual advertisements with regard to not only distance but the extent to which an intermediate object might absorb the sound, for example, a pane of soundproof glass versus a thinly constructed wall.


The asset tagged to receive an advertisement may be movable and rotatable and may be programmed to dynamically orientate towards the user camera as the user manipulates around the game environment. Ad campaigns may be interleaved with special programming. Special programming may influence ad campaigns, variables in the tags relating to the ad campaigns, or may relate to the game environment. Special programming may influence (e.g., terminate or replace) an ad campaign or modify variables or functions contained in an ad campaign or tag. Special programming may accommodate for dynamic reconfiguration and reuse of an advertising asset. For example, special programming ay be used to communicate special messages, game messages, forum messages, facilitate chat and so forth. Special programming may also be used to transfer control of the advertising asset to the game environment so that the advertising real estate can be used to convey game information and other information.


Notwithstanding the providing of detailed descriptions of exemplary embodiments, it is to be understood that the present invention may be embodied in various forms. Therefore, specific details disclosed herein are not to be interpreted as limiting, but rather as a basis for the claims and as a representative basis for teaching one skilled in the art to employ the present invention in virtually any appropriately detailed system, structure, method, process, or manner.

Claims
  • 1. A method for determining a line-of-sight path to media content, the method comprising: storing information regarding a plurality of digital content files in memory, each digital content file associated with a specified level of partial obstruction;associating at least one of the digital content files with an obstruction probe, wherein a size of the obstruction probe corresponds to the specified level of partial obstruction;defining a portion of a digital environment as a media content area, wherein the digital environment includes one or more objects;evaluating a line-of-sight associated with the at least one digital content file at the defined media content area by identifying whether the obstruction probe traveling along the line-of-sight collides with one or more of the objects; andprocessing interaction with the at least one digital content file at the defined media content area based on the evaluated line-of-sight.
  • 2. The method of claim 1, wherein the obstruction probe is a spherical object, and wherein the size corresponds to a radius of the spherical object.
  • 3. The method of claim 1, wherein the line-of-sight corresponds to a path between a surface of the defined media content area and a viewpoint in the digital environment.
  • 4. The method of claim 3, wherein the viewpoint is associated with at least one of an avatar of a user and a camera.
  • 5. The method of claim 4, further comprising adjusting the camera based on the evaluated line-of-sight.
  • 6. The method of claim 1, further comprising adjusting the size of the obstruction probe when a different digital content file is placed in the media content area, the adjusted size corresponding to the specified level of obstruction associated with the different digital content file.
  • 7. The method of claim 1, wherein evaluating the line-of-sight is based on an amount of the obstruction probe that collides with the one or more objects.
  • 8. The method of claim 1, further comprising receiving a setting sent over a communication network from a server, wherein the setting specifies the level of partial obstruction associated with the at least one digital content file.
  • 9. The method of claim 1, wherein processing the interaction with the at least one digital content file at the defined media content area includes establishing an impression based on the evaluated line-of-sight.
  • 10. The method of claim 1, further comprising updating the evaluation of the line-of-sight based on movement by at least one of the media content area or one of the objects.
  • 11. A system for determining a line-of-sight path to media content, the system comprising: memory that stores information regarding a plurality of digital content files, each file associated with a specified level of partial obstruction; anda processor that executes instructions stored in the memory, wherein the processor executes the instructions to: associate at least one of the digital content files with an obstruction probe, wherein a size of the obstruction probe corresponds to the specified level of partial obstruction;define a portion of a digital environment as a media content area, wherein the digital environment includes one or more objects;evaluate a line-of-sight associated with the at least one digital content file at the defined media content area by identifying whether the obstruction probe traveling along the line-of-sight collides with one or more of the objects; andprocess interaction with the at least one digital content file at the defined media content area based on the evaluated line-of-sight.
  • 12. The system of claim 11, wherein the obstruction probe is a spherical object, and wherein the size corresponds to a radius of the spherical object.
  • 13. The system of claim 11, wherein the line-of-sight corresponds to a path between a surface of the defined media content area and a viewpoint in the digital environment.
  • 14. The system of claim 13, wherein the viewpoint is associated with at least one of an avatar of a user and a camera.
  • 15. The system of claim 14, wherein the processor executes further instructions to adjust the camera based on the evaluated line-of-sight.
  • 16. The system of claim 11, wherein the processor executes further instructions to adjust the size of the obstruction probe when a different digital content file is placed in the media content area, the adjusted size corresponding to the specified level of obstruction associated with the different digital content file.
  • 17. The system of claim 11, wherein the processor evaluates the line-of-sight based on an amount of the obstruction probe that collides with the one or more objects.
  • 18. The system of claim 11, further comprising a communication interface that communicates over a communication network to receive a setting sent from a server, wherein the setting specifies the level of partial obstruction associated with the at least one digital content file.
  • 19. The system of claim 11, wherein the processor processes the interaction with the at least one digital content file at the defined media content area by establishing an impression based on the evaluated line-of-sight.
  • 20. The system of claim 11, wherein the processor executes further instructions to update the evaluation of the line-of-sight based on movement by at least one of the media content area or one of the objects.
  • 21. A non-transitory computer-readable storage medium, having embodied thereon a program executable by a processor to perform a method for determining a line-of-sight path to media content, the method comprising: storing information regarding a plurality of digital content files in memory, each digital content file associated with a specified level of partial obstruction;associating at least one of the digital content files with an obstruction probe, wherein a size of the obstruction probe corresponds to the specified level of partial obstruction;defining a portion of a digital environment as a media content area, wherein the digital environment includes one or more objects;evaluating a line-of-sight associated with the at least one digital content file at the defined media content area by identifying whether the obstruction probe traveling along the line-of-sight collides with one or more of the objects; andprocessing interaction with the at least one digital content file at the defined media content area based on the evaluated line-of-sight.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application is a continuation and claims the priority benefit of U.S. patent application Ser. No. 16/594,354 filed Oct. 7, 2019, now U.S. Pat. No. 10,789,611, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 15/992,014 filed May 29, 2018, now U.S. Pat. No. 10,467,651, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 15/285,928 filed Oct. 5, 2016, now U.S. Pat. No. 9,984,388, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 14/336,452, filed Jul. 21, 2014, now U.S. Pat. No. 9,466,074, which is a continuation of U.S. patent application Ser. No. 13/939,178 filed Jul. 10, 2013, now U.S. Pat. No. 8,795,076, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 11/241,229, filed Sep. 30, 2005, now U.S. Pat. No. 8,574,074 the disclosures of which are incorporated herein by reference. The present application is related to U.S. patent application Ser. No. 09/780,995 filed Feb. 9, 2001 and entitled “In-Contents Advertising Method, In-Content Advertising Server, and Program-Transferring Medium for Realizing In-Contents Advertising,” which claims the priority benefit of Japanese patent application number 2000-241861 filed Jul. 4, 2000 and Japanese patent application number 2000-375096 filed Dec. 8, 2000. The present application is also related to U.S. patent application Ser. No. 10/268,495 filed Oct. 9, 2002 and entitled “System and Method for Camera Navigation,” which claims the priority benefit of U.S. provisional patent application No. 60/328,488 filed Oct. 10, 2001. This application is further related to U.S. patent application Ser. No. 11/240,655 filed Sep. 30, 2005 and entitled “Targeted Advertising.” The disclosure of all the aforementioned applications is incorporated by reference.

US Referenced Citations (895)
Number Name Date Kind
4521021 Dixon Jun 1985 A
4542897 Melton et al. Sep 1985 A
4734690 Waller Mar 1988 A
4807158 Blanton et al. Feb 1989 A
4905168 McCarthy et al. Feb 1990 A
4926255 Von Kohorn May 1990 A
4969036 Bhanu et al. Nov 1990 A
5014234 Edwards, Jr. May 1991 A
5083271 Thacher et al. Jan 1992 A
5105184 Pirani et al. Apr 1992 A
5227874 Von Kohorn Jul 1993 A
5231568 Cohen et al. Jul 1993 A
5283731 Lalonde et al. Feb 1994 A
5305195 Murphy Apr 1994 A
5305389 Palmer Apr 1994 A
5319454 Schutte Jun 1994 A
5347632 Filepp et al. Sep 1994 A
5373440 Cohen et al. Dec 1994 A
5377997 Wilden et al. Jan 1995 A
5446919 Wilkins Aug 1995 A
5462275 Lowe et al. Oct 1995 A
5497479 Hornbuckle Mar 1996 A
5512935 Majeti et al. Apr 1996 A
5526041 Glatt Jun 1996 A
5539450 Handelman Jul 1996 A
5548645 Ananda Aug 1996 A
5564038 Grantz et al. Oct 1996 A
5565909 Thibadeau et al. Oct 1996 A
5572643 Judson Nov 1996 A
5574447 Roylance Nov 1996 A
5592212 Handelman Jan 1997 A
5630757 Gagin et al. May 1997 A
5636346 Saxe Jun 1997 A
5646992 Subler et al. Jul 1997 A
5663757 Morales Sep 1997 A
5684526 Yoshinobu Nov 1997 A
5697844 Von Kohorn Dec 1997 A
5699497 Erdahl et al. Dec 1997 A
5707289 Watanabe et al. Jan 1998 A
5712979 Graber et al. Jan 1998 A
5721827 Logan et al. Feb 1998 A
5724521 Dedrick Mar 1998 A
5737619 Judson Apr 1998 A
5748867 Cosman et al. May 1998 A
5751956 Kirsch May 1998 A
5758068 Brandt et al. May 1998 A
5758257 Herz et al. May 1998 A
5762553 Takasugi et al. Jun 1998 A
5771347 Grantz Jun 1998 A
5774170 Hite et al. Jun 1998 A
5794210 Goldhaber et al. Aug 1998 A
5798519 Vock et al. Aug 1998 A
5805815 Hill Sep 1998 A
5822523 Rothschild et al. Oct 1998 A
5848396 Gerace Dec 1998 A
5855008 Goldhaber et al. Dec 1998 A
5857149 Suzuki Jan 1999 A
5860073 Ferrel et al. Jan 1999 A
5867208 McLaren Feb 1999 A
5876286 Lee Mar 1999 A
5878135 Blatter et al. Mar 1999 A
5879235 Kaneko et al. Mar 1999 A
5883958 Ishiguro et al. Mar 1999 A
5903816 Broadwin et al. May 1999 A
5910987 Ginter et al. Jun 1999 A
5911582 Redford Jun 1999 A
5916024 Von Kohorn Jun 1999 A
5917725 Thacher et al. Jun 1999 A
5929850 Broadwin et al. Jul 1999 A
5937037 Kamel et al. Aug 1999 A
5946646 Schena et al. Aug 1999 A
5946664 Ebisawa Aug 1999 A
5948061 Merriman et al. Sep 1999 A
5964660 James et al. Oct 1999 A
5970143 Schneier et al. Oct 1999 A
5974398 Hanson et al. Oct 1999 A
5987511 Elixmann et al. Nov 1999 A
5991735 Gerace Nov 1999 A
6005602 Matthews, III Dec 1999 A
6012984 Roseman Jan 2000 A
6015348 Lambright et al. Jan 2000 A
6016348 Blatter et al. Jan 2000 A
6020883 Herz et al. Feb 2000 A
6024643 Begis Feb 2000 A
6026368 Brown et al. Feb 2000 A
6029046 Khan et al. Feb 2000 A
6036601 Heckel Mar 2000 A
6047289 Thorne et al. Apr 2000 A
6088722 Herz et al. Jul 2000 A
6089975 Dunn Jul 2000 A
6108637 Blumenau Aug 2000 A
6113494 Lennert Sep 2000 A
6119098 Guyot et al. Sep 2000 A
6149519 Osaki et al. Nov 2000 A
6151631 Ansell et al. Nov 2000 A
6165070 Nolte et al. Dec 2000 A
6179713 James et al. Jan 2001 B1
6181988 Schneider et al. Jan 2001 B1
6188398 Collins-Rector et al. Feb 2001 B1
6196920 Spaur et al. Mar 2001 B1
6199082 Ferrel et al. Mar 2001 B1
6216129 Eldering Apr 2001 B1
6236975 Boe et al. May 2001 B1
6238290 Tarr et al. May 2001 B1
6251017 Leason et al. Jun 2001 B1
6263360 Arnold et al. Jul 2001 B1
6264555 Glazman et al. Jul 2001 B1
6264560 Goldberg et al. Jul 2001 B1
6267672 Vance Jul 2001 B1
6267675 Lee Jul 2001 B1
6275854 Himmel et al. Aug 2001 B1
6275989 Broadwin et al. Aug 2001 B1
6298348 Eldering Oct 2001 B1
6308328 Bowcutt et al. Oct 2001 B1
6312337 Edwards et al. Nov 2001 B1
6320495 Sporgis Nov 2001 B1
6324519 Eldering Nov 2001 B1
6343990 Rasmussen et al. Feb 2002 B1
6346045 Rider et al. Feb 2002 B2
6366701 Chalom et al. Apr 2002 B1
6371850 Sonoda Apr 2002 B1
6379251 Auxier et al. Apr 2002 B1
6381362 Deshpande et al. Apr 2002 B1
6385592 Angles et al. May 2002 B1
6385596 Wiser et al. May 2002 B1
6390922 Vange et al. May 2002 B1
6393574 Kashiwagi et al. May 2002 B1
6394899 Walker May 2002 B1
6400996 Berg et al. Jun 2002 B1
6411936 Sanders Jun 2002 B1
6434614 Blumenau Aug 2002 B1
6443843 Walker et al. Sep 2002 B1
6446130 Grapes Sep 2002 B1
6446261 Rosser Sep 2002 B1
6456234 Johnson Sep 2002 B1
6457010 Eldering et al. Sep 2002 B1
6460036 Herz Oct 2002 B1
6468155 Zucker et al. Oct 2002 B1
6470138 Um et al. Oct 2002 B1
6484148 Boyd Nov 2002 B1
6484149 Jammes et al. Nov 2002 B1
6489955 Newhall, Jr. Dec 2002 B1
6496826 Chowdhury et al. Dec 2002 B1
6496857 Dustin et al. Dec 2002 B1
6513160 Dureau Jan 2003 B2
6516338 Landsman et al. Feb 2003 B1
6529940 Humble Mar 2003 B1
6530840 Cuomo et al. Mar 2003 B1
6532448 Higginson et al. Mar 2003 B1
6539375 Kawasaki Mar 2003 B2
6539544 Ebisawa Mar 2003 B2
6553178 Abecassis Apr 2003 B2
6560578 Eldering May 2003 B2
6563523 Suchocki et al. May 2003 B1
6564217 Bunney et al. May 2003 B2
6590578 Nishiumi et al. Jul 2003 B2
6595859 Lynn Jul 2003 B2
6606746 Zdepski et al. Aug 2003 B1
6611812 Hurtado et al. Aug 2003 B2
6611813 Bratton Aug 2003 B1
6611957 Ebisawa Aug 2003 B2
6612932 Stern Sep 2003 B2
6615039 Eldering Sep 2003 B1
6616533 Rashkovskiy Sep 2003 B1
6625578 Spaur et al. Sep 2003 B2
6632138 Serizawa et al. Oct 2003 B1
6640097 Corrigan et al. Oct 2003 B2
6640335 Ebisawa Oct 2003 B2
6640336 Ebisawa Oct 2003 B1
6645068 Kelly et al. Nov 2003 B1
6654725 Langheinrich et al. Nov 2003 B1
6656050 Busch et al. Dec 2003 B2
6659861 Faris et al. Dec 2003 B1
6663105 Sullivan et al. Dec 2003 B1
6669562 Shiino Dec 2003 B1
6669564 Young et al. Dec 2003 B1
6680746 Kawai et al. Jan 2004 B2
6683941 Brown et al. Jan 2004 B2
6684194 Eldering et al. Jan 2004 B1
6687608 Sugimoto et al. Feb 2004 B2
6697792 Bunney et al. Feb 2004 B2
6698020 Zigmond et al. Feb 2004 B1
6699127 Lobb et al. Mar 2004 B1
6701363 Chiu et al. Mar 2004 B1
6704930 Eldering et al. Mar 2004 B1
6709335 Bates et al. Mar 2004 B2
6712702 Goldberg et al. Mar 2004 B2
6714236 Wada et al. Mar 2004 B1
6714723 Abecassis Mar 2004 B2
6714917 Eldering et al. Mar 2004 B1
6716103 Eck et al. Apr 2004 B1
6718551 Swix et al. Apr 2004 B1
6721748 Knight et al. Apr 2004 B1
6731238 Johnson May 2004 B2
6738078 Duncombe May 2004 B1
6745011 Hendrickson et al. Jun 2004 B1
6757740 Parekh et al. Jun 2004 B1
6758746 Hunter et al. Jul 2004 B1
6758754 Lavanchy et al. Jul 2004 B1
6758755 Kelly et al. Jul 2004 B2
6764395 Guyett Jul 2004 B1
6764403 Gavin Jul 2004 B2
6771290 Hoyle Aug 2004 B1
6783460 Galyean, III et al. Aug 2004 B2
6785902 Zigmond et al. Aug 2004 B1
6799327 Reynolds et al. Sep 2004 B1
6814663 Edwards et al. Nov 2004 B2
6820277 Eldering et al. Nov 2004 B1
6826614 Hanmann et al. Nov 2004 B1
6827645 Morita et al. Dec 2004 B2
6840861 Jordan et al. Jan 2005 B2
6863612 Willis Mar 2005 B2
6874683 Keronen et al. Apr 2005 B2
6882978 Ebisawa Apr 2005 B2
6890256 Walker et al. May 2005 B2
6895170 Lambert et al. May 2005 B1
6912398 Domnitz Jun 2005 B1
6928414 Kim Aug 2005 B1
6941574 Broadwin et al. Sep 2005 B1
6942575 Mergler Sep 2005 B2
6948062 Clapper Sep 2005 B1
6954728 Kusumoto Oct 2005 B1
6955605 Young et al. Oct 2005 B2
6964608 Koza Nov 2005 B1
6967566 Weston et al. Nov 2005 B2
6968567 Gordon et al. Nov 2005 B1
6970834 Martin et al. Nov 2005 B2
6970915 Partovi et al. Nov 2005 B1
6973664 Fries Dec 2005 B2
6987221 Platt Jan 2006 B2
6995788 James Feb 2006 B2
7028082 Rosenberg et al. Apr 2006 B1
7072849 Filepp et al. Jul 2006 B1
7076445 Cartwright Jul 2006 B1
7086187 Bandak Aug 2006 B2
7136871 Ozer et al. Nov 2006 B2
7305442 Lundy Dec 2007 B1
7362999 Petschke et al. Apr 2008 B2
7363643 Drake et al. Apr 2008 B2
7370002 Heckerman et al. May 2008 B2
7370073 Yen et al. May 2008 B2
7386127 Bar-On Jun 2008 B2
7401140 Goulden et al. Jul 2008 B2
7421454 DeShan et al. Sep 2008 B2
7437368 Kolluri et al. Oct 2008 B1
7466823 Vestergaard et al. Dec 2008 B2
7487112 Barnes, Jr. Feb 2009 B2
7594189 Walker et al. Sep 2009 B1
7707485 Laksono Apr 2010 B2
7852222 Johnson et al. Dec 2010 B2
8005713 Sanz-Pastor et al. Aug 2011 B1
8024766 Addington Sep 2011 B2
8060407 Delker et al. Nov 2011 B1
8074076 Courtois Dec 2011 B2
8175921 Kopra May 2012 B1
8191088 Edwards et al. May 2012 B2
8267783 van Datta Sep 2012 B2
8272964 van Datta Sep 2012 B2
8302030 Soroca et al. Oct 2012 B2
8406739 Hull et al. Mar 2013 B2
8574074 van Datta Nov 2013 B2
8626584 van Datta Jan 2014 B2
8645992 Russell Feb 2014 B2
8676900 Yruski Mar 2014 B2
8751310 van Datta Jun 2014 B2
8763090 Capati Jun 2014 B2
8763157 Navar Jun 2014 B2
8769558 Navar Jul 2014 B2
8795076 van Datta Aug 2014 B2
9015747 Russell Apr 2015 B2
9129301 van Datta Sep 2015 B2
9195991 van Datta Nov 2015 B2
9367862 Yruski Jun 2016 B2
9466074 van Datta Oct 2016 B2
9474976 Capati et al. Oct 2016 B2
9525902 Navar Dec 2016 B2
9531686 Navar Dec 2016 B2
9864998 Yruski Jan 2018 B2
9873052 van Datta Jan 2018 B2
9984388 van Datta May 2018 B2
10042987 Navar Aug 2018 B2
10046239 van Datta Aug 2018 B2
10298703 Capati May 2019 B2
10390101 Rusell Aug 2019 B2
10410248 Yruski Sep 2019 B2
10467651 van Datta Nov 2019 B2
10657538 Yruski May 2020 B2
10789611 van Datta Sep 2020 B2
11004089 Yruski May 2021 B2
20010011226 Greer et al. Aug 2001 A1
20010013009 Greening et al. Aug 2001 A1
20010014915 Blumenau Aug 2001 A1
20010021994 Nash Sep 2001 A1
20010025245 Flickinger et al. Sep 2001 A1
20010025254 Park Sep 2001 A1
20010025274 Zehr et al. Sep 2001 A1
20010027412 Son Oct 2001 A1
20010032125 Bhan et al. Oct 2001 A1
20010032132 Moran Oct 2001 A1
20010032133 Moran Oct 2001 A1
20010032137 Bennett et al. Oct 2001 A1
20010032333 Flickinger Oct 2001 A1
20010034643 Acres Oct 2001 A1
20010034762 Jacobs et al. Oct 2001 A1
20010037232 Miller Nov 2001 A1
20010039210 St-Denis Nov 2001 A1
20010047297 Wen Nov 2001 A1
20010049620 Blasko Dec 2001 A1
20010052123 Kawai Dec 2001 A1
20020004743 Kutaragi et al. Jan 2002 A1
20020004744 Muyres et al. Jan 2002 A1
20020007307 Miller et al. Jan 2002 A1
20020007310 Long Jan 2002 A1
20020010626 Agmoni Jan 2002 A1
20020010628 Burns Jan 2002 A1
20020010757 Granik et al. Jan 2002 A1
20020013174 Murata Jan 2002 A1
20020013725 Takakura et al. Jan 2002 A1
20020018076 Gianola Feb 2002 A1
20020018982 Conroy Feb 2002 A1
20020019774 Kanter Feb 2002 A1
20020022476 Go Feb 2002 A1
20020022516 Forden Feb 2002 A1
20020023000 Bollay Feb 2002 A1
20020026345 Juels Feb 2002 A1
20020026355 Mitsuoka et al. Feb 2002 A1
20020026638 Eldering et al. Feb 2002 A1
20020032608 Kanter Mar 2002 A1
20020032626 DeWolf et al. Mar 2002 A1
20020032906 Grossman Mar 2002 A1
20020044687 Federman Apr 2002 A1
20020046087 Hey Apr 2002 A1
20020046095 Wallace Apr 2002 A1
20020046102 Dohring et al. Apr 2002 A1
20020049679 Russell et al. Apr 2002 A1
20020049968 Wilson et al. Apr 2002 A1
20020051521 Patrick May 2002 A1
20020055833 Sterling May 2002 A1
20020055876 Gabler May 2002 A1
20020056107 Schlack May 2002 A1
20020059577 Lu et al. May 2002 A1
20020059590 Kitsukawa et al. May 2002 A1
20020059610 Ellis May 2002 A1
20020061778 Acres May 2002 A1
20020067730 Hinderks et al. Jun 2002 A1
20020069240 Berk Jun 2002 A1
20020069405 Chapin et al. Jun 2002 A1
20020072965 Merriman et al. Jun 2002 A1
20020072966 Eldering et al. Jun 2002 A1
20020073235 Chen et al. Jun 2002 A1
20020077906 Remler Jun 2002 A1
20020077985 Kobata et al. Jun 2002 A1
20020078441 Drake et al. Jun 2002 A1
20020078444 Krewin et al. Jun 2002 A1
20020080968 Olsson Jun 2002 A1
20020082077 Johnson et al. Jun 2002 A1
20020082910 Kontogouris Jun 2002 A1
20020082913 Li Jun 2002 A1
20020082941 Bird Jun 2002 A1
20020083435 Blasko et al. Jun 2002 A1
20020083439 Eldering Jun 2002 A1
20020083441 Flickinger et al. Jun 2002 A1
20020083442 Eldering Jun 2002 A1
20020083443 Eldering et al. Jun 2002 A1
20020083444 Blasko et al. Jun 2002 A1
20020083445 Flickinger et al. Jun 2002 A1
20020083451 Gill et al. Jun 2002 A1
20020087402 Zustak Jul 2002 A1
20020087403 Meyers et al. Jul 2002 A1
20020087887 Busam et al. Jul 2002 A1
20020087973 Hamilton et al. Jul 2002 A1
20020087975 Schlack Jul 2002 A1
20020087980 Eldering et al. Jul 2002 A1
20020094868 Tuck et al. Jul 2002 A1
20020095676 Knee et al. Jul 2002 A1
20020098891 Graham et al. Jul 2002 A1
20020099600 Merriman et al. Jul 2002 A1
20020099611 De Souza et al. Jul 2002 A1
20020099653 De Souza et al. Jul 2002 A1
20020100040 Bull Jul 2002 A1
20020107073 Binney Aug 2002 A1
20020107075 Stephan Aug 2002 A1
20020107730 Bernstein Aug 2002 A1
20020109680 Orbanes et al. Aug 2002 A1
20020111154 Eldering et al. Aug 2002 A1
20020111172 DeWolf et al. Aug 2002 A1
20020111825 Martin et al. Aug 2002 A1
20020111865 Middleton, III et al. Aug 2002 A1
20020112035 Carey et al. Aug 2002 A1
20020112233 Cantu Bonilla et al. Aug 2002 A1
20020112240 Basco et al. Aug 2002 A1
20020112249 Hendricks et al. Aug 2002 A1
20020112250 Koplar et al. Aug 2002 A1
20020114466 Tanaka et al. Aug 2002 A1
20020116284 Steelman et al. Aug 2002 A1
20020120506 Hagen Aug 2002 A1
20020120574 Ezaki Aug 2002 A1
20020120589 Aoki Aug 2002 A1
20020122052 Reich et al. Sep 2002 A1
20020123928 Eldering et al. Sep 2002 A1
20020129362 Chang et al. Sep 2002 A1
20020129368 Schlack et al. Sep 2002 A1
20020133398 Geller et al. Sep 2002 A1
20020136407 Denning et al. Sep 2002 A1
20020138493 Shapiro et al. Sep 2002 A1
20020143639 Beckett et al. Oct 2002 A1
20020143652 Beckett Oct 2002 A1
20020143782 Headings et al. Oct 2002 A1
20020143901 Lupo et al. Oct 2002 A1
20020144262 Plotnick et al. Oct 2002 A1
20020144263 Eldering et al. Oct 2002 A1
20020147633 Rafizadeh Oct 2002 A1
20020147638 Banerjee et al. Oct 2002 A1
20020147645 Alao et al. Oct 2002 A1
20020152117 Cristofalo et al. Oct 2002 A1
20020155878 Lert, Jr. et al. Oct 2002 A1
20020155891 Okada et al. Oct 2002 A1
20020157002 Messerges et al. Oct 2002 A1
20020159304 Morita et al. Oct 2002 A1
20020161625 Brito-Valladares et al. Oct 2002 A1
20020161639 Goldstein Oct 2002 A1
20020164977 Link, II et al. Nov 2002 A1
20020164999 Johnson Nov 2002 A1
20020165026 Perkins et al. Nov 2002 A1
20020165764 Wade et al. Nov 2002 A1
20020173349 Ach, III Nov 2002 A1
20020173359 Gallo et al. Nov 2002 A1
20020175936 Tenembaum Nov 2002 A1
20020178442 Williams Nov 2002 A1
20020178445 Eldering et al. Nov 2002 A1
20020178447 Plotnick et al. Nov 2002 A1
20020184047 Plotnick et al. Dec 2002 A1
20020184086 Linde Dec 2002 A1
20020184088 Rosenberg Dec 2002 A1
20020184130 Blasko Dec 2002 A1
20020184642 Lude et al. Dec 2002 A1
20020185825 Miyamoto et al. Dec 2002 A1
20020193066 Connelly Dec 2002 A1
20020194058 Eldering Dec 2002 A1
20020194585 Connelly Dec 2002 A1
20020194590 Pong Dec 2002 A1
20020194598 Connelly Dec 2002 A1
20020194607 Connelly Dec 2002 A1
20030004810 Eldering Jan 2003 A1
20030009762 Hooper et al. Jan 2003 A1
20030014307 Heng Jan 2003 A1
20030014312 Fleisher Jan 2003 A1
20030014414 Newman Jan 2003 A1
20030014754 Chang Jan 2003 A1
20030018527 Filepp et al. Jan 2003 A1
20030018797 Dunning et al. Jan 2003 A1
20030028433 Merriman et al. Feb 2003 A1
20030033405 Perdon et al. Feb 2003 A1
20030035075 Butler et al. Feb 2003 A1
20030036944 Lesandrini et al. Feb 2003 A1
20030046148 Rizzi et al. Mar 2003 A1
20030048293 Werkhoven Mar 2003 A1
20030054888 Walker et al. Mar 2003 A1
20030060247 Goldberg et al. Mar 2003 A1
20030066092 Wagner et al. Apr 2003 A1
20030070167 Holtz et al. Apr 2003 A1
20030073496 D'Amico et al. Apr 2003 A1
20030074252 Chandler-Pepelnjak et al. Apr 2003 A1
20030074661 Krapf et al. Apr 2003 A1
20030076347 Barrett et al. Apr 2003 A1
20030079226 Barrett Apr 2003 A1
20030084449 Chane et al. May 2003 A1
20030084456 Ryan et al. May 2003 A1
20030093311 Knowlson May 2003 A1
20030100375 Wakae et al. May 2003 A1
20030101449 Bentolila et al. May 2003 A1
20030101451 Bentolila et al. May 2003 A1
20030101454 Ozer et al. May 2003 A1
20030103644 Klayh Jun 2003 A1
20030104867 Kobayashi et al. Jun 2003 A1
20030110131 Alain et al. Jun 2003 A1
20030110171 Ozer et al. Jun 2003 A1
20030110499 Knudson et al. Jun 2003 A1
20030115074 Freeman et al. Jun 2003 A1
20030115318 Wueste Jun 2003 A1
20030115587 Kendall et al. Jun 2003 A1
20030120940 Vataja Jun 2003 A1
20030126150 Chan Jul 2003 A1
20030135513 Quinn et al. Jul 2003 A1
20030139966 Sirota et al. Jul 2003 A1
20030144044 Piarsky Jul 2003 A1
20030144048 Silva Jul 2003 A1
20030148625 Ho et al. Aug 2003 A1
20030149618 Sender et al. Aug 2003 A1
20030149623 Chen Aug 2003 A1
20030149938 McElfresh et al. Aug 2003 A1
20030149975 Eldering et al. Aug 2003 A1
20030154126 Gehlot et al. Aug 2003 A1
20030158872 Adams Aug 2003 A1
20030163369 Arr Aug 2003 A1
20030163482 Bunney et al. Aug 2003 A1
20030171988 Sugihara Sep 2003 A1
20030171990 Rao et al. Sep 2003 A1
20030172374 Vinson et al. Sep 2003 A1
20030172376 Coffin, III Sep 2003 A1
20030177490 Hoshino et al. Sep 2003 A1
20030182567 Barton et al. Sep 2003 A1
20030182663 Gudorf et al. Sep 2003 A1
20030187719 Brocklebank Oct 2003 A1
20030190961 Seidman Oct 2003 A1
20030191690 McIntyre et al. Oct 2003 A1
20030191742 Yonezawa et al. Oct 2003 A1
20030195021 Yamashita et al. Oct 2003 A1
20030195801 Takakura et al. Oct 2003 A1
20030195837 Kostic et al. Oct 2003 A1
20030199292 Greenberg Oct 2003 A1
20030200452 Tagawa et al. Oct 2003 A1
20030204632 Willebeek-LeMair et al. Oct 2003 A1
20030208680 Byrne et al. Nov 2003 A1
20030212608 Cliff Nov 2003 A1
20030215211 Coffin, III Nov 2003 A1
20030216961 Barry Nov 2003 A1
20030221100 Russ et al. Nov 2003 A1
20030221113 Kupka et al. Nov 2003 A1
20030226141 Krasnow et al. Dec 2003 A1
20030226142 Rand Dec 2003 A1
20030229893 Sgaraglino Dec 2003 A1
20040002380 Brosnan et al. Jan 2004 A1
20040003396 Babu Jan 2004 A1
20040014454 Burgess et al. Jan 2004 A1
20040015397 Barry et al. Jan 2004 A1
20040015398 Hayward Jan 2004 A1
20040015608 Ellis et al. Jan 2004 A1
20040019521 Birmingham Jan 2004 A1
20040025174 Cerrato Feb 2004 A1
20040030595 Park Feb 2004 A1
20040034536 Hughes Feb 2004 A1
20040034686 Guthrie Feb 2004 A1
20040039648 Candelore et al. Feb 2004 A1
20040039796 Watkins Feb 2004 A1
20040043817 Willis Mar 2004 A1
20040043819 Willis Mar 2004 A1
20040044567 Willis Mar 2004 A1
20040044569 Roberts et al. Mar 2004 A1
20040044571 Bronnimann et al. Mar 2004 A1
20040044574 Cochran et al. Mar 2004 A1
20040054589 Nicolas et al. Mar 2004 A1
20040057348 Shteyn et al. Mar 2004 A1
20040059625 Schrader Mar 2004 A1
20040060060 Carr Mar 2004 A1
20040064833 Lee et al. Apr 2004 A1
20040068483 Sakurai et al. Apr 2004 A1
20040068552 Kotz et al. Apr 2004 A1
20040073482 Wiggins et al. Apr 2004 A1
20040076404 Nakano et al. Apr 2004 A1
20040078263 Altieri Apr 2004 A1
20040078266 Kim Apr 2004 A1
20040078292 Blumenau Apr 2004 A1
20040078809 Drazin Apr 2004 A1
20040083133 Nicholas et al. Apr 2004 A1
20040088170 Nakanishi et al. May 2004 A1
20040088583 Yoon et al. May 2004 A1
20040102248 Young et al. May 2004 A1
20040103024 Patel et al. May 2004 A1
20040103429 Carlucci et al. May 2004 A1
20040107136 Nemirofsky et al. Jun 2004 A1
20040110565 Levesque Jun 2004 A1
20040111317 Ebisawa Jun 2004 A1
20040111484 Young et al. Jun 2004 A1
20040116183 Prindle Jun 2004 A1
20040117272 Shehab Jun 2004 A1
20040121835 Willis et al. Jun 2004 A1
20040121842 Willis et al. Jun 2004 A1
20040126747 Fujisawa et al. Jul 2004 A1
20040133467 Siler Jul 2004 A1
20040133480 Domes Jul 2004 A1
20040133518 Dryall Jul 2004 A1
20040133909 Ma Jul 2004 A1
20040137980 Aenlle Jul 2004 A1
20040139465 Matthews, III et al. Jul 2004 A1
20040140352 Walker et al. Jul 2004 A1
20040143478 Ward Jul 2004 A1
20040143495 Koenig Jul 2004 A1
20040148221 Chu Jul 2004 A1
20040148424 Berkson et al. Jul 2004 A1
20040148625 Eldering et al. Jul 2004 A1
20040152517 Hardisty et al. Aug 2004 A1
20040152518 Kogo Aug 2004 A1
20040153360 Schumann Aug 2004 A1
20040153363 Stehling Aug 2004 A1
20040153385 Allibhoy et al. Aug 2004 A1
20040153453 Brodie et al. Aug 2004 A1
20040158858 Paxton Aug 2004 A1
20040162758 Willis Aug 2004 A1
20040162759 Willis Aug 2004 A1
20040163101 Swix et al. Aug 2004 A1
20040163111 Palazzo et al. Aug 2004 A1
20040163134 Willis Aug 2004 A1
20040168063 Revital et al. Aug 2004 A1
20040168188 Bennington et al. Aug 2004 A1
20040168202 Ebihara Aug 2004 A1
20040169678 Oliver Sep 2004 A1
20040172324 Merriman et al. Sep 2004 A1
20040172331 Merriman et al. Sep 2004 A1
20040172332 Merriman et al. Sep 2004 A1
20040172343 Allibhoy et al. Sep 2004 A1
20040176170 Eck et al. Sep 2004 A1
20040176995 Fusz Sep 2004 A1
20040177001 Salinas Sep 2004 A1
20040181808 Schaefer et al. Sep 2004 A1
20040186766 Fellenstein et al. Sep 2004 A1
20040186771 Squires Sep 2004 A1
20040186993 Risan et al. Sep 2004 A1
20040192351 Duncan Sep 2004 A1
20040193488 Khoo et al. Sep 2004 A1
20040193902 Vogler et al. Sep 2004 A1
20040194123 Fredlund et al. Sep 2004 A1
20040194128 McIntyre et al. Sep 2004 A1
20040201629 Bates et al. Oct 2004 A1
20040204238 Aoki Oct 2004 A1
20040204247 Walker et al. Oct 2004 A1
20040205157 Bibelnieks et al. Oct 2004 A1
20040205508 Wecker et al. Oct 2004 A1
20040205807 Wilcoxson et al. Oct 2004 A1
20040210472 Lew et al. Oct 2004 A1
20040210489 Jackson et al. Oct 2004 A1
20040210661 Thompson Oct 2004 A1
20040210824 Shoff et al. Oct 2004 A1
20040219977 Ebisawa Nov 2004 A1
20040220850 Ferrer et al. Nov 2004 A1
20040221018 Ji Nov 2004 A1
20040224772 Canessa et al. Nov 2004 A1
20040225562 Turner Nov 2004 A1
20040225715 Gottfried Nov 2004 A1
20040230593 Rudin et al. Nov 2004 A1
20040230994 Urdang et al. Nov 2004 A1
20040234932 Hughes et al. Nov 2004 A1
20040236585 Kohnke et al. Nov 2004 A1
20040243455 Smith Dec 2004 A1
20040243466 Trzybinski et al. Dec 2004 A1
20040243470 Ozer et al. Dec 2004 A1
20040243623 Ozer et al. Dec 2004 A1
20040248649 Arai et al. Dec 2004 A1
20040249786 Dabney et al. Dec 2004 A1
20040252051 Johnson Dec 2004 A1
20040254831 Dean Dec 2004 A1
20040254957 Hyotyniemi et al. Dec 2004 A1
20040255148 Monteiro et al. Dec 2004 A1
20040259553 Delaney et al. Dec 2004 A1
20040260609 Loeb et al. Dec 2004 A1
20040261125 Ellis et al. Dec 2004 A1
20040266535 Reeves Dec 2004 A1
20040266537 Morris Dec 2004 A1
20040267611 Hoerenz Dec 2004 A1
20040267806 Lester Dec 2004 A1
20050005242 Hoyle Jan 2005 A1
20050015267 Barringer et al. Jan 2005 A1
20050021387 Gottfurcht Jan 2005 A1
20050021396 Pearch et al. Jan 2005 A1
20050021397 Cui et al. Jan 2005 A1
20050021403 Ozer et al. Jan 2005 A1
20050021465 Segerstrom Jan 2005 A1
20050021470 Martin et al. Jan 2005 A1
20050021853 Parekh et al. Jan 2005 A1
20050022019 Medvinsky et al. Jan 2005 A1
20050027587 Latona et al. Feb 2005 A1
20050027595 Ha et al. Feb 2005 A1
20050027699 Awadallah et al. Feb 2005 A1
20050028188 Latona et al. Feb 2005 A1
20050028195 Feinleib et al. Feb 2005 A1
20050032577 Blackburn et al. Feb 2005 A1
20050033700 Vogler et al. Feb 2005 A1
20050038698 Lukose et al. Feb 2005 A1
20050038702 Merriman et al. Feb 2005 A1
20050050070 Sheldon Mar 2005 A1
20050055725 Stewart Mar 2005 A1
20050060264 Shrock et al. Mar 2005 A1
20050075155 Sitrick Apr 2005 A1
20050075172 Coleman Apr 2005 A1
20050076051 Carobus et al. Apr 2005 A1
20050088320 Kovach Apr 2005 A1
20050091107 Blum Apr 2005 A1
20050091108 Frost Apr 2005 A1
20050091111 Green et al. Apr 2005 A1
20050096975 Moshe May 2005 A1
20050096983 Werkhoven May 2005 A1
20050097622 Zigmond et al. May 2005 A1
20050101386 Lavanchy et al. May 2005 A1
20050102177 Takayama May 2005 A1
20050102202 Linden et al. May 2005 A1
20050107158 Kanisawa et al. May 2005 A1
20050108095 Perlmutter May 2005 A1
20050113170 McHugh May 2005 A1
20050114526 Aoyama May 2005 A1
20050125286 Crippen et al. Jun 2005 A1
20050125528 Burke et al. Jun 2005 A1
20050130725 Creamer et al. Jun 2005 A1
20050143174 Goldman et al. Jun 2005 A1
20050144063 Spector Jun 2005 A1
20050144073 Morrisroe et al. Jun 2005 A1
20050144251 Slate Jun 2005 A1
20050149396 Horowitz et al. Jul 2005 A1
20050149397 Morgenstern et al. Jul 2005 A1
20050153760 Varley Jul 2005 A1
20050154640 Kolluri et al. Jul 2005 A1
20050154717 Watson et al. Jul 2005 A1
20050155056 Knee et al. Jul 2005 A1
20050155083 Oh et al. Jul 2005 A1
20050160442 Kaplowitz Jul 2005 A1
20050164757 Ebisawa Jul 2005 A1
20050165640 Kotorov Jul 2005 A1
20050165644 Beyda et al. Jul 2005 A1
20050171865 Beardow Aug 2005 A1
20050177413 Blumberg et al. Aug 2005 A1
20050177430 Willis Aug 2005 A1
20050177431 Willis et al. Aug 2005 A1
20050177461 Rosefelt et al. Aug 2005 A1
20050177853 Williams et al. Aug 2005 A1
20050178940 Granick Aug 2005 A1
20050179685 Kake et al. Aug 2005 A1
20050182693 Alivandi Aug 2005 A1
20050182737 Brown Aug 2005 A1
20050185825 Hoshino et al. Aug 2005 A1
20050192071 Matsuno et al. Sep 2005 A1
20050192864 Ganz Sep 2005 A1
20050193411 Funston Sep 2005 A1
20050193425 Sull et al. Sep 2005 A1
20050195157 Kramer et al. Sep 2005 A1
20050202385 Coward et al. Sep 2005 A1
20050203804 Suzuki et al. Sep 2005 A1
20050203811 David Sep 2005 A1
20050203849 Benson Sep 2005 A1
20050204381 Ludvig et al. Sep 2005 A1
20050216346 Kusumoto et al. Sep 2005 A1
20050216348 Martin et al. Sep 2005 A1
20050216581 Blumenau et al. Sep 2005 A1
20050216932 Danker Sep 2005 A1
20050222908 Altberg et al. Oct 2005 A1
20050223100 Chen et al. Oct 2005 A1
20050227749 Bender et al. Oct 2005 A1
20050228797 Koningstein et al. Oct 2005 A1
20050235030 Lauckhart et al. Oct 2005 A1
20050235199 Adams Oct 2005 A1
20050235310 Bies Oct 2005 A1
20050235318 Grauch et al. Oct 2005 A1
20050240476 Bigott Oct 2005 A1
20050240961 Jerding et al. Oct 2005 A1
20050246736 Beyda et al. Nov 2005 A1
20050247769 Potter et al. Nov 2005 A1
20050251539 Parekh et al. Nov 2005 A1
20050254366 Amar Nov 2005 A1
20050255914 McHale et al. Nov 2005 A1
20050256768 Robinson Nov 2005 A1
20050261062 Lewin et al. Nov 2005 A1
20050261962 Chuah Nov 2005 A1
20050266906 Stevens Dec 2005 A1
20050266907 Weston et al. Dec 2005 A1
20050270537 Mian et al. Dec 2005 A1
20050273618 Takemura et al. Dec 2005 A1
20050283395 Lesandrini et al. Dec 2005 A1
20050283401 Swix et al. Dec 2005 A1
20050283797 Eldering et al. Dec 2005 A1
20050286860 Conklin Dec 2005 A1
20050288999 Lerner et al. Dec 2005 A1
20060007312 James Jan 2006 A1
20060009199 Mittal Jan 2006 A1
20060031405 Goldman et al. Feb 2006 A1
20060031551 Agresta et al. Feb 2006 A1
20060080702 Diez et al. Apr 2006 A1
20060085517 Kaurila Apr 2006 A1
20060085816 Funk et al. Apr 2006 A1
20060090186 Santangelo et al. Apr 2006 A1
20060129917 Volk et al. Jun 2006 A1
20060130095 Willis et al. Jun 2006 A1
20060135232 Willis Jun 2006 A1
20060143650 Tanikawa et al. Jun 2006 A1
20060150249 Gassen et al. Jul 2006 A1
20060167747 Goodman Jul 2006 A1
20060193471 Stehle Aug 2006 A1
20060195859 Konig et al. Aug 2006 A1
20060195860 Eldering et al. Aug 2006 A1
20060195902 King et al. Aug 2006 A1
20060200381 Elkholy et al. Sep 2006 A1
20060206912 Klarfeld et al. Sep 2006 A1
20060212347 Fang et al. Sep 2006 A1
20060212350 Ellis et al. Sep 2006 A1
20060217110 Othmer Sep 2006 A1
20060224951 Burke et al. Oct 2006 A1
20060230141 Willis Oct 2006 A1
20060242667 Peterson et al. Oct 2006 A1
20060242703 Abeni Oct 2006 A1
20060248209 Chiu Nov 2006 A1
20060248569 Lienhart et al. Nov 2006 A1
20060248595 Kelly et al. Nov 2006 A1
20060253323 Phan et al. Nov 2006 A1
20060253330 Maggio et al. Nov 2006 A1
20060265503 Jones et al. Nov 2006 A1
20060268667 Jellison, Jr. et al. Nov 2006 A1
20060287915 Boulet et al. Dec 2006 A1
20060294566 Zlattner Dec 2006 A1
20070027771 Collins et al. Feb 2007 A1
20070038508 Jain et al. Feb 2007 A1
20070038516 Apple et al. Feb 2007 A1
20070038931 Allaire et al. Feb 2007 A1
20070050254 Driscoll Mar 2007 A1
20070050256 Walker et al. Mar 2007 A1
20070055980 Mageid et al. Mar 2007 A1
20070061204 Ellis et al. Mar 2007 A1
20070061711 Bodin et al. Mar 2007 A1
20070061838 Grubbs et al. Mar 2007 A1
20070066287 Papulov Mar 2007 A1
20070072676 Baluja Mar 2007 A1
20070073756 Manhas et al. Mar 2007 A1
20070078706 Datta et al. Apr 2007 A1
20070078712 Ott et al. Apr 2007 A1
20070078714 Ott Apr 2007 A1
20070078989 van Datta Apr 2007 A1
20070079326 Datta et al. Apr 2007 A1
20070079331 Datta et al. Apr 2007 A1
20070079335 McDonough Apr 2007 A1
20070083611 Farago et al. Apr 2007 A1
20070089151 Moore et al. Apr 2007 A1
20070094081 Yruski Apr 2007 A1
20070094082 Yruski Apr 2007 A1
20070094083 Yruski Apr 2007 A1
20070094363 Yruski Apr 2007 A1
20070101360 Gutta et al. May 2007 A1
20070118425 Yruski May 2007 A1
20070130012 Yruski Jun 2007 A1
20070130594 Hidary et al. Jun 2007 A1
20070146812 Lawton Jun 2007 A1
20070150919 Morishita Jun 2007 A1
20070157220 Cordray et al. Jul 2007 A1
20070162945 Mills Jul 2007 A1
20070168288 Bozeman Jul 2007 A1
20070174471 Van Rossum Jul 2007 A1
20070225047 Bakos Sep 2007 A1
20070244760 Bodnar et al. Oct 2007 A1
20070276926 LaJoie et al. Nov 2007 A1
20070294740 Drake et al. Dec 2007 A1
20070294773 Hydrie et al. Dec 2007 A1
20070299935 Plastina et al. Dec 2007 A1
20080046917 de Heer Feb 2008 A1
20080046948 Verosub Feb 2008 A1
20080097872 Peckover Apr 2008 A1
20080102947 Hays et al. May 2008 A1
20080104106 Rosenberg et al. May 2008 A1
20080109844 Baldeschwieler et al. May 2008 A1
20080114861 Gildred May 2008 A1
20080120407 Chen et al. May 2008 A1
20080126476 Nicholas et al. May 2008 A1
20080127244 Zhang May 2008 A1
20080137645 Skog Jun 2008 A1
20080140239 Rosenberg et al. Jun 2008 A1
20080140717 Rosenberg et al. Jun 2008 A1
20080141372 Massey et al. Jun 2008 A1
20080195664 Maharajh et al. Aug 2008 A1
20080207137 Maharajh et al. Aug 2008 A1
20080307454 Ahanger et al. Dec 2008 A1
20090083788 Russell Mar 2009 A1
20090094160 Webster et al. Apr 2009 A1
20090183081 Rodriguez Jul 2009 A1
20090199230 Kumar et al. Aug 2009 A1
20090204481 Navar Aug 2009 A1
20090254430 Cherenson Oct 2009 A1
20100022310 van Datta Jan 2010 A1
20100030640 van Datta Feb 2010 A1
20100043022 Kaftan Feb 2010 A1
20100169467 Shukla et al. Jul 2010 A1
20100169910 Collins et al. Jul 2010 A1
20100269138 Krikorian et al. Oct 2010 A1
20110004669 Navar Jan 2011 A1
20110010545 Kill et al. Jan 2011 A1
20110015975 Yruski et al. Jan 2011 A1
20110029383 Engel et al. Feb 2011 A1
20110041161 Capati Feb 2011 A1
20110125582 Datta et al. May 2011 A1
20110138058 Ishida Jun 2011 A1
20110307339 Russell Dec 2011 A1
20120078714 Gross Mar 2012 A1
20130232000 van Datta Sep 2013 A1
20130232001 van Datta Sep 2013 A1
20130297411 van Datta Nov 2013 A1
20140019229 van Datta Jan 2014 A1
20140019249 Nicholas et al. Jan 2014 A1
20140089081 Yruski Mar 2014 A1
20140215224 Navar Jul 2014 A1
20140304328 Capati Oct 2014 A1
20140324576 van Datta Oct 2014 A1
20140337882 Navar Nov 2014 A1
20150294368 Russell Oct 2015 A1
20160027053 van Datta Jan 2016 A1
20160292736 Yruski Oct 2016 A1
20170091804 van Datta Mar 2017 A1
20170164030 Navar Jun 2017 A1
20170206341 Navar Jul 2017 A1
20170208145 Capati Jul 2017 A1
20180225676 Yruski Aug 2018 A1
20180374116 van Datta Dec 2018 A1
20200034875 van Datta Jan 2020 A1
Foreign Referenced Citations (39)
Number Date Country
9959097 Nov 1999 AU
2106122 Mar 1994 CA
2250680 Apr 2000 CA
1 653 819 Aug 2005 CN
103279874 Sep 2013 CN
0 337 539 Oct 1989 EP
0 405 776 Jan 1991 EP
0 620 688 Oct 1994 EP
0 625 760 Nov 1994 EP
0 743 595 Oct 1996 EP
0 905 928 Mar 1999 EP
2 141 907 Jan 1985 GB
2 194 369 Mar 1988 GB
12-20925 Sep 1989 JP
63-35569 Dec 1994 JP
81-17445 May 1996 JP
81-73634 Jul 1996 JP
82-80934 Oct 1996 JP
2001-111921 Apr 2001 JP
2001-321556 Nov 2001 JP
2002-259433 Sep 2002 JP
2002-358455 Dec 2002 JP
2002-366971 Dec 2002 JP
2003-248844 Sep 2003 JP
2004-102475 Apr 2004 JP
2004-298469 Oct 2004 JP
WO 199314462 Jul 1993 WO
WO 199319427 Sep 1993 WO
WO 199322017 Nov 1993 WO
WO 199323125 Nov 1993 WO
WO 199512442 May 1995 WO
WO 199512853 May 1995 WO
WO 9851384 Nov 1998 WO
WO 2003032127 Apr 2003 WO
WO 2004100010 Nov 2004 WO
WO 2005086969 Sep 2005 WO
WO 2007041022 Apr 2007 WO
WO 2007041028 Apr 2007 WO
WO 2007130681 Nov 2007 WO
Non-Patent Literature Citations (194)
Entry
EP 18207669.5, First Examination Report dated Jan. 11, 2021.
Andreaux. J.-P.; Copy Protection system for digital home networks; Mar. 2004; IEEE, vol. 21, Issue: 2; pp. 100-108.
Business Wire, “Juno launches America's first free Internet e-mail service; Initial advertisers include Land's End, Miramax and Snappie,” Apr. 19, 1996.
Business Wire, “RTIME Announces First 100-Person Twitch Game for Internet; “RTIME Rocks!” Demonstrates the Power of the RTIME Interactive Networking Engine to Support Large Scale, High Performance, Internet Game Play,” Apr. 14, 1997.
Cohen, Josh, “A General Overview of Two New Technologies for Playing Protected Content on Portable or Networked Devices,” Microsoft Windows Media, Jun. 2004, 1-8.
Courtois N et al: An Algebraic Masking Method to Protect AES Against Power Attacks, ‘Online! XP002344150 Retrieved from the Internet: URL:eprint.iacr.org/2005/204.pdf>’ retrieved on Sep. 8, 2005!
Fontijn, Willem; AmbientDB: P2P Data Management Middleware for Ambient Intelligence; Year: 2004; IEEE; pp. 1-5.
Microsoft Corporation, “A Technical Overview of Windows Media DRM 10 for Devices,” MicrosoftWindows Media, Sep. 2004, 1-16.
Microsoft Corporation, “Architecture of Windows Media Rights Manager,” www.microsoft.com/windows/windowsmedia/howto/articles/drmarchitecture.aspc, May 2004.
Pricewaterhousecoopers, “Lab Online Ad Measurement Study,” Dec. 2001.
Recording Industry Association of America, “Frequently Asked Questions—Webcasting,” www.riaa.com/issues/licensing/webcasting_faq.asp. (acc. 2004).
Statement in Accordance with the Notice from the European Patent Office dated Oct. 1, 2007 Concerning Business Methods Nov. 1, 2007, XP002456252.
U.S. Copyright Office, “The Digital Millennium Copyright Act of 1998,” Oct. 1998, 1-18.
What TV Ratings Really Mean (And Other Frequently-Asked Questions). Nielsen Media Research. Web. <http:!/documents.chelmsford. k 12. ma.us/dsweb/GeUDocument-14983/nielsenmedia.htm>, Jun. 2005.
PCT/US06/037018, International Preliminary Report on Patentability dated Apr. 1, 2008.
PCT/US06/037018, International Search Report and Written Opinion dated Aug. 7, 2007.
PCT/US06/036958, International Preliminary Report on Patentability dated Apr. 1, 2008.
PCT/US06/036958, International Search Report and Written Opinion dated Apr. 27, 2007.
PCT/US07/11059, International Preliminary Report on Patentability dated Nov. 11, 2008.
PCT/US07/11059, International Search Report and Written Opinion dated May 30, 2008.
EP 06815173.7, First Examination Report dated Feb. 23, 2016.
EP 06815173.7, Extended European Search Report dated Oct. 5, 2011.
EP 18207669.5, Extended European Search Report dated Mar. 25, 2019.
JP 2009-509786, Decision of Refusal dated Oct. 30, 2012.
JP 2009-509786, Decision of Refusal dated Aug. 2, 2011.
JP 2009-509786, Decision of Refusal dated Jul. 28, 2011.
JP 2013-039681, Decision of Refusal dated Feb. 3, 2015.
JP 2013-039681, Notification of Reason for Refusal dated Feb. 12, 2014.
CN 200780016268.2, First Office Action dated Jan. 4, 2012.
CN 201310051520.0, First Office Action dated Sep. 1, 2015.
EP 07776856.2, Extended European Search Report dated Jun. 9, 2011.
U.S. Appl. No. 11/241,229 Final Office Action dated Apr. 23, 2010.
U.S. Appl. No. 11/241,229 Office Action dated Nov. 19, 2009.
U.S. Appl. No. 13/939,178 Office Action dated Oct. 10, 2013.
U.S. Appl. No. 14/336,452 Office Action dated Jan. 8, 2016.
U.S. Appl. No. 15/285,928 Office Action dated Sep. 13, 2017.
U.S. Appl. No. 15/992,014 Office Action dated Feb. 13, 2019.
U.S. Appl. No. 16/594,354 Office Action dated Feb. 4, 2020.
U.S. Appl. No. 12/571,204 Office Action dated Feb. 28, 2012.
U.S. Appl. No. 12/571,225 Office Action dated Feb. 2, 2012.
U.S. Appl. No. 11/240,655 Final Office Action dated Nov. 14, 2013.
U.S. Appl. No. 11/240,655 Office Action dated Aug. 5, 2013.
U.S. Appl. No. 11/240,655 Final Office Action dated Jan. 27, 2010.
U.S. Appl. No. 11/240,655 Office Action dated Apr. 16, 2009.
U.S. Appl. No. 13/857,080 Office Action dated Jul. 28, 2017.
U.S. Appl. No. 13/857,080 Final Office Action dated Feb. 24, 2017.
U.S. Appl. No. 13/857,080 Office Action dated Aug. 2, 2016.
U.S. Appl. No. 13/857,080 Final Office Action dated Aug. 19, 2015.
U.S. Appl. No. 13/857,080 Office Action dated Apr. 29, 2015.
U.S. Appl. No. 13/857,082 Final Office Action dated Feb. 28, 2017.
U.S. Appl. No. 13/857,082 Office Action dated Aug. 18, 2016.
U.S. Appl. No. 13/857,082 Final Office Action dated Aug. 11, 2015.
U.S. Appl. No. 13/857,082 Office Action dated Apr. 16, 2015.
U.S. Appl. No. 12/190,323 Final Office Action dated Feb. 25, 2013.
U.S. Appl. No. 12/190,323 Office Action dated May 7, 2012.
U.S. Appl. No. 12/190,323 Office Action dated Jun. 8, 2011.
U.S. Appl. No. 12/190,323 Final Office Action dated Nov. 14, 2011.
U.S. Appl. No. 14/691,404 Office Action dated Sep. 18, 2018.
U.S. Appl. No. 14/691,404 Final Office Action dated Jan. 30, 2018.
U.S. Appl. No. 14/691,404 Office Action dated Sep. 21, 2017.
U.S. Appl. No. 14/691,404 Final Office Action dated Mar. 22, 2017.
U.S. Appl. No. 14/691,404 Office Action dated Oct. 27, 2016.
U.S. Appl. No. 14/691,404 Final Office Action dated Mar. 25, 2016.
U.S. Appl. No. 14/691,404 Office Action dated Nov. 13, 2015.
U.S. Appl. No. 13/191,398 Final Office Action dated Jun. 10, 2014.
U.S. Appl. No. 13/191,398 Office Action dated Dec. 3, 2013.
U.S. Appl. No. 13/191,398 Final Office Action dated Jun. 7, 2013.
U.S. Appl. No. 13/191,398 Office Action dated Mar. 22, 2012.
U.S. Appl. No. 11/535,370 Final Office Action dated Jun. 8, 2010.
U.S. Appl. No. 11/535,307 Office Action dated Dec. 10, 2009.
U.S. Appl. No. 11/535,307 Final Action dated Sep. 8, 2009.
U.S. Appl. No. 11/535,307 Office Action dated Apr. 16, 2009.
U.S. Appl. No. 13/013,789 Office Action dated Feb. 19, 2019.
U.S. Appl. No. 13/013,789 Final Office Action dated Oct. 3, 2018.
U.S. Appl. No. 13/013,789 Office Action dated May 24, 2018.
U.S. Appl. No. 13/013,789 Final Office Action dated Aug. 25, 2017.
U.S. Appl. No. 13/013,789 Office Action dated May 4, 2017.
U.S. Appl. No. 13/013,789 Final Office Action dated Jun. 17, 2016.
U.S. Appl. No. 13/013,789 Office Action dated Feb. 12, 2016.
U.S. Appl. No. 13/013,789 Final Office Action dated Jul. 28, 2014.
U.S. Appl. No. 13/013,789 Office Action dated Dec. 20, 2013.
U.S. Appl. No. 13/013,789 Final Office Action dated Feb. 27, 2013.
U.S. Appl. No. 13/013,789 Office Action dated Oct. 9, 2012.
U.S. Appl. No. 11/452,848 Final Office Action dated Apr. 7, 2015.
U.S. Appl. No. 11/452,848 Office Action dated Oct. 23, 2014.
U.S. Appl. No. 11/452,848 Final Office Action dated Jun. 5, 2014.
U.S. Appl. No. 11/452,848 Office Action dated Nov. 18, 2013.
U.S. Appl. No. 11/452,848 Final Office Action dated Feb. 15, 2011.
U.S. Appl. No. 11/452,848 Office Action dated Sep. 15, 2010.
U.S. Appl. No. 11/452,848 Final Office Action dated Apr. 21, 2010.
U.S. Appl. No. 11/452,848 Office Action dated Oct. 20, 2009.
U.S. Appl. No. 11/452,848 Final Office Action dated Jul. 9, 2009.
U.S. Appl. No. 11/452,848 Office Action dated Jan. 27, 2009.
U.S. Appl. No. 14/028,327 Final Office Action dated Mar. 19, 2015.
U.S. Appl. No. 14/028,327 Office Action dated Oct. 8, 2014.
U.S. Appl. No. 14/028,327 Final Office Action dated Jun. 9, 2014.
U.S. Appl. No. 14/028,327 Office Action dated Nov. 7, 2013.
U.S. Appl. No. 14/875,682 Office Action dated Jul. 31, 2018.
U.S. Appl. No. 14/875,682 Office Action dated Feb. 22, 2018.
U.S. Appl. No. 14/875,682 Office Action dated Jan. 26, 2017.
U.S. Appl. No. 14/875,682 Final Office Action dated Jul. 18, 2016.
U.S. Appl. No. 14/875,682 Office Action dated Jan. 29, 2016.
U.S. Appl. No. 12/782,678 Final Office Action dated Jul. 31, 2013.
U.S. Appl. No. 12/782,678 Office Action dated Jan. 7, 2013.
U.S. Appl. No. 12/782,678 Office Action dated Oct. 4, 2012.
U.S. Appl. No. 14/308,313 Final Office Action dated Oct. 23, 2015.
U.S. Appl. No. 14/308,313 Office Action dated Apr. 27, 2015.
U.S. Appl. No. 15/333,932 Final Office Action dated Mar. 22, 2018.
U.S. Appl. No. 15/333,932 Office Action dated Aug. 14, 2017.
U.S. Appl. No. 11/586,990 Office Action dated Mar. 18, 2016.
U.S. Appl. No. 11/586,990 Final Office Action dated Dec. 8, 2014.
U.S. Appl. No. 11/586,990 Office Action dated Aug. 12, 2014.
U.S. Appl. No. 11/586,990 Final Office Action dated Apr. 7, 2014.
U.S. Appl. No. 11/586,990 Office Action dated Nov. 20, 2013.
U.S. Appl. No. 11/586,990 Final Office Action dated Apr. 10, 2013.
U.S. Appl. No. 11/586,990 Office Action dated Nov. 23, 2012.
U.S. Appl. No. 11/586,990 Final Office Action dated Feb. 14, 2011.
U.S. Appl. No. 11/586,990 Office Action dated Sep. 15, 2010.
U.S. Appl. No. 11/588,036 Final Office Action dated Jun. 19, 2018.
U.S. Appl. No. 11/588,036 Office Action dated Mar. 16, 2018.
U.S. Appl. No. 11/588,036 Final Office Action dated Mar. 15, 2017.
U.S. Appl. No. 11/588,036 Office Action dated Aug. 31, 2016.
U.S. Appl. No. 11/588,036 Final Office Action dated Aug. 4, 2015.
U.S. Appl. No. 11/588,036 Office Action dated Jan. 15, 2015.
U.S. Appl. No. 11/588,036 Final Office Action dated Apr. 15, 2014.
U.S. Appl. No. 11/588,036 Office Action dated Jan. 6, 2014.
U.S. Appl. No. 11/588,036 Final Office Action dated Oct. 4, 2012.
U.S. Appl. No. 11/588,036 Office Action dated Apr. 27, 2012.
U.S. Appl. No. 11/588,036 Final Office Action dated Feb. 17, 2011.
U.S. Appl. No. 11/588,036 Office Action dated Sep. 14, 2010.
U.S. Appl. No. 11/586,958 Final Office Action dated Jan. 18, 2017.
U.S. Appl. No. 11/586,958 Office Action dated Jun. 23, 2016.
U.S. Appl. No. 11/586,958 Final Office Action dated Aug. 4, 2015.
U.S. Appl. No. 11/586,958 Office Action dated Jan. 14, 2015.
U.S. Appl. No. 11/586,958 Final Office Action dated Mar. 12, 2014.
U.S. Appl. No. 11/586,958 Office Action dated Nov. 6, 2013.
U.S. Appl. No. 11/586,958 Final Office Action dated Feb. 14, 2011.
U.S. Appl. No. 11/586,958 Office Action dated Sep. 30, 2010.
U.S. Appl. No. 11/586,989 Final Office Action dated Dec. 9, 2010.
U.S. Appl. No. 11/586,989 Office Action dated May 11, 2010.
U.S. Appl. No. 11/586,989 Office Action dated Mar. 30, 2009.
U.S. Appl. No. 14/091,327 Office Action dated Mar. 12, 2015.
U.S. Appl. No. 15/180,615 Final Office Action dated Nov. 1, 2018.
U.S. Appl. No. 15/180,615 Office Action dated Apr. 20, 2018.
U.S. Appl. No. 15/180,615 Final Office Action dated May 19, 2017.
U.S. Appl. No. 15/180,615 Office Action dated Nov. 2, 2016.
U.S. Appl. No. 11/586,959 Final Office Action dated Jan. 29, 2016.
U.S. Appl. No. 11/586,959 Office Action dated Jul. 9, 2015.
U.S. Appl. No. 11/586,959 Final Office Action dated Dec. 8, 2014.
U.S. Appl. No. 11/586,959 Office Action dated Jul. 31, 2014.
U.S. Appl. No. 11/586,959 Office Action dated Feb. 12, 2014.
U.S. Appl. No. 11/586,959 Final Office Action dated Aug. 30, 2013.
U.S. Appl. No. 11/586,959 Office Action dated May 8, 2013.
U.S. Appl. No. 11/586,959 Final Office Action dated Oct. 5, 2012.
U.S. Appl. No. 11/586,959 Office Action dated Apr. 27, 2012.
U.S. Appl. No. 11/586,959 Final Office Action dated Feb. 14, 2011.
U.S. Appl. No. 11/586,959 Office Action dated Oct. 1, 2010.
U.S. Appl. No. 12/370,531 Office Action dated Aug. 1, 2013.
U.S. Appl. No. 12/370,531 Final Office Action dated Aug. 3, 2011.
U.S. Appl. No. 12/370,531 Office Action dated Nov. 16, 2011.
U.S. Appl. No. 12/370,531 Final Office Action dated Aug. 1, 2011.
U.S. Appl. No. 12/370,531 Office Action dated Feb. 2, 2011.
U.S. Appl. No. 14/315,694 Office Action dated Mar. 25, 2016.
U.S. Appl. No. 14/315,694 Final Office Action dated Oct. 27, 2015.
U.S. Appl. No. 14/315,694 Office Action dated Apr. 10, 2015.
U.S. Appl. No. 15/385,688 Office Action dated Sep. 4, 2020.
U.S. Appl. No. 15/385,688 Final Office Action dated Feb. 24, 2020.
U.S. Appl. No. 15/385,688 Office Action dated Jun. 13, 2019.
U.S. Appl. No. 15/385,688 Final Office Action dated Dec. 14, 2018.
U.S. Appl. No. 15/385,688 Office Action dated Jun. 5, 2018.
U.S. Appl. No. 15/385,688 Final Office Action dated Jan. 18, 2018.
U.S. Appl. No. 11/588,236 Office Action dated Sep. 9, 2009.
U.S. Appl. No. 11/588,236 Office Action dated Mar. 5, 2009.
U.S. Appl. No. 12/703,188 Office Action dated Apr. 5, 2017.
U.S. Appl. No. 12/703,188 Final Office Action dated Sep. 7, 2016.
U.S. Appl. No. 12/703,188 Office Action dated Apr. 1, 2016.
U.S. Appl. No. 12/703,188 Office Action dated Apr. 1, 2015.
U.S. Appl. No. 12/703,188 Final Office Action dated Jul. 14, 2014.
U.S. Appl. No. 12/703,188 Office Action dated Nov. 21, 2013.
U.S. Appl. No. 12/703,188 Final Office Action dated Oct. 12, 2012.
U.S. Appl. No. 12/703,188 Office Action dated Apr. 6, 2012.
U.S. Appl. No. 15/866,308 Office Action dated Oct. 28, 2020.
U.S. Appl. No. 15/866,308 Final Office Action dated Feb. 13, 2020.
U.S. Appl. No. 15/866,308 Office Action dated Sep. 17, 2019.
U.S. Appl. No. 10/924,009 Supplemental Final Office Action dated Feb. 4, 2009.
U.S. Appl. No. 10/924,009 Final Office Action dated Dec. 5, 2008.
U.S. Appl. No. 10/924,009 Office Action dated Jun. 30, 2008.
U.S. Appl. No. 12/717,108 Final Office Action dated Jan. 31, 2012.
U.S. Appl. No. 12/717,108 Final Office Action dated Jul. 20, 2011.
U.S. Appl. No. 12/717,108 Office Action dated Feb. 9, 2011.
U.S. Appl. No. 14/242,664 Office Action dated Feb. 29, 2016.
U.S. Appl. No. 14/242,664 Office Action dated Aug. 31, 2015.
U.S. Appl. No. 15/391,522 Office Action dated Nov. 27, 2017.
U.S. Appl. No. 15/866,308 Final Office Action dated Mar. 10, 2021.
Related Publications (1)
Number Date Country
20210004863 A1 Jan 2021 US
Continuations (6)
Number Date Country
Parent 16594354 Oct 2019 US
Child 17028338 US
Parent 15992014 May 2018 US
Child 16594354 US
Parent 15285928 Oct 2016 US
Child 15992014 US
Parent 14336452 Jul 2014 US
Child 15285928 US
Parent 13939178 Jul 2013 US
Child 14336452 US
Parent 11241229 Sep 2005 US
Child 13939178 US