Method and system for automatic insertion of interactive TV triggers into a broadcast data stream

Information

  • Patent Grant
  • 10609451
  • Patent Number
    10,609,451
  • Date Filed
    Wednesday, September 6, 2017
    7 years ago
  • Date Issued
    Tuesday, March 31, 2020
    4 years ago
Abstract
A method and system are disclosed for automatic insertion of interactive television (TV) triggers into a broadcast data stream. The automatic insertion of the Interactive TV trigger can be based on recognition of one or more elements within the broadcast data stream. In particular, a recognized element can trigger the insertion of the interactive TV trigger into the broadcast data stream. Furthermore, the interactive TV trigger can be pre-inserted into the stored broadcast content having the recognized element prior to being broadcast.
Description
FIELD OF THE INVENTION

The present invention pertains to the field of interactive television. More particularly, the present invention relates to a method and system for automatic insertion of interactive television (TV) triggers into a broadcast data stream.


BACKGROUND OF THE INVENTION

Today, advancements in television systems provide a wide variety of services and application for television users. One such advancement is interactive television (“iTV”). In an iTV environment, a TV user can interact with a broadcast or service being provided on the TV. For example, a broadcast may include enhanced, interactive content (“interactive content”) such as a Universal Resource Locator (URL) address in which a TV user can select to access a website on the Internet or World Wide Web at the selected URL address.


In current iTV systems, if a broadcaster wants to broadcast programs with interactive content, the broadcaster must manually modify each program to add the interactive content prior to broadcasting. As such, a disadvantage of using current iTV is the substantial effort required to add interactive content to broadcast programs. Another disadvantage of using current iTV systems is the unwanted delay in deploying interactive content with broadcast programs.


SUMMARY OF THE INVENTION

A method and system are disclosed for automatic insertion of interactive television (TV) triggers into a broadcast data stream. In one embodiment, the automatic insertion of the Interactive TV trigger is based on recognition of one or more elements within the broadcast data stream. In particular, a recognized element can trigger the insertion of the interactive TV trigger into the broadcast data stream. In another embodiment, the interactive TV trigger is pre-inserted into the stored broadcast content having the recognized element prior to being broadcast.


Other features of the present invention will be apparent from the accompanying drawings and from the detailed description which follows.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention is illustrated by way of example, and not limitation, in the figures of the accompanying drawings, in which like references indicate similar elements and in which:



FIG. 1A illustrates an exemplary interactive television system having an insertion platform architecture in which the present invention can be practiced;



FIG. 1B illustrates an exemplary screen shot of a TV broadcast including an interactive TV trigger;



FIG. 2 illustrates a simplified block diagram of the set-top box of FIG. 1A having an insertion module according to one embodiment;



FIG. 3 illustrates a block diagram of the insertion platform architecture of FIG. 1A according to one embodiment;



FIG. 4 illustrates an internal block diagram of an insertion server of FIG. 3 according to one embodiment;



FIG. 5 illustrates a flow diagram of an operation for inserting content into a broadcast data stream according to one embodiment;



FIG. 6A illustrates a broadcast data stream having a media asset according to one embodiment;



FIG. 6B illustrates a look-up table for determining if a broadcast stream is to be inserted with an interactive element according to one embodiment; and



FIG. 7 illustrates a flow diagram of an operation to insert an interactive element into a broadcast data stream by checking a media asset in the broadcast data stream with attributes of the interactive element.





DETAILED DESCRIPTION

A method and system are described for automatic insertion of interactive television (TV) triggers into a broadcast data stream. In one embodiment, the automatic insertion of the Interactive TV trigger is based on recognition of one or more elements within the broadcast data stream. In particular, a recognized element can trigger the insertion of the interactive TV trigger into the broadcast data stream. In another embodiment, the interactive TV trigger is pre-inserted into the stored broadcast content having the recognized element prior to being broadcast.


The following embodiments allow interactive content (e.g., an interactive TV trigger) to be automatically inserted into a broadcast data stream avoiding manual modification of the broadcast data stream. The following embodiments allow for easy and rapid deployment and insertion of interactive content into a broadcast data stream.


Furthermore, in the following embodiments, an interactive TV trigger can refer to any number of types of interactive content supported by a number of interactive content standards. For example, an interactive TV trigger can be based on an enhanced, interactive content standard such as Advanced Television Enhancement Forum (ATVEF) standard for Transport Type A or Transport Type B, Internet Protocol (IP) Multicasting standard, Multi Protocol Encapsulation for Digital Video Broadcasting (DVB), or Broadcast Hypertext Markup Language (HTML) standards being developed for Digital Television (DTV).



FIG. 1A illustrates an exemplary interactive television system 100 having insertion platform architecture 110 in which the present invention can be practiced. Referring to FIG. 1A, interactive television system 100 includes a set-top box 106 connected to a TV 104. Set-top box 106 and TV 104 can receive inputs from a remote controller 122. TV 104 is shown to have a network 102 connection capabilities. That is, TV 104 can be connected to network 102 via set-top box 106. In one embodiment, TV 104 can receive and display TV broadcast 108 with interactive TV trigger 112. Interactive TV trigger 112 can be used to retrieve information from web server 118. For example, a user can select or access interactive TV trigger 112 to retrieve a web page from web server 118 for display on TV 104. Set-top box 106 can also be connected to network 102 through insertion platform architecture 110 or directly through an input/output (I/O) interface such as a telephone line.


Set-top box 106 is a receiver for TV 104. Set-top box 106 is compliant for enhanced, interactive television content. For example, set-top box 106 can be compliant for the Advanced Television Enhancement Forum (ATVEF) standard for Transport Type A or Transport Type B delivery of enhanced, interactive television content. ATVEF defines a standard to create and deliver enhanced content for a number of mediums including analog (NTSC) or digital (ATSC) mediums from a number of sources such as, for example, terrestrial, cable, or satellite sources. TV 104 is a display device. TV 104 can support analog, Digital Video Broadcasting (DVB), Advanced Television Systems Committee (ATSC) or any of the other known TV standards.


Set-top box receives input signals 120 (e.g., television signals) to display on TV 104. Input signals 120 can be broadcast signals from a plurality of programming sources. For example, set-top box 106 can receive broadcast signals as input signals 120 from a terrestrial, cable, or satellite source. Input signals 120 can be analog or digital signals. Set-top box 106 can provide on-screen displays (OSDs) or graphical user interfaces (GUIs) to facilitate interactive services, e.g., accessing content on the Internet. As shown in FIG. 1A, set-top box 106 can be controlled by a user of remote controller 122. Alternatively, set-top box 106 can be controlled by other types of input devices such as, for example, an infrared (IR) keyboard.


Insertion platform architecture 110 can be a computing system having multiple sub-systems for performing the content insertion techniques described herein. Alternatively, insertion platform architecture 110 can include hardware and/or software modules operating within set-top box 106 to perform the content insertion techniques described herein. For purposes of illustration, insertion platform architecture 110 is an external system to set-top box 106.


Insertion platform architecture 110 provides input for set-top box 106. In one embodiment, insertion platform architecture 110 provides set-top box 106 with a TV broadcast along 108 with interactive TV trigger 112 as input for input signals 120. As will be described in further detail below, insertion platform architecture automatically inserts interactive TV trigger 112 into TV broadcast 108. In one embodiment, insertion platform architecture 110 automatically inserts interactive TV trigger 112 into TV broadcast 108 based on the elements contained within TV broadcast 108. For instance, insertion platform architecture 110 can associate elements within TV broadcast 108 with a specific interactive element (e.g., interactive TV trigger 112), which is to be inserted into TV broadcast 108. In an alternative embodiment, insertion platform architecture 110 pre-inserts the interactive TV trigger 112 into TV broadcast having recognized elements.


Insertion platform architecture 110 can include on or more databases to store attributes that are used to match with elements within TV broadcast 108. In another embodiment, insertion platform architecture 110 sends interactive TV trigger 112 to insertion module 107 within set-top box 106, and insertion module 107 automatically inserts interactive TV trigger 112 into TV broadcast 108 received by set-top box 106.


Network 102 can represent a network such as the Internet hosting the World Wide Web (WWW). The WWW allows for a uniform way of accessing information on the Internet using HTML compliant browsers. Network 102 can be other type of networks such as, for example, a local area network (LAN) or a wide area network (WAN). Network 102 can also represent wired or wireless networks. Although one web server 102 is shown in FIG. 1A, any number of web servers can be connected to network 102. Furthermore, other types of network devices can also be connected to network 102, which can provide content for TV 104, such as, for example, a network router, bridge, gateway, or other like network devices.


Remote controller 122 is a control device for a user to provide inputs (e.g., infrared (IR) or radio frequency (RF) signals to set-top box 106 and/or TV 104. Remote controller 122 can include alphanumeric keys, options keys, functions keys, and other like keys to operate set-top box 106 or TV 104. In one embodiment, a user can interact with interactive TV trigger 112 using remote controller 122. In particular, a user can access and navigate through a browser (not shown) operating on TV 104 by pressing selectively certain buttons or keys on remote controller 122. Interactive television system 100 can also be implemented in numerous configurations. For example, TV 104 can have the functionality of set-top box 106 contained internally. In addition, TV 104 can also be a computing device, which can display television signals.



FIG. 1B illustrates an exemplary screen shot 150 of TV broadcast 108 including interactive TV trigger 112. Interactive TV trigger 112 can be automatically inserted into TV broadcast 112 by insertion platform architecture 110. Alternatively, set-top box 106 can automatically insert interactive TV trigger 112 into TV broadcast 108. Referring to FIG. 1B, TV 104 displays TV broadcast 108 with interactive TV trigger 112 as Universal Resource Location (URL) content (“http://www.xyz.com”). The URL location is exemplary for a website location on the Internet related to TV broadcast 108. For example, TV broadcast 108 may be a clothing store commercial and the exemplary URL location may the location of a website owned and operated by the clothing store.


In one embodiment, interactive TV trigger 112 is an ATVEF trigger. An ATVEF trigger is a data mechanism to alert receivers (e.g., set-top box 106) of incoming content enhancements. In particular, ATVEF triggers include information about enhancements that are available to the user. For example, interactive TV trigger 112 can be an ATVEF trigger that includes a URL location as shown in FIG. 1B. Interactive TV trigger 112 can be broadcasted to set-top box 106 with TV broadcast 108. In one embodiment, interactive TV trigger 112 is enhanced content, which may be already stored within set-top box 106. In another embodiment, interactive TV trigger 112 may include readable description of content such as, for example, “press the browse button for more information about the product being advertised,” which can be displayed on TV 104 by set-top box 106. Interactive TV trigger 112 can also include JavaScript code. For example, interactive TV trigger 112 can be used to execute another piece of JavaScript code within a web page on the Internet.



FIG. 2 illustrates a simplified block diagram of set-top box 106 of FIG. 1A having insertion module 107 according to one embodiment. For purposes of clarity, only the basic components of set-top box 106 are shown in block diagram form. Set-top box 106 is configured to provide enhanced, interactive television content services.


Referring to FIG. 2, set-top box 106 includes a central processing unit (CPU) 234 coupled to memory devices 238, input/output (I/O) interfaces 236, decoder 232, and insertion module 107. Decoder 202 can receive inputs signals 120. In one embodiment, one of the input signals 120 is TV broadcast 108 with interactive TV trigger 112. Interactive TV trigger 112 can be enhanced, interactive television content. Decoder 232 can receive input signals 120 as analog (NTSC) or digital (ATSC) signals from a number sources including terrestrial, cable, and satellite sources. Decoder 232 decodes and outputs a TV signal to TV 104. The TV signal can include TV broadcast 108 and/or interactive TV trigger 112.


CPU 234 is the central control mechanism for set-top box 106. CPU 234 can execute code or instructions stored in memory devices 238 or stored in insertion module 107 or external storage devices via network 102. For example, I/O interfaces 236 may include a connection to network 102 provided by a dial-up modem. I/O interfaces 236 can also include an (IR) or (RF) interface for receiving inputs from remote controller 122. CPU 234 can also execute code or instructions to authorize decoder 232 to output automatically interactive TV trigger 112 with TV broadcast 108 from insertion platform architecture 110 on TV 104. In one embodiment, CPU 234 can execute software modules within insertion module 107 to insert automatically interactive TV trigger 112 into TV broadcast 108 within set-top box 106 for display on TV 104.


Insertion module 107 includes hardware and/or software modules to process interactive TV triggers from insertion platform architecture 110. Insertion module 107 can automatically insert interactive TV trigger 112 into TV broadcast 108 in which interactive TV trigger 112 was received from insertion platform architecture 110. In one embodiment, insertion module 107 can process a pre-stored interactive TV trigger 112 within set-top box 106 and automatically insert interactive TV trigger 112 into TV broadcast 108. Insertion module 107 can also automatically insert interactive TV trigger 112 into TV broadcast based on a specific pattern within TV broadcast 108. That is, if a specific pattern is recognized, insertion module 107 can insert interactive TV trigger 112 into TV broadcast 108.


Memory devices 238 can include a machine-readable medium that provides (i.e., stores and/or transmits) information in a form readable, e.g., by CPU 234. Memory devices 208 may include a read only memory (ROM), random access memory (RAM), magnetic disk storage media, optical storage media, or flash memory devices. The code or instructions stored in memory devices 238 can be represented by carrier wave signals, infrared signals, digital signals, and by other like signals.



FIG. 3 illustrates a block diagram of insertion platform architecture 110 of FIG. 1A according to one embodiment. Insertion platform architecture 110 provides details of the subsystems and networks to insert automatically interactive TV triggers into a TV broadcast. The subsystems can include hardware and/or software modules to insert content as described herein. For purposes of illustration, insertion platform architecture 110 is shown communicating with a single set-top box 106. Insertion platform architecture 110, however, can be implemented to communicate with a plurality of set-top boxes.


Referring to FIG. 3, set-top box 106 can receive a TV broadcast from a plurality of sources such a satellite source 324, wireless Multipoint Microwave Distribution System (MMDS) source 323, x digital subscriber loop (DSL) source 322, cable modem source 321, and a video server and tape machine source 316 via broadcast network 312. Insertion platform architecture 110 includes insertion server 303A, audio/video (A/V) decoder 313, repository 304A, properties, rules & logic 304B, pattern engine 305, insertion server 303B, A/V decoder 315, existing content file server 316A, existing content databases 316B, and multiplexer/encoder 314.


Two insertion servers 303A and 303B can be used to insert automatically interactive TV trigger 112 into TV broadcast 108. Insertion servers 303A and 303B can be a general purpose computing system, workstation, or client server. In one embodiment, insertion server 303A provides automatic dynamic insertion of interactive content into a broadcast data stream from audio/video (A/V) decoder 313 based on a fine personalization that is performed at the set-top box 106. Thus, in one embodiment, insertion server 303A can be considered to operate within the context or environment of set-top box 106, having access to viewer or profile information therein. In one embodiment, insertion server 303B provides automatic pre-insertion of interactive content into existent content stored in existing content databases 316B, which is controlled by existing content file server 316A. In another embodiment, insertion server 303B provides automatic dynamic insertion of interactive content into a broadcast data stream provided by multiplexer/encoder 314 based on coarse or fine personalization. The insertion and personalization process for insertion servers 303A and 303B will be discussed in further detail with respect to FIG. 4.


Insertion servers 303A and 303B interface with A/V decoder 313, A/V decoder 315 and multiplexer/encoder 314, respectively. In one embodiment, multiplexer/encoder 314 can provide ATVEF, VBI Encoding, MPEG Multiplexing, or IP encapsulation services to insert interactive content (e.g., interactive TV trigger 112) provided insertion server 303B into a broadcast data stream (e.g., TV broadcast 108). In one embodiment, the formatting and structure of the content is based on ATVEF broadcast and delivery standards. Insertion server 303A and 303B manage the delivery and insertion of interactive content into broadcast data streams content, triggers and media to the TV & data broadcast (“TV broadcast”).


Repository 304A is a facility to access database 304B storing meta-data, properties, rules & logic and pattern engine 305 to drive the interactive content insertion process for insertion servers 303A and 303B. Repository 304A can be a general purpose computing system, workstation, or client server. Repository communicates and provides information and data to insertion server 303A and insertion server 303B. In one embodiment, repository 304 provides interactive keys, attributes and interactive content to insertion servers 303A and 303B. The keys, attributes and interactive content can be stored in database 304B.


In one embodiment, database 304B stores interactive elements and attributes and a dictionary to associate interactive elements with pre-determined media assets in a broadcast data stream based on attributes of the interactive elements. That is, if a media asset is found in a broadcast data stream, the associated interactive element can be inserted into the broadcast data stream. Such a process will be explained in the following embodiments. Pattern engine 305 is a hardware and/or software module used to identify, locate or find the pattern or type of asset in filtering a broadcast data stream. Alternatively, pattern engine 305 can be a separate server similar to repository 304A. For example, pattern engine 305 can supply instructions to insertion servers 303A and 303B to insert interactive TV trigger 112 into TV broadcast 108 if it contains a media asset having a pattern “A” based on the meta-data, properties and rules stored in database 304B.



FIG. 4 illustrates an internal block diagram 400 an insertion server of FIG. 3 according to one embodiment. Referring to FIG. 4, block diagram 400 can represent internal components for both insertion server 303A and 303B of FIG. 3, which includes an interactivity engine 402 coupled with an audio/video pattern recognition filters 411 and insertion engine 401.


Audio/video pattern recognition filters 411 are hardware and/or software modules used to recognize specific media assets with a broadcast data stream. For example, audio/video pattern recognition filters 411 can filter broadcast data stream 600 for a specific media asset 602 such as that shown in FIG. 6A. Media asset 602 can have a specific pattern, e.g., a pattern “A.” Audio/video pattern recognition filters 411 can receive instructions to recognize pattern “A” for media asset 602 (e.g., pattern “A”) within broadcast data stream 600. If the pattern is found, audio/video pattern recognition filters 411 informs interactivity engine 402 that the pattern was found.


Interactivity engine 402 includes a look-up table 405 having a plurality of entries 620-1 through 620-N such as that shown in FIG. 6B. Interactivity engine 402 will associate the found pattern with one of a plurality of interactive elements 630-1 through 630-N within the entries by matching using the corresponding attributes 625-1 through 625-N. Such attributes can include broadcast information type (e.g. classical, rock, history, art, etc.), names of producers of the program, names of cast members, channel number, names of broadcast advertisers or sponsors, or other types of personalization information. Media asset 602 can thus include information (“pattern”) having the same type of information or data.


In one embodiment, interactivity engine 402 will inform insertion engine 401 to insert the matching interactive element with media asset 602 into broadcast data stream 600. Insertion engine 401 can retrieve the interactive element (e.g., interactive TV trigger 112) from repository 304A or directly from interactivity engine 402. Insertion engine 401 can then insert the interactive element into broadcast data stream 600. In one embodiment, the interactive element is an ATVEF trigger. Insertion engine 401 can be used to insert such an interactive element at several points of delivery of a broadcast data stream. For example, insertion engine 401 can be used to insert interactive elements into existing stored media (e.g. for legacy broadcast assets), to insert interactive elements at the head-end prior to delivery, or to insert interactive elements at the client side (e.g., set-top box 106) by using insertion module 107 within set-top box 106 to assist in decoding the interactive element with the broadcast data stream.



FIG. 5 illustrates a flow diagram of an operation 500 for inserting content into a broadcast data stream according to one embodiment. Initially, operation 500 begins at operation 502.


At operation 502, one or more elements in a broadcast data stream are recognized. For example, a media asset 602 within broadcast data stream 600 can be recognized.


At operation 504, an interactive TV trigger is automatically inserted into the broadcast data stream based on the recognized elements. For example, interactive TV trigger 112 can be inserted into broadcast data stream 600 by one of the insertion servers 303A or 303B.



FIG. 7 illustrates a flow diagram of an operation 700 to insert an interactive element into a broadcast data stream by checking a media asset in the broadcast data stream with attributes of the interactive element. Initially, operation 700 begins at operation 702.


At operation 702, a media asset in a broadcast stream is recognized. For example, audio/video pattern recognition filters 411 is instructed to recognize media asset 602 within broadcast data stream 600.


At operation 704, a check is made if the media asset matches with an interactive element. For example, interactivity engine uses look-up table 405 to check if media asset 602 has a matching interactive element within look-up table 405.


At operation 706, if the media asset does not match an interactive element, the broadcast data stream is passed on to a receiver. For example, insertion server 303A or insertion 303B passes broadcast data stream 600 to set-top box 106.


At operation 708, if the media asset does match an interactive element, the matched interactive element is automatically inserted into the broadcast data stream. For example, insertion engine 401 within an insertion server will insert an interactive element, e.g., interactive TV trigger 112 into broadcast data stream 600.


Thus, a method and system for automatic insertion of interactive TV triggers into a broadcast data stream have been described. Although the present invention has been described with reference to specific exemplary embodiments, it will be evident that various modifications and changes may be made to these embodiments without departing from the broader spirit and scope of the invention as set forth in the claims. Accordingly, the specification and drawings are to be regarded in an illustrative sense rather than a restrictive sense.

Claims
  • 1. A method comprising: receiving, by a client device from a server, media content comprising a plurality of patterns and comprising a first interactive trigger associated with a first pattern of the plurality of patterns, wherein the plurality of patterns comprises at least one of an audio pattern or a video pattern;inserting, into the received media content, a second interactive trigger associated with a second pattern of the plurality of patterns; andcausing, after the inserting, the received media content to be output via a display.
  • 2. The method of claim 1, wherein at least one of the first interactive trigger or the second interactive trigger comprises an Advanced Television Enhancement Forum (ATVEF) trigger.
  • 3. The method of claim 1, wherein the inserting of the second interactive trigger is based on profile information stored in the client device.
  • 4. The method of claim 1, wherein the client device comprises a set top box.
  • 5. The method of claim 1, further comprising: determining, in the received media content, the second pattern of the plurality of patterns, wherein the inserting of the second interactive trigger into the received media content comprises automatically inserting the second interactive trigger into the received media content based on the determining of the second pattern in the media content.
  • 6. The method of claim 1, further comprising: identifying and locating the second pattern in the received media content.
  • 7. An apparatus comprising: a processor and memory storing instructions that, when executed by the processor, cause the apparatus to: receive, from a server, media content comprising a plurality of patterns and comprising a first interactive trigger associated with a first pattern of the plurality of patterns, wherein the plurality of patterns comprises at least one of an audio pattern or a video pattern;insert, into the media content, a second interactive trigger associated with a second pattern of the plurality of patterns; andcause, after the second interactive trigger is inserted into the media content, the media content to be output via a display.
  • 8. The apparatus of claim 7, wherein at least one of the first interactive trigger or the second interactive trigger comprises an Advanced Television Enhancement Forum (ATVEF) trigger.
  • 9. The apparatus of claim 7, wherein the instructions, when executed by the processor, cause the apparatus to insert the second interactive trigger into the media content based on profile information stored in the apparatus.
  • 10. The apparatus of claim 7, wherein the apparatus comprises a set top box.
  • 11. The apparatus of claim 7, wherein the instructions, when executed by the processor, further cause the apparatus to: determine, in the media content, the second pattern of the plurality of patterns; andinsert the second interactive trigger into the media content automatically based on the determining of the second pattern in the media content.
  • 12. The apparatus of claim 7, wherein the instructions, when executed by the processor, further cause the apparatus to identify and locate the second pattern in the media content.
  • 13. A system comprising: a client computer; anda server computer configured to:insert, into media content, a first interactive trigger associated with a first pattern of a plurality of patterns, andsend the media content including the first interactive trigger to the client computer, wherein the plurality of patterns comprises at least one of an audio pattern or a video pattern; andwherein the client computer comprises a processor and memory storing instructions that, when executed by the processor, cause the client computer to:receive the media content sent by the server computer,insert a second interactive trigger associated with a second pattern of the plurality of patterns into the media content, andcause, after the second interactive trigger is inserted into the media content, the media content to be output via a display.
  • 14. The system of claim 13, wherein the server computer is further configured to: determine the first pattern in the media content, andinsert the first interactive trigger into the media content automatically based on the determining of the first pattern in the media content.
  • 15. The system of claim 13, wherein the instructions, when executed by the processor, further cause the client computer to: determine the second pattern in the media content, andinsert the second interactive trigger into the media content automatically based on the determining of the second pattern in the media content.
  • 16. The system of claim 13, wherein the instructions, when executed by the processor, further cause the client computer to insert the second interactive trigger into the media content based on profile information stored in the client computer.
  • 17. The system of claim 13, further comprising a second client computer, wherein the server computer is further configured to send the media content including the first interactive trigger to the second client computer.
  • 18. The system of claim 17, wherein the second client computer comprises a second processor and second memory storing second instructions that, when executed by the second processor, cause the second client computer to: receive the media content sent by the server computer;insert, into the media content, a third interactive trigger associated with a third pattern of the plurality of patterns, andcause, after the third interactive trigger is inserted into the media content, the media content to be output to a second display.
  • 19. The system of claim 18, wherein the second instructions, when executed by the second processor, further cause the second client computer to insert the third interactive trigger into the media content based on second profile information stored in the second client computer.
  • 20. The system of claim 13, wherein the instructions, when executed by the processor, further cause the client computer to identify and locate the second pattern in the media content.
RELATED APPLICATION

This application is related to and claims priority to U.S. Provisional Application No. 60/199,686, entitled, “METHOD AND SYSTEM FOR TRANSFORMING CONTENT FOR EXECUTION ON MULTIPLE PLATFORMS,” filed on Apr. 24, 2000, the disclosure of which is hereby incorporated herein by reference.

US Referenced Citations (268)
Number Name Date Kind
4647974 Butler et al. Mar 1987 A
5469206 Strubbe et al. Nov 1995 A
5485221 Banker et al. Jan 1996 A
5563648 Menand et al. Oct 1996 A
5579057 Banker et al. Nov 1996 A
5592551 Lett et al. Jan 1997 A
5600364 Hendricks et al. Feb 1997 A
5631694 Aggarwal et al. May 1997 A
5671377 Bleidt et al. Sep 1997 A
5689799 Dougherty et al. Nov 1997 A
5706434 Kremen et al. Jan 1998 A
5708845 Wistendahl et al. Jan 1998 A
5727159 Kikinis Mar 1998 A
5742680 Wilson Apr 1998 A
5774660 Brendel et al. Jun 1998 A
5774664 Hidary et al. Jun 1998 A
5774666 Portuesi Jun 1998 A
5809242 Shaw et al. Sep 1998 A
5818438 Howe et al. Oct 1998 A
5819269 Uomini Oct 1998 A
5826258 Gupta et al. Oct 1998 A
5848396 Gerace Dec 1998 A
5872588 Aras et al. Feb 1999 A
5881284 Kubo Mar 1999 A
5895477 Orr et al. Apr 1999 A
5907322 Kelly et al. May 1999 A
5925595 Seitz et al. Jul 1999 A
5929849 Kikinis Jul 1999 A
5937331 Kalluri et al. Aug 1999 A
5961603 Kunkel et al. Oct 1999 A
5969715 Dougherty et al. Oct 1999 A
5982445 Eyer et al. Nov 1999 A
5991306 Burns et al. Nov 1999 A
6002394 Schein et al. Dec 1999 A
6006256 Zdepski et al. Dec 1999 A
6008836 Bruck et al. Dec 1999 A
6014134 Bell et al. Jan 2000 A
6014502 Moraes Jan 2000 A
6018764 Field et al. Jan 2000 A
6018768 Ullman et al. Jan 2000 A
6025837 Matthews, III et al. Feb 2000 A
6029045 Picco et al. Feb 2000 A
6049333 LaJoie et al. Apr 2000 A
6055566 Kikinis Apr 2000 A
6061056 Menard et al. May 2000 A
6067107 Travaille et al. May 2000 A
6076734 Dougherty et al. Jun 2000 A
6101510 Stone et al. Aug 2000 A
6122658 Chaddha Sep 2000 A
6169543 Wehmeyer Jan 2001 B1
6177931 Alexander et al. Jan 2001 B1
6182072 Leak et al. Jan 2001 B1
6188398 Collins-Rector et al. Feb 2001 B1
6205582 Hoarty Mar 2001 B1
6209130 Rector, Jr. et al. Mar 2001 B1
6219042 Anderson et al. Apr 2001 B1
6230319 Britt, Jr. et al. May 2001 B1
6233546 Datig May 2001 B1
6233592 Schnelle et al. May 2001 B1
6237022 Bruck et al. May 2001 B1
6240555 Shoff et al. May 2001 B1
6253238 Lauder et al. Jun 2001 B1
6263501 Schein et al. Jul 2001 B1
6275989 Broadwin et al. Aug 2001 B1
6295058 Hsu et al. Sep 2001 B1
6297853 Sharir et al. Oct 2001 B1
6301584 Ranger Oct 2001 B1
6308206 Singh Oct 2001 B1
6327574 Kramer et al. Dec 2001 B1
6338094 Scott et al. Jan 2002 B1
6345292 Daugherty et al. Feb 2002 B1
6357042 Srinivasan et al. Mar 2002 B2
6363525 Dougherty et al. Mar 2002 B1
6373609 Mizrahi Apr 2002 B1
6415303 Meier et al. Jul 2002 B1
6424717 Pinder et al. Jul 2002 B1
6430624 Jamtgaard et al. Aug 2002 B1
6434747 Khoo et al. Aug 2002 B1
6446261 Rosser Sep 2002 B1
6457103 Challenger et al. Sep 2002 B1
6459427 Mao et al. Oct 2002 B1
6460180 Park et al. Oct 2002 B1
6471392 Holl et al. Oct 2002 B1
6473609 Schwartz et al. Oct 2002 B1
6473858 Shimomura et al. Oct 2002 B1
6496981 Wistendahl et al. Dec 2002 B1
6510557 Thrift Jan 2003 B1
6513019 Lewis Jan 2003 B2
6513160 Dureau Jan 2003 B2
6515349 Terashima Feb 2003 B2
6516349 Lieberman Feb 2003 B1
6518950 Dougherty et al. Feb 2003 B1
6530082 Del Sesto et al. Mar 2003 B1
6536041 Knudson et al. Mar 2003 B1
6542593 Bowman-Amuah Apr 2003 B1
6546488 Dillon et al. Apr 2003 B2
6547754 Evans et al. Apr 2003 B1
6553409 Zhang et al. Apr 2003 B1
6564251 Katariya et al. May 2003 B2
6564263 Bergman et al. May 2003 B1
6571392 Zigmond et al. May 2003 B1
6574793 Ngo et al. Jun 2003 B1
6574795 Carr Jun 2003 B1
6577346 Perlman Jun 2003 B1
6584153 Gordon et al. Jun 2003 B1
6601103 Goldschmidt Iki et al. Jul 2003 B1
6614914 Rhoads et al. Sep 2003 B1
6615408 Kaiser et al. Sep 2003 B1
6637032 Feinleib Oct 2003 B1
6640145 Hoffberg et al. Oct 2003 B2
6643652 Helgeson et al. Nov 2003 B2
6651218 Adler et al. Nov 2003 B1
6654754 Knauft et al. Nov 2003 B1
6668378 Leak et al. Dec 2003 B2
6675350 Abrams et al. Jan 2004 B1
6676350 McSherry et al. Jan 2004 B1
6681393 Bauminger et al. Jan 2004 B1
6698020 Zigmond et al. Feb 2004 B1
6698023 Levitan Feb 2004 B2
6718551 Swix et al. Apr 2004 B1
6725461 Dougherty et al. Apr 2004 B1
6730624 Motoki et al. May 2004 B2
6735487 Marshall et al. May 2004 B1
6742043 Moussa et al. May 2004 B1
6745368 Boucher et al. Jun 2004 B1
6750919 Rosser Jun 2004 B1
6756997 Ward, III et al. Jun 2004 B1
6757708 Craig et al. Jun 2004 B1
6779154 Nussbaum et al. Aug 2004 B1
6785902 Zigmond et al. Aug 2004 B1
6792616 Jerding et al. Sep 2004 B1
6799326 Boylan, III et al. Sep 2004 B2
6799327 Reynolds et al. Sep 2004 B1
6810526 Menard et al. Oct 2004 B1
6826597 Lonnroth et al. Nov 2004 B1
6839903 Shintani et al. Jan 2005 B1
6847969 Mathai et al. Jan 2005 B1
6925595 Whitledge et al. Aug 2005 B1
6934963 Reynolds et al. Aug 2005 B1
6938270 Blackketter et al. Aug 2005 B2
6941474 Boies et al. Sep 2005 B2
6941574 Broadwin et al. Sep 2005 B1
7028327 Dougherty et al. Apr 2006 B1
7062777 Alba et al. Jun 2006 B2
7079176 Freeman et al. Jul 2006 B1
7080039 Marsh Jul 2006 B1
7100183 Kunkel et al. Aug 2006 B2
7103605 Hazi et al. Sep 2006 B1
7103904 Blackketter et al. Sep 2006 B1
7143042 Sinai et al. Nov 2006 B1
7143428 Bruck et al. Nov 2006 B1
7168084 Hendricks et al. Jan 2007 B1
7222155 Gebhardt et al. May 2007 B1
7222163 Girouard et al. May 2007 B1
7287018 Lennon Oct 2007 B2
7394816 Bill Jul 2008 B1
7500188 Trapani et al. Mar 2009 B1
7500195 Sahota et al. Mar 2009 B2
7512799 Chow et al. Mar 2009 B1
7530016 Sahota et al. May 2009 B2
7631338 Del Sesto et al. Dec 2009 B2
7634787 Gebhardt et al. Dec 2009 B1
7650376 Blumenau Jan 2010 B1
7653923 Flickinger Jan 2010 B2
7668963 Miner et al. Feb 2010 B1
7702995 Sahota et al. Apr 2010 B2
7783968 Sahota et al. Aug 2010 B2
7930631 Sahota et al. Apr 2011 B2
7941564 Gebhardt et al. May 2011 B2
8032651 Miner et al. Oct 2011 B2
8079045 Krapf et al. Dec 2011 B2
8296792 Sahota et al. Oct 2012 B2
8407744 Moon et al. Mar 2013 B2
8447840 Fong et al. May 2013 B1
8650480 Sahota et al. Feb 2014 B2
8667387 Sahota et al. Mar 2014 B2
8667530 Sahota et al. Mar 2014 B2
8935719 Sargent Jan 2015 B2
8973056 Ellis et al. Mar 2015 B2
20010001160 Shoff et al. May 2001 A1
20010003212 Marler et al. Jun 2001 A1
20010003828 Peterson et al. Jun 2001 A1
20010011265 Cuan et al. Aug 2001 A1
20010023436 Srinivasan Sep 2001 A1
20010027491 Terretta et al. Oct 2001 A1
20010037500 Reynolds et al. Nov 2001 A1
20010047298 Moore et al. Nov 2001 A1
20010047518 Sahota et al. Nov 2001 A1
20010051031 Hada et al. Dec 2001 A1
20010054031 Lee et al. Dec 2001 A1
20010056460 Sahota et al. Dec 2001 A1
20020010928 Sahota Jan 2002 A1
20020011265 Ganachaud Jan 2002 A1
20020046237 Yokokura Apr 2002 A1
20020046273 Lahr et al. Apr 2002 A1
20020049749 Helgeson et al. Apr 2002 A1
20020056087 Berezowski et al. May 2002 A1
20020059644 Andrade et al. May 2002 A1
20020075403 Barone et al. Jun 2002 A1
20020104086 Tomsen et al. Aug 2002 A1
20020112237 Kelts Aug 2002 A1
20020131511 Zenoni Sep 2002 A1
20020136448 Bortolussi et al. Sep 2002 A1
20020138849 Blackketter Sep 2002 A1
20020184627 Alba et al. Dec 2002 A1
20030023971 Martinolich et al. Jan 2003 A1
20030028873 Lemmons Feb 2003 A1
20030131311 McNamara Jul 2003 A1
20030149581 Chaudhri et al. Aug 2003 A1
20040006747 Tyler Jan 2004 A1
20040133910 Gordon et al. Jul 2004 A1
20050071877 Navarro Mar 2005 A1
20050108633 Sahota et al. May 2005 A1
20050108634 Sahota et al. May 2005 A1
20050108637 Sahota et al. May 2005 A1
20050114757 Sahota et al. May 2005 A1
20050132420 Howard et al. Jun 2005 A1
20050149968 Konig et al. Jul 2005 A1
20050193410 Eldering Sep 2005 A1
20050216932 Danker Sep 2005 A1
20060130120 Brandyberry et al. Jun 2006 A1
20060143655 Ellis et al. Jun 2006 A1
20060168624 Carney et al. Jul 2006 A1
20060236342 Kunkel et al. Oct 2006 A1
20070091919 Sandoval Apr 2007 A1
20070136753 Bovenschulte et al. Jun 2007 A1
20070157249 Cordray et al. Jul 2007 A1
20070162961 Tarrance et al. Jul 2007 A1
20080010659 Zenith Jan 2008 A1
20080109843 Ullah May 2008 A1
20080215428 Ramer et al. Sep 2008 A1
20080319852 Gardner et al. Dec 2008 A1
20090119723 Tinsman May 2009 A1
20090193456 Ahn et al. Jul 2009 A1
20090199114 Lewis et al. Aug 2009 A1
20090217329 Riedl et al. Aug 2009 A1
20090254931 Pizzurro et al. Oct 2009 A1
20100161826 Miner et al. Jun 2010 A1
20100171741 Brill et al. Jul 2010 A1
20100174783 Zarom Jul 2010 A1
20100215340 Pettit et al. Aug 2010 A1
20100251279 Piard et al. Sep 2010 A1
20100268773 Hunt et al. Oct 2010 A1
20100333153 Sahota et al. Dec 2010 A1
20110119478 Jackson May 2011 A1
20110191667 Sahota et al. Aug 2011 A1
20120041825 Kasargod et al. Feb 2012 A1
20120084812 Thompson et al. Apr 2012 A1
20120117584 Gordon May 2012 A1
20120189272 Kunigita et al. Jul 2012 A1
20130014192 Sahota et al. Jan 2013 A1
20130051770 Sargent Feb 2013 A1
20130117406 Chevillat et al. May 2013 A1
20130144725 Li et al. Jun 2013 A1
20130173742 Thomas et al. Jul 2013 A1
20130179925 Woods et al. Jul 2013 A1
20130205343 Kirby Aug 2013 A1
20130325603 Shamir et al. Dec 2013 A1
20140053223 Vorobyov et al. Feb 2014 A1
20140130085 Sahota et al. May 2014 A1
20140181855 Fife et al. Jun 2014 A1
20140189500 Sahota et al. Jul 2014 A1
20140201769 Neumeier et al. Jul 2014 A1
20140282653 Ariantaj et al. Sep 2014 A1
20140282735 Davis et al. Sep 2014 A1
20140379855 Sahota et al. Dec 2014 A1
20150156564 Sargent Jun 2015 A1
20150256903 Walker Sep 2015 A1
Foreign Referenced Citations (15)
Number Date Country
60217091 Jun 2007 DE
0915621 May 1999 EP
0942595 Sep 1999 EP
1269750 Jan 2003 EP
1286541 Feb 2003 EP
1381961 Jan 2004 EP
1826981 Aug 2007 EP
1954054 Aug 2008 EP
2000078043 Dec 2000 WO
2001058159 Aug 2001 WO
0182614 Nov 2001 WO
0182613 Nov 2001 WO
2002086746 Oct 2002 WO
03017129 Feb 2003 WO
200506758 Jan 2005 WO
Non-Patent Literature Citations (42)
Entry
“Spyglass Prism: Concepts and Applications”, Spyglass Inc, 1997, pp. 1-8.
Puder, A., et al, “Position Papers: System Support for Knowledge-Based Trading in Open Service Markets”, Proceedings of the 7th Workshop on ACM SIGOPS European Workshop: Systems Support for Worldwide Applications, Sep. 1996, pp. 289-296.
Ensink, Brian, et al, “XML Based Adaptation of the Composite Approach for Database Integration”, Proceedings of the 37th Annual Southeast Regional Conference (CD-ROM), Apr. 1999, pp. 1-6.
Arens, Yigal, et al, “Intelligent Caching: Selecting, Representing, and Reusing Data in an Information Server”, Proceedings of the Third International Conference on Information and Knowledge Management, Nov. 1994, pp. 433-438.
Spencer, Ken, “Using XML to Build Internet Solutions”, Windows IT Pro, http://www.winntmag.com/Articles/Print.cfm? ArticleiD=5056, Apr. 1999.
Bayardo, R J., et al., “InfoSieuth: Agent-Based Semantic Integration of Information in Open and Dynamic Environments”, ACM SIGMOD Record, (Jun. 1997), vol. 26, Issue 2, pp. 195-206.
Calvary, Gaelle, et al., “From Single-User Architectural Design to PAC': a Generic Software Architecture Model for CSCW”, Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, (Mar. 1997), pp. 342-349.
Wegner, Lutz, et al, “Applications: A Visual Interface for Synchronous Collaboration and Negotiated Transactions”, Proceedings of the Workshop on Advanced Visual Interfaces, May 1996, pp. 156-165.
Bulterman, Dick C., “Embedded Video in Hypermedia Documents: Supporting Integration and Adaptive Control”, ACM Transactions on Information Systems (TOIS), vol. 13, Issue 4, Oct. 1995, pp. 440-470.
Hijiri, Toshiki, et al, “A Spatial Hierarchical Compression Method for 3D Streaming Animation”, Proceedings of the Fifth Symposium on Virtual Reality Modeling Language (Web3D-VRML), Feb. 2000, pp. 95-101.
Glushko, Robert J., et al, “An XML Framework for Agent-Based E-Commerce”, Communications of the ACM, vol. 42, No. 3, Mar. 1999, pp. 106-114.
Claypool, Kajal, et al., “OQL_SERF: An ODMG Implementation of the Template-Based Schema Evolution Framework”, Proceedings of the 1998 Conference of the Centre for Advanced Studies on Collaborative Research, (Nov. 1998), pp. 1-15.
Fraisse. S, et al., “Generating Hypermedia Form Specifications by Sketching Multimedia Templates”, ACM, (1996), pp. 353-363.
Iyengar, Arun , et al., “Improving Web Server Performance by Caching Dynamic Data”, proceedings of the US EN IX Symposium on Internet Technologies and Systems, (Dec. 1997), pp. 1-13.
Puerta, Angel, et al., “Towards a General Computational Framework for Model-Based Interface Development”, Proceedings of the 4th International Conference on Intelligent User Interfaces, (Dec. 1998), pp. 171-178.
Shim, Simons., et al., “Template Based Synchronized Multimedia Integration Language Authoring Tool”, SPIE Conference on Internet Imaging, (Dec. 1999), vol. 3964, pp. 134-142.
“XHTML 1.0: The Extensible HyperText Markup Language, A Reformulation of HTML 4.0 in XML 1.0”, www.w3.org, downloaded from http://www.w3.org!TR/19991WD-html-in-xml-199902241, Feb. 24, 1999, pp. 1-15.
Barta, Robert, et al, “Syndication with JML”, Proceedings of the 2000 ACM Symposium on Applied Computing—vol. 2, ACM Press, Mar. 2000, pp. 962-970.
Kuchling, Andrew, “XML, the eXtensible Markup Language”, Linux Journal, Specialized Systems Consultants, Inc., Nov. 1998, pp. 1-3.
Microsoft Computer Dictionary, Third Edition, Microsoft Press, 1997. p. 505.
Smith, John R., et al. “Scalable Multimedia Delivery for Pervasive Computing”, Proceedings of the Seventh ACM International Conference on Multimedia (Part 1), Oct. 1999, pp. 131-140.
Microsoft Computer Dictionary, Third Edition, Microsoft Press, 1997, p. 368.
International Preliminary Examination Report, PCT/US01/13244, dated Feb. 18, 2003 (11 pages).
“Advanced Television Enhancement Forum Specification (ATVEF)”, ATVEF, 34 pages, Feb. 2, 1999.
Palmer, Jonathan W., et al, “Digital Newspapers Explore Marketing on the Internet”, Communications of the ACM, vol. 12, Issue 9, Sep. 1999, pp. 32-40.
Hodes, Todd D., et al, “Composable Ad Hoc Location-Based Services for Heterogeneous Mobile Clients”, Wireless Networks, vol. 5, Issue 5, Oct. 1999, pp. 411-427.
International Search Report for International Application No. PCTIUS01/13244 (Metatv, Inc.), dated Sep. 10, 2001.
International Search Report for International Application No. PCT/US01/13243 (Metatv, Inc.) dated Oct. 10, 2001.
Morris S: “The MHP Tutorial”, Internet Citation, 2002, XP002281053, retrieved from the Internet: URL: http://www.mhp-interactive.org/tutorials/mhp/printerfriendly/ (retrieved on May 19, 2004).
Extended European Search Report—EP12181319—dated Dec. 19, 2012.
Steinhorn, Jason and Mark Kohler, “Enhancing TV With ATVEF” Embedded Systems Programming, Oct. 1999, pp. 55-64.
International Search Report for International Application No. PCT/US01/13241 (Metatv, Inc.), dated Sep. 10, 2001.
Advanced Television Systems Committee, Inc., “ATSC Standard: Programming Metadata Communication Protocol, Revision B,” Document A/76B, 52 pages (Jan. 14, 2008).
Heather Wilner, “Verizon FiOS Brings You the 2012 London Summer Olympic Games,” 3 pages, available at <http://forums.verizon.com/t5/Verizon-at-Home/Verizon-FiOS-Brings-You-the-2012-London-Summer-Olympic-Games/ba-p/462267> (Jul. 23, 2012, accessed Mar. 5, 2014).
Extended European Search Report—EP 14159424.2—dated May 13, 2014.
Extended European Search Report—EP Appl. 15158005.7—dated Jul. 29, 2015.
European Office Action—EP App 14159424.2—dated Sep. 27, 2016.
Oct. 30, 2018—European Summons to Oral Proceedings—EP 14159424.2.
Sep. 17, 2018—European Office Action—EP 15158005.7.
May 14, 2018—Canadian Office Action—CA 2,787,562.
Feb. 8, 2019—Canadian Office Action—2,787,562.
Oct. 8, 2019—Canadian Office Action—CA 2,787,562.
Related Publications (1)
Number Date Country
20180213282 A1 Jul 2018 US
Provisional Applications (1)
Number Date Country
60199686 Apr 2000 US
Continuations (1)
Number Date Country
Parent 09841644 Apr 2001 US
Child 15696847 US