Interactive user interface for television applications

Abstract
An iTV user interface includes a navigable menu area through which a user can select from among a variety of services and a content display area in which formatted data associated with a selected menu service are displayed to the user. The formatted data is displayed in a rotating or ticker fashion having both a passive mode, in which the formatted data is displayed in a sequence independent of any selection operations by a television viewer, and active mode, in which the formatted data is displayed in a sequence that depends on selection operations by the television viewer. Advertisements and/or promotions may be interspersed among the content items. The navigable menu area and the content display area may be displayed to the viewer at the same time as television programming is displayed on the television.
Description
FIELD OF THE INVENTION

The present invention relates to a user interface for use with video systems, and in particular interactive television (iTV) systems, that provides for both information display and navigation.


BACKGROUND

In the past few years, home entertainment has evolved from purely one-way radio and television (TV) broadcasts to include two-way communication that encompasses interactive games, video on demand (VOD), educational fare and consumer-oriented services (e.g., interactive shopping). For the most part, these services have been offered through familiar cable and satellite television distribution systems.


Broadly speaking, there are two paradigms for using cable and satellite systems in this fashion. One is computer-centric: using an installed satellite or cable feed (perhaps with a cable or telephone network back channel) as a high-bandwidth conduit to bring online computing services, such as Internet access, into the home. The second is television-centric: adding interactive functions to TV programming.


In general, systems deployed using the television-centric model have been grouped under the label interactive television or iTV. iTV services are usually delivered to subscribers' homes through an existing cable system, along with other, free and pay-for content. One example of available iTV services is a series of interactive TV portal screens, collectively called a “walled garden”. A walled garden is a collection of user content targeted toward a specific viewing audience. Delivered to the television via a local decoder system (sometimes called a set-top box), viewers navigate the walled garden using a remote control, exploring and interacting with offerings such as banking services, news, sports, weather, entertainment, and shopping. Network operators typically gain revenue from walled gardens by taking a percentage of the transactions or charging for space in TV malls. Walled gardens extend the network operator's brand identity while maintaining the well-known names that provide the content. Thus, the service provider can create a cohesive set of up-to-date content to provide viewers with a rich interactive experience through their television set. These walled garden services can be enhanced through the provision of on-demand video services being made available related to the walled garden content. In this manner various special interest ‘virtual’ channels that include on-demand video services, such as home and garden, travel, history, geography, sports, educational and news channels can be made available to the home viewer.


As indicated above, accessing iTV services sometimes requires the installation of a special decoder system (a set-top box) that may include a dedicated remote control unit. Other systems incorporate the iTV decoder system in a set-top box that is also used to decode and/or decrypt the television signals. In either case, the set-top box and its associated remote control provide an interface for the user to access the iTV services, usually through some form of on-screen menu that is displayed in response to user command inputs. The on-screen menu may or may not be overlaid on top of the television picture. Common examples of on-screen menus include audio/video program guides, electronic catalogs, and advertising information.


Clearly, an important component of any walled garden or other iTV service is the interactivity provided through the combination of the set-top box and its remote control and the on-screen menu. Collectively, the set-top box, the remote control and the visual information display make up the user interface for the iTV services. To date, however, these user interfaces have not been satisfactory.


SUMMARY OF THE INVENTION

The present invention relates to user interfaces for interactive television (iTV) and other applications, which in one embodiment serves as an information presentation architecture for enabling users of iTV services to access a range of broadcast, on-demand programming and other content (which may or may not be related to a current broadcast and/or channel context of the television viewer), offered through an iTV network and made available via applications that may execute on a receiver such as a set-top box. The interactive user interface can be displayed at the same time as television programming, either as an overlay to television programming or with the television programming displayed in a scaled window.


In one embodiment, the present invention provides a content collection, distribution and display system that includes a server component configured to capture content for display to a television viewer by harvesting data from multiple data sources according to content collection rules specified in a configuration file that identifies locations from which the content can be harvested, and to produce a set of data files that include the harvested data, business rules and display rules for such data. The system also includes a distribution component coupled to receive the data files produced by the server component and to transmit the data files to a receiver utilizing a transmission channel (e.g., an in-band or out-of-band component of a satellite and/or cable television distribution network). Finally, the system includes a client component configured to receive the data files via the transmission channel and to populate a user interface, displayable to the television viewer simultaneously with other television programming via the receiver, with the harvested data as formatted data subject to the business rules and display rules.


The user interface may include a navigable menu area through which a user can select from among a variety of interactive television services. That is, the menu area provides access to a variety of services (e.g., video on demand services, game services, television programming related content services, managed content services, special interest services, self-provisioning services, chat services, email services, and television programming guide services) into which the content can be categorized. The user interface may also include a content display area in which those portions of the formatted data associated with a selected menu service are displayed to the television viewer. In some cases, this content is displayed in a rotating or ticker fashion (having both a passive mode, in which the formatted data is displayed in a sequence independent of any selection operations by a television viewer, and active mode, in which the formatted data is displayed in a sequence that depends on selection operations by the television viewer), with advertisements and/or promotions interspersed among the content items. In this way individual content items are displayed for configurable periods of time and then automatically replaced with new content items. The navigable menu area and the content display area may be displayed to the viewer at the same time as television programming is displayed on the television.


Within the user interface display, the displayed categories (e.g., news, weather, sports, entertainment, travel, special interest, demographic, world, national, regional, channel and program information, and local information) and services may change dynamically as new services are introduced and/or new business rules are defined by a network operator. The user interface may also include a content user interface component (referred to below as a content bar) that displays designated content (e.g., date/time/temperature information) regardless of what information is displayed in the content display area. Some embodiments of the present invention also provide for the display of communication alerts (such as e-mail alerts) via the user interface.


The content pieces, advertising and promotions may be organized and/or delivered according to configurable business rules defining content distribution across an iTV network. In this way, the content pieces, advertising and promotions, and preferred placement of elements of content, including tiles, within displayed categories may be dynamically controlled. The advertisements and/or promotions may be linked to services within the iTV network, and/or to an interactive capability to self-provision a related service. The business rules may be configured according to time of day, television channel, channel content, preferred placement, viewer demographic, viewer behavior, navigation context, and/or content category.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention is illustrated by way of example, and not limitation, in the figures of the accompanying drawings in which:



FIGS. 1A and 1B illustrate examples of on-screen menu and ticker portions of user interfaces for iTV services configured in accordance with embodiments of the present invention.



FIG. 2 illustrates one example of a content collection, distribution and display system having server and client components configured in accordance with an embodiment of the present invention.



FIG. 3 illustrates another example of a content collection, distribution and display system having server and client components configured in accordance with an embodiment of the present invention.



FIG. 4A illustrates a further example of an on-screen menu portion of a user interface for iTV services configured in accordance with an embodiment of the present invention in which a menu overlay is used.



FIG. 4B illustrates a further example of an on-screen menu portion of a user interface for iTV services configured in accordance with an embodiment of the present invention in which scaled television video is used.



FIG. 4C illustrates yet another example of an on-screen menu portion of a user interface for iTV services configured in accordance with an embodiment of the present invention in which ticker state has changed from that shown in FIG. 4A.



FIGS. 5A and 5B illustrate an example of a TV Assistant configured in accordance with an embodiment of the present invention, and which becomes populated by content that is specific to a currently tuned broadcast TV channel.



FIG. 6 illustrates an example of a client component of a TV Assistant configured in accordance with an embodiment of the present invention.



FIG. 7 illustrates an example of a client data package component of a TV Assistant client configured in accordance with an embodiment of the present invention.



FIG. 8 illustrates an example of a business rules component TV Assistant client configured in accordance with an embodiment of the present invention.



FIG. 9 illustrates an example of a template layout component of a TV Assistant client configured in accordance with an embodiment of the present invention.



FIG. 10 illustrates an example of a client display engine of a TV Assistant client configured in accordance with an embodiment of the present invention.





DETAILED DESCRIPTION

Described herein is a user interface for iTV applications. The present user interface may be embodied as a television application that provides navigation, information and communication services together, over full-screen television or embedded/scaled television programming. The application serves as a user interface to iTV services (e.g., walled garden services) and content and may be accessed through operation of a remote control device or other operator controls of a set-top box. In one embodiment, the remote control device may have a dedicated input mechanism (e.g., a physical or virtual button or a programmable set of keystrokes, etc.) to allow for rapid access to the on-screen menu portion of the user interface. The present application interface uniquely combines information and navigation features. For example, a subset of offered services and content may be provided directly through an on-screen or other menu and links may be provided to allow the user access to expanded information and services in the context defined by the subset.


As indicated, the on-screen menu may be presented as an overlay over full-screen broadcast television. In other cases, the menu may be presented as part of a special screen (e.g., allowing for scaled television video presentation) or as a separate picture-in-picture. Regardless of the format of presentation, the menu acts as a user interface to provide access to network operator services (e.g., iTV and/or MSO services), timely pieces of information, promotional or advertising material, and in some cases will still allow a user to view regular television programming while the menu is being displayed.



FIGS. 1A and 1B illustrate examples of user interfaces configured in accordance with the present invention. The user interfaces depicted in these examples include an overlay menu referred to as a TV Assistant 10. In other embodiments, TV Assistant 10 is displayed together with a scaled version of the television program video. As indicated above, the TV Assistant 10 may appear on a television screen (not shown in detail in this view) whenever a user inputs an appropriate command sequence using an associated remote control unit. In one embodiment, the appearance/disappearance of the TV Assistant 10 is associated with user commands applied through a MENU button (or some other designated button or command input device) of a TV or set-top box remote control while watching television.


As shown in the illustration, key pieces of information (e.g., weather information, news headlines, sports scores, stock quotes, etc.) are displayed through TV Assistant 10. Components of TV Assistant 10 include a main menu 12, which serves as the main menu to various interactive services (e.g., walled garden content, VOD, games, etc.); and an information content area 14, which displays updated content pieces including but not limited to news headlines, sports scores, stock information, etc.


TV Assistant 10 can also provide access to extended content. That is, additional information related to the content pieces displayed in content area 14 can be seamlessly accessed by allowing the user to navigate from the TV Assistant 10 to any other iTV service including but not limited to walled gardens/portal application (e.g., retrieving the full story for a headline that appears in the TV Assistant's content area 14 can be accomplished by selecting an appropriate link using the remote control). Such access may be provided through a navigation feature such as “GO” button 16 or other graphical element.


Network operator branding may be provided through a graphical element 18. Alternatively or in addition, content provider branding may be provided in a similar fashion. Further, a Time/Date/Current Temperature display may be incorporated in a content bar 20. Navigation items that provide an exit back to regular television programming and/or to further information or assistance (Help) may also be provided as separate selection elements or as part of the content bar 20. Preferably, content and other items displayed in the content bar 20 are invariant with respect to the content displayed in the information content area 14. As shown, the TV Assistant 10 occupies only a portion of the total available television screen, leaving room for the regular television picture 22 to be displayed for user enjoyment.


In addition to these features, other information or services may be displayed/accessed through TV Assistant 10. For example, basic advertising through interstitial or persistent advertisements can be made available. Advanced advertising providing a receptacle 23 for targeted ads can also be accommodated, as shown in FIG. 1B. Communication services such as instant messaging or chat capabilities can be accessed through appropriate menu options and, of course, the TV Assistant 10 can be customized (e.g., through a set-up option 25) for particular content display. For example, the TV Assistant 10 provides a service operator and/or a user with the ability to customize the content types that populate the TV Assistant's content area 14 and/or lower content bar 20, such as selected stock quotes, weather forecasts for specific cities, etc.


In addition, alerts may be set, providing the ability to notify a user of incoming mail (e.g., e-mail or instant messages) or pre-selected content alerts such as stock price targets, etc. The TV Assistant 10 also provides a vehicle for promotions: That is, a network operator has the opportunity to promote various services, offers, etc., through this interface. Self-provisioning is also provided for through an interface for adding interactive services or programming packages.


Broadcast partner integration that is context sensitive to the broadcast channel to which the television is tuned provides the ability for the TV Assistant 10 to be populated by content that is specific to the currently tuned broadcast TV channel. The TV Assistant 10 also provides an interface for commerce services, allowing uses the ability to complete simple transactions. Of course, the TV Assistant 10 also acts as an interface for displaying program specific information synchronized to the broadcast TV program.


The menu choices on both the main menu 12 and ticker menu 24 may be configured subject to service provider requirements and the following is only intended to illustrate one possible implementation. In the example shown, the Interactive option (labeled ITV) provides access to the walled garden home (e.g., a Web Browser). The Games option provides access to various interactive games services. The Guide option provides access to an Electronic Program Guide and the VOD option launches video on demand service.


As indicated above, the Help button 26 provides access to a Help page in the walled garden. The Back to TV option 28 allows a user to exit from the TV Assistant 10 to full screen TV viewing. The lower content bar 20 may display Date/Time/Current Temperature, and in one embodiment the time is updated every 5 seconds at the same time a new ticker item is displayed in content area 14.


The ticker menu 24 acts to highlight the information provided in content area 14. That is, the ticker menu 24 provides a quick visual indication of the type of content currently being accessed/displayed. The ticker rotates through content items within the designated menu categories every 5 seconds (or other predetermined time period). For example, when accessing sports content, the ticker may cause new scores from different sporting events to be displayed in the content area 14 every 5 seconds. By default the ticker component always begins by cycling through the News items (or other user programmed menu selection) and then moves on to the other menu options.


If the user does not select any of the items in the ticker menu 24 he/she remains in what is referred to as “passive” ticker mode. In the passive ticker mode, the ticker cycles through all of the News items first, displaying associated content in content area 14, then automatically moves to the next category when all (or a subset of the) current News items have been displayed. The ticker continues in this fashion, cycling through all (or a network operator defined subset) of the content items in each ticker category until all categories have been exhausted and then repeats the procedure. The currently selected ticker category may be graphically represented with a background color matching the background color of the ticker menu 24. Therefore, when the ticker automatically changes from News to Local, the background color of the navigation button (shown as a highlighted portion of the ticker menu 24) changes automatically to reflect this category change.


In addition to the passive mode, TV Assistant 10 also provides an “active” ticker mode. The user has the option of selecting any of the ticker menu items using the remote control and entering active ticker mode. At that point, the user is presented with the ticker content for the selected category (e.g., the user may be presented with local news items or community messages when he/she selects the Local ticker category) and these content items will continue to repeat within content area 14 (e.g., changing every 5 seconds or other predetermined time period) until the user selects another ticker menu category. The active ticker category navigation button may be graphically represented with a background color matching the background color of the ticker. Additionally, the navigation button is also complemented with a graphic icon (a diamond in this example) to indicate that the user is in the active ticker state.


Text or graphic based ads/promotions can be presented in an interstitial type display as content rotates through the TV Assistant's content area 14. One display point for text-based promotions would be as an additional content ticker item in-between category changes when in passive ticker mode. In active mode, a promotion could appear as the first or last (or other) ticker item in the rotation.


When selected, the GO button 16 will link the user to the appropriate page in the walled garden that corresponds to the current content item displayed in the content area 14. For example, selecting “GO” while a news headline is being displayed will take the user to the story page for that headline in the walled garden. Of course, other selection mechanisms may be used.


The various buttons and other graphical items in this interface may change color to reflect state. For example:

    • a. Unselected: White text on top of a background color.
    • b. Selected: White text on top of a different background color which matches the color of the ticker content area 14, thus drawing a visual connection between the category button and the ticker contents.
    • c. Selection Focus: White text on top of a unique background color representing the element with current focus. The user will be able to navigate to any selectable element. This may be represented by the background color of the element changing from the “unselected” background color to the “selection focus” color upon key press.


As shown in FIG. 2, the TV Assistant application utilizes both a server-side component and a client-side component. The server-side component is responsible for delivering data on a scheduled basis. Data delivery mechanisms may include, but are not limited to, one or more of the following: Internet Protocol (IP) messages, such as through in-band or out-of-band data pathways, or though an IF gateway product; or in-band or out-of-band carousel data. Such data may also be encapsulated within MPEG2 packets for transmission through the in-band broadcast system. The server-side component provides data from multiple disparate sources using a TV Assistant server 30. The server 30 extracts, collects, integrates, and deploys content and commerce from a variety of sources 32 such as web servers, raw television feeds, data sources (such as file servers, etc.) into new rich television-centric services. It allows existing Internet brands, content, and commerce services to be introduced to a new audience of viewers. Content is dynamically harvested, re-purposed and enriched for the television without affecting the existing technology infrastructure or publishing resources of the content provider.


As shown, content from the various sources 32 is retrieved by the TV Assistant server 30 and formatted according to instructions provided via a configuration file TVAConfig.xml. This configuration file (which can specify “business rules” for content collection, update, etc.) is created by the network or content service provider using a TV Assistant Administrator component 34. In this example, the file is an extensible markup language (XML) file, however, any convenient programming language may be used. TVAConfig.xml is used to create the selected contents and the delivery location of the TVAContents.txt file (discussed below). It allows the operator of a given head-end, through a web interface or other command interface, to specify information regarding content such as location, syndication sources, number of items displayed per source, etc. It also allows the operator to indicate where the TVAContents.txt file 36 is to be deployed, including, where appropriate, the IP address of an Applet Verification Server (AVS) 42 in the head-end and associated attributes such as start/end time, interval, and refresh rate. An example of the TVAConfig.xml file is:

















<TVA_ServletConfig>



 <City>



  <Name>San Francisco</Name>



  <State>CA</State>



  <Zip>94111</Zip>



  <WorldNews_Url> (insert local URL for world news)



  </WorldNews_Url>



  <LocalNews_Url> (insert local URL for local news)



  </LocalNews_Url>



  <Sports_Url> (insert local URL for sports) </Sports_Url>



  <Weather_Url> (insert local URL for weather) </Weather_Url>



  <WorldNews_Count>5</WorldNews_Count>



  <LocalNews_Count>10</LocalNews_Count>



  <Sports_Count>8</Sports_Count>



  <Weather_Count>3</Weather_Count>



  <AVS_IP> (insert IP address of applet verification server)



  <AVS_IP>



 </City>



</TVA_ServletConfig>










In one embodiment of the present invention, a server-side servlet named TVAGenerator.java resident at the TV Assistant server 30 is responsible for capturing content by harvesting data from the multiple data sources 32 via a syndication server (not shown in detail) as specified in the TVAConfig file. One example of a syndication server suitable for this purpose is described in detail in U.S. patent application Ser. No. 09/841,017, entitled “METHOD AND SYSTEM FOR TRANSFORMING CONTENT FOR EXECUTION ON MULTIPLE PLATFORMS”, filed Apr. 23, 2001 and assigned to the assignee of the present invention, the complete disclosure of which is hereby incorporated by reference. In addition, the TVAGenerator file formats the captured contents into a data file, TVAContents.txt, 36 uploads TVAContents.txt to an appropriate broadcast file server (BFS) 38 (e.g., as specified in the TVAConfig file), by using a TVAFileUploader utility 46 to transfer the file through a verification system (if necessary) such as Liberate's AVS 42.


In this example, TVAFileUploader 46 is a Java utility that uses HTTP POSTs to upload the data file 36. This file uploader is executed each time the TVAContents.txt file is updated and generated. The file uploader reads the TVAContents.txt file 36 into a byte array and delivers it over TCP/IP to the broadcast file server 38.


TVAContents.txt is the data file 36 that contains the content to be displayed in the TV Assistant 10. This file gets generated on a schedule, typically every 30 minutes but this time period is configurable by the network operator, to acquire the latest data, and is stored in the local file system of the server 30 where TVAGenerator.java runs. The broadcast server 38 is responsible for delivering data file 36 (TVAContents.txt) to the client-side application using the selected delivery mechanism (e.g., TCP/IP over an existing signaling channel or via a data carousel channel, etc.). The TV Assistant server 30 thus creates a file (TVAContents) 36 for distribution to a set-top box 40 via a network operator's existing content delivery platform. In this example, the file 36 is a text file but this is a design choice that may vary depending upon the system configuration and implementation.


After receiving the data file 36, the set-top box 40 is responsible for reading the file contents and displaying the corresponding content in the manner described above. In the present example of the client—server architecture of the TV assistant application, the client application is programmed in Java (TVA.java) and is utilized in conjunction with a commercially available broadcast file server 38, e.g., the Liberate Compact platform, and a Motorola DCT2000 set-top box 40. Other systems configured in accordance with the present invention may be implemented using other languages such as C, C++, etc. on other set-top boxes 40 and/or TV middleware platforms 38.


The client-side component is an application that can be accessed in one of several ways depending on the set-top platform being used. For example, the client application (i.e., the actual client-side computer software) may be flashed onto the set-top box 40 as a stand-alone application or as an extension (or customization) of some set-top box resident application, or TV middleware system; or a broadcast application that is carouseled to the set-top box 40 over an in-band or out-of-band data channel from the broadcast server 38. Advantages of deploying the TV Assistant as a resident application or an extension/customization of a resident application of a set-top box 40 include:

    • a. Instant access via the MENU (or other) button on the remote control;
    • b. Accessibility over any broadcast TV channel;
    • c. No latency in load time as a result of being resident in the set-top box memory;
    • d. Ability to launch other services such as VOD, Walled Garden, Mail, Self-provisioning, etc.


The TVA.java file instantiates itself in the set-top box 40 and specifies the user interface menu's on-screen size and location. This allows the TV assistant 10 to position itself in a specific region of the TV screen, for example as specified by x-y coordinates. Upon receipt of the TVAContents.txt file 36, the TVA.java file reads the data contained in that file 36 to populate the TV Assistant 10. In some cases, this involves reading in the latest information from TVAContents.txt based on a schedule through a timer event. TVA.java formats the information and displays the content for each timer event. This may include examining sports score strings to determine game status and, if the game is final, displaying winning team/score in a different color. Refreshing of content is accomplished by periodically looking for new TVAContents.txt files transmitted by the data carousel server 44 associated with broadcast server 38. The TVA.java file also enables navigation to any selectable element in the TV Assistant 10 (including other applications such as VOD, etc.) and displays the following button states: unselected, selected, and selected/focus.


As shown in FIG. 6, the TV Assistant client component 81 itself includes four components (a data package component 82, a business rules component 84, a template layout component 86 and a client display engine 88), which cooperate to render the TV Assistant 10 as seen by the television viewer. That is, these four components accept, interpret and present the content delivered by the broadcast server 38 to the television viewer.


The client data package component 82 is shown in detail in FIG. 7 and is responsible for receiving and parsing data elements and preparing them for display. The data package component 82 communicates with the distribution system to retrieve data packages (including set-top data 83, server data 85 and middleware data 87) as they are received, or as demanded by the client display engine 88. The data package component 82 includes a deserializer 89 and data parser 90, which together deserialize and interpret any optimized XML and binary data received from the server-side packaging component. Data package component 82 communicates with the client display engine 88 that creates and manages the data objects representing the data that is received in packages. Business rules metadata and template layouts are deserialized and reconstructed into objects for the business rules and template layout components 84 and 86, respectively. The data package component 82 communicates with the set-top box operating system interfaces to resolve references to set-top related data 83, such as time and date; and communicates with the resident middleware system to resolve references and requests for middleware data 87, such as current channel information. Other data (such as images, content for the ticker and content bar, VOD metadata and the like) is treated as server data 85 for the TV Assistant 10.


The business rules component 84 is shown in detail in FIG. 8 and is responsible for constructing business rules objects and applying these business rules to template objects as directed by the template layout component 86. In one embodiment, business rules are described in XML format and define placement or other objectives that may be based on criteria such as: time of day, current channel, current broadcast content, viewer demographic, viewer behavior, navigation context and/or current content category.


As shown in the illustration, the business rules themselves may, in one embodiment, comprise rules associated with the following data categories 85:

    • i. Advertising and Promotional rules
    • ii. TV channel association rules
    • iii. Timing rules (for rotation of content, ads and promotions)
    • iv. Category placement rules
    • v. Preferred placement rules
    • vi. VOD metadata rules
    • vii. Self provisioning rules


The business rules component 84 receives the business rules data from the data package component 82, interprets them, and provides instructions that will eventually be applied to layout objects by the template layout component 86 and the client display engine 88. The business rules component 84 determines what interface elements 87 (data, images, ads, menu choices, etc.) are shown and when, and in what order and sequence those elements should be displayed. In one embodiment this may include displaying an advertisement for a sports product when the viewer is tuned to a sports channel. In another embodiment this may include providing links to a specific video when a user is on a specific channel at a specific time.


An example that illustrates the business rules metadata that is authored and/or generated by the server to control these display parameters is given below:

















<?xml version=“1.0” encoding=“UTF-8”?>



<meta-business-rules>



  <ad-location-list start=“1” end=“9”>



    <location id=“1”>



      <campaign id=“1” start=“1” end=“4”/>



      <campaign id=“11” start=“5” end=“7”/>



      <campaign id=“1” start=“8” end=“9”/>



    </location>



    <location id=“2”>



      <campaign id=“12” start=“1” end=“3”/>



      <campaign id=“3” start=“3” end=“5”/>



      <campaign id=“12” start=“6” end=“9”/>



    </location>



      <location id=“3”>



      <campaign id=“5” start=“1” end=“3”/>



    <campaign id=“15” start=“4” end=“7”/>



  </location>



  </ad-location-list>



  <ad-campaign-list start=“1” end=“9”>



  <campaign id=“1” displayrule=“1”>



    <ad id=“1” adtype=“1” href=“String” src=“String”



displayruledata=“1”/>



    <ad id=“7” adtype=“1” href=“String” src=“String”



displayruledata=“2”/>



    <ad id=“4” adtype=“1” href=“String” src=“String”



displayruledata=“3”/>



  </campaign>



  <campaign id=“3” displayrule=“2”>



    <ad id=“8” adtype=“2” href=“String” src=“String”



displayruledata=“1”/>



    <ad id=“2” adtype=“2” href=“String” src=“String”



displayruledata=“2”/>



    <ad id=“5” adtype=“2” href=“String” src=“String”



displayruledata=“3”/>



  </campaign>



</ad-campaign-list>



<channel-association-list start=“1” end=“9”>



  <content-type id=“1”>



    <channel id=“3” start=“1” end=“4” />



    <channel id=“4” start=“1” end=“5” />



    <channel id=“11” start=“3” end=“9”/>



    <channel id=“13” start=“1” end=“9”/>



  </content-type >



  <content-type id=“2”>



    <channel id=“3” start=“1” end=“3”/>



    <channel id=“5” start=“2” end=“9”/>



  </content-type >



</channel-association-list>



<link-association-list start=“1” end=“9”>



  <content-link-type id=“1”>



    <channel-type id=“3” start=“1” end=“4” >



      <application id=21233 />



    </channel-type>



    <channel-type id=“3” start=“5” end=“9” >



      <application id=21244 />



    </channel-type>



    <channel-type id=“4” start=“1” end=“9” >



      <application id=21413 />



    </channel-type>



  </content-link-type >



  <content-link-type id=“2”>



    <channel-type id=“3” start=“1” end=“3”>



      <service id=11233 />



    </channel-type>



    <channel-type id=“3” start=“4” end=“9”>



      <service id=11222 />



    </channel-type>



    <channel-type id=“7” start=“3” end=“9”>



      <application id=21255 />



    </channel-type>



    <channel-type id=“11” start=“1” end=“9”>



      <service id=21215 />



    </channel-type>



  </content-link-type >



</link-association-list>



<content-rotation-list start=“1” end=“9”>



  <content-type id=“1”>



    <channel-type id=“2” start=“1” end=“9”>



      <rotation displayrule=“2” displayruledata=“7” />



    </channel-type>



    <channel-type id=“3” start=“1” end=“3”>



      <rotation displayrule=“1” displayruledata=“5” />



    </channel-type>



    <channel-type id=“3” start=“4” end=“7”>



      <rotation displayrule=“0” />



    </channel-type>



    <channel-type id=“3” start=“8” end=“9”>



      <rotation nodisplay=“true” />



    </channel-type>



    <channel-type id=“4” start=“5” end=“9”>



      <rotation displayrule=“1” displayruledata=“5” />



    </channel-type>



    <channel-type id=“0” start=“1” end=“9”>



      <rotation displayrule=“2” displayruledata=“4” />



    </channel-type>



  </content-type >



  <content-type id=“2”>



    <channel-type id=“2” start=“1” end=“9”>



      <rotation displayrule=“1” displayruledata=“3” />



    </channel-type>



    <channel-type id=“3” start=“1” end=“9”>



      <rotation displayrule=“1” displayruledata=“5” />



    </channel-type>



  </content-type >



</content-rotation-list>



<preferred-placement-list start=“1” end=“9”>



  <provider id=“1233”>



    <channel id=“2” start=“1” end=“9”>



      <placement displayrule=“2” displayruledata=“1” />



    </channel >



    <channel id=“3” start=“1” end=“4”>



      <placement displayrule=“1” displayruledata=“1” />



    </channel >



  </provider>



  <provider id=“1453”>



    <channel id=“4” start=“1” end=“9”>



      <placement displayrule=“1” displayruledata=“4” />



    </channel >



    <channel id=“7” start=“3” end=“6”>



      <placement displayrule=“1” displayruledata=“4” />



    </channel >



  </provider>



  <service id=“3253”>



    <channel id=“4” start=“1” end=“9”>



      <placement displayrule=“2” displayruledata=“4” />



    </channel >



  </service>



  <provider-type id=“1”>



    <channel-type id=“2” start=“1” end=“9”>



      <placement displayrule=“3” displayruledata=“4” />



    </channel-type>



    <channel-type id=“3” start=“1” end=“3”>



      <placement displayrule=“2” displayruledata=“2” />



    </channel-type>



    <channel-type id=“1” start=“1” end=“9”>



      <placement nodisplay=“true” />



    </channel-type>



    <channel-type id=“0” start=“1” end=“9”>



      <placement displayrule=“2” displayruledata=“4” />



    </channel-type>



  </provider-type >



  </preferred-placement-list>



</meta-business-rules>










The template layout component 86 is shown in detail in FIG. 9 and receives template definitions and data (content, ads, promotions, and VOD metadata) from the data package component 82, layout factors and rules from the business rules component 84 (which collectively are organized as content placement information 89 and object components 91) and prepares the interface for display by the client display engine 88. The template layout component 86 identifies all the aspects of the TV Assistant interface and orders them into a viewable user interface. It is responsible for creating a logical navigation structure for each screen, positioning of defined objects on a screen, and communicating information between objects and screen using the client display engine interfaces.


The client display engine 88 is shown in detail in FIG. 10 and constructs and manages the objects (via object manager 91) that are used throughout the TV Assistant client 81. The client display engine 88 provides a set of interfaces for the other components of the client to manipulate these objects. It receives data and instructions from the data package component 82, the business rules component 84 and the template layout component 86 and generates the interface display on the set top box 40 for the viewer. The client display engine 88 manages the layout and navigation elements of the user interface via layout manager 93 and can provide links to other applications. It is responsible for monitoring user activity and generating the appropriate response. It includes a scripting and event model 97, which in one embodiment may be similar to Javascript, that provides a procedural language that a template author may use to program the display object model and business rules object model, and to respond programmatically to user events. It also includes a run-time cache object (page manager 95) that stores information about objects and pages and forthcoming pages for faster performance and reduced latency during page transitions. Other modules include a VOD sessions manager 99 and an application manager 100.


The client component, TVA.java, thus includes java classes and utilities for image manipulation, swapping, button functionality and text control. All of these assets along with all image assets, which in one embodiment are optimized through Liberate's VQ compression tool, are packaged into a file, TVA.jar, and either uploaded to the BFS 38 on a schedule or, ideally, flashed to the set-top box 40 a master application for best performance and ease of access.


In operation, the data file 36 (TVAContents.txt) is continuously broadcast to the set-top box 40 over either an in-band or out-of-band data channel. On the Liberate Compact platform, the data file 36 is delivered in an out-of-band channel in order for TV Assistant 10 to work over any analog or digital broadcast channel. A typical out-of-band channel provides approximate bandwidths of 2 Mb/sec downstream and 300 K/sec upstream.


The size of the data file 36 is entirely dependent upon the amount of information that populates the TV Assistant 10. Below are a few potential scenarios:


Scenario 1 (Approximately 3K File Size)

    • 5 news headlines
    • 5 sports headlines/scores
    • 5 local headlines
    • 15 sports scores
    • 1 city weather forecast (3 day)
    • 5 promotional headlines


Scenario 2 (Approximately 5K File Size)

    • 10 news headlines
    • 10 sports headlines
    • 5 local headlines
    • 5 entertainment headlines
    • 20 sports scores
    • 20 stock quotes
    • 5 city weather forecast
    • 5 promotional headlines


Having thus described the individual components of the present content collection, distribution and display system, it is helpful to discuss an example of how this architecture is used. In the following example, specific details are provided for the creation, optimization, and automation of a TV Assistant specific to a platform configuration comprising Liberate's Compact 3.2.47 with MAC-J “Elmo” running on a Motorola DCT 2000 set-top box. These details are provided so that those of ordinary skill in the art can more fully appreciate certain implementations of the materials discussed above, however, this example should not be read as limiting the broader scope of the present invention to these implementation details.


Using authoring tools provided, for example, in the TV Assistant Administrator component 34, the application author 48 (e.g., an iTV network operator) generates the necessary layout and data templates for the TV Assistant application, and deploys them to the application server 30. The layout templates (TVA.xml) 50 may be bundled with any bitmap images used for buttons, user interface controls, etc. This allows the network operator to customize the look and feel of the TV Assistant 10 as it will ultimately be displayed to the television viewer.


The business rules and configuration data file (TVAConfig.xml) are also authored and specify the rules for organizing and placing content, promotions and advertising within the TV Assistant. In this example, dynamic data generation templates (used to generate dynamic text data, image and video references) are created as Java Server Pages (DATA[N].jsp) 54 that conform to the Application Frameworks XML language developed by the assignee of the present invention, which provides a set of executable tags to access dynamic data harvested by a catalog engine 56. When executed, these templates 54 create a set of named data packages (TVA.dat) 58. Further details regarding this process may be found in the above-cited U.S. patent application Ser. No. 09/841,017.


The TVA.xml file 50 is run through a packager component 60 of the TV Assistant server 30, which uses extensible stylesheet language transformations (XSLT) to transform the XML information to a highly optimized, binary XML format file (TVA.mac) 62. These templates reference the contents of data packages via predetermined hierarchical package name references. The packager 60 then transports the TVA.mac file 62 directly to the Liberate broadcast server 38, where the Liberate Mediacast server 44 carousels TVA.mac to the set-top box 40 via an in-band channel. This process only needs to occur once, or as often as the application designer wishes to make changes to the layout and application behavior.


With data templates already residing on the TV Assistant server 30, processes need to occur to execute and create the dynamic data packages with their text data, image and video reference information. This process is initiated by specifying the data generation business rules within the TV Assistant Admin component 34 for scheduling data packages. In some cases, this may be done through a web-type interface (Admin Manager) 78 associated with server 30. According to the specified schedules, the packager 60 will initiate execution of the data templates (DATA[N].jsp) 54 via an Application Frameworks component 64 of server 30.


The DATA[N].jsp pages are broken into separate file types: those containing text data, including video url references, and those containing the image assets related to that data (backgrounds, photos, ads, etc.). When the DATA[N].jsp files are executed by the Application Frameworks component 64, the latest text, video references and image data are retrieved from the catalog engine 56. Independently, the catalog engine 56 is configured to harvest, process and store related data and metadata from content feeds 32, asset distribution interface (ADI) packages, and/or the Internet. If desired, an advertising server 66 may be invoked to resolve any advertising and/or promotion resources that are being provisioned for the given period as configured by the TV Assistant Admin 34. Dynamic business rules defined by the network service provider to be applied by the TV Assistant application client are retrieved and packaged with the data.


The DATA[N].jsp pages are processed by the Application Frameworks component 64 and the dynamic data sets returned to the packager 60. The packager transforms the text and video reference portion of the data into a binary format file (TVA.dat) 66. The packager then generates TVA.html files 68 for the image portion of the data, and these files refer back to the image data. Finally, the packager generates an TVA.cdf file 70 containing references to the TVA.dat and TVA.html files 58 and 60 and posts this to the data carousel (DC) server 44.


The DC server 44 parses the TVA.cdf file 70 and downloads all the necessary data files referenced therein. It will also render the images referenced by the TVA.html files 68 into MPEG I-Frames and P-Frames, as required. The DC server 44 creates objects in Mediacast Signaling Protocol (MSP—a protocol for data broadcasting to set top boxes that allows any number of clients to access a carousel without increasing the downstream bandwidth required) format and sends them to the multimedia streamer (MMS) component 72 of broadcast server 38 for transport to set-top box 40 via the distribution network 74. At this point the necessary data and transport stream files will be spinning continuously on the carousel over an in-hand channel. Additionally, the application client, which provides the execution platform and APIs used to implement the TV Assistant application, could itself be included in this carousel if it has not already been flashed onto the set-top box 40.


A system administrator for the network service provider can access the Admin Manager web interface 78 for the TV Assistant server 30 at any time to complete a number of tasks, including the ability to: modify configurations for the TV Assistant application, modify advertising and promotion campaign settings, change ticker settings, change business rules settings, change content harvesting sources, specify schedules and business rules for creating data packages, and a host of other activities.


Having examined the content collection and distribution architecture, some further examples of the on-screen menu portion of the present user interface may be described. As indicated above, the TV Assistant may be displayed as an overlay menu simultaneously with television video. FIG. 4A provides an example of such a display. In this example, the TV Assistant 10 is displayed over regular television program video 22, but that video is still viewable. Components of the TV Assistant 10 have been organized in a different fashion than was the case in FIG. 1, for example the content bar 20 now appears at the upper left of TV Assistant 10 and the number and types of main menu 12 categories and ticker categories 24 have changed, but the functionality provided by TV assistant 10 has not changed.


Referring now to FIG. 4B, the same features, functionality and layout for TV Assistant 10 are presented as was the case for the FIG. 4A embodiment, with the exception that the television video 22 is now displayed in a scaled fashion. That is, the television video 22 is displayed in a window framed by TV Assistant 22, but the image so displayed is a scaled representation of what would otherwise occupy the entire television screen if the TV Assistant 10 were not displayed. Such image scaling may be accomplished using conventional picture-in-picture or split screen image presentation technology as exists in many television receives and/or set-top box units.



FIG. 4C illustrates an example of TV Assistant 10 when the ticker state has changed. Compare the illustration set forth in FIG. 4A to that in FIG. 4C. Notice that the ticker state has changed from “Spotlight” to “Weather” and the corresponding content in content area 14 has changed to reflect current weather conditions. This is one example of how content is rotated within the content area 14. The ticker menu 24 now has the “Weather” menu highlighted to reflect the ticker state, as described above. Similar state changes could be reflected if the scaled television video representation was used.


Earlier it was noted that the TV Assistant provided a mechanism for broadcast partner integration that is context sensitive to the broadcast channel to which the television is tuned. That is, the TV Assistant may be populated by content that is specific to the currently tuned broadcast TV channel. An example of such functionality is shown in FIGS. 5A and 5B.



FIG. 5A illustrates an example of the TV Assistant 10, this time shown without any television video. This example illustrates one configuration of TV Assistant 10 as might be displayed in connection with a television channel for which there is no corresponding context sensitive information which the broadcaster or network operator wishes to provide access. In other words, the TV Assistant 10 is displayed in a basic form as an informational overly with basic information about the current channel (e.g., channel number, broadcaster call letters, etc.) and the lower content bar 20.


When the viewer changes channels to a channel for which there is context sensitive information which the broadcaster or network operator wishes to provide access to, however, the content displayed within TV Assistant 10 changes, as shown in FIG. 5B. In this example, a more advanced interface is presented with broadcaster specific content displayed in the content area 14. For example, news and weather interactive content services may be made available to the viewer via the TV Assistant user interface. Rather than being presented as text highlights as in the previous examples, in this case such interactive content is signaled using graphical elements such as tiles 80, which can be selected by a viewer using a remote control similar to the way in which one might use a computer's cursor control device to select a hyperlink displayed in a web browser. In this way, the TV Assistant user interface may extend the television broadcast content by presenting an interactive content area that includes time or context sensitive messages or other items (e.g., in the form of a tiled grid or as textual references) based on business rules such as time of day, channel selected, and so on. This presents viewers with additional opportunities to select on-demand content supplemental to the broadcast or promoted by the broadcaster, programmer, advertiser or network operator.


Thus, an interactive user interface for television applications has been described. Content for display through the TV Assistant application is retrieved and packaged for delivery using the TV Assistant server 30. The data package is delivered via various protocols including TCP/IP or out-of-band broadcast to populate the content area 14 of the on-screen menu. Selecting any sub category, e.g., News, Sports, Weather, of main menu 12 changes the displayed content to that topic. Selecting the “GO” button 16 drives the user to the related page in the walled garden for more information on that particular headline/item.

Claims
  • 1. An apparatus comprising: one or more processors; andmemory storing executable instructions that, when executed by the one or more processors, cause the apparatus to: cause display, within a guide menu and in a first repeatedly-rotating sequence, of a plurality of content items associated with a plurality of menu categories; andbased on receiving a selection associated with a particular menu category of the plurality of menu categories, cause display, within the guide menu and in a second repeatedly-rotating sequence, of a set of content items, of the plurality of content items, associated with the particular menu category.
  • 2. The apparatus of claim 1, wherein the first repeatedly-rotating sequence is based on a sequential display order that causes display of each content item associated with a first menu category before causing display of each content item associated with a second menu category.
  • 3. The apparatus of claim 1, wherein the particular menu category is for news content.
  • 4. The apparatus of claim 1, wherein the particular menu category is for sports content.
  • 5. The apparatus of claim 1, wherein the particular menu category is for weather content.
  • 6. The apparatus of claim 1, wherein the plurality of menu categories is for a plurality of types of content, and wherein, for each of the plurality of menu categories, a menu category is for a type of content that is different from each other type of content of the plurality of types of content.
  • 7. The apparatus of claim 1, wherein the executable instructions, when executed by the one or more processors, cause the apparatus to: cause display of an advertisement between two content items of the plurality of content items.
  • 8. The apparatus of claim 1, wherein the executable instructions, when executed by the one or more processors, cause the apparatus to: cause display of an advertisement between two content items of the set of content items.
  • 9. An apparatus comprising: one or more processors; andmemory storing executable instructions that, when executed by the one or more processors, cause the apparatus to:cause display, within a guide menu and in a first repeatedly-rotating sequence, of a first set of content items, of a plurality of content items, associated with a first menu category of a plurality of menu categories, wherein the plurality of content items are associated with the plurality of menu categories;based on causing display of a last content item from the first set of content items, cause display, within the guide menu and in the first repeatedly-rotating sequence, of a second set of content items, of the plurality of content items, associated with a second menu category; andbased on receiving a selection associated with the second menu category, cause display, within the guide menu and in a second repeatedly-rotating sequence, of the second set of content items.
  • 10. The apparatus of claim 9, wherein the first repeatedly-rotating sequence is based on a sequential display order that causes display of each content item associated with the first menu category before causing display of each content item associated with the second menu category.
  • 11. The apparatus of claim 9, wherein the second menu category is for news content.
  • 12. The apparatus of claim 9, wherein the second menu category is for sports content.
  • 13. The apparatus of claim 9, wherein the second menu category is for weather content.
  • 14. The apparatus of claim 9, wherein the plurality of menu categories is for a plurality of types of content, and wherein, for each of the plurality of menu categories, a menu category is for a type of content, of the plurality of types of content, that is different from each other type of content of the plurality of types of content.
  • 15. One or more non-transitory computer-readable media storing executable instructions that, when executed, cause an apparatus to: cause display, within a guide menu and in a first repeatedly-rotating sequence, of a plurality of content items associated with a plurality of menu categories; andbased on receiving a selection associated with a particular menu category of the plurality of menu categories, cause display, within the guide menu and in a second repeatedly-rotating sequence, of a set of content items, of the plurality of content items, associated with the particular menu category.
  • 16. The one or more non-transitory computer-readable media of claim 15, wherein the first repeatedly-rotating sequence is based on a sequential display order that causes display of each content item associated with a first menu category before causing display of each content item associated with a second menu category.
  • 17. The one or more non-transitory computer-readable media of claim 15, wherein the particular menu category is for news content.
  • 18. The one or more non-transitory computer-readable media of claim 15, wherein the particular menu category is for sports content.
  • 19. The one or more non-transitory computer-readable media of claim 15, wherein the particular menu category is for weather content.
  • 20. The one or more non-transitory computer-readable media of claim 15, wherein the plurality of menu categories is for a plurality of types of content, and wherein, for each of the plurality of menu categories, a menu category is for a type of content, of the plurality of types of content, that is different from each other type of content of the plurality of types of content.
RELATED APPLICATIONS

This application is a continuation of and claims priority to U.S. application Ser. No. 13/736,283, filed Jan. 8, 2013, which is a divisional of and claims priority to U.S. application Ser. No. 10/247,901, filed Sep. 19, 2002, issued Jan. 29, 2013 as U.S. Pat. No. 8,365,230, which is a non-provisional of and claims priority to both U.S. provisional application No. 60/395,277, filed Jul. 11, 2002, and U.S. provisional application No. 60/323,658, filed Sep. 19, 2001. Each of the above-mentioned applications is incorporated herein by reference in its entirety.

US Referenced Citations (453)
Number Name Date Kind
5287489 Nimmo et al. Feb 1994 A
5321750 Nadan Jun 1994 A
5353121 Young et al. Oct 1994 A
5485221 Banker et al. Jan 1996 A
5521841 Arman et al. May 1996 A
5530939 Mansfield, Jr. et al. Jun 1996 A
5583563 Wanderscheid et al. Dec 1996 A
5589892 Knee et al. Dec 1996 A
5592551 Lett et al. Jan 1997 A
5594509 Florin et al. Jan 1997 A
5613057 Caravel Mar 1997 A
5621456 Florin et al. Apr 1997 A
5657072 Aristides et al. Aug 1997 A
5659793 Escobar et al. Aug 1997 A
5666645 Thomas et al. Sep 1997 A
5675752 Scott et al. Oct 1997 A
5694176 Bruette et al. Dec 1997 A
5737552 Lavallee et al. Apr 1998 A
5802284 Karlton et al. Sep 1998 A
5826102 Escobar et al. Oct 1998 A
5844620 Coleman et al. Dec 1998 A
5850218 LaJoie et al. Dec 1998 A
5852435 Vigneaux et al. Dec 1998 A
5860073 Ferrel et al. Jan 1999 A
5883677 Hofmann Mar 1999 A
5892902 Clark Apr 1999 A
5892905 Brandt et al. Apr 1999 A
5905492 Straub et al. May 1999 A
5929849 Kikinis Jul 1999 A
5945987 Dunn Aug 1999 A
5960194 Choy et al. Sep 1999 A
5990890 Etheredge Nov 1999 A
5996025 Day et al. Nov 1999 A
6002394 Schein et al. Dec 1999 A
6005561 Hawkins et al. Dec 1999 A
6008803 Rowe et al. Dec 1999 A
6008836 Bruck et al. Dec 1999 A
6016144 Blonstein et al. Jan 2000 A
6025837 Matthews, III et al. Feb 2000 A
6038560 Wical Mar 2000 A
6049823 Hwang Apr 2000 A
6061695 Slivka et al. May 2000 A
6067108 Yokote et al. May 2000 A
6088722 Herz et al. Jul 2000 A
6091411 Straub et al. Jul 2000 A
6094237 Hashimoto Jul 2000 A
6141003 Chor et al. Oct 2000 A
6148081 Szymanski et al. Nov 2000 A
6162697 Singh et al. Dec 2000 A
6169543 Wehmeyer Jan 2001 B1
6172677 Stautner et al. Jan 2001 B1
6177931 Alexander et al. Jan 2001 B1
6191781 Chaney et al. Feb 2001 B1
6195692 Hsu Feb 2001 B1
6205582 Hoarty Mar 2001 B1
6219839 Sampsell Apr 2001 B1
6239795 Ulrich et al. May 2001 B1
6240555 Shoff et al. May 2001 B1
6281940 Sciammarella Aug 2001 B1
6292187 Gibbs et al. Sep 2001 B1
6292827 Raz Sep 2001 B1
6295057 Rosin et al. Sep 2001 B1
6314569 Chernock et al. Nov 2001 B1
6317885 Fries Nov 2001 B1
6345305 Beck et al. Feb 2002 B1
6405239 Addington et al. Jun 2002 B1
6415438 Blackketter et al. Jul 2002 B1
6421067 Kamen et al. Jul 2002 B1
6426779 Noguchi et al. Jul 2002 B1
6442755 Lemmons et al. Aug 2002 B1
6477705 Yuen et al. Nov 2002 B1
6486920 Arai et al. Nov 2002 B2
6522342 Gagnon et al. Feb 2003 B1
6529950 Lumelsky et al. Mar 2003 B1
6530082 Del Sesto et al. Mar 2003 B1
6532589 Proehl et al. Mar 2003 B1
6564263 Bergman et al. May 2003 B1
6567104 Andrew et al. May 2003 B1
6571392 Zigmond et al. May 2003 B1
6591292 Morrison et al. Jul 2003 B1
6621509 Eiref et al. Sep 2003 B1
6636887 Augeri Oct 2003 B1
6658661 Arsenault et al. Dec 2003 B1
6678891 Wilcox et al. Jan 2004 B1
6684400 Goode et al. Jan 2004 B1
6698020 Zigmond et al. Feb 2004 B1
6704359 Bayrakeri et al. Mar 2004 B1
6731310 Craycroft et al. May 2004 B2
6745367 Bates et al. Jun 2004 B1
6760043 Markel Jul 2004 B2
6763522 Kondo et al. Jul 2004 B1
6766526 Ellis Jul 2004 B1
6806887 Chernock et al. Oct 2004 B2
6857128 Borden, IV et al. Feb 2005 B1
6886029 Pecus et al. Apr 2005 B1
6904610 Bayrakeri et al. Jun 2005 B1
6910191 Segerberg et al. Jun 2005 B2
6918131 Rautila et al. Jul 2005 B1
6963880 Pingte et al. Nov 2005 B1
7028327 Dougherty et al. Apr 2006 B1
7065785 Shaffer et al. Jun 2006 B1
7080400 Navar Jul 2006 B1
7103904 Blackketter et al. Sep 2006 B1
7114170 Harris et al. Sep 2006 B2
7134072 Lovett et al. Nov 2006 B1
7152236 Wugofski et al. Dec 2006 B1
7162694 Venolia Jan 2007 B2
7162697 Markel Jan 2007 B2
7174512 Martin et al. Feb 2007 B2
7177861 Tovinkere et al. Feb 2007 B2
7197715 Valeria Mar 2007 B1
7207057 Rowe Apr 2007 B1
7213005 Mourad et al. May 2007 B2
7221801 Jang et al. May 2007 B2
7237252 Billmaier Jun 2007 B2
7293275 Krieger et al. Nov 2007 B1
7305696 Thomas et al. Dec 2007 B2
7313806 Williams et al. Dec 2007 B1
7337457 Pack et al. Feb 2008 B2
7360232 Mitchell Apr 2008 B2
7363612 Satuloori et al. Apr 2008 B2
7406705 Crinon et al. Jul 2008 B2
7440967 Chidlovskii Oct 2008 B2
7464344 Carmichael et al. Dec 2008 B1
7472137 Edelstein et al. Dec 2008 B2
7490092 Sibley et al. Feb 2009 B2
7516468 Deller et al. Apr 2009 B1
7523180 DeLuca et al. Apr 2009 B1
7587415 Gaurav et al. Sep 2009 B2
7624416 Vandermolen et al. Nov 2009 B1
7640487 Amielh-Caprioglio et al. Dec 2009 B2
7702315 Engstrom et al. Apr 2010 B2
7703116 Moreau et al. Apr 2010 B1
7721307 Hendricks et al. May 2010 B2
7743330 Hendricks et al. Jun 2010 B1
7752258 Lewin et al. Jul 2010 B2
7861259 Barone, Jr. Dec 2010 B2
7913286 Sarachik et al. Mar 2011 B2
7958528 Moreau et al. Jun 2011 B2
7975277 Jerding et al. Jul 2011 B1
8006262 Rodriguez et al. Aug 2011 B2
8032914 Rodriguez Oct 2011 B2
8156533 Crichton Apr 2012 B2
8220018 de Andrade et al. Jul 2012 B2
8266652 Roberts et al. Sep 2012 B2
8296805 Tabatabai et al. Oct 2012 B2
8365230 Chane et al. Jan 2013 B2
8381259 Khosla Feb 2013 B1
8434109 Kamimaeda et al. Apr 2013 B2
8448208 Moreau et al. May 2013 B2
8660545 Redford et al. Feb 2014 B1
8699862 Sharifi et al. Apr 2014 B1
8793256 McIntire et al. Jul 2014 B2
8850495 Pan Sep 2014 B2
8863196 Patil et al. Oct 2014 B2
8938675 Holladay et al. Jan 2015 B2
8943533 de Andrade et al. Jan 2015 B2
8973063 Spilo et al. Mar 2015 B2
9021528 Moreau et al. Apr 2015 B2
9363560 Moreau et al. Jun 2016 B2
9473548 Chakrovorthy et al. Oct 2016 B1
9516253 De Andrade et al. Dec 2016 B2
10149014 Chane Dec 2018 B2
20010014206 Artigalas et al. Aug 2001 A1
20010027563 White et al. Oct 2001 A1
20010049823 Matey Dec 2001 A1
20010056573 Kovac et al. Dec 2001 A1
20010056577 Gordon et al. Dec 2001 A1
20020010928 Sahota Jan 2002 A1
20020016969 Kimble Feb 2002 A1
20020023270 Thomas et al. Feb 2002 A1
20020026642 Augenbraun et al. Feb 2002 A1
20020032905 Sherr et al. Mar 2002 A1
20020035573 Black et al. Mar 2002 A1
20020041104 Graf et al. Apr 2002 A1
20020042915 Kubischta et al. Apr 2002 A1
20020042920 Thomas et al. Apr 2002 A1
20020046099 Frengut et al. Apr 2002 A1
20020059094 Hosea et al. May 2002 A1
20020059586 Carney et al. May 2002 A1
20020059629 Markel May 2002 A1
20020067376 Martin et al. Jun 2002 A1
20020069407 Fagnani et al. Jun 2002 A1
20020070978 Wishoff et al. Jun 2002 A1
20020078444 Krewin et al. Jun 2002 A1
20020078449 Gordon et al. Jun 2002 A1
20020083450 Kamen et al. Jun 2002 A1
20020100041 Rosenberg et al. Jul 2002 A1
20020107973 Lennon et al. Aug 2002 A1
20020108121 Alao et al. Aug 2002 A1
20020108122 Alao et al. Aug 2002 A1
20020120609 Lang et al. Aug 2002 A1
20020124254 Kikinis Sep 2002 A1
20020144268 Khoo et al. Oct 2002 A1
20020144269 Connelly Oct 2002 A1
20020144273 Reto Oct 2002 A1
20020147645 Alao et al. Oct 2002 A1
20020152477 Goodman et al. Oct 2002 A1
20020156839 Peterson et al. Oct 2002 A1
20020156890 Carlyle et al. Oct 2002 A1
20020162120 Mitchell Oct 2002 A1
20020169885 Alao et al. Nov 2002 A1
20020170059 Hoang Nov 2002 A1
20020171691 Currans et al. Nov 2002 A1
20020171940 He et al. Nov 2002 A1
20020184629 Sie et al. Dec 2002 A1
20020188944 Noble Dec 2002 A1
20020194181 Wachtel Dec 2002 A1
20020196268 Wolff et al. Dec 2002 A1
20020199187 Gissin et al. Dec 2002 A1
20020199190 Su Dec 2002 A1
20030001880 Holtz et al. Jan 2003 A1
20030005444 Crinon et al. Jan 2003 A1
20030005453 Rodriguez et al. Jan 2003 A1
20030014752 Zaslaysky et al. Jan 2003 A1
20030014753 Beach et al. Jan 2003 A1
20030018755 Masterson et al. Jan 2003 A1
20030023970 Panabaker Jan 2003 A1
20030025832 Swart et al. Feb 2003 A1
20030028871 Wang et al. Feb 2003 A1
20030028873 Lemmons Feb 2003 A1
20030041104 Wingard et al. Feb 2003 A1
20030051246 Wilder et al. Mar 2003 A1
20030056216 Wugofski et al. Mar 2003 A1
20030056218 Wingard et al. Mar 2003 A1
20030058948 Kelly et al. Mar 2003 A1
20030061028 Dey et al. Mar 2003 A1
20030066081 Barone et al. Apr 2003 A1
20030067554 Klarfeld et al. Apr 2003 A1
20030068046 Lindqvist et al. Apr 2003 A1
20030070170 Lennon Apr 2003 A1
20030079226 Barrett Apr 2003 A1
20030084443 Laughlin et al. May 2003 A1
20030084444 Ullman et al. May 2003 A1
20030084449 Chane et al. May 2003 A1
20030086694 Davidsson May 2003 A1
20030093790 Logan et al. May 2003 A1
20030093792 Labeeb et al. May 2003 A1
20030097657 Zhou et al. May 2003 A1
20030110500 Rodriguez Jun 2003 A1
20030110503 Perkes Jun 2003 A1
20030115219 Chadwick Jun 2003 A1
20030115612 Mao et al. Jun 2003 A1
20030126601 Roberts et al. Jul 2003 A1
20030132971 Billmaier et al. Jul 2003 A1
20030135464 Mourad et al. Jul 2003 A1
20030135582 Allen et al. Jul 2003 A1
20030140097 Schloer Jul 2003 A1
20030151621 McEvilly et al. Aug 2003 A1
20030158777 Schiff et al. Aug 2003 A1
20030172370 Satuloori et al. Sep 2003 A1
20030177501 Takahashi et al. Sep 2003 A1
20030182663 Gudorf et al. Sep 2003 A1
20030189668 Newnam et al. Oct 2003 A1
20030204814 Elo et al. Oct 2003 A1
20030204846 Breen et al. Oct 2003 A1
20030204854 Blackketter et al. Oct 2003 A1
20030207696 Willenegger et al. Nov 2003 A1
20030226141 Krasnow et al. Dec 2003 A1
20030229899 Thompson et al. Dec 2003 A1
20040003402 McKenna Jan 2004 A1
20040003404 Boston et al. Jan 2004 A1
20040019900 Knightbridge et al. Jan 2004 A1
20040019908 Williams et al. Jan 2004 A1
20040022271 Fichet et al. Feb 2004 A1
20040024753 Chane et al. Feb 2004 A1
20040025180 Begeja et al. Feb 2004 A1
20040031015 Ben-Romdhane et al. Feb 2004 A1
20040031058 Reisman Feb 2004 A1
20040031062 Lemmons Feb 2004 A1
20040039754 Harple Feb 2004 A1
20040073915 Dureau Apr 2004 A1
20040078814 Allen Apr 2004 A1
20040107437 Reichardt et al. Jun 2004 A1
20040107439 Hassell et al. Jun 2004 A1
20040111465 Chuang et al. Jun 2004 A1
20040128699 Delpuch et al. Jul 2004 A1
20040133923 Watson et al. Jul 2004 A1
20040136698 Mock Jul 2004 A1
20040168186 Rector et al. Aug 2004 A1
20040172648 Xu et al. Sep 2004 A1
20040189658 Dowdy Sep 2004 A1
20040194136 Finseth et al. Sep 2004 A1
20040199578 Kapczynski et al. Oct 2004 A1
20040221306 Noh Nov 2004 A1
20040224723 Farcasiu Nov 2004 A1
20040225751 Urali Nov 2004 A1
20040226051 Carney et al. Nov 2004 A1
20050005288 Novak Jan 2005 A1
20050015796 Bruckner et al. Jan 2005 A1
20050015804 LaJoie et al. Jan 2005 A1
20050028208 Ellis et al. Feb 2005 A1
20050086172 Stefik Apr 2005 A1
20050125835 Wei Jun 2005 A1
20050149972 Knudson Jul 2005 A1
20050155063 Bayrakeri et al. Jul 2005 A1
20050160458 Baumgartner Jul 2005 A1
20050259147 Nam et al. Nov 2005 A1
20050262542 DeWeese et al. Nov 2005 A1
20050283800 Ellis et al. Dec 2005 A1
20050287948 Hellwagner et al. Dec 2005 A1
20060004743 Murao et al. Jan 2006 A1
20060059525 Jerding et al. Mar 2006 A1
20060068818 Leitersdorf et al. Mar 2006 A1
20060080707 Laksono Apr 2006 A1
20060080716 Nishikawa et al. Apr 2006 A1
20060104511 Guo et al. May 2006 A1
20060105793 Gutowski et al. May 2006 A1
20060125962 Shelton et al. Jun 2006 A1
20060143191 Cho et al. Jun 2006 A1
20060156336 Knudson et al. Jul 2006 A1
20060195865 Fablet Aug 2006 A1
20060200842 Chapman et al. Sep 2006 A1
20060206470 McIntyre Sep 2006 A1
20060206912 Klarfeld et al. Sep 2006 A1
20060233514 Weng et al. Oct 2006 A1
20060248572 Kitsukama et al. Nov 2006 A1
20070019001 Ha Jan 2007 A1
20070050343 Siddaramappa et al. Mar 2007 A1
20070064715 Lloyd et al. Mar 2007 A1
20070083538 Roy et al. Apr 2007 A1
20070112761 Xu et al. May 2007 A1
20070211762 Song et al. Sep 2007 A1
20070214123 Messer et al. Sep 2007 A1
20070214488 Nguyen et al. Sep 2007 A1
20070220016 Estrada et al. Sep 2007 A1
20070239707 Collins et al. Oct 2007 A1
20070250901 McIntire et al. Oct 2007 A1
20070260700 Messer Nov 2007 A1
20070261072 Boulet et al. Nov 2007 A1
20070271587 Rowe Nov 2007 A1
20080037722 Klassen Feb 2008 A1
20080060011 Kelts Mar 2008 A1
20080071770 Schloter et al. Mar 2008 A1
20080092201 Agarwal et al. Apr 2008 A1
20080113504 Lee et al. May 2008 A1
20080126109 Cragun et al. May 2008 A1
20080133504 Messer et al. Jun 2008 A1
20080148317 Opaluch Jun 2008 A1
20080163304 Ellis Jul 2008 A1
20080183681 Messer et al. Jul 2008 A1
20080183698 Messer et al. Jul 2008 A1
20080189740 Carpenter et al. Aug 2008 A1
20080196070 White et al. Aug 2008 A1
20080204595 Rathod et al. Aug 2008 A1
20080208796 Messer et al. Aug 2008 A1
20080208839 Sheshagiri et al. Aug 2008 A1
20080221989 Messer et al. Sep 2008 A1
20080235209 Rathod et al. Sep 2008 A1
20080235393 Kunjithapatham et al. Sep 2008 A1
20080235725 Hendricks Sep 2008 A1
20080250010 Rathod et al. Oct 2008 A1
20080256097 Messer et al. Oct 2008 A1
20080266449 Rathod et al. Oct 2008 A1
20080276278 Krieger et al. Nov 2008 A1
20080282294 Carpenter et al. Nov 2008 A1
20080288641 Messer et al. Nov 2008 A1
20080288644 Gilfix et al. Nov 2008 A1
20080301320 Morris Dec 2008 A1
20080301732 Archer et al. Dec 2008 A1
20080317233 Rey et al. Dec 2008 A1
20090006315 Mukherjea et al. Jan 2009 A1
20090019485 Ellis et al. Jan 2009 A1
20090024629 Miyauchi Jan 2009 A1
20090025054 Gibbs et al. Jan 2009 A1
20090083257 Bargeron et al. Mar 2009 A1
20090094113 Berry et al. Apr 2009 A1
20090094632 Newnam et al. Apr 2009 A1
20090094651 Damm et al. Apr 2009 A1
20090123021 Jung et al. May 2009 A1
20090133025 Malhotra et al. May 2009 A1
20090164904 Horowitz et al. Jun 2009 A1
20090183210 Andrade Jul 2009 A1
20090222872 Schlack Sep 2009 A1
20090228441 Sandvik Sep 2009 A1
20090240650 Wang et al. Sep 2009 A1
20090249427 Dunnigan et al. Oct 2009 A1
20090271829 Larsson et al. Oct 2009 A1
20090288132 Hegde Nov 2009 A1
20090292548 Van Court Nov 2009 A1
20100023966 Shahraray et al. Jan 2010 A1
20100077057 Godin et al. Mar 2010 A1
20100079670 Frazier et al. Apr 2010 A1
20100175084 Ellis et al. Jul 2010 A1
20100180300 Carpenter et al. Jul 2010 A1
20100223640 Reichardt et al. Sep 2010 A1
20100250190 Zhang et al. Sep 2010 A1
20100251284 Ellis et al. Sep 2010 A1
20100257548 Lee et al. Oct 2010 A1
20110055282 Roving Mar 2011 A1
20110058101 Earley et al. Mar 2011 A1
20110087348 Wong Apr 2011 A1
20110093909 Roberts et al. Apr 2011 A1
20110131204 Bodin et al. Jun 2011 A1
20110176787 DeCamp Jul 2011 A1
20110209180 Ellis et al. Aug 2011 A1
20110211813 Marks Sep 2011 A1
20110214143 Rits et al. Sep 2011 A1
20110219386 Hwang et al. Sep 2011 A1
20110219419 Reisman Sep 2011 A1
20110225417 Maharajh et al. Sep 2011 A1
20110246495 Mallinson Oct 2011 A1
20110247042 Mallinson Oct 2011 A1
20120002111 Sandoval et al. Jan 2012 A1
20120054811 Spears Mar 2012 A1
20120117151 Bill May 2012 A1
20120192226 Zimmerman et al. Jul 2012 A1
20120227073 Hosein et al. Sep 2012 A1
20120233646 Coniglio et al. Sep 2012 A1
20120295686 Lockton Nov 2012 A1
20120324002 Chen Dec 2012 A1
20120324494 Burger et al. Dec 2012 A1
20120324495 Matthews, III et al. Dec 2012 A1
20120324518 Thomas et al. Dec 2012 A1
20130014155 Clarke et al. Jan 2013 A1
20130040623 Chun et al. Feb 2013 A1
20130051770 Sargent Feb 2013 A1
20130103446 Bragdon et al. Apr 2013 A1
20130110769 Ito May 2013 A1
20130111514 Slavin et al. May 2013 A1
20130170813 Woods et al. Jul 2013 A1
20130176493 Khosla Jul 2013 A1
20130198642 Carney et al. Aug 2013 A1
20130262997 Markworth et al. Oct 2013 A1
20130298038 Spivack et al. Nov 2013 A1
20130316716 Tapia et al. Nov 2013 A1
20130326570 Cowper et al. Dec 2013 A1
20130332839 Frazier et al. Dec 2013 A1
20130332852 Castanho et al. Dec 2013 A1
20130347018 Limp et al. Dec 2013 A1
20130347030 Oh et al. Dec 2013 A1
20140006951 Hunter Jan 2014 A1
20140009680 Moon et al. Jan 2014 A1
20140032473 Enoki et al. Jan 2014 A1
20140068648 Green et al. Mar 2014 A1
20140075465 Petrovic et al. Mar 2014 A1
20140089423 Jackels Mar 2014 A1
20140089967 Mandalia et al. Mar 2014 A1
20140129570 Johnson May 2014 A1
20140149918 Asokan et al. May 2014 A1
20140150022 Oh et al. May 2014 A1
20140237498 Ivins Aug 2014 A1
20140267931 Gilson et al. Sep 2014 A1
20140279852 Chen Sep 2014 A1
20140280695 Sharma et al. Sep 2014 A1
20140282122 Mathur Sep 2014 A1
20140325359 Vehovsky et al. Oct 2014 A1
20140327677 Walker Nov 2014 A1
20140359662 Packard et al. Dec 2014 A1
20140365302 Walker Dec 2014 A1
20140373032 Merry et al. Dec 2014 A1
20150026743 Kim et al. Jan 2015 A1
20150263923 Kruglick Sep 2015 A1
Foreign Referenced Citations (23)
Number Date Country
0624039 Nov 1994 EP
0963115 Dec 1999 EP
1058999 Dec 2000 EP
1080582 Mar 2001 EP
2323489 Sep 1998 GB
2448874 Nov 2008 GB
2448875 Nov 2008 GB
9963757 Dec 1999 WO
2000011869 Mar 2000 WO
0033576 Jun 2000 WO
0110115 Feb 2001 WO
0182613 Nov 2001 WO
02063426 Aug 2002 WO
02063471 Aug 2002 WO
02063851 Aug 2002 WO
02063878 Aug 2002 WO
03009126 Jan 2003 WO
2003026275 Mar 2003 WO
2007115224 Oct 2007 WO
2008053132 May 2008 WO
2011053271 May 2011 WO
2012094105 Jul 2012 WO
2012154541 Nov 2012 WO
Non-Patent Literature Citations (45)
Entry
Fernando Pereira, “The MPEG-4 Book”, Prentice Hall, Jul. 10, 2002.
Michael Adams, “Open Cable Architecture”, Cisco Press, Dec. 3, 1999.
Andreas Kraft and Klaus Hofrichter, “An Approach for Script-Based Broadcast Application Production”, Springer-Verlag Berlin Heidelberg, pp. 74-82, 1999.
Mark Riehl, “XML and Perl”, Sams, Oct. 16, 2002.
MetaTV, Inc., PCT/US02/29917 filed Sep. 19, 2002, International Search Report dated Apr. 14, 2003; ISA/US; 6 pages.
Sylvain Devillers, “Bitstream Syntax Definition Language: an Input to MPEG-21 Content Representation”, Mar. 2001, ISO, ISO/IEC JTC1/SC29/WG11 MPEG01/M7053.
Shim, et al., “A SMIL Based Graphical Interface for Interactive TV”, Internet Tech. Laboratory Dept. of Comp. Engineering, San Jose State University, pp. 257-266, 2003.
Yoon, et al., “Video Gadget: MPET-7 Based Audio-Visual Content Indexing and Browsing Engine”, LG Electronics Institute of Technology, 2001, pp. 59-68.
Watchwith webpage; http://www.watchwith.com/content_owners/watchwith_plalform_components.jsp (last visited Mar. 12, 2013).
Matt Duffy; TVplus App reveals content click-through rates north of 10% across sync enabled programming; http://www.tvplus.com/blog/TVplus-App-reveals-content-click-through-rates-north-of-10-Percent-across-sync-enabled-programming (retrieved from the Wayback Machine on Mar. 12, 2013).
“In Time for Academy Awards Telecast, Companion TV App Umami Debuts First Real-Time Sharing of a TV Program's Images”; Umami News; http:www.umami.tv/2012-02-23.html (retrieved from the Wayback Machine on Mar. 12, 2013).
European Patent Application No. 09175979.5—Office Action dated Dec. 13, 2011.
Canadian Patent Application No. 2,685,833—Office Action dated Jan. 20, 2012.
Li, Y. et al. “Reliable Video Clock Time Recognition”, Pattern Recognition, 2006, 1CPR 1006, 18th International Conference on Pattern Recognition, 4 pages.
European Search Report dated Mar. 1, 2010.
Salton et al., Computer Evaluation of Indexing and Text Processing Journal of the Association for Computing Machinery, vol. 15, No. 1, Jan. 1968, pp. 8-36.
Smith, J.R. et al., An Image and Video Search Engine for the World-Wide Web Storage and Retrieval for Image and Video Databases 5, San Jose, Feb. 13-14, 1997, Proceedings of Spie, Belingham, Spie, US, vol. 3022, Feb. 13, 1997, pp. 84-95.
Kontothoanassis, Ledonias et al. “Design, Implementation, and Analysis of a Multimedia Indexing and Delivery Server”, Technical Report Series, Aug. 1999, Cambridge Research Laboratory.
Messer, Alan et al., “SeeNSearch: A context Directed Search Facilitator for Home Entertainment Devices”, Paper, Samsung Information Systems America Inc., San Jose, CA, 2008.
Boulgouris N. V. et al., “Real-Time Compressed-Domain Spatiotemporal Segmentation and Ontologies for Video Indexing and Retrieval”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 14, No. 5, pp. 606-621, May 2004.
Changsheng Xu et al., “Using Webcast Text for Semantic Event Detection in Broadcast Sports Video”, IEEE Transactions on Multimedia, vol. 10, No. 7, pp. 1342-1355, Nov. 2008.
Liang Bai et al., “Video Semantic Content Analysis based on Ontology”, International Machine Vision and Image Processing Conference, pp. 117-124, Sep. 2007.
Koskela M. et al., “Measuring Concept Similarities in Multimedia Ontologies: Analysis and Evaluations”, IEEE Transactions on Multimedia, vol. 9, No. 5, pp. 912-922, Aug. 2007.
Steffan Staab et al., “Semantic Multimedia”, Reasoning Web; Lecture Notes in Computer Science, pp. 125-170, Sep. 2008.
European Search Report for Application No. 09180776.8, dated Jun. 7, 2010, 9 pages.
European Search Report, EP 09 18 0762, completion date Mar. 22, 2010.
European Search Report dated Jun. 4, 2010.
EP Application No. 09 179 987.4-1241—Office Action dated Feb. 15, 2011.
European Application No. 09 175 979.5—Office Action dated Apr. 11, 2011.
Boronat F et al: “Multimedia group and inter-stream synchronization techniques: A comparative study”, Information Systems. Pergamon Press. Oxford. GB. vol. 34. No. 1. Mar. 1, 2009 (Mar. 1, 2009). pp. 108-131. XP025644936.
Extended European Search Report—EP14159227.9—dated Sep. 3, 2014.
Canadian Office Action—CA 2,685,833—dated Jan. 22, 2015.
European Extended Search Report—EP 13192112A—dated May 11, 2015.
CA Response to Office Action—CA Appl. 2,685,833—Submitted Jul. 17, 2015.
Response to European Office Action—European Appl. 13192112.4—submitted Dec. 9, 2015.
CA Office Action—CA App 2,685,833—dated Jan. 27, 2016.
European Office Action—EP App 14159227.9—dated Jul. 12, 2016.
Agnieszka Zagozdzinnska et al. “Tridaq Systems in HEP Experiments at LHC Accelerator” Kwartalnik Elektroniki I Telekomunikacji, vol. 59, No. 4, Oct. 2013.
CA Office Action—CA Application 2685833—dated Feb. 8, 2017.
Nov. 29, 2017—Canadian Office Action—CA 2,685,833.
Feb. 19, 2018—European Summons to Oral Proceedings—EP 14159227.9.
Mar. 9, 2018—European Office Action—EP 13192112.4.
Jul. 31, 2018—European Decision to Refuse—14159227.9.
Sep. 5, 2019—Canadian Office Action—CA 2,685,833.
Nov. 6, 2019—Canadian Office Action—CA 2,832,800.
Related Publications (1)
Number Date Country
20190158923 A1 May 2019 US
Provisional Applications (2)
Number Date Country
60395277 Jul 2002 US
60323658 Sep 2001 US
Divisions (1)
Number Date Country
Parent 10247901 Sep 2002 US
Child 13736283 US
Continuations (1)
Number Date Country
Parent 13736283 Jan 2013 US
Child 16160375 US