Set top box health and configuration

Information

  • Patent Grant
  • 10015483
  • Patent Number
    10,015,483
  • Date Filed
    Friday, June 30, 2017
    7 years ago
  • Date Issued
    Tuesday, July 3, 2018
    6 years ago
Abstract
Disclosure is directed to determining a condition of the television receiver and communicating information regarding that condition to end-points that are external to the television receiver. The television receiver may create a two-dimensional barcode with information regarding the receiver condition and output the two-dimensional barcode for display on a display device. Once displayed on the display device, the two-dimensional barcode may be scanned into a mobile device where the information regarding the receiver condition may be stored, processed and/or forwarded to another location or device.
Description
TECHNICAL FIELD

Embodiments discussed herein are generally directed to determining a condition of the television receiver and communicating information regarding that condition by a two-dimensional barcode that is displayed on a display device associated with the television receiver.


BACKGROUND

Cable television, satellite television, and other providers of program services typically provide customers with a television receiver that is adapted to receive and output program service content. The television receiver (one example of which is a set-top box) may, from time to time, malfunction or fail due to aging, normal wear and tear, or other causes. These malfunctions or failures may be an inconvenience to customers whose service may be interrupted while the television receiver is being repaired or replaced. In some cases, a customer may program his or her television receiver with recoding timers or certain customizable settings that may be lost when the receiver malfunctions or fails. Accordingly, there is a need to provide a user with a mechanism for anticipating and/or mitigating the effects of receiver failure or malfunction.


SUMMARY

Embodiments discussed herein are generally directed to determining a condition of the television receiver and communicating information regarding that condition to end-points that are external to the television receiver. The receiver condition may be, for example, an error condition encountered in the operation of the receiver, or a receiver configuration or setting programmed by the user. The television receiver may create a two-dimensional barcode with information regarding the receiver condition and output the two-dimensional barcode for display on a display device. Once displayed on the display device, the two-dimensional barcode may be scanned into a mobile device where the information regarding the receiver condition may be stored, processed and/or forwarded to another location or device. For error conditions, the mobile device may forward the two-dimensional barcode or information encoded in the two-dimensional barcode to an appropriate party who may effect repairs. For receiver configurations or settings, the mobile device may store the configuration or settings for later user by the television receiver.


One embodiment is directed to a method of managing a television receiver, comprising: sensing a condition of a television receiver; in response to sensing the condition of the television receiver, creating a two-dimensional barcode with information corresponding to the condition of the television receiver; and outputting the two-dimensional barcode form the television receiver for display on a display device; wherein, when the two-dimensional barcode is scanned from the display device into a mobile device, the two-dimensional barcode conveys the information corresponding to the condition of the television receiver to the mobile device.


Another embodiment is directed to a television receiver, comprising: a tuner operable to receive a program service transmission having a plurality of channels, the tuner operable to select one of the channels and to prepare the channel to be output in a data signal to a display device; a memory connected to the tuner, the memory including a tangible storage medium operable to store computer readable data and instructions; a processor connected to the memory operable to run computer executable code stored in the memory; a sensing module configured to execute on the processor to sense a condition of a television receiver; an encoding module configured to execute on the processor to create a two-dimensional barcode with information corresponding to the condition of the television receiver; and an output module configured to execute on the processor to output the two-dimensional barcode from the television receiver for display on the display device; wherein, when the two-dimensional barcode is scanned from the display device into a mobile device, the two-dimensional barcode conveys the information corresponding to the condition of the television receiver to the mobile device.


Another embodiment is directed to a method, comprising: scanning a two-dimensional barcode into a mobile device, the two-dimensional barcode being displayed on a display device associated with television receiver; and transmitting information regarding the error condition from the mobile device to a service provider; wherein the service provider utilizes the information regarding the error condition to initiate repair of the television receiver.


Another embodiment is directed to a method, comprising: scanning a two-dimensional barcode into a mobile device, the two dimensional barcode being displayed on a display device associated with television receiver; decoding the two-dimensional barcode at the mobile device to determine one or more configuration settings for the television receiver; and storing the one or more configuration settings for later use in reconfiguring the television receiver.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic diagram showing a sample system, components and implementations discussed herein;



FIG. 2 is a flow chart that illustrates a method executed by the television receiver illustrated in FIG. 1;



FIG. 3A-3C are illustrations of outputs shown on a display device in connection with the operation of the method illustrated in FIG. 2;



FIG. 4 is a flow chart that illustrates a method executed by the mobile device illustrated in FIG. 1;



FIG. 5 is a flow chart that illustrates another method executed by the television receiver illustrated in FIG. 1;



FIG. 6 is a flow chart that illustrates another method executed by the mobile device illustrated in FIG. 1;



FIG. 7A-7B are illustrations of output shown on a display device in connection with the operation of the method illustrated in FIG. 5; and



FIG. 8 is a flow chart that illustrates another method executed by the television receiver illustrated in FIG. 1.





DETAILED DESCRIPTION


FIG. 1 is a schematic illustration of a general operating environment showing components and features of embodiments discussed herein. FIG. 1 includes a television receiver, generally identified by reference numeral 108. Embodiments discussed herein are generally directed to determining a condition of the television receiver 108 and communicating information regarding that condition to end-points that are external to the television receiver 108. In certain instances, communicating information regarding receiver conditions is directed towards mitigating the inconveniences associated with malfunctions and/or failures of the television receiver 108. For example, in order to expedite repairs of the television receiver 108, the receiver may communicate information regarding error conditions that are encountered or otherwise sensed in the receiver's operations. In order to facilitate recovery from a catastrophic failure, the television receiver 108 may communicate information regarding settings or configurations that have been programmed by the user. The television receiver 108 may also communicate information regarding user-programmed configurations or settings in order to facilitate transfer of these configurations or settings into a different television receiver 108.


The receiver 108 depicted in FIG. 1 may be configured to communicate with or receive signals from a service provider 104 that broadcasts, transmits, or otherwise delivers a content service to a receiver 108. The receiver 108 can include a set-top box (STB), a digital video recorder (DVR), a cable receiver, a general purpose computing device, and so on. The receiver 108 may also include a cable modem that receives streaming video and/or audio. Generally, a “receiver” may be any device capable of receiving video and/or audio content included in a broadcast or other content service transmission from a service provider 104.


The receiver 108 may be associated with an individual, business or other entity, user or subscriber that receives a content service transmission from the service provider 104. Generally the terms “user” and/or “subscriber” refer to an individual or company who receives a content service transmission. This may include those who have purchased a subscription to the content service transmission. Alternatively or additionally, the terms “user” and/or “subscriber” may refer to individuals who have been given access to the content service transmission through promotional offers and/or other non-fee-based agreements.


As used herein, a “service provider” may include any service that provides a content transmission to a receiver 108 such as, without limitation, a satellite television service, a direct television service or a cable television service, or a streaming video delivered across a network such as the Internet. Accordingly, a “content service transmission” encompasses transmission of information across a cable network (for example from a cable headend to a cable receiver), an Internet or other computer-accessible medium (including a local area network, wide-area network, and so on), including Internet protocol television transmissions, a wireless network such as a radio frequency or infrared network, and so on.


In connection with embodiments that operate in the context of a satellite television service, the service provider 104 may provide a content service transmission through an uplink center. Such a satellite television service may utilize a direct broadcast satellite (DBS) system, which can incorporate packetized transmission signals according to an appropriate standard, such as the MPEG-2 and/or MPEG-4 standards. The uplink center may include a transmitter or other equipment operable to transmit a modulated signal having data representing audio and/or visual content. The modulated signal may be received at a satellite, which in turn retransmits the modulated signal to be received at one or more terrestrial locations. The retransmitted signal may be received from the satellite at one or more satellite dishes, which are typically associated with one or more receivers 108. In connection with embodiments that operate in the context of a cable television service, the service provider 104 may provide a content service transmission to a headend, which, in turn, delivers the content service transmission to the receiver 108.


The receiver 108 may include a tuner 124 operable to receive the content service transmission signal from the service provider 104 and a decoder 128 to decode the received signal. The decoder 128 may be programmed to decrypt, demodulate, demultiplex or otherwise decode some or all of the received signals in accordance with purchases and selections made by a user. Output from the decoder 128 may be directed to an audio visual (A/V) processing module or other signal output portion, which may process the video and audio streams using digital-to-analog conversion techniques, or compressed digital to uncompressed digital conversion techniques, to produce one or more output signals. The output signals may be sent to a display device 140, such as a television or monitor in order to display content to a user.


The receiver 108 may include or be associated with a recorder 136, such as a digital video recorder (DVR). The recorder 136 may be integrated into the receiver 108 or may be a stand alone device. The recorder 136 may be operated by a user who programs the device or receiver function to record a particular program at a specified time. When the program occurs, the recorder 136 will record and store the program, which can then be viewed later. In addition to this functionality, the recorder 136 may buffer a certain amount of content during a live transmission. Buffering a live transmission allows a user to pause and/or rewind the content of the transmission and to then display the content in a non-live or delayed manner.


The receiver 108 may include or be associated with a memory or other storage device 152, such as magnetic or optical storage. The storage device 152 may be operable to store data received from the decoded content transmission signal. The storage device 152 may be volatile or non-volatile memory implemented using any suitable technique or technology such as, for example, random access memory (RAM), disk storage, flash memory, solid state and so on. The storage device 152 may be located within the receiver 108 or separately from the receiver 108. The storage device 152 may removable. The stored data set may include audio and/or visual content to be transmitted and output through a display device, such as a television or monitor. Generally, audio/visual content may include still images, video images, animation and/or audio. Portable Network Graphics (PNG) or other appropriate formats, such as for example, Tagged Image File Format (TIFF), Joint Photographic Experts Group (JPEG), Motion Picture Experts Group (MPEG)-2, MPEG-4 may be used to display an image or video.


The receiver 108 may additionally include a processor 132 operable to run executable code in connection with various functions associated with the receiver 108. For example, the processor 132 may display graphics, images, animations or other content through an output device, such as a television or monitor. The storage device 152 may store an application, file, or other data that is useable by the processor 132. As used herein, an application includes processor executable code that may be run to carry out one or more functions associated with the receiver 108. “Processor executable code” includes any computer-readable media or commands that may be ultimately interpreted by a processor, such as HTML or XML files that are rendered into user-viewable applications by an application executed by the processor 132.


The processor 132 may also perform such tasks as executing commands received from a user. User commands may be sent to the receiver 108 through a user input device 144 such as remote or other wireless device. As used herein, a “user input device” may include any device operable to receive input from a user and to convey the input to the receiver 108. In one embodiment, the user input device 144 may be a hand-held device having a number of buttons or keys that when actuated by a user cause the user input device to covey information to the receiver 108 using a suitable communication means, such as an infrared signal. The user input device 144 may include a pointing device or functionality that allows the user to control the position of a cursor that is displayed on an output device. For example, the user input device 144 may include a track ball or glide plane that may be manipulated to control cursor movements. The user input device 144 may include a motion sensor or accelerometer that allows a user to control displayed items or graphics, such as a cursor, through movements of his or her hand or arm that cause a displacement of the user input device 144. It should be appreciated that other input devices such as a computer mouse or touch screen may be used and other communication means, wired or wireless, may be used. Information sent to the receiver 108 may include, for example, a command to change the output channel. Commands sent to the receiver 108 may be entered through a dedicated display menu.


The storage device 152 may store units of processor executable code in the form of one or more modules configured to implement certain functions described herein. Embodiments discussed herein are directed to determining a condition of a television receiver 104 and communicating that condition to a mobile device 156. In determining the condition of the television receiver 104, the storage device 152 may include a condition sensing module 110 that is configured to detect a receiver condition such as an error that occurs during the operation of the receiver 104. In some instances, receiver condition sensed by the condition sensing module 110 may be a user-configurable setting such as a favorite channel list or recordings timers.


In communicating with the mobile device 156, the television receiver 108 typically encodes a two-dimensional barcode with at least information regarding the receiver condition and outputs the two-dimensional barcode for display on a display device 140. As used herein, a “two-dimensional barcode” is a machine readable representation of data that includes a pattern of bars, squares, dots, hexagons or other geometric forms. The geometric forms are arranged within the two-dimensional barcode to form a message according to a predefined symbology that defines the meaning of various symbols. In some implementations, the two-dimensional barcode may arrange symbols in a grid or matrix. In other implementations, the symbols of the two-dimensional barcode may be arranged in a circular patterns. In still other implementations, steganography techniques may be used to encode symbols within an image in such a way that a viewer is not aware of the symbols. One example of a two-dimensional barcode is QR code, developed by the Denso-Wave corporation of Japan. Another example of a two-dimensional barcode is Aztec code, developed by Andrew Longacre, Jr. and Robert Hussey.


The receiver 104 may include an encoding module 112 configured to create a two-dimensional barcode with one or more codes corresponding to the receiver condition sensed by the condition sensing module 110. Once the encoding module 112 create the appropriate information in a two-dimensional barcode, an output module 116 may output the two-dimensional barcode from the television receiver 104 for display on the display device 140. Once displayed on the display device 140, the two-dimensional barcode may be scanned into the mobile device 156 where the information regarding the receiver condition may be stored, processed and/or forwarded to another location or device. In accordance with certain embodiments, the output module 116 may be configured to output the two-dimensional barcode to a printer device. For example, the output module 116 may print the two-dimensional barcode if a television or other display device 140 is unavailable (e.g., the display device 140 may be broken or malfunctioning). Once the two-dimensional barcode is printed to a piece of paper, the mobile device 156 may scan the paper to aquire the two-dimensional bar code.


Generally, the mobile device 156 may include a processor 164 operable to load and store various processor-executable modules stored in the storage device 160. The mobile device 148 may additionally include an input device, such a keypad or keyboard, and an output device, such as an LED or LCD screen. In addition to processor-executable code, the storage device 160 may store data associated with two-dimensional barcodes that are scanned into the mobile device 156 from the receiver 108. In this regard, the mobile device 156 may include a scanner 168 that is operable to scan or otherwise read a two-dimensional barcode. In one embodiment, the scanner 168 may be a photo-sensor or laser scanner that operates by sweeping a beam of light across a surface that displays the two-dimensional barcode. In other embodiments, the scanner 168 may be charge-coupled device, such a digital camera, that is operable to capture an image of the two-dimensional barcode. In connection with capturing an image of the two-dimensional barcode, the mobile device 156 may be configured to extract information contained in the two-dimensional barcode using digital signal processing or other appropriate techniques.


The mobile device 156 may be operable to transfer content across a network 176 to a database 180 or to the service provider 104. Network transactions may be conducted through the operation of a network interface 172, such as a modem network interface card, cable plug or jack, and so on. Using the interface, the mobile device 156 may communicate over a network 176, such as a packet switched network or a circuit switched network, one example of which is the public switched telephone network (PSTN). It should be appreciated that the network 176 may be any type of network capable of data communication, such as, for example, a local or wide area network or the Internet. The mobile device 156 may communicate through a network connection using any suitable communication protocol such as TCP/IP.


The type of receiver condition that is communicated by the television receiver 104 may determine the type of action taken by the mobile device 156 once the mobile device 156 scans the two-dimensional barcode. In one instance, the television receiver 104 may communicate an error condition that has occurred or is occurring in the operations of the television receiver 104. In this instance, the mobile device 156 may forward the two-dimensional barcode or information encoded in the two-dimensional barcode to an appropriate party who may effect repairs. In another instance, the television receiver 104 may communicate one or more configuration settings for the television receiver or a user input device 144 associated with the television receiver 108. In this instance, the mobile device 156 may store the configuration settings for later user by the television receiver 108.


The receiver 108 may additionally include a configuration module 120 that may be utilized in those instances where configuration settings for the television receiver 108 or the user input device 144 are communicated to the mobile device 156. Specifically, the configuration module 120 may re-configure the television receiver 108 and/or the user input device 144 according to settings that have been previously stored by the mobile device 156. For example, the configuration module 120 may be utilized to recover from a catastrophic failure by reloading configuration settings lost in the failure. Here, the configuration module may connect to a storage medium used by the mobile device 156 to store configuration settings and then download the stored settings for use in reconfiguring the television receiver 108 or the user input device 144.



FIG. 2 is a flow chart that illustrates a method of communicating an error condition that is present in a television receiver 108. Initially, in operation 204, the condition sensing module 110 receives an indication of an error condition occurring in the television receiver 108. The error condition may be, for example, the failure of a particular part or component of the television receiver 108. Following operation 204, operation 208 may be executed.


In operation 208, the condition sensing module 110 determines an error code that corresponds to the error condition sensed in operation 204. The condition sensing module 110 may be configured to include a list or table of various error codes that classify the various errors which are known to occur in a television receiver 108. The error codes used in operation 208 may be used to facilitate rapid and accurate recognition of the error condition by technicians or other personnel who will be called upon to repair the television receiver 108. The error codes may be specific to a particular service provider who provides content service transmission to the television receiver 108. Alternatively, the error codes may be a generally recognized scheme of error classification recognized by various service providers within the cable and/or satellite television industry. Following operation 208, operation 212 may be executed.


In operation 212, the encoding module 112 creates a two-dimensional barcode with the error code corresponding to the error condition as determined or classified in operation 208. Following operation 212, operation 216 may be executed. In operation 216, the encoding module 112 may incorporate routing information into the two-dimensional barcode. Following operation 216, operation 220 may be executed. In operation 220, the output module 116 may output the two-dimensional barcode which is created in operations 212 and 216 from the television receiver 108 for display on the display device 140. The routing information incorporated in operation 216 may be used by the mobile device 156 in order to properly direct or otherwise route the error code information from the mobile device to the service provider 104. In one embodiment, the routing information may include a uniform resource locater (URL), which may be used by the mobile device 156 to access a particular website through a browser or other web interface operating on the mobile device 156. Through the website accessed through the URL, the mobile device 156 may enter an error code such that the error code is transmitted across the network 176 to be received by the service provider 104.



FIGS. 3A-3C are illustrations of output mechanisms such as graphical user interfaces and/or prompts that may be used to communicate an error condition from a television receiver using a two-dimensional barcode. FIG. 3A includes a graphical illustration of output shown or visible on the display device 140. In FIG. 3A, the display device 140 shows program content 304, which is received by the television receiver 108 from the service provider 104 and output on the display device 140. Additionally in FIG. 3A, the display device 140 shows a “pop up” or icon 308 that is overlaid on top of the program content 304. The icon 308 includes an error message that informs the user or viewer that the television receiver 108 has detected an internal error or error condition. The icon 308 may include a message that specifies the manner in which the user may take action to learn more about the error condition and/or to mitigate the error condition. For example, the icon 308 may include a message that says, for example, press info or select to receive more information.



FIG. 3B is an illustration of output from the display device 140 that may be displayed in response to a user selecting or otherwise responding to the icon displayed in FIG. 3A. The menu displayed in FIG. 3B may be a graphical user interface or other menu that takes up a substantial portion of the display device 140. In certain embodiments, a portion of the surface area of the display device 140 may be reserved for a scaled down or smaller window that displays the normal program content. As shown in FIG. 3B, the window 312 includes program contents otherwise visible during display of normal programming. The error detection menu 314 shown in FIG. 3B may include error message 316 that provides more detailed information concerning the error detected within the television receiver 108. (It should be appreciated that in certain embodiments all information related to the error condition may displayed to the user through one or more messages that are overlaid over broadcast content.) Additionally, the message 316 may include directions or other instructions for a user to scan a two-dimensional barcode 320 that is displayed on the display device 140. As discussed in connection with FIG. 2, the two-dimensional barcode 320 may include an encoding of an error code corresponding to the error condition and/or routing information usable to direct the error code to the proper service provider 104 through network transactions initiated across the network 176. In connection with the display of the two-dimensional barcode 320, a user may scan the two-dimensional barcode 320 into his or her mobile device 156. As described above, scanning the two-dimensional barcode 320 may include photographing the two-dimensional barcode 320 using a camera or scanner 168 incorporated into the mobile device 156.


In certain instances, the error condition sensed in operation 204 may be sufficiently critical to the function of the television receiver 108 that the condition sensing module 110 may determine that a failure of the television receiver is imminent. In such circumstances, the output module 116 may be configured to stop other programming content from being output to the display device 140 and instead output only a two-dimensional barcode encoded with the error condition determined in operation 204. Such an output is shown in FIG. 3C. As shown in FIG. 3C, the display device 140 includes a two-dimensional barcode 320 displayed by itself in a central location on the display device 140. In response to such a display, the user may scan or photograph the two-dimensional barcode 320 and communicate the information contained therein to the service provider 104, as appropriate.



FIG. 4 is a flow chart that illustrates a method of receiving the two-dimensional barcode 320 from a television receiver 108 into a mobile device 156 and for communicating the information contained therein to the service provider 104. Initially, in operation 404, the mobile device 156 may scan the displayed two-dimensional barcode 320. In operation 404 a specifically designed two-dimensional barcode scanner or a camera may scan or photograph the two-dimensional barcode as it is displayed on the display device 104. Following operation 404 operation, 408 may be executed.


In operation 408, the mobile device 156 may connect to the service provider 104 using routing information contained in the two-dimensional barcode received in operation 408. Connecting to the service provider may include initiating a network transaction across the network 176 using TCP/IP protocol or other appropriate methods of network communication. Additionally, operation 408 may include formatting a message which is to be sent across the network 176. Such messages may include e-mail or MMS-type messaging formats. Following operation 408, operation 412 may be executed. In operation 412, the mobile device 156 may transmit error codes from the mobile device to the service provider 104. Alternatively or in combination, operation 412 may include transmitting a scanned image of the two-dimensional barcode such as in an email or MMS message.



FIG. 5 is a flow chart that illustrates a method of communicating a configuration setting associated with a television receiver 108. Initially, in operation 504, the condition sensing module 110 senses the state or other configuration of the television receiver 108. As used herein, the configuration of the television receiver may include such user-programmable or adjustable settings as favorite channels, specific recording timers, and so on. Operation 504 may additionally include sensing a configuration of a programmable remote control or other input device 144. Specifically, the input device 144 may include certain buttons or other input mechanisms that are programmed by a user to have a specific function. Such user-programmed characteristics of the television receiver 108 and/or remote 144 are sensed and recorded by the condition sensing module 110 in operation 504. Following operation 504, operation 508 may be executed.


In operation 508, the encoding module 112 may create the television receiver 108 and/or remote 144 configuration into a two-dimensional barcode. Following operation 508, operation 512 may be executed. In operation 512, the output module 116 may output the two-dimensional barcode created in operation 508 from the television receiver 108 to the display device 140.



FIG. 6 is a flow chart that illustrates a method of receiving configuration information and storing that configuration information by a mobile device 176. Initially, in operation 604, the mobile device 156 scans the two-dimensional barcode into the mobile device 156. Operation 604 may include the use of a scanner or camera 168 that scans or photographs the two-dimensional barcode as it is displayed on the display device 140. Following operation 604, operation 608 may be executed.


In operation 608, the mobile device 156 may connect to a storage medium. The storage medium in one embodiment may include the storage device 160 incorporated internally into the mobile device 156. In other embodiments, the storage medium may be a database 180 that is accessed by the mobile device 156 through a network transaction occurring across the network 176. Following operation 608, operation 612 may be executed. In operation 612, the mobile device 156 may store the configuration information in the storage medium accessed in operation 608.



FIGS. 7A and 7B are illustrations of graphical user interfaces or menus used to program various features of the television receiver 108 and/or remote control 144, and to store those configurations in a location that is separate from the television receiver 104. FIG. 7a includes a configuration menu 708 that includes a window 712 which includes a scaled or reduced version of the program content typically output from the display device 140. The configuration menu 720 additionally includes a number of sub menus that may be used to program various features of the television receiver 104. For example, the configuration menu 702 may include favorite channels sub-menu 704 that may be used to compile a list of channels that the user typically views during the day. The favorite channels menu 704 may be accessed when needed to view the user's favorite channels in an efficient and speedy manner. The configuration menu 702 may additionally include a recording timers sub-menu 708. The recording timers sub-menu 708 may include a listing of certain events or programs that the user has programmed or scheduled to be recorded by the recorder 136. The configuration menu 702 may additionally include a remote settings sub-menu 716. The remote settings sub-menu 716 may be used to program certain buttons or other input mechanisms on the user input device 144 with specific user defined functionality. The configuration menu 702 may additionally include a message 724 that instructs the user to select a certain button or enter a certain command to generate a two-dimensional barcode that will preserve the various settings established through the use of the configuration menu 702.



FIG. 7B includes a subsequent menu or screen utilized by the configuration menu 702. The screen or menu shown in FIG. 7B is displayed in response to a user selecting or instructing the generation of the two-dimensional barcode. The menu 702 shown in FIG. 7B includes a message that informs the user that the two-dimensional barcode has been generated and/or instructs the user to scan the two-dimensional barcode 720 into his or her mobile device 156.



FIG. 8 is a flow chart illustrating method of preserving the configurations associated with television receiver and/or remote control. Initially, in operation 804, the configuration sensing module 110 receives an indication that the television receiver 108 and/or remote 144 is to be reconfigured. Reconfiguring may occur, for example, when the television receiver 108 is reprogrammed or updated in a way that erases previous settings. Alternatively, the receiver 108 may have experienced a catastrophic failure that requires reconfiguring. In other instances, the service provider 104 may in fact replace the user's existing television receiver 108 with a more updated or improved receiver. Following operation 804, operation 808 may be executed.


In operation 808, the television receiver 108 establishes a connection between a television receiver 108 and a configuration data source. The configuration data source may, in one embodiment, be the storage device 160 associated with the mobile device 156. In other embodiments, the configuration data source may be the database 180. Establishing the connection may include initiating a network connection across the network 176, if the receiver 108 has such functionality. In other embodiments, the connection may be established through a USB port or other communication port associated with the receiver 108.


Following operation 808, operation 812 may be executed. In operation 812, the television receiver 108 downloads the configuration data into the television receiver 108. Following operation 812, operation 816 may be executed. In operation 816, the television receiver may configure itself or the remote 144 according to the configuration data acquired in operation 812.


By way of example and not limitation, some system elements described herein such as the condition sensing module 110, the encoding module 112, the output module 116 and the configuration module 120 are depicted in the figures as processor executable software or code elements that are stored in a stored in a storage device 152. However, it should be appreciated that some system designs consistent with the teachings described herein may implement separate modules within a television receiver that serve the functions implemented by the illustrated condition sensing module 110, the encoding module 112, the output module 116 and the configuration module 120. Such modules may include hardware and/or software elements that are implemented apart from the storage device 152. In some embodiments, such separate modules may utilize the storage device 152 for such purposes as loading and/or storing data.


While embodiments are discussed herein in connection with a exemplary satellite or cable broadcast system, it should be appreciated that embodiments may be used in connection other types of networks or content delivery mechanisms. Generally, the disclosure includes content delivered from a provider to a receiver across or over a network. The network across which content may be delivered may include satellite or cable system. Alternatively, the network may include a local area network, wide area network or the Internet. In connection with certain embodiments, a receiver may include a general purpose computer operable to receive data or other content across a network, such as a wide area network of the internet. In such embodiments, the computer may be configured so that a provider can access a web site, a file transfer protocol (FTP) site, a file sharing system or site, and so on. Moreover, the order of method operations, such those shown in FIG. 2, FIG. 4, FIG. 5, FIG. 6 and FIG. 8, described herein is by way of example and limitation. Certain implementations may reorder method operations without departing from the spirit and scope of the disclosure.

Claims
  • 1. A method of managing a television receiver, comprising: sensing a condition of a television receiver;in response to sensing the condition of the television receiver, creating a two-dimensional barcode with information corresponding to configuration settings of the television receiver, the information comprising settings information corresponding to user-adjustable settings;outputting the two-dimensional barcode from the television receiver for display on a display device;wherein, when the two-dimensional barcode is scanned from the display device into a mobile device, the two-dimensional barcode conveys the information corresponding to the configuration settings to the mobile device, so that the mobile device obtains the information for later use by the television receiver;subsequently communicating, by the television receiver with the mobile device, to transfer the information corresponding to the configuration settings from the mobile device to the television receiver; andself-configuring, by the television receiver, based at least in part on the information corresponding to the configuration settings transferred from the mobile device to the television receiver, including the settings information corresponding to the user-adjustable settings.
  • 2. The method of claim 1, further comprising: receiving a request to reconfigure the television receiver according to the information corresponding to the configuration settings retained by the mobile device;where the communicating, by the television receiver with the mobile device, is responsive to the request and comprises establishing a connection from the television receiver to a storage medium on which the mobile device has stored the information corresponding to the configuration settings.
  • 3. The method of claim 1, wherein the condition corresponds to one or more user-adjustable settings for the television receiver, and the sensing is executed in response to receiving a user request at the television receiver to save the one or more user-adjustable settings.
  • 4. The method of claim 1, wherein the condition of the television receiver is an error condition occurring during an operation of the television receiver, and the method further comprises: determining that the error condition will result in an imminent failure of an internal component of the television receiver;wherein the creating the two-dimensional barcode with information corresponding to configuration settings of the television receiver is in response to the determining.
  • 5. The method of claim 4, further comprising: developing a second two-dimensional barcode based on a corresponding error code, wherein the developing includes encoding the second two-dimensional barcode with routing information; andoutputting the second two-dimensional barcode from the television receiver for display on the display device;wherein, when the second two-dimensional barcode is scanned from the display device into the mobile device, the second two-dimensional barcode conveys the routing information to the mobile device, so that the mobile device obtains the information to access a remote site with the routing information and direct error information regarding the error condition to a service provider capable of mitigating the error condition.
  • 6. The method of claim 5, wherein: the routing information includes a uniform resource locator configured to access a website through a browser operating on the mobile device; andthe website is configured to communicate the error condition to the service provider using information obtained by scanning the second two-dimensional barcode.
  • 7. The method of claim 5, wherein the routing information causes the mobile device to automatically launch an application that is configured to communicate with the service provider.
  • 8. The method of claim 1, wherein the two-dimensional barcode is additionally created with second information corresponding to configuration settings of a user input device configured to communicate to the television receiver, and, when the two-dimensional barcode is scanned from the display device into the mobile device, the two-dimensional barcode conveys the second information to the mobile device so that the mobile device obtains the second information for later use in input device reconfiguration.
  • 9. A method, comprising: scanning a two-dimensional barcode into a mobile device, the two-dimensional barcode: being displayed on a display device associated with a television receiver;and created with information corresponding to configuration settings of a television receiver in response to a condition of the television receiver, the information comprising settings information corresponding to user-adjustable settings;wherein, when the two-dimensional barcode is scanned from the display device into the mobile device, the two-dimensional barcode conveys the information corresponding to the configuration settings to the mobile device, so that the mobile device obtains the information for later use by the television receiver;decoding the two-dimensional barcode at the mobile device to determine the information corresponding to the configuration settings to the mobile device;storing the information corresponding to the configuration settings for later use in reconfiguring the television receiver; andsubsequently communicating with the television receiver to transfer the information corresponding to the configuration settings from the mobile device to the television receiver to reconfigure the television receiver based at least in part on the information, including the settings information corresponding to the user-adjustable settings.
  • 10. The method of claim 9, wherein the information corresponding to the configuration settings includes a configuration setting for a remote control device associated with the television receiver.
  • 11. The method of claim 9, wherein the storing the information corresponding to the configuration settings comprises: establishing a network connection between the mobile device and a storage device that is remote from the mobile device;transmitting the storing the information corresponding to the configuration settings from the mobile device to the storage device across the network connection; andstoring the information corresponding to the configuration settings at the storage device.
  • 12. The method of claim 9, further comprising: receiving a request to reconfigure the television receiver according to the information corresponding to the configuration settings retained by the mobile device;where the communicating comprises establishing a connection between the television receiver and a storage medium on which the mobile device has stored the information corresponding to the configuration settings.
  • 13. A television receiver, comprising: a tuner operable to receive a program service transmission having a plurality of channels, the tuner operable to select one of the channels and to prepare the channel to be output in a data signal to a display device;a memory connected to the tuner, the memory including a tangible storage medium operable to store computer readable data and instructions;a processor connected to the memory operable to run computer executable code stored in the memory;a sensing module configured to execute on the processor to sense a condition of the television receiver;an encoding module configured to execute on the processor to create, in response to sensing the condition of the television receiver, a two-dimensional barcode with information corresponding to configuration settings of the television receiver, the information comprising settings information corresponding to user-adjustable settings;an output module configured to execute on the processor to output the two-dimensional barcode from the television receiver for display on the display device;wherein, when the two-dimensional barcode is scanned from the display device into a mobile device, the two-dimensional barcode conveys the information corresponding to the configuration settings to the mobile device, so that the mobile device obtains the information for later use by the television receiver; anda configuration module configured to execute on the processor to: subsequently communicate with the mobile device to transfer the information corresponding to the configuration settings from the mobile device to the television receiver; andconfiguring the television receiver, based at least in part on the information corresponding to the configuration settings transferred from the mobile device to the television receiver, including the settings information corresponding to the user-adjustable settings.
  • 14. The television receiver of claim 13, wherein the communicating comprises connecting to a storage medium used by the mobile device to store the information corresponding to the configuration settings, to download the settings information corresponding to the user-adjustable settings, and wherein the configuring the television receiver comprises setting the television receiver according to the settings information.
  • 15. The television receiver of claim 13, further comprising: a recorder operable to a record and store a program associated with a channel of the program service transmission;wherein the settings information includes a recording timer configured to cause the recorder to record a particular program.
  • 16. The television receiver of claim 13, the configuration module to receive a request to reconfigure the television receiver according to the information corresponding to the configuration settings retained by the mobile device, where the communicating is responsive to the request and comprises establishing a connection from the television receiver to a storage medium on which the mobile device has stored the information corresponding to the configuration settings.
  • 17. The television receiver of claim 13, wherein: the condition corresponds to one or more user-adjustable settings for the television receiver, andthe sensing is executed in response to receiving a user request at the television receiver to save the one or more user-adjustable settings.
  • 18. The television receiver of claim 13, wherein: the condition of the television receiver is an error condition occurring during an operation of the television receiver;the television receiver is configured to determine whether the error condition will result in an imminent failure of an internal component of the television receiver; andthe creating the two-dimensional barcode with information corresponding to configuration settings of the television receiver is in response to determining that the error condition will result in an imminent failure of an internal component of the television receiver.
  • 19. The television receiver of claim 18, wherein: the encoding module is further configured to develop a second two-dimensional barcode based on a corresponding error code, wherein the developing includes encoding the second two-dimensional barcode with routing information; andthe output module is further configured to output the second two-dimensional barcode from the television receiver for display on the display device;wherein, when the second two-dimensional barcode is scanned from the display device into the mobile device, the second two-dimensional barcode conveys the routing information to the mobile device, so that the mobile device obtains the information to access a remote site with the routing information and direct error information regarding the error condition to a service provider capable of mitigating the error condition.
  • 20. The television receiver of claim 19, wherein: the routing information includes a uniform resource locator configured to access a website through a browser operating on the mobile device; andthe website is configured to communicate the error condition to the service provider using information obtained by scanning the second two-dimensional barcode.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. Nonprovisional patent application Ser. No. 13/037,302, which was filed on Feb. 28, 2011, and entitled “Set Top Box Health and Configuration,” which is hereby incorporated by reference in its entirety.

US Referenced Citations (306)
Number Name Date Kind
4809325 Hayashi et al. Feb 1989 A
4837414 Edamula Jun 1989 A
5500681 Jones Mar 1996 A
5510603 Hess et al. Apr 1996 A
5581636 Skinger Dec 1996 A
5602377 Beller et al. Feb 1997 A
5703349 Meyerson et al. Dec 1997 A
5929849 Kikinis Jul 1999 A
5959285 Schuessler Sep 1999 A
5963265 Bae et al. Oct 1999 A
5978855 Metz et al. Nov 1999 A
6006990 Ye et al. Dec 1999 A
6058238 Ng May 2000 A
6263502 Morrison et al. Jul 2001 B1
6438751 Voyticky et al. Aug 2002 B1
6512919 Ogasawara Jan 2003 B2
6556273 Wheeler et al. Apr 2003 B1
6704929 Ozer et al. Mar 2004 B1
6766956 Boylan et al. Jul 2004 B1
6983304 Sato Jan 2006 B2
7046161 Hayes May 2006 B2
7206029 Cohen-Solal Apr 2007 B2
7206409 Antonellis et al. Apr 2007 B2
7221405 Basson et al. May 2007 B2
7244404 Rosenberg et al. Jul 2007 B2
7328848 Xia et al. Feb 2008 B2
7349668 Ilan et al. Mar 2008 B2
7369180 Xing May 2008 B2
7373652 Bayrakeri et al. May 2008 B1
7387250 Muni Jun 2008 B2
7394519 Mossman et al. Jul 2008 B1
7424976 Muramatsu Sep 2008 B2
7443449 Momosaki et al. Oct 2008 B2
7487527 Ellis et al. Feb 2009 B2
7587601 Levy et al. Sep 2009 B2
7604172 Onogi Oct 2009 B2
7612748 Tateuchi Nov 2009 B2
7624417 Dua Nov 2009 B2
7624916 Sato et al. Dec 2009 B2
7673297 Arsenault et al. Mar 2010 B1
7797430 Ichieda Sep 2010 B2
7818675 Maruyama et al. Oct 2010 B2
7841531 Onogi Nov 2010 B2
8010977 Hogyoku Aug 2011 B2
8045054 Bishop et al. Oct 2011 B2
8186572 Herzig May 2012 B2
8292166 Gomez et al. Oct 2012 B2
8364018 McArdle Jan 2013 B2
8380993 Chen et al. Feb 2013 B2
8386339 Minnick et al. Feb 2013 B2
8408466 Gratton Apr 2013 B2
8427455 Matsuda Apr 2013 B2
8430302 Minnick et al. Apr 2013 B2
8439257 Beals et al. May 2013 B2
8443407 Gaede et al. May 2013 B2
8468610 Beals et al. Jun 2013 B2
8511540 Anguiano Aug 2013 B2
8534540 Gratton et al. Sep 2013 B2
8550334 Gratton et al. Oct 2013 B2
8553146 Kennedy Oct 2013 B2
8746554 Gomez et al. Jun 2014 B2
8786410 Beals et al. Jul 2014 B2
8827150 Gratton et al. Sep 2014 B2
8833640 Martch et al. Sep 2014 B2
8856853 Casagrande et al. Oct 2014 B2
8875173 Kilaru et al. Oct 2014 B2
8886172 Gomez Nov 2014 B2
8931031 Schaefer Jan 2015 B2
9092830 Gomez et al. Jul 2015 B2
9148686 Gerhards et al. Sep 2015 B2
9280515 Gaede et al. Mar 2016 B2
9329966 Dugan et al. May 2016 B2
9367669 Gratton Jun 2016 B2
9571888 Casagrande et al. Feb 2017 B2
9596500 Gomez et al. Mar 2017 B2
9652108 Mountain May 2017 B2
20010037297 McNair Nov 2001 A1
20010052133 Pack et al. Dec 2001 A1
20020011521 Lahey et al. Jan 2002 A1
20020027612 Brill et al. Mar 2002 A1
20020049980 Hoang Apr 2002 A1
20020065728 Ogasawara May 2002 A1
20020082931 Siegel et al. Jun 2002 A1
20020112250 Koplar et al. Aug 2002 A1
20020195495 Melick et al. Dec 2002 A1
20030018711 Imanishi Jan 2003 A1
20030050854 Showghi et al. Mar 2003 A1
20030077065 Scholten et al. Apr 2003 A1
20030112974 Levy Jun 2003 A1
20030121978 Rubin et al. Jul 2003 A1
20030125092 Burnhouse et al. Jul 2003 A1
20030151562 Kulas Aug 2003 A1
20030172374 Vinson et al. Sep 2003 A1
20040005900 Zilliacus Jan 2004 A1
20040019691 Daymond et al. Jan 2004 A1
20040026508 Nakajima et al. Feb 2004 A1
20040044532 Karstens Mar 2004 A1
20040046790 Agarwal et al. Mar 2004 A1
20040049672 Nollet et al. Mar 2004 A1
20040260850 Yu et al. Dec 2004 A1
20050009564 Hayaashi et al. Jan 2005 A1
20050011958 Fukasawa et al. Jan 2005 A1
20050015800 Holcomb Jan 2005 A1
20050015815 Shoff et al. Jan 2005 A1
20050055281 Williams Mar 2005 A1
20050059339 Honda et al. Mar 2005 A1
20050097618 Arling et al. May 2005 A1
20050107135 Deeds et al. May 2005 A1
20050125301 Muni Jun 2005 A1
20050149967 Hanley et al. Jul 2005 A1
20050163483 Rassool Jul 2005 A1
20050180804 Andrew et al. Aug 2005 A1
20050203854 Das Sep 2005 A1
20050258250 Melick et al. Nov 2005 A1
20050262548 Shimojo et al. Nov 2005 A1
20050264694 Ilan et al. Dec 2005 A1
20060064700 Ludvig et al. Mar 2006 A1
20060065733 Lee et al. Mar 2006 A1
20060071076 Tamayama Apr 2006 A1
20060079247 Ritter Apr 2006 A1
20060086796 Onogi Apr 2006 A1
20060090179 Hsu et al. Apr 2006 A1
20060095286 Kimura May 2006 A1
20060124742 Rines et al. Jun 2006 A1
20060196950 Kiliccote Sep 2006 A1
20060203339 Kleinberger et al. Sep 2006 A1
20060208088 Sekiguchi Sep 2006 A1
20060265731 Matsuda Nov 2006 A1
20070008344 Medina Jan 2007 A1
20070011709 Katz et al. Jan 2007 A1
20070016934 Okada et al. Jan 2007 A1
20070016936 Okada et al. Jan 2007 A1
20070017350 Uehara Jan 2007 A1
20070019215 Yu Jan 2007 A1
20070063050 Attia et al. Mar 2007 A1
20070073585 Apple et al. Mar 2007 A1
20070143788 Abernathy et al. Jun 2007 A1
20070159522 Neven Jul 2007 A1
20070174198 Kasahara Jul 2007 A1
20070192723 Anzelde et al. Aug 2007 A1
20070200335 Tuschel Aug 2007 A1
20070205596 Mizuno et al. Sep 2007 A1
20070206020 Duffield et al. Sep 2007 A1
20070256118 Nomura et al. Nov 2007 A1
20070288594 Philyaw et al. Dec 2007 A1
20080022323 Koo Jan 2008 A1
20080059998 McClenny et al. Mar 2008 A1
20080062164 Bassi et al. Mar 2008 A1
20080073434 Epshteyn et al. Mar 2008 A1
20080077324 Hatano et al. Mar 2008 A1
20080082684 Gaos et al. Apr 2008 A1
20080092154 Hogyoku Apr 2008 A1
20080112615 Obrea et al. May 2008 A1
20080156879 Melick et al. Jul 2008 A1
20080182561 Kim et al. Jul 2008 A1
20080189185 Matsuo et al. Aug 2008 A1
20080200153 Fitzpatrick et al. Aug 2008 A1
20080200160 Fitzpatrick et al. Aug 2008 A1
20080201078 Fitzpatrick et al. Aug 2008 A1
20080244675 Sako et al. Oct 2008 A1
20080263621 Austerlitz et al. Oct 2008 A1
20080267537 Thuries Oct 2008 A1
20080281624 Shibata Nov 2008 A1
20080288460 Poniatowski et al. Nov 2008 A1
20080288600 Clark Nov 2008 A1
20080307348 Jones et al. Dec 2008 A1
20090029725 Kindberg Jan 2009 A1
20090031071 Chiu Jan 2009 A1
20090031373 Hogyoku Jan 2009 A1
20090070699 Birkill et al. Mar 2009 A1
20090083808 Morrison Mar 2009 A1
20090088213 Rofougaran Apr 2009 A1
20090094546 Anzelde et al. Apr 2009 A1
20090108057 Mu et al. Apr 2009 A1
20090113334 Chakra et al. Apr 2009 A1
20090116074 Wilsher May 2009 A1
20090154759 Koskinen et al. Jun 2009 A1
20090157511 Spinnell et al. Jun 2009 A1
20090157530 Nagamoto et al. Jun 2009 A1
20090172780 Sukeda et al. Jul 2009 A1
20090179852 Refai et al. Jul 2009 A1
20090180025 Dawson Jul 2009 A1
20090196456 Bisti Aug 2009 A1
20090200367 Arnouse Aug 2009 A1
20090212112 Li et al. Aug 2009 A1
20090212113 Chiu et al. Aug 2009 A1
20090234570 Sever Sep 2009 A1
20090250512 Deck et al. Oct 2009 A1
20090254954 Jeong Oct 2009 A1
20090293088 Mukerji et al. Nov 2009 A1
20090293110 Koga Nov 2009 A1
20090294538 Wihlborg et al. Dec 2009 A1
20090300673 Bachet et al. Dec 2009 A1
20090303036 Sahuguet Dec 2009 A1
20090307232 Hall Dec 2009 A1
20090307719 Clark et al. Dec 2009 A1
20090312105 Koplar Dec 2009 A1
20090320066 Soldan et al. Dec 2009 A1
20100001072 Onogi Jan 2010 A1
20100004984 Beyabani Jan 2010 A1
20100017457 Jumpertz et al. Jan 2010 A1
20100020970 Liu et al. Jan 2010 A1
20100026721 Park et al. Feb 2010 A1
20100031162 Wiser et al. Feb 2010 A1
20100036936 Cox et al. Feb 2010 A1
20100053339 Aaron et al. Mar 2010 A1
20100081375 Rosenblatt et al. Apr 2010 A1
20100089996 Koplar Apr 2010 A1
20100096448 Melick et al. Apr 2010 A1
20100103018 Yoon et al. Apr 2010 A1
20100114715 Schuster et al. May 2010 A1
20100129057 Kulkarni May 2010 A1
20100131373 Kubasov May 2010 A1
20100131900 Spetalnick May 2010 A1
20100131970 Falcon May 2010 A1
20100131983 Shannon et al. May 2010 A1
20100138344 Wong Jun 2010 A1
20100149187 Slavin et al. Jun 2010 A1
20100154035 Damola et al. Jun 2010 A1
20100161437 Pandey Jun 2010 A1
20100163613 Bucher et al. Jul 2010 A1
20100169935 Abbruzzese Jul 2010 A1
20100188514 Sato et al. Jul 2010 A1
20100201894 Nakayama et al. Aug 2010 A1
20100205628 Davis et al. Aug 2010 A1
20100217663 Ramer et al. Aug 2010 A1
20100225653 Sao Sep 2010 A1
20100261454 Shenfield et al. Oct 2010 A1
20100262924 Kalu Oct 2010 A1
20100262992 Casagrande Oct 2010 A1
20100272193 Khan et al. Oct 2010 A1
20100272420 Soohoo et al. Oct 2010 A1
20100275010 Ghirardi Oct 2010 A1
20100279710 Dicke et al. Nov 2010 A1
20100295868 Zahnert et al. Nov 2010 A1
20100301115 Berkun Dec 2010 A1
20100313231 Okamoto et al. Dec 2010 A1
20100319041 Ellis Dec 2010 A1
20100327060 Moran et al. Dec 2010 A1
20110000958 Herzig Jan 2011 A1
20110007630 Almhana et al. Jan 2011 A1
20110030068 Imai Feb 2011 A1
20110039573 Hardie Feb 2011 A1
20110058516 Small et al. Mar 2011 A1
20110061003 Miyazawa et al. Mar 2011 A1
20110065451 Danado et al. Mar 2011 A1
20110087539 Rubinstein et al. Apr 2011 A1
20110107374 Roberts et al. May 2011 A1
20110107386 de los Reyes May 2011 A1
20110138408 Adimatyam et al. Jun 2011 A1
20110208710 Lesavich Aug 2011 A1
20110258058 Carroll et al. Oct 2011 A1
20110264527 Fitzpatrick et al. Oct 2011 A1
20110264530 Santangelo et al. Oct 2011 A1
20110282727 Phan et al. Nov 2011 A1
20110296046 Arya et al. Dec 2011 A1
20110314485 Abed Dec 2011 A1
20110321114 Newell Dec 2011 A1
20120075529 Wong et al. Mar 2012 A1
20120084206 Mehew et al. Apr 2012 A1
20120096499 Dasher et al. Apr 2012 A1
20120117232 Brown et al. May 2012 A1
20120127110 Amm et al. May 2012 A1
20120128267 Dugan et al. May 2012 A1
20120130835 Fan et al. May 2012 A1
20120130851 Minnick et al. May 2012 A1
20120131416 Dugan et al. May 2012 A1
20120137318 Kilaru et al. May 2012 A1
20120138671 Gaede et al. Jun 2012 A1
20120139826 Beals et al. Jun 2012 A1
20120139835 Morrison et al. Jun 2012 A1
20120142322 Gomez Jun 2012 A1
20120151293 Beals Jun 2012 A1
20120151524 Kilaru et al. Jun 2012 A1
20120153015 Gomez et al. Jun 2012 A1
20120153017 Bracalente et al. Jun 2012 A1
20120155838 Gerhards et al. Jun 2012 A1
20120158919 Aggarwal et al. Jun 2012 A1
20120159563 Gomez et al. Jun 2012 A1
20120168493 Worms Jul 2012 A1
20120168510 Gratton Jul 2012 A1
20120169928 Casagrande et al. Jul 2012 A1
20120175416 Gomez et al. Jul 2012 A1
20120181329 Gratton et al. Jul 2012 A1
20120182320 Beals et al. Jul 2012 A1
20120188112 Beals et al. Jul 2012 A1
20120188442 Kennedy Jul 2012 A1
20120198572 Beals et al. Aug 2012 A1
20120199643 Minnick et al. Aug 2012 A1
20120206648 Casagrande et al. Aug 2012 A1
20120215830 Anguiano Aug 2012 A1
20120217292 Gratton et al. Aug 2012 A1
20120217293 Martch et al. Aug 2012 A1
20120218470 Schaefer Aug 2012 A1
20120218471 Gratton Aug 2012 A1
20120222055 Schaefer et al. Aug 2012 A1
20120222071 Gaede et al. Aug 2012 A1
20120222081 Schaefer et al. Aug 2012 A1
20120293327 Mountain Nov 2012 A1
20120311623 Davis et al. Dec 2012 A1
20130068838 Gomez et al. Mar 2013 A1
20130239157 Gaede et al. Sep 2013 A1
20140046661 Bruner Feb 2014 A1
20140076963 Gratton et al. Mar 2014 A1
20140158762 Gomez et al. Jun 2014 A1
20160066050 Gerhards et al. Mar 2016 A1
Foreign Referenced Citations (63)
Number Date Country
2 634 951 Jan 2010 CA
1352765 Jun 2002 CN
1571503 Jan 2005 CN
1636371 Jul 2005 CN
1675930 Sep 2005 CN
1839398 Sep 2006 CN
101 227 581 Jul 2008 CN
101253504 Aug 2008 CN
101 355 685 Jan 2009 CN
101 409 027 Apr 2009 CN
101 873 467 Oct 2010 CN
101 894 113 Nov 2010 CN
101 895 772 Nov 2010 CN
23 36 711 Feb 1975 DE
10 2007 038 810 Feb 2009 DE
1 021 035 Jul 2000 EP
1 383 071 Jan 2004 EP
1 555 808 Jul 2005 EP
1 724 695 Nov 2006 EP
1 757 222 Feb 2007 EP
1 768 400 Mar 2007 EP
2 079 051 Jul 2009 EP
2 131 289 Dec 2009 EP
2 439 936 Apr 2012 EP
2 643 769 Oct 2013 EP
2 565 748 Dec 1985 FR
2 929 467 Oct 2009 FR
2 044 446 Oct 1980 GB
2 165 129 Apr 1986 GB
2 311 451 Sep 1997 GB
2 325 765 Dec 1998 GB
2 471 567 Jan 2011 GB
4698CHENP2013 Jun 2016 IN
7734CHENP2013 Jun 2016 IN
2000-222116 Aug 2000 JP
2002-215768 Aug 2002 JP
3929450 Jun 2007 JP
2007-213548 Aug 2007 JP
2008-244556 Oct 2008 JP
2008-276716 Nov 2008 JP
2009-140204 Jun 2009 JP
2004 0087776 Oct 2004 KR
299433 Mar 1997 TW
200915193 Apr 2009 TW
200926075 Jun 2009 TW
201032139 Sep 2010 TW
201038061 Oct 2010 TW
201043039 Dec 2010 TW
9527275 Oct 1995 WO
9741690 Nov 1997 WO
0106593 Jan 2001 WO
0118589 Mar 2001 WO
0158146 Aug 2001 WO
2004019442 Mar 2004 WO
2005109338 Nov 2005 WO
2007009005 Jan 2007 WO
2009056897 May 2009 WO
2009057651 May 2009 WO
2009116954 Sep 2009 WO
2009144536 Dec 2009 WO
2010149161 Dec 2010 WO
2011009055 Jan 2011 WO
2012071176 May 2012 WO
Non-Patent Literature Citations (272)
Entry
European Search Report for EP 11 846 858.6 dated Jun. 12, 2017, all pages.
Jung, Eui-Hyun et al., “A Robust Digital Watermarking System Adopting 2D Barcode against Digital Piracy on P2P Network,” IJCSNS International Journal of Computer Science and Network Security, vol. 6, No. 10, Oct. 2006, 6 pages.
“Android App Reviews & Showcase Just a Tapp Away,” Android Tapp, 10pp. Found online at http://www.androidtapp.com/download-the-weather-channel-android-app-from-your-tv/, Oct. 22, 2010.
O'Sullivan, “Can Mobile Barcodes Work on TV?,” India and Asia Pacific Mobile Industry Reports, Sep. 2009, 4 pp. Found online at http://gomonews.com/can-mobile-barcodes-work-on-tv/, Feb. 5, 2013.
“FOX TV Uses QR Codes,” 2D Barcode Strategy, Sep. 2010, 6 pp. Found online at http://www.2dbarcodestrategy.com/2010/09/fox-tv-uses-qr-codes.html, Oct. 22, 2010.
“FOX's Fringe Uses QR Code,” 2D Barcode Strategy, Oct. 2010, 4 pp. Found on the Internet at http://www.2dbarcodestrategy.com/2010/10/foxs-fringe-uses-qr-code.html, Oct. 22, 2010.
“Mobile Paths: QR Codes Come to TV,” Mobile Behavior: An Omnicom Group Company, Sep. 2010, 8 pp. Found online at http://www.mobilebehavior.com/2010/09/27/mobile-paths-qr-codes-come-to-tv, Oct. 22, 2010.
“What Can I Do with the QR Barcode,” Search Magnet Local-QR Barcode Technology, 2 pp. Found online at http://www.searchmagnetlocal.com/qr_barcode_technology.html, Oct. 22, 2010.
Byford, D., “Universal Interactive Device,” International Business Machines Corporation, Jun. 1998, 1 page.
Costedio, K., “Bluefly QR Codes Debut on TV,” 2 pp. Found online at http://www.barcode.com/Mobile-Barcode-news/bluefly-qr-codes-debut-on-tv.html, Oct. 22, 2010.
First Examination Report from European Patent Office dated May 18, 2015 for EP 11849496.2, 7 pages.
Examination Search Report from the European Patent Office dated Dec. 4, 2015 for EP 12707418.5, 8 pages.
Extended European Search Report for EP 12152690.9 dated Jun. 19, 2012, 9 pages.
European Office Action for EP 12716751.8 dated Nov. 11, 2015, 4 pages.
Extended European Search Report for EP 11842890.3 dated Mar. 26, 2014, 8 pages.
European Office Action for EP 11842890.3 dated Mar. 13, 2015, 8 pages.
European Office Action for EP 11842890.3 dated May 9, 2016, all pages.
Extended European Search Report for EP 11850819.1 dated Mar. 17, 2014, 5 pages.
Office Action for EP 11850819.1 dated Nov. 12, 2015, 4 pages.
Office Action for EP 12705768.5 dated May 25, 2016, all pages.
Extended European Search Report for EP 11846486 dated Mar. 26, 2014, 5 pages.
Extended European Search Report for EP 11852630 dated Jun. 30, 2014, 7 pages.
Office Action for Korean Patent Application No. 10-2013-7015610 dated Oct. 21, 2016, all pages.
Office Action for Korean Patent Application No. 10-2013-7015610 dated Feb. 21, 2017, all pages.
Office Action from European Patent Office for Application No. 11852630.0 dated May 12, 2015, 7 pages.
European Search Report for EP 11844504 dated Feb. 24, 2015, 10 pages.
European Search Report for EP 11855065 dated Mar. 6, 2014, 6 pages.
Notice of Allowance for Canadian Application 2,818,585 dated Apr. 3, 2017, 1 page.
Notice of Allowance for Canadian Application 2,828,463 dated Apr. 4, 2017, 1 page.
Office Action for CA 2,818,585 dated Nov. 28, 2016, all pages.
Office Action for EP 11811502.1 dated Aug. 29, 2016, all pages.
Office Action for EP 11855065 dated Mar. 13, 2015, 6 pages.
First Office Action by the Mexican Institute of Industrial Property for Mexican Patent Application No. MX/a/2013/009881 dated Aug. 14, 2014, 3 pages.
Notice of Allowance by the Mexican Institute of Industrial Property for Mexican Patent Application No. MX/a/2013/009881 dated Jan. 12, 2015, 1 page.
Publication of Brazil appln No. BR 11 2013 012218-8 on Aug. 9, 2016, 1 page.
Gao, J., et al., “A 2D Barcode-Based Mobile Payment System,” Multimedia and Ubiquitous Engineering, 2009, 10 pp. Found online at http://ieeexplore.ieee.org/Xplore/login.jsp?url=http%3A%2F%2Fieeexplore.ieee.org%2Fie . . . , Oct. 22, 2010.
International Search Report and Written Opinion of PCT/US2011/059977 dated Mar. 19, 2012, 7 pages.
International Preliminary Report on Patentability of PCT/US2011/059977 dated Jun. 6, 2013, 6 pages.
International Preliminary Report on Patentability of PCT/US2012/048032 dated Apr. 3, 2014, 6 pages.
International Search Report and Written Opinion of PCT/US11/60002 dated Feb. 15, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/060002 dated Jun. 6, 2013, 6 pages.
International Search Report and Written Opinion of PCT/US11/60094 dated Mar. 30, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/060094 dated Jun. 20, 2013, 6 pages.
International Search Report and Written Opinion of PCT/US11/60104 dated Mar. 29, 2012, 9 pages.
International Search Report and Written Opinion of PCT/US2011/60121 dated Feb. 14, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/060121 dated Jun. 20, 2013, 6 pages.
International Search Report and Written Opinion of PCT/US11/61074 dated Jan. 6, 2012, 11 pages.
International Search Report and Written Opinion of PCT/US11/61211 dated Mar. 29, 2012, 8 pages.
International Preliminary Report on Patentability for PCT/US2011/061211 dated Jul. 4, 2013, 7 pages.
International Search Report and Written Opinion of PCT/US11/61773 dated Feb. 21, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/061773 dated Jun. 6, 2013, 6 pages.
International Search Report and Written Opinion of PCT/US11/61778 dated Mar. 2, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/061778 dated Jul. 11, 2013, 6 pages.
International Search Report and Written Opinion of PCT/US11/63111 dated Apr. 4, 2012, 9 pages.
International Preliminary Report on Patentability of PCT/US2011/063111 dated Jun. 13, 2013, 8 pages.
International Search Report and Written Opinion of PCT/US2011/064709 dated Apr. 10, 2012, 8 pages.
International Search Report and Written Opinion of PCT/US2011/060098 dated Mar. 29, 2012, 10 pages.
International Preliminary Report on Patentability for PCT/US2011/060098 dated Jun. 13, 2013, 9 pages.
International Search Report and Written Opinion of PCT/US2011/063308 dated Mar. 29, 2012, 10 pages.
International Preliminary Report on Patentability for PCT/US2011/063308 dated Jul. 18, 2013, 9 pages.
International Search Report and Written Opinion of PCT/US2011/068161 dated Jun. 14, 2012, 19 pages.
International Preliminary Report on Patentability of PCT/US2011/068161 dated Jul. 25, 2013, 13 pages.
International Search Report and Written Opinion of PCT/US2011/068176 dated Mar. 29, 2012, 15 pages.
International Search Report and Written Opinion of PCT/US2012/021657 dated May 23, 2012, 12 pages.
International Search Report of PCT/US2012/022581 dated Oct. 8, 2012, 18 pages.
International Preliminary Report on Patentability for PCT/US2012/022581 dated Aug. 8, 2013, 12 pages.
International Search Report and Written Opinion of PCT/US2012/022405 dated Apr. 19, 2012, 11 pages.
International Preliminary Report on Patentability for PCT/US2012/022405 dated Aug. 8, 2013, 7 pages.
International Search Report and Written Opinion of PCT/US2012/024923 dated May 22, 2012, 12 pages.
International Preliminary Report on Patentability for PCT/US2012/024923 dated Aug. 29, 2013, 8 pages.
International Search Report and Written Opinion of PCT/US2012/024956 dated Jun. 11, 2012, 10 pages.
International Preliminary Report on Patentability for PCT/US2012/024956 dated Aug. 29, 2013, 7 pages.
International Search Report and Written Opinion of PCT/US2012/025502 dated Jun. 8, 2012, 13 pages.
International Preliminary Report on Patentability of PCT/US2012/025502 dated Sep. 6, 2013, 9 pages.
International Search Report and Written Opinion of PCT/US2012/025607 dated Jun. 8, 2012, 13 pages.
International Preliminary Report on Patentability for PCT/US2012/025607 dated Sep. 12, 2013, 8 pages.
International Search Report and Written Opinion of PCT/US2012/025634 dated May 7, 2012, 8 pages.
International Preliminary Report on Patentability for PCT/US2012/025634 dated Sep. 6, 2013, 5 pages.
International Search Report and Written Opinion of PCT/US2012/026373 dated Jun. 13, 2012, 14 pages.
International Preliminary Report on Patentability for PCT/US2012/026373 dated Sep. 12, 2013, 10 pages.
International Search Report and Written Opinion of PCT/US2012/026624 dated Aug. 29, 2012, 14 pages.
International Preliminary Report on Patentability for PCT/US2012/026624 dated Sep. 12, 2013, 12 pages.
International Search Report and Written Opinion of PCT/US2012/026722 dated Jun. 28, 2012, 11 pages.
International Search Report and Written Opinion of PCT/US2012/048032, dated Oct. 16, 2012, 14 pages.
International Search Report and Written Opinion of PCT/US2011/060109 dated Feb. 14, 2012, 7 pages.
International Preliminary Report on Patentability for PCT/US2011/060109 dated Jun. 20, 2013, 7 pages.
First Examination Report from European Patent Office dated Feb. 4, 2015 for EP 12716751.8, 4 pages.
First Office Action for CN 201180065044.7 dated Feb. 13, 2015 by the State Intellectual Property Office (SIPO), 4 pages.
First Office Action with Search Report for CN 201280013891.3 dated Jan. 15, 2016, 13 pages.
Second Office Action CN 201280013891.3 dated Aug. 12, 2016, all pages.
Second Office Action for CN 201180065044.7 dated Sep. 9, 2015 by the State Intellectual Property Office (SIPO), 23 pages.
Office Action from European Patent Office for Application No. 12716728.6 dated Feb. 26, 2015, 5 pages.
Notice of Allowance and search report for ROC (Taiwan) Patent Application No. 101106288 received May 29, 2015, 9 pages.
Office Action of the Intellectual Property Office for ROC Patent App. No. 101101486 dated Aug. 5, 2014, 4 pages.
Office Action of the Intellectual Property Office for ROC Patent App. No. 100143194 dated Sep. 23, 2014, 10 pages.
Office Action of the Intellectual Property Office for ROC Patent App. No. 100142978 dated Sep. 23, 2014, 9 pages.
Office Action from State Intellectual Property Office for CN Appln. No. 201180056242.7 received Jun. 17, 2015, 10 pages.
Second Office Action from State Intellectual Property Office for CN Appln. No. 201180056242.7 dated Jan. 26, 2016, all pages.
Third Office Action from State Intellectual Property Office for CN Appln. No. 201180056242.7 dated Jul. 28, 2016, all pages.
First Office Action and Search Report from State Intellectual Property Office for CN Appln. No. 201180064527.5 dated Oct. 23, 2015, 10 pages.
Second Office Action from State Intellectual Property Office for CN Appln. No. 201180064527.5 dated Jun. 12, 2016, all pages.
(Translation) Rejection Decision for CN Appln. No. 201180064527.5 dated Oct. 9, 2016, all pages.
The First Office Action dated Sep. 11, 2014 for Mexican Patent Application No. MX/a/2013/007672 is not translated into English, 2 pages.
The Second Office Action dated Jun. 1, 2015 for Mexican Patent Application No. MX/a/2013/007672 is not translated into English, 2 pages.
Office Action dated Mar. 2, 2017 for KR 10-2013-7020865, all pages.
Notice to Grant received Jun. 9, 2017 for KR 10-2013-7020865, all pages.
Notice of Allowance dated Nov. 10, 2015 for Mexican Patent Application No. MX/a/2013/007672, 1 page.
The First Office Action dated Jul. 13, 2015 for Mexican Patent Application No. MX/a/2013/009791 is not translated into English, 2 pages.
Notice of Allowance for Mexican Patent Application No. MX/a/2013/009791 dated Mar. 15, 2016, 1 page.
Office Action dated Nov. 12, 2014 for Mexican Patent Application No. MX/a/2013/009794, 2 pages.
Office Action dated Oct. 17, 2016 for European Patent Appln. No. 12701638.4, all pages.
Notice of Allowance dated Feb. 18, 2015 for Mexican Patent Application No. MX/a/2013/009794, 1 page.
The First Office Action for Mexican Patent Application No. MX/a/2013/006262 is not translated into English. This document is from prosecution of the corresponding foreign matter for which we do not have a translation. dated Aug. 7, 2014, 2 pages.
Office Action dated Feb. 10, 2015 for Mexican Patent Application No. MX/a/2013/006770, 2 pages.
Office Action dated Feb. 6, 2015 for Mexican Patent Application No. MX/a/2013/006520, 2 pages.
Office Action dated Jan. 28, 2015 for Mexican Patent Application No. MX/a/2013/006973, 9 pages.
Notice of Allowance for Mexican Patent Application No. MX/a/2013/006973 dated Sep. 4, 2015, 1 page.
Office Action dated Dec. 5, 2014 for Mexican Patent Application No. MX/a/2013/009882, 2 pages.
Office Action for European Patent App. 12704473.3 dated Apr. 29, 2016, all pages.
The Second Office Action dated Apr. 22, 2015 for Mexican Patent Application No. MX/a/2013/009883, 2 pages.
Supplementary European Search Report for EP 11843423 completed Mar. 23, 2016, 8 pages.
Supplementary European Search Report for EP 11843045 completed Mar. 31, 2016, all pages.
Kato et al, “2D barcodes for mobile phones”, Mobile Technology, Applications and Systems, 2005 2nd International Conference on Guangzhou, China Nov. 15-17, 2005, Piscataway, NJ, USA, IEEE, Piscataway, NJ, USA, Nov. 15, 2005, pp. 8pp. 8, XP031887368, DOI: 10.1109/MTAS.2005.207166; ISBN: 978-981-05-4573-4, 8 pages.
Liu, Yue et al., “Recognition of QR code with mobile phones,” Control and Decision Conference, 2008. CCDC 2008. Jul. 2-4, 2008, pp. 203, 206.
Ngee, S., “Data Transmission Between PDA and PC Using WiFi for Pocket Barcode Application,” Thesis, University Teknologi Malaysia, May 2007, 126 pp. Found online at http://eprints.utm.my/6421/1/SeahYeowNgeeMFKE20007TTT.pdf, Oct. 22, 2010.
Olson, E., “Bar Codes add Detail on Items in TV Ads,” New York Times, Sep. 2010, 3 pp. Found online at http:www.nytimes.com/2010/09/27/business/media/27bluefly.html?src=busin, Oct. 22, 2010.
Publication of BR 11 2014 020007-6 A2 on Jun. 20, 2017, 1 page.
Publication of PCT/US2011/059977 by the India Controller General of Patents Designs and Trademarks as India Patent Publication No. 4694/CHENP/2013 A on Sep. 5, 2014,1 page.
Publication of PCT/US2012/025634 by the India General Patents Designs and Trademarks as India Patent Publication No. 6967/CHENP/2013 A dated Aug. 1, 2014, 1 page.
Rekimoto, J., et al., “Augment-able Reality: Situated Communication Through Physical and Digital Spaces,” Sony Computer Science Laboratory, 2002, 8 pp. Found online at Citeseer: 10.1.1.20.34[1].pdf, Oct. 22, 2010.
Schmitz, A., et al., “Ad-Hoc Multi-Displays for Mobile Interactive Applications,” 31st Annual Conference of the European Association for Computer Graphics (Eurographics 2010), May 2010, vol. 29, No. 2, 8 pages.
Silverstein, B., “QR Codes and TV Campaigns Conned,” ReveNews, Sep. 2010, 5 pp. Found online at http://www.revenews.com/barrysilverstein/qr-codes-and-tv-campaigns-connect/, Oct. 22, 2010.
Smith, L., “QR Barcodes Make History on Global TV,” 3 pp. Found online at http://lindsaysmith.com/worlds-first-mobio-mini-telethon/, Oct. 22, 2010.
Yamanari, T., et al., “Electronic Invisible Code Display Unit for Group Work on Reminiscence Therapy,” Proceedings of the International MultiConference of Engineers and Computer Scientists 2009, vol. 1, IMECS 2009, Mar. 2009, 6 pp. Retrieved from the Internet: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.145.6904&rep1&type=pdf.
Yang, C., et al., “Embedded Digital Information Integrated by Video-on-Demand System,” Proceedings of the Fourth International Conference on Networked Computing and Advanced Information Management, IEEE Computer Society, 2008, 6 pages.
First Office Action including Search Report from the State Intellectual Property Office for CN Patent Appln. No. 201280014034.5 dated Apr. 5, 2016, all pages.
First Office Action and Search Report from the State Intellectual Property Office (SIPO) for CN 201180056249.9 dated Feb. 3, 2016, all pages.
Notice of Decision to Grant for KR 10-2013-7024307 dated Apr. 14, 2017, 2 pages.
Second Office Action issued by State Intellectual Property Office (SIPO) for CN 201180056249.9 dated Feb. 4, 2017, all pages.
Office Action and Search Report from the State Intellectual Property Office for CN Pat. Appln. No. 201180066584.7 dated Jul. 10, 2015, 12 pages.
Second Office Action issued by State Intellectual Property Office for CN Pat. Appln. No. 201180066584.7 dated Jan. 11, 2016, 5 pages.
Office Action and Search Report for ROC (Taiwan) Patent Application No. 10014870 dated May 7, 2014, issued in the corresponding foreign application, 9 pages.
Office Action for Korean Patent Application No. 10-2013-7020207 dated Dec. 21, 2016, all pages.
Decision to Grant for Korean Patent Application No. 10-2013-7020207 dated Mar. 9, 2017, all pages.
Office Action and Search Report for ROC (Taiwan) Patent Application No. 100149344 dated Jan. 23, 2015, 8 pages.
Search Report for Patent Application ROC (Taiwan) Patent Application No. 100149344 dated Oct. 28, 2015, 1 page.
Office Action and Search Report for ROC (Taiwan) Pat. Appln. No. 101106313 received May 12, 2016, all pages.
Office Action and Search Report for Roc (Taiwan) Pat. Appln. No. 101106313 received Nov. 4, 2016, all pages.
Office Action and Search Report for ROC (Taiwan) Patent Application No. 100142966 dated May 27, 2014, 6 pages.
Office Action for European Patent Application No. 12707435.9 dated Mar. 12, 2015, 6 pages.
Office Action for European Patent Application No. 12719817.4 dated Jun. 23, 2014 issued in the corresponding foreign application, 5 pages.
First Office Action for CN 201280010873 dated Mar. 2, 2016, all pages.
Notice of Decision to Grant for CN 201280010873 dated Mar. 25, 2016, all pages. (not in English).
Notice of Allowance for Canadian Application 2,822,214 dated Nov. 28, 2016, 1 page.
U.S. Appl. No. 14/179,336, filed Feb. 12, 2014, Non-Final Office Action dated May 22, 2014, 14 pages.
U.S. Appl. No. 14/179,336, filed Feb. 12, 2014 Final Office Action dated Dec. 1, 2014, 30 pages.
U.S. Appl. No. 14/179,336, filed Feb. 12, 2014 Notice of Allowance dated Feb. 18, 2015, 15 pages.
U.S. Appl. No. 13/302,717, filed Nov. 22, 2011 Non Final Rejection dated Jun. 16, 2016, all pages.
U.S. Appl. No. 13/302,717, filed Nov. 22, 2011 Non Final Rejection dated Dec. 2, 2015, 27 pages.
U.S. Appl. No. 13/302,717, filed Nov. 22, 2011 Final Rejection dated May 8, 2015, 44 pages.
U.S. Appl. No. 13/302,717, filed Nov. 22, 2011 Non-Final Rejection dated Dec. 18, 2014, 71 pages.
U.S. Appl. No. 12/958,073, filed Dec. 1, 2010, Office Action dated Aug. 31, 2012, 12 pages.
U.S. Appl. No. 12/958,073, filed Dec. 1, 2010, Notice of Allowance dated Jan. 17, 2013, 17 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Non-Final Office Action dated Mar. 9, 2012, 17 pages.
U.S. Appl. No. 12/964,478, filed Dec. 9, 2010, Non-Final Office Action dated Mar. 26, 2013, 19 pages.
U.S. Appl. No. 12/964,478, filed Dec. 9, 2010, Final Office Action dated Sep. 16, 2013, 12 pages.
U.S. Appl. No. 12/971,349, filed Dec. 17, 2010, Office Action dated Nov. 10, 2011, 9 pages.
U.S. Appl. No. 12/971,349, filed Dec. 17, 2010, Final Office Action dated Jan. 20, 2012, 10 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Notice of Allowance dated Jul. 16, 2014, 15 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Final Rejection dated Oct. 30, 2012, 17 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Non-Final Office Action dated Mar. 25, 2013, 17 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Non-Final Office Action dated Jul. 12, 2013, 22 pages.
U.S. Appl. No. 12/961,369, filed Dec. 6, 2010, Non-Final Office Action dated Feb. 13, 2014, 21 pages.
U.S. Appl. No. 12/971,349, filed Dec. 7, 2010, Notice of Allowance dated Oct. 2, 2013, 24 pages.
U.S. Appl. No. 12/971,349, filed Dec. 7, 2010), Final Rejection dated Oct. 24, 2012, 11 pages.
U.S. Appl. No. 12/971,349, filed Dec. 17, 2010, Office Action dated Jul. 16, 2012, 11 pages.
U.S. Appl. No. 12/981,244, filed Dec. 29, 2010, Office Action dated Dec. 21, 2012, 23 pages.
U.S. Appl. No. 12/981,244, filed Dec. 29, 2010, Final Office Action dated Oct. 30, 2013, 10 pages.
U.S. Appl. No. 12/981,244, filed Dec. 29, 2010, Notice of Allowance dated Mar. 25, 2014, 17 pages.
U.S. Appl. No. 12/984,385, filed Jan. 4, 2011, Notice of Allowance dated Nov. 28, 2012, 11 pages.
U.S. Appl. No. 12/984,385, filed Jan. 4, 2011, Office Action dated Jul. 12, 2012, 16 pages.
U.S. Appl. No. 12/986,721, filed Jan. 7, 2011, Office Action dated Mar. 16, 2012, 6 pages.
U.S. Appl. No. 12/986,721, filed Jan. 7, 2011, Notice of Allowance dated Jun. 21, 2012, 7 pages.
U.S. Appl. No. 12/953,227, filed Nov. 23, 2010, Final Office Action dated May 24, 2013, 17 pages.
U.S. Appl. No. 12/953,227, filed Nov. 23, 2010, Office Action dated Oct. 7, 2012, 31 pages.
U.S. Appl. No. 12/953,227, filed Nov. 23, 2010, Non Final Office action dated Mar. 24, 2015, 39 pages.
U.S. Appl. No. 12/953,227, filed Nov. 23, 2010, Final Office Action dated Nov. 6, 2015, 26 pages.
U.S. Appl. No. 12/953,227, filed Nov. 23, 2010, Notice of Allowance dated May 9, 2017, all pages.
U.S. Appl. No. 13/015,382, filed Jan. 27, 2011, Office Action dated Nov. 13, 2012, 7 pages.
U.S. Appl. No. 13/015,382, filed Jan. 27, 2011, Notice of Allowance dated Feb. 22, 2013, 12 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Office Action dated Nov. 2, 2012, 18 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Final Office Action dated Jun. 27, 2013, 13 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Non-Final Office Action dated Nov. 3, 2014, 33 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Final Office Action dated May 13, 2015, 34 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Non-Final Office Action dated Dec. 14, 2015, 27 pages.
U.S. Appl. No. 13/016,483, filed Jan. 28, 2011 Final Office Action dated Jul. 5, 2016, all pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011 Non Final Rejection dated Feb. 17, 2015, 57 pages.
U.S. Appl. No. 12/953,273, filed Nov. 23, 2010, Notice of Allowance, dated Oct. 18, 2012, 11 pages.
U.S. Appl. No. 12/965,645, filed Dec. 10, 2010, Non-Final Office Action, dated Jul. 19, 2013, 20 pages.
U.S. Appl. No. 12/965,645, filed Dec. 10, 2010, Final Office Action, dated Mar. 18, 2014, 24 pages.
U.S. Appl. No. 12/965,645, filed Dec. 10, 2010, Notice of Allowance, dated Jun. 20, 2014, 35 pages.
U.S. Appl. No. 12/973,431, filed Dec. 20, 2010, Non-Final Rejection dated May 15, 2013, 30 pages.
U.S. Appl. No. 12/973,431, filed Dec. 20, 2010, Final Office Action dated Aug. 27, 2013, 11 pages.
U.S. Appl. No. 12/973,431, filed Dec. 20, 2010 Non-Final Rejection dated Dec. 19, 2014, 30 pages.
U.S. Appl. No. 12/973,431, filed Dec. 20, 2010 Notice of Allowance dated May 28, 2015, 20 pages.
U.S. Appl. No. 13/007,317, filed Jan. 14, 2011, Office action dated Dec. 19, 2012, 29 pages.
U.S. Appl. No. 13/010,557, filed Jan. 20, 2011, Final Rejection dated Jan. 16, 2014, 17 pages.
U.S. Appl. No. 13/010,557, filed Jan. 20, 2011, Non-Final Rejection dated Aug. 5, 2013, 17 pages.
U.S. Appl. No. 13/014,591, Notice of Allowance dated May 24, 2013, 32 pages.
U.S. Appl. No. 13/020,678, filed Feb. 3, 2011, Office Action dated Jul. 30, 2012, 15 pages.
U.S. Appl. No. 13/020,678, filed Feb. 3, 2011, Notice of Allowance dated Jan. 3, 2013, 13 pages.
U.S. Appl. No. 13/007,317, Notice of Allowance dated May 13, 2013, 16 pages.
U.S. Appl. No. 13/028,030, filed Feb. 15, 2011, Office Action dated Jan. 11, 2013, 14 pages.
U.S. Appl. No. 13/028,030, filed Feb. 15, 2011, Final Office Action dated Jul. 11, 2014, 43 pages.
U.S. Appl. No. 13/028,030, filed Feb. 15, 2011, Non-Final Office Action dated Feb. 6, 2015, 56 pages.
U.S. Appl. No. 13/028,030, filed Feb. 15, 2011, Final Office Action dated Jul. 17, 2015, 63 pages.
U.S. Appl. No. 13/031,115, Notice of Allowance dated Apr. 16, 2013, 24 pages.
U.S. Appl. No. 13/034,482, filed Feb. 24, 2011 Notice of Allowance dated Aug. 29, 2014, 45 pages.
U.S. Appl. No. 13/034,482, filed Feb. 24, 2011, Final Office Action dated Apr. 25, 2013, 19 pages.
U.S. Appl. No. 13/034,482, filed Feb. 24, 2011, Office Action dated Oct. 19, 2012, 11 pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011, Office Action dated Oct. 30, 2012, 11 pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011, Final Rejection dated Mar. 29, 2013, 20 pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011, Non Final Rejection dated Mar. 6, 2014, 20 pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011 Final Rejection dated Aug. 27, 2014, 38 pages.
U.S. Appl. No. 13/035,474, filed Feb. 25, 2011 Non Final Rejection dated Sep. 11, 2015, 65 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010 Non-Final Office Action dated May 14, 2015, 21 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010 Final Office Action dated Dec. 3, 2014, 19 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010, Non-Final Office Action dated Jun. 6, 2014, 19 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010, Final Office Action dated Apr. 18, 2013, 14 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010, Non-Final Office Action dated Dec. 6, 2012, 17 pages.
U.S. Appl. No. 12/960,285, filed Dec. 3, 2010, Notice of Allowance dated Nov. 18, 2015, 31 pages.
U.S. Appl. No. 13/006,270, filed Jan. 13, 2011, Non-Final Office Action dated Oct. 8, 2013, 20 pages.
U.S. Appl. No. 13/006,270, filed Jan. 13, 2011, Final Office Action dated May 9, 2014, 41 pages.
U.S. Appl. No. 13/006,270, filed Jan. 13, 2011 Non-Final Office Action dated Sep. 12, 2014, 41 pages.
U.S. Appl. No. 13/006,270, filed Jan. 13, 2011, Final Office Action dated Mar. 23, 2014, 51 pages.
U.S. Appl. No. 13/028,030, filed Feb. 15, 2011 Non-Final Office Action dated Dec. 17, 2013, 60 pages.
U.S. Appl. No. 13/035,525, filed Feb. 25, 2011, Office Action dated Jul. 18, 2012, 15 pages.
U.S. Appl. No. 13/035,525, filed Feb. 25, 2011, Final Office Action dated Jan. 31, 2013, 26 pages.
U.S. Appl. No. 13/035,525, filed Feb. 25, 2011, Non-Final Office Action dated May 15, 2013, 15 pages.
U.S. Appl. No. 13/035,525, filed Feb. 25, 2011, Final Office Action dated Sep. 12, 2013, 21 pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Office Action dated Mar. 1, 2013, 20 pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Final Office Action dated Oct. 16, 2013, 28 pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Final Office Action dated May 4, 2015, 54 pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Non-Final Office Action dated Jan. 12, 2016, 62 pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Final Office Action dated Jul. 12, 2016, all pages.
U.S. Appl. No. 13/037,302, filed Feb. 28, 2011, Notice of Allowance dated Feb. 16, 2017, all pages.
U.S. Appl. No. 13/037,312, filed Feb. 28, 2011, Office Action dated Aug. 15, 2012, 9 pages.
U.S. Appl. No. 13/037,312, filed Feb. 28, 2011, Notice of Allowance dated Jun. 13, 2013, 10 pages.
U.S. Appl. No. 13/037,312, filed Feb. 28, 2011, Final Office Action dated Feb. 28, 2013, 18 pages.
U.S. Appl. No. 13/037,316, filed Feb. 28, 2011, Office Action dated Jan. 30, 2013, 21 pages.
U.S. Appl. No. 13/037,316, filed Feb. 28, 2011, Final Office Action dated Aug. 28, 2013, 13 pages.
U.S. Appl. No. 13/037,333, filed Feb. 28, 2011, Notice of Allowance dated Jan. 18, 2013, 27 pages.
U.S. Appl. No. 13/192,287, filed Jul. 27, 2011, Notice of Allowance dated Dec. 14, 2015, 14 pages.
U.S. Appl. No. 13/192,287, filed Jul. 27, 2011, Final Office Action dated Jan. 28, 2014, 18 pages.
U.S. Appl. No. 13/192,287, filed Jul. 27, 2011, Non Final Office Action dated Jun. 13, 2013, 22 pages.
U.S. Appl. No. 13/673,480, filed Nov. 9, 2012, Office Action dated Jan. 16, 2013, 27 pages.
U.S. Appl. No. 13/673,480, filed Nov. 9, 2012 Final Office Action dated Sep. 9, 2013, 10 pages.
U.S. Appl. No. 13/673,480, filed Nov. 9, 2012 Notice of Allowance dated Nov. 12, 2013, 16 pages.
U.S. Appl. No. 13/475,794, filed May 18, 2012 Non-Final Office Action dated Sep. 18, 2013, 19 pages.
U.S. Appl. No. 13/475,794, filed May 18, 2012 Non-Final Office Action dated Nov. 21, 2014, 33 pages.
U.S. Appl. No. 13/475,794, filed May 18, 2012 Final Office Action dated Jun. 1, 2015, 45 pages.
U.S. Appl. No. 13/475,794, filed May 18, 2012 Non Final Office Action dated Jul. 29, 2016, all pages.
U.S. Appl. No. 13/475,794, filed May 18, 2012 Notice of Allowance dated Jan. 5, 2017, all pages.
U.S. Appl. No. 13/864,474, filed Apr. 17, 2013 Non Final Office Action dated Aug. 11, 2015, 59 pages.
U.S. Appl. No. 13/864,474, filed Apr. 17, 2013 Final Office Action dated Nov. 20, 2015, all pages.
U.S. Appl. No. 13/864,474, filed Apr. 17, 2013 Non Final Office Action dated Mar. 23, 2016, all pages.
U.S. Appl. No. 13/864,474, filed Apr. 17, 2013 Notice of Allowance dated Feb. 16, 2017, all pages.
U.S. Appl. No. 13/968,611, filed Aug. 16, 2013, Notice of Allowance dated May 2, 2014, 40 pages.
U.S. Appl. No. 13/968,611, filed Aug. 16, 2013, Non-Final Office Action dated Jan. 17, 2014, 21 pages.
U.S. Appl. No. 14/852,787, filed Sep. 14, 2015, Non-Final Office Action dated Sep. 14, 2016, all pages.
U.S. Appl. No. 14/852,787, filed Sep. 14, 2015, Final Office Action dated Jan. 13, 2017, all pages.
Third Office Action from State Intellectual Property Office for CN Appln. No. CN 201280013891.3 dated Dec. 30, 2016, all pages.
Related Publications (1)
Number Date Country
20170366806 A1 Dec 2017 US
Continuations (1)
Number Date Country
Parent 13037302 Feb 2011 US
Child 15639871 US