Advancements in media delivery systems and media-related technologies continue to increase at a rapid pace. Increasing demand for media has influenced the advances made to media-related technologies. Computer systems have increasingly become an integral part of the media-related technologies. Computer systems may be used to carry out several media-related functions. The widespread access to media has been accelerated by the increased use of computer networks, including the Internet and cloud networking.
Many businesses use one or more computer networks to deliver media between the various computers connected to the networks. Users of computer technologies continue to demand increased access to media and an increase in the efficiency of these technologies. Improving the efficiency of computer technologies is always desirable to anyone who uses and relies on computers.
With the wide-spread use of computers and mobile devices has come an increased presence of automation and home security products. Advancements in automation allow users to control aspects of their home using mobile computing devices. However, current implementations of automation may involve multiple devices in a piecemeal fashion.
According to at least one embodiment, a computer-implemented method for integration of a set top box and an automation system is described. In one configuration, subscriber program content may be provided. An aspect of the premises may be monitored via one or more sensors. A monitor channel may be provided to display the monitored aspect of the premises. The monitor channel may be displayed in a channel guide among channels of the subscriber program content.
In some embodiments, the method may include tracking a plurality of entities associated with the premises and providing a tracking channel among channels of the subscriber program content. The tracking channel may display a map of the tracked plurality of entities. In some cases, the method may include modifying a native application programming interface (API) of the set top box in order to provide the monitor channel in the channel guide among the channels of the subscriber program content. Upon receiving a user command, the set top box may be configured to record the monitor channel for a specified duration on a specified date.
In one embodiment, a predetermined change related to the one or more sensors may be detected. A notification may be displayed on a display connected to the set top box. A user input may be received in reply to the displaying of the notification. The notification may be displayed in relation to a presently viewed channel of the subscriber program content. In some embodiments, the method includes pausing a channel from the subscriber program content and displaying the notification on at least a portion of the paused channel.
In one embodiment, upon detecting a first user input, an automation user interface may be displayed. The automation user interface may include information regarding the one or more sensors. Upon detecting a second user input in relation to the automation user interface, an aspect of at least one of the one or more sensors may be adjusted. In some embodiments, the first and second user inputs may include a voice command input, a video gesture input (e.g., a predetermined pattern of movement made by the user and detected by a camera), tactile gesture input (e.g., a predetermined pattern of movement made by the user on a touch pad), a mobile computing device input, a remote control device input, and the like.
In one example, subscriber program content may be provided. An action of a user may be monitored. Upon determining the monitored action of the user matches a predetermined user action, an aspect of at least one of the one or more sensors may be adjusted based on a current state of the one or more sensors. In one embodiment, the method includes detecting the action of the user via a camera connected to the set top box. In some cases the method includes identifying a pattern associated with the monitored action of the user, detecting an initiation of the identified pattern associated with the monitored action of the user, and/or upon detecting the initiation of the identified pattern, adjusting an aspect of the one or more sensors.
A computing device configured to integrate a set top box and an automation system is also described. The device may include a processor and memory in electronic communication with the processor. The memory may store instructions that may be executable by the processor to provide subscriber program content, monitor an aspect of a premises using one or more sensors, and provide a monitor channel to display the monitored aspect of the premises, the monitor channel being displayed in a channel guide among channels of the subscriber program content.
A computer-program product to integrate a set top box and an automation system is also described. The computer-program product may include a non-transitory computer-readable medium that stores instructions. The instructions may be executable by the processor to provide subscriber program content, monitor an aspect of a premises using one or more sensors, and provide a monitor channel to display the monitored aspect of the premises, the monitor channel being displayed in a channel guide among channels of the subscriber program content.
Features from any of the above-mentioned embodiments may be used in combination with one another in accordance with the general principles described herein. These and other embodiments, features, and advantages will be more fully understood upon reading the following detailed description in conjunction with the accompanying drawings and claims.
The accompanying drawings illustrate a number of exemplary embodiments and are a part of the specification. Together with the following description, these drawings demonstrate and explain various principles of the instant disclosure.
While the embodiments described herein are susceptible to various modifications and alternative forms, specific embodiments have been shown by way of example in the drawings and will be described in detail herein. However, the exemplary embodiments described herein are not intended to be limited to the particular forms disclosed. Rather, the instant disclosure covers all modifications, equivalents, and alternatives falling within the scope of the appended claims.
The systems and methods described herein relate to automation. More specifically, the systems and methods described herein relate to integrating automation in a subscriber-based media content set top box, such as a satellite and/or cable digital video recorder (DVR). Currently, when a user is viewing programming delivered from a subscription content provider and wants to check a video camera feed or another automation input, the user typically exits the set top box and switches to a separate video input to view the automation input. The systems and methods described herein, however, allow a user to view a video camera feed from the subscriber-based media content set top box.
In some configurations, the device 105 may include a user interface 135, application 140, and automation module 145. Although the components of the device 105 are depicted as being internal to the device 105 it is understood that one or more of the components may be external to the device 105 and connect to device 105 through wired and/or wireless connections. In some embodiments, application 140 may be located on mobile computing device 155 in order to allow a user to interface with a function of device 105 and/or automation module 145.
In some embodiments, device 105 may communicate with service provider 110 via network 115. Example of networks 115 include cloud networks, local area networks (LAN), wide area networks (WAN), virtual private networks (VPN), wireless networks (using 802.11, for example), cellular networks (using 3G and/or LTE, for example), etc. In some configurations, the network 115 may include the internet. It is noted that in some embodiments, the device 105 may not include an automation module 145. For example, the device 105 may include the application 140 that allows the device 105 to interface with the automation controller 160 via the automation module 145 located on service provider 110. In some embodiments, the device 105, the automation controller 160, and the service provider 110 may include an automation module 145 where at least a portion of the functions of the automation module 145 are performed separately and/or concurrently on the device 105, automation controller 160, and/or the service provider 110. Likewise, in some embodiments, a user may access the functions of device 105 and/or automation controller 160 (directly or through device 105 via automation module 145) from mobile computing device 155. For example, in some embodiments, mobile computing device 155 includes a mobile application that interfaces with one or more functions of device 105, automation controller 160, automation module 145, and/or service provider 110.
In some embodiments, the service provider 110 may be coupled to database 120. For example, device 105 may access program content 150 in database 120 over the network 115 via service provider 110. Database 120 may be internal or external to the service provider 110. In one example, the device 105 may be coupled to database 120 via network 115.
Automation module 145 may allow a user to control (either directly or via automation controller 160), from a subscription-content media set top box, an aspect of the premises, including security, locking or unlocking a door, checking the status of a door, locating a person or item, controlling lighting, thermostat, cameras, and the like. In some configurations, application 140 may enable device 105 to interface with automation controller 160 via automation module 145 to provide automation content to device 105 and/or mobile computing device 155. Thus, application 140, via the automation module 145, may allow users to control aspects of their home. Further details regarding the automation module 145 are discussed below.
In some embodiments, service provider 110 may provide subscriber program content (cable/satellite television programming, for example) to a user via a set top box located in a home, office, etc. of the user. Examples of set top boxes include cable set top boxes, satellite set top boxes, DVRs, personal video recorders (PVRs), and the like. Monitoring module 205 may be configured to monitor an entity, such as a security camera, a door lock, a door status, a smoke alarm (or other type of sensor), a home appliance, utility equipment, energy/utility usage, and the like. Home appliances may include a refrigerator, oven, microwave oven, stove, dishwasher, washer, dryer, and the like. Utility equipment may include a garage door opening system, heating ventilation air conditioning (HVAC) equipment, culinary water equipment (water heater, water softener, water meter, fire alarm, in-home sprinkler system, etc.), telephony equipment, irrigation water equipment (lawn sprinkler system, etc.), natural gas system (carbon monoxide sensor, gas meter, gas detection system, etc.), and the like. Monitoring module 205 may be configured to detect a predetermined change related to the entity or the triggering of a predetermined threshold related to the entity (carbon monoxide levels exceed a certain level, for example).
In one embodiment, notification module 210, in conjunction with the user interface 135, may display a notification on a display connected to the set top box. Notification module 210 may display the notification in relation to a presently viewed channel of the subscriber program content. In some cases, notification module 210 may display the notification in at least a portion of a viewing area of the channel. For example, a user watching content from a satellite DVR on a television may receive a notification displayed on at least a portion of the screen of the television. In some embodiments, entity control module 225, in conjunction with user interface 135, may receive a user input in response to the notification. For example, a notification relating to an oven (e.g., an oven timer), a washer/dryer cycle, a dishwasher wash cycle, water softener operation (e.g., time for a refill of the water softener), furnace operation (e.g., time to replace a furnace air filter), water heater, air conditioner, refrigerator, microwave oven, current energy use, energy use patterns, water use, and the like, may be displayed in conjunction with the content provided by a subscriber set top box. In one example, a user may set a timer for a meal cooking in an oven. With one minute left on the timer, the notification module 210 may display a notification on a channel presently being viewed by the user. When the notification is displayed, the content currently being viewed by the user on the channel may be paused. When a response to the notification is received or after a predetermined amount of time has lapsed, the content may resume.
In some embodiments, notification 210 may adapt a notification based on certain conditions. Notification module 210 may determine whether to display a notification on a display based on notification rules associated with a current state of device 105. For example, notification 210 may determine whether to display the notification based on the content provided by a set top box that a user is currently viewing. For instance, if a user is watching a particular sports event (e.g., WORLD SERIES®, SUPERBOWL®, etc.) notification 210 may block the notification from being displayed. In some cases, notification module 210 may alter the notification based on the current state of the device 105. For one type of programming and set top box content, notification module 210 may show scrolling text across an edge of the television screen. For another type of programming, notification module 210 may play a sound (e.g., a tone, a combination of tones, a recorded voice notification, a text to speech notification, and the like) based on the type of notification. In some embodiments, notification module 210 may alter the destination of the notification based on the type of notification and/or the type of programming currently being viewed by the user. For example, if the notification rules require minimum notification for a certain type of programming, the notification module 210 may display in alternative modes of communication. For instance, notification module 210 may email or text message a notification to a user based on the type of notification, the type of programming the user is currently watching, whether programming is currently being watched, and/or the current location of the user, etc.
In some cases, notification 210 may determine whether to display a notification in relation to programming currently being watched by the user and how to handle the programming based on the type of notification and/or the type of programming. For example, automation module 145-a may pause the programming for one type of notification, show a pop up message for another type of notification without pausing the programming, split the screen between a window for a security camera view and a window for the programming for another type of notification, and so forth. Additionally, or alternatively, notification module 210 may pause the programming for one type of notification based on one type of programming, block the notification for the same type of notification based on another type of programming, and continue showing the programming for the same type of notification based on a third type of programming.
The user may reply to the timer notification by dismissing the notification. Additionally, or alternatively, the user may respond by changing the time left on the timer. For example, the user may add additional time (e.g., 30 seconds) when the timer expires. In some embodiments, the notification module 210 may provide two or more options for the user to select (e.g., dismiss the notification, add 30 seconds to the timer, add 60 seconds to the timer, and the like), receive a selection from the user, and update the timer accordingly. In some embodiments, the automation module 145 may interact with one or more sensors associated with the oven, such as an oven camera, an oven temperature probe, etc. The notification module 210, may include the current state of the one or more sensors associated with the oven in the timer notification. Upon receiving a notification that the remaining time on the timer has expired, in some embodiments, the notification module 210 may notify the user that the timer has expired.
In some embodiments, as mentioned above, upon detecting a predetermined change related to the entity, media module 230 may pause subscriber program content currently provided on a channel. For example, the automation module 145-a may pause and/or record a portion of the content the user was viewing when the notification is displayed. In some cases, the notification may include a picture in picture notification, a text bar, scrolling text, a pop up, or drop down notification message. In some cases, the presently viewed channel may be paused upon the notification module 210 displaying the notification.
In one embodiment, the monitoring module 205 may monitor an entity associated with a home of a user in relation to content provided by a subscriber set top box. In one example, upon detecting a first user input, automation module 145-a, in conjunction with the execution of application 140, may be configured to display an automation user interface (e.g., user interface 135). The automation user interface may include access to information regarding the one or more sensors. Upon detecting a second user input in relation to the automation user interface, entity control module 225 may adjust an aspect of at least one of the one or more sensors.
User inputs comprise at least one of a voice command input, a video gesture input, tactile gesture input, a mobile computing device input, and a remote control device input. In some embodiments, automation module 145-a, in conjunction with the application 140 and user interface 135, may remotely interface with the automation user interface via a mobile application executing on a mobile computing device. For example, a user may press a menu button on a remote control configured to control a satellite DVR. A menu may appear that include access to automation functions, which may be provided as an automation application (e.g., application 140) executable from and within the satellite DVR. Upon selecting the automation function, the user may view, modify, or set a function related to automation.
In one embodiment, media module 230 may provide subscriber program content via a set top box. Monitoring module 205 may monitor both an entity associated with the premises and an action of the user. Thus, automation module 145-a may be configured to correlate a current state of an entity with a detected action of the user. Upon detecting that the monitored action of the user matches a predetermined user action, entity control module 225 may be configured to adjust an aspect of at least one of the one or more sensors based on the current state of the one or more sensors. In some embodiments, monitoring module 205, in conjunction with camera 125, may detect the action of the user via a camera connected to the set top box (e.g., camera 125). For example, monitoring module 205 may perform facial recognition and other recognition algorithms to detect and recognize an identity of a user. In some embodiments, a satellite DVR may filter available content based on user recognition. For example, a user may allow a child of a certain age to watch only certain programming appropriate to the child's age. Upon recognizing that the child is operating the satellite DVR, the automation module 145 may limit the content provided by the satellite DVR to the child to only age-appropriate content. In some embodiments, automation controller 160 may monitor a current state of device 105 (e.g., a satellite DVR is on, off, presently switched off, the time at which it is turned on or off, program is currently being watched, etc.). Based on the present state of the satellite DVR, automation controller 160 may execute one or more automation commands automatically. For example, automation controller 160 may detect that a user switches off the satellite DVR at or after a certain time (e.g., at or after 11:00 P.M.). In response, automation controller 160 may lock a door or window that it detects is currently unlocked, warn the user of a door or window currently open, turn on or off one or more certain lights in or outside the home, adjust a temperature on a thermostat, and the like.
In some embodiments, pattern detection module 215 may be configured to identify a pattern associated with the monitored action of the user. In some embodiments, pattern detection module 215 may detect an initiation of the identified pattern associated with the monitored action of the user. Upon detecting the initiation of the identified pattern, entity control module 225 may adjust an aspect of at least one of the one or more sensors. For example, pattern detection module 215 may detect that a certain user dims the lights to a certain level, watches a certain program at certain times each week, and routinely turns off the dishwasher when it is running during the certain times. Upon recognizing the pattern, automation module 145 may automatically turn on the television, tune a satellite and/or cable set top box to a particular channel, dim the lights to a predetermined level, and pause the dishwasher cycle, if it is currently running. Upon detecting the end period of the satellite content, automation module 145-a may resume the dishwasher cycle.
In one embodiment, media module 230 may provide subscriber program content via a set top box located in a premises. Monitoring module 205 may monitor an aspect of the premises. Media module 230, in conjunction with the user interface 135, may provide a monitor channel to display the monitored aspect of the home to the user. The monitor channel may be displayed in a channel guide among channels of the subscriber program content. For example, a subscription to cable or satellite media content may include a channel guide that includes a list of selectable channels and their respective content. Among the list of channels provided by subscription, automation module 145-a may append one or more automation channels that are selectable and interact with user commands in a similar manner as the subscription channels in the channel guide.
In one embodiment, entity control module 225 may track a plurality of entities associated with a premises. Tracked entities may include persons, pets, objects, etc., allowing the media set top box to track a person, a pet, and/or an item (e.g., set of keys, a mobile computing device, a remote control, and the like) using one or more technologies related to global positioning satellite (GPS) system, radio frequency identification (RFID), wireless (e.g., 802.11, etc.), near-field communications (NFC), and the like. Media module 230 may provide a tracking channel, similar to the monitor channel described above, to display a map of the tracked plurality of entities among channels of the subscriber program content. In some embodiments, automation module 145-a may provide an automation preview channel. The automation preview channel may provide a real-time preview of two or more automation channels. Media module 230 may access a native application programming interface (API) of the set top box in order to provide the monitoring, tracking, and other automation channels in the channel guide among the channels of the subscriber program content. In some embodiments, media module 230 may modify an aspect of the set top box API in order to provide the automation channels in the channel guide among the subscription channels. Upon receiving a user command, media module 230, in conjunction with the user interface 135, may schedule the set top box to record the monitor channel, tracking channel, or other automation channels for a specified duration on a specified date. In some embodiments, a recording of an automation channel may be event-driven, such as recording an automation channel upon detecting motion at a security camera communicatively connected to the set top box. In some embodiments, upon detecting motion at a front door camera, automation module 145-a may pause the content a user is currently viewing from the media set top box and switch to a channel on the set top box dedicated to displaying the content of the front door camera. In some embodiments, automation module 145-a may split the screen to show the current content provided by a media provider and the front door camera channel. Alternatively, automation module 145-a may display the front door camera in a picture in picture element over the current channel.
At box 305, subscriber program content may be provided to a user. At box 310, an aspect of a premises may be monitored via one or more sensors (e.g., automation and/or security sensors). At block 315, a monitor channel may be provided to display the monitored aspect of the premises. The monitor channel may be displayed in a channel guide among channels of the subscriber program content.
At step 405, subscriber program content may be provided to a user via a processor in a set top box. At step 410, an entity associated with a premises may be monitored by the set top box. At step 415, a predetermined change related to the entity may be detected. At step 420, a notification may be displayed on a display connected to the set top box. At step 425, a user input may be received in response to the displayed notification.
At block 505, an entity associated with a premises of a user may be monitored. At block 510, upon detecting a first user input, a building automation user interface may be displayed. The building automation user interface may include information regarding at least one of the one or more sensors. At block 515, upon detecting a second user input in relation to the automation user interface, an aspect of at least one of the one or more sensors may be adjusted. In some embodiments, the first and second user inputs may include a voice command input, a video gesture input (a predetermined pattern of movement made by the hand and detected by a camera, for example), tactile gesture input (a predetermined pattern of movement made by the hand on a touch pad, for example), a mobile computing device input, a remote control device input, and the like.
At step 605, subscriber program content may be provided to a user. At block 610, an entity associated with a premises may be monitored. At block 615, an action of the user may be monitored. At block 620, upon detecting the monitored action of the user matches a predetermined user action, an aspect of at least one of the one or more sensors may be adjusted based on a current state of the one or more sensors.
Bus 705 allows data communication between central processor 710 and system memory 715, which may include read-only memory (ROM) or flash memory (neither shown), and random access memory (RAM) (not shown), as previously noted. The RAM is generally the main memory into which the operating system and application programs are loaded. The ROM or flash memory can contain, among other code, the Basic Input-Output system (BIOS) which controls basic hardware operation such as the interaction with peripheral components or devices. For example, the automation module 145-b to implement the present systems and methods may be stored within the system memory 715. Applications (e.g., application 140) resident with controller 700 are generally stored on and accessed via a non-transitory computer readable medium, such as a hard disk drive (e.g., fixed disk 775) or other storage medium. Additionally, applications can be in the form of electronic signals modulated in accordance with the application and data communication technology when accessed via interface 785.
Storage interface 780, as with the other storage interfaces of controller 700, can connect to a standard computer readable medium for storage and/or retrieval of information, such as a fixed disk drive 775. Fixed disk drive 775 may be a part of controller 700 or may be separate and accessed through other interface systems. Network interface 785 may provide a direct connection to a remote server via a direct network link to the Internet via a POP (point of presence). Network interface 785 may provide such connection using wireless techniques, including digital cellular telephone connection, Cellular Digital Packet Data (CDPD) connection, digital satellite data connection, or the like. In some embodiments, one or more sensors (e.g., motion sensor, smoke sensor, glass break sensor, door sensor, window sensor, carbon monoxide sensor, and the like) connect to controller 700 wirelessly via network interface 785.
Many other devices or subsystems (not shown) may be connected in a similar manner (e.g., entertainment system, computing device, remote cameras, wireless key fob, wall mounted user interface device, cell radio module, battery, alarm siren, door lock, lighting system, thermostat, home appliance monitor, utility equipment monitor, and so on). Conversely, all of the devices shown in
Moreover, regarding the signals described herein, those skilled in the art will recognize that a signal can be directly transmitted from a first block to a second block, or a signal can be modified (e.g., amplified, attenuated, delayed, latched, buffered, inverted, filtered, or otherwise modified) between the blocks. Although the signals of the above described embodiment are characterized as transmitted from one block to the next, other embodiments of the present systems and methods may include modified signals in place of such directly transmitted signals as long as the informational and/or functional aspect of the signal is transmitted between blocks. To some extent, a signal input at a second block can be conceptualized as a second signal derived from a first signal output from a first block due to physical limitations of the circuitry involved (e.g., there will inevitably be some attenuation and delay). Therefore, as used herein, a second signal derived from a first signal includes the first signal or any modifications to the first signal, whether due to circuit limitations or due to passage through other circuit elements which do not change the informational and/or final functional aspect of the first signal.
While the foregoing disclosure sets forth various embodiments using specific block diagrams, flowcharts, and examples, each block diagram component, flowchart step, operation, and/or component described and/or illustrated herein may be implemented, individually and/or collectively, using a wide range of hardware, software, or firmware (or any combination thereof) configurations. In addition, any disclosure of components contained within other components should be considered exemplary in nature since many other architectures can be implemented to achieve the same functionality.
The process parameters and sequence of steps described and/or illustrated herein are given by way of example only and can be varied as desired. For example, while the steps illustrated and/or described herein may be shown or discussed in a particular order, these steps do not necessarily need to be performed in the order illustrated or discussed. The various exemplary methods described and/or illustrated herein may also omit one or more of the steps described or illustrated herein or include additional steps in addition to those disclosed.
Furthermore, while various embodiments have been described and/or illustrated herein in the context of fully functional computing systems, one or more of these exemplary embodiments may be distributed as a program product in a variety of forms, regardless of the particular type of computer-readable media used to actually carry out the distribution. The embodiments disclosed herein may also be implemented using software modules that perform certain tasks. These software modules may include script, batch, or other executable files that may be stored on a computer-readable storage medium or in a computing system. In some embodiments, these software modules may configure a computing system to perform one or more of the exemplary embodiments disclosed herein.
The foregoing description, for purpose of explanation, has been described with reference to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the present systems and methods and their practical applications, to thereby enable others skilled in the art to best utilize the present systems and methods and various embodiments with various modifications as may be suited to the particular use contemplated.
Unless otherwise noted, the terms “a” or “an,” as used in the specification and claims, are to be construed as meaning “at least one of.” In addition, for ease of use, the words “including” and “having,” as used in the specification and claims, are interchangeable with and have the same meaning as the word “comprising.” In addition, the term “based on” as used in the specification and the claims is to be construed as meaning “based at least upon.”
The present Application is a continuation of U.S. patent application Ser. No. 14/291,835, filed May 30, 2014, titled “SET TOP BOX AUTOMATION,” which claims the benefit of U.S. Provisional Patent Application No. 61/834,409, titled “SYSTEMS AND METHODS FOR SET TOP BOX HOME AUTOMATION,” filed on Jun. 12, 2013, which is assigned to the assignee hereof, the disclosures of which are expressly incorporated herein in their entireties by this reference.
Number | Name | Date | Kind |
---|---|---|---|
5252947 | Marciano | Oct 1993 | A |
5289362 | Liebl et al. | Feb 1994 | A |
5398057 | Tapp | Mar 1995 | A |
5572438 | Ehlers et al. | Nov 1996 | A |
5657076 | Tapp | Aug 1997 | A |
5815219 | Okuizumi | Sep 1998 | A |
5822012 | Jeon et al. | Oct 1998 | A |
6529230 | Chong | Mar 2003 | B1 |
6792319 | Bilger | Sep 2004 | B1 |
6868292 | Ficco et al. | Mar 2005 | B2 |
6868293 | Schurr et al. | Mar 2005 | B1 |
6874691 | Hildebrand | Apr 2005 | B1 |
7061393 | Buckingham et al. | Jun 2006 | B2 |
7119675 | Khandelwal et al. | Oct 2006 | B2 |
7124397 | Mathur et al. | Oct 2006 | B1 |
7129833 | Albert | Oct 2006 | B2 |
7181293 | Rothman et al. | Feb 2007 | B2 |
7260610 | Grooters et al. | Aug 2007 | B2 |
7298871 | Lee et al. | Nov 2007 | B2 |
7327251 | Corbett, Jr. | Feb 2008 | B2 |
7342488 | Wolfe et al. | Mar 2008 | B2 |
7403598 | Tyroler et al. | Jul 2008 | B2 |
7403838 | Deen et al. | Jul 2008 | B2 |
7425977 | Sakai | Sep 2008 | B2 |
7461343 | Kates | Dec 2008 | B2 |
7558671 | Dossas et al. | Jul 2009 | B2 |
7561977 | Horst et al. | Jul 2009 | B2 |
7565227 | Richard et al. | Jul 2009 | B2 |
7616123 | Ridder et al. | Nov 2009 | B2 |
7650319 | Hoffberg et al. | Jan 2010 | B2 |
7664239 | Groff et al. | Feb 2010 | B2 |
7719440 | Delp et al. | May 2010 | B2 |
7743096 | Bouilloux-Lafont et al. | Jun 2010 | B2 |
7746223 | Howarter et al. | Jun 2010 | B2 |
7786891 | Owens et al. | Aug 2010 | B2 |
7924149 | Mendelson | Apr 2011 | B2 |
7962485 | Trandal et al. | Jun 2011 | B1 |
7978059 | Petite et al. | Jul 2011 | B2 |
7990985 | Chen | Aug 2011 | B2 |
8006268 | Sloo | Aug 2011 | B2 |
8024073 | Imes et al. | Sep 2011 | B2 |
8042048 | Wilson et al. | Oct 2011 | B2 |
8049613 | Poder | Nov 2011 | B2 |
8091794 | Siddaramanna et al. | Jan 2012 | B2 |
8095233 | Shankar et al. | Jan 2012 | B1 |
8098281 | Croak et al. | Jan 2012 | B1 |
8099195 | Imes et al. | Jan 2012 | B2 |
8108076 | Imes et al. | Jan 2012 | B2 |
8159334 | McClenny et al. | Apr 2012 | B2 |
8166498 | Walter | Apr 2012 | B2 |
8181199 | Jeong | May 2012 | B2 |
8200186 | Ashley, Jr. et al. | Jun 2012 | B2 |
8234676 | Klein et al. | Jul 2012 | B2 |
8239073 | Fausak et al. | Aug 2012 | B2 |
8242900 | Bennett, III et al. | Aug 2012 | B2 |
8244102 | Bhogal et al. | Aug 2012 | B2 |
8289182 | Vogel et al. | Oct 2012 | B2 |
8295990 | Venkatakrishnan et al. | Oct 2012 | B2 |
8299889 | Kumar et al. | Oct 2012 | B2 |
8299939 | Gudlavenkatasiva et al. | Oct 2012 | B2 |
8301270 | Quail | Oct 2012 | B2 |
8316413 | Crabtree | Nov 2012 | B2 |
8321885 | Pino, Jr. et al. | Nov 2012 | B2 |
8332055 | Veillette | Dec 2012 | B2 |
8374728 | Kim | Feb 2013 | B2 |
8386082 | Oswald | Feb 2013 | B2 |
8390462 | Belz et al. | Mar 2013 | B2 |
8397264 | Ansari et al. | Mar 2013 | B2 |
8400322 | Acedo et al. | Mar 2013 | B2 |
8411208 | McRae | Apr 2013 | B2 |
8413204 | White et al. | Apr 2013 | B2 |
8423000 | Dhuna | Apr 2013 | B2 |
8429707 | Patel et al. | Apr 2013 | B2 |
8442501 | Rao et al. | May 2013 | B1 |
8447289 | Rao et al. | May 2013 | B2 |
8467908 | Broniak et al. | Jun 2013 | B2 |
8472937 | Rao et al. | Jun 2013 | B1 |
8473111 | Shankar et al. | Jun 2013 | B1 |
8489242 | Valluri et al. | Jul 2013 | B2 |
8498749 | Imes et al. | Jul 2013 | B2 |
8504921 | Wilson et al. | Aug 2013 | B2 |
8510255 | Fadell et al. | Aug 2013 | B2 |
20020082748 | Enga et al. | Jun 2002 | A1 |
20020094111 | Puchek et al. | Jul 2002 | A1 |
20020097165 | Hulme et al. | Jul 2002 | A1 |
20030070182 | Mead | Apr 2003 | A1 |
20040010808 | deCarmo | Jan 2004 | A1 |
20040024717 | Sneeringer | Feb 2004 | A1 |
20040025179 | Russ et al. | Feb 2004 | A1 |
20040034873 | Zenoni | Feb 2004 | A1 |
20040049785 | Grzeczkowski et al. | Mar 2004 | A1 |
20040059815 | Buckingham et al. | Mar 2004 | A1 |
20040083128 | Buckingham et al. | Apr 2004 | A1 |
20040148632 | Park et al. | Jul 2004 | A1 |
20040212498 | Peterson et al. | Oct 2004 | A1 |
20040254683 | Kim | Dec 2004 | A1 |
20050086341 | Enga et al. | Apr 2005 | A1 |
20050108091 | Sotak et al. | May 2005 | A1 |
20050159863 | Howard et al. | Jul 2005 | A1 |
20050228883 | Brown | Oct 2005 | A1 |
20060022816 | Yukawa | Feb 2006 | A1 |
20060080408 | Istvan et al. | Apr 2006 | A1 |
20060168618 | Choi | Jul 2006 | A1 |
20060190974 | Lee | Aug 2006 | A1 |
20060202832 | Reznik et al. | Sep 2006 | A1 |
20060270421 | Phillips et al. | Nov 2006 | A1 |
20060277576 | Acharya et al. | Dec 2006 | A1 |
20060293100 | Walter | Dec 2006 | A1 |
20070103835 | Sorenson | May 2007 | A1 |
20070112939 | Wilson et al. | May 2007 | A1 |
20070126873 | Xu et al. | Jun 2007 | A1 |
20070130593 | Choi et al. | Jun 2007 | A1 |
20070192486 | Wilson | Aug 2007 | A1 |
20070239317 | Bogolea et al. | Oct 2007 | A1 |
20070256105 | Tabe | Nov 2007 | A1 |
20080001773 | Rye et al. | Jan 2008 | A1 |
20080007621 | Ying et al. | Jan 2008 | A1 |
20080011864 | Tessier et al. | Jan 2008 | A1 |
20080055423 | Ying et al. | Mar 2008 | A1 |
20080088438 | Aninye et al. | Apr 2008 | A1 |
20080092199 | McCarthy et al. | Apr 2008 | A1 |
20080129821 | Howarter et al. | Jun 2008 | A1 |
20080168503 | Sparrell | Jul 2008 | A1 |
20080195254 | Jung et al. | Aug 2008 | A1 |
20080242319 | Paschetto et al. | Oct 2008 | A1 |
20080309759 | Wilson et al. | Dec 2008 | A1 |
20090007170 | Everitt et al. | Jan 2009 | A1 |
20090138805 | Hildreth | May 2009 | A1 |
20090150925 | Henderson | Jun 2009 | A1 |
20100050215 | Wester | Feb 2010 | A1 |
20100070101 | Benes et al. | Mar 2010 | A1 |
20100076615 | Daniel et al. | Mar 2010 | A1 |
20100121968 | Clark | May 2010 | A1 |
20100138007 | Clark et al. | Jun 2010 | A1 |
20100168924 | Tessier et al. | Jul 2010 | A1 |
20100173677 | Fu | Jul 2010 | A1 |
20100192171 | Dozoretz et al. | Jul 2010 | A1 |
20100211972 | Howarter et al. | Aug 2010 | A1 |
20100241284 | Maeda et al. | Sep 2010 | A1 |
20100250015 | Flikkema | Sep 2010 | A1 |
20100262313 | Chambers et al. | Oct 2010 | A1 |
20100289644 | Slavin et al. | Nov 2010 | A1 |
20100319034 | Mountain | Dec 2010 | A1 |
20100332235 | David | Dec 2010 | A1 |
20110015797 | Gilstrap | Jan 2011 | A1 |
20110022242 | Bukhin | Jan 2011 | A1 |
20110030021 | Campagna et al. | Feb 2011 | A1 |
20110032423 | Jing et al. | Feb 2011 | A1 |
20110046798 | Imes et al. | Feb 2011 | A1 |
20110046799 | Imes et al. | Feb 2011 | A1 |
20110046805 | Bedros et al. | Feb 2011 | A1 |
20110061080 | Hill et al. | Mar 2011 | A1 |
20110088003 | Swink et al. | Apr 2011 | A1 |
20110093876 | Belz et al. | Apr 2011 | A1 |
20110153101 | Thomas et al. | Jun 2011 | A1 |
20110173668 | Rooks et al. | Jul 2011 | A1 |
20110181412 | Alexander et al. | Jul 2011 | A1 |
20110187930 | Crabtree | Aug 2011 | A1 |
20110190952 | Goldstein | Aug 2011 | A1 |
20110202189 | Venkatakrishnan et al. | Aug 2011 | A1 |
20110202195 | Finch et al. | Aug 2011 | A1 |
20110211584 | Mahmoud | Sep 2011 | A1 |
20110264276 | Kressner et al. | Oct 2011 | A1 |
20110264294 | Chen | Oct 2011 | A1 |
20110296459 | Medina et al. | Dec 2011 | A1 |
20110299412 | Diab et al. | Dec 2011 | A1 |
20110307101 | Imes et al. | Dec 2011 | A1 |
20110317622 | Arsenault | Dec 2011 | A1 |
20110321073 | Yarvis et al. | Dec 2011 | A1 |
20120017012 | Bartholomay et al. | Jan 2012 | A1 |
20120020518 | Taguchi | Jan 2012 | A1 |
20120053739 | Brian et al. | Mar 2012 | A1 |
20120053740 | Venkatakrishnan et al. | Mar 2012 | A1 |
20120053741 | Beyerle et al. | Mar 2012 | A1 |
20120059932 | Messer et al. | Mar 2012 | A1 |
20120065787 | Broniak et al. | Mar 2012 | A1 |
20120065791 | Besore et al. | Mar 2012 | A1 |
20120083937 | Kong et al. | Apr 2012 | A1 |
20120084808 | Howarter et al. | Apr 2012 | A1 |
20120092163 | Hart | Apr 2012 | A1 |
20120101637 | Imes et al. | Apr 2012 | A1 |
20120109395 | Finch et al. | May 2012 | A1 |
20120117503 | Hofrichter et al. | May 2012 | A1 |
20120131672 | O'Reirdan et al. | May 2012 | A1 |
20120166004 | Park et al. | Jun 2012 | A1 |
20120166007 | Jeong et al. | Jun 2012 | A1 |
20120185801 | Madonna et al. | Jul 2012 | A1 |
20120215725 | Imes et al. | Aug 2012 | A1 |
20120221718 | Imes et al. | Aug 2012 | A1 |
20120232701 | Carty et al. | Sep 2012 | A1 |
20120239202 | Voysey | Sep 2012 | A1 |
20120265586 | Mammone | Oct 2012 | A1 |
20120271475 | Wang et al. | Oct 2012 | A1 |
20120272261 | Reynolds et al. | Oct 2012 | A1 |
20120272267 | Klein et al. | Oct 2012 | A1 |
20120273581 | Kolk et al. | Nov 2012 | A1 |
20120291068 | Khushoo et al. | Nov 2012 | A1 |
20120323368 | White, III et al. | Dec 2012 | A1 |
20120323382 | Kamel et al. | Dec 2012 | A1 |
20130006425 | Bell et al. | Jan 2013 | A1 |
20130006437 | Verfuerth et al. | Jan 2013 | A1 |
20130010120 | Nnoruka | Jan 2013 | A1 |
20130013123 | Ozaki | Jan 2013 | A1 |
20130033561 | Kwon et al. | Feb 2013 | A1 |
20130036440 | Eyer | Feb 2013 | A1 |
20130060393 | Somasundaram et al. | Mar 2013 | A1 |
20130090746 | Quail | Apr 2013 | A1 |
20130110295 | Zheng et al. | May 2013 | A1 |
20130113822 | Putrevu et al. | May 2013 | A1 |
20130117442 | Mountain | May 2013 | A1 |
20130125177 | Pino et al. | May 2013 | A1 |
20130145386 | Belz et al. | Jun 2013 | A1 |
20130147623 | Somasundaram et al. | Jun 2013 | A1 |
20130158728 | Lee et al. | Jun 2013 | A1 |
20130165152 | Nichols | Jun 2013 | A1 |
20130167035 | Imes et al. | Jun 2013 | A1 |
20130173071 | Lee et al. | Jul 2013 | A1 |
20130182905 | Myers et al. | Jul 2013 | A1 |
20130184880 | McMahon | Jul 2013 | A1 |
20130195285 | De la Fuente et al. | Aug 2013 | A1 |
20130204454 | Choi et al. | Aug 2013 | A1 |
20130330059 | Matsumoto | Dec 2013 | A1 |
20140123208 | Plagemann et al. | May 2014 | A1 |
Number | Date | Country |
---|---|---|
1020040067511 | Jul 2004 | KR |
2010147688 | Dec 2010 | WO |
Entry |
---|
Supplementary European Search Report for International Application No. EP 14811212, dated Oct. 27, 2016. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2014/041352, dated Oct. 22, 2014. |
Nguyen, V. , “Samsung Smart TV Voice, Gesture and Face Recognition Hands-On”, retrieved from the internet: URL:http://www.slashgear.com/samsung-smart-tv-voice-gesture-and-face-recognition-hands-on-24229664/ [retrieved on Oct. 17, 2016], May 24, 2012. |
Number | Date | Country | |
---|---|---|---|
20180124450 A1 | May 2018 | US |
Number | Date | Country | |
---|---|---|---|
61834409 | Jun 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14291835 | May 2014 | US |
Child | 15858126 | US |