This relates generally to video, including broadcast and streaming television, movies and interactive games.
Television may be distributed by broadcasting television programs using radio frequency transmissions of analog or digital signals. In addition, television programs may be distributed over cable and satellite systems. Finally, television may be distributed over the Internet using streaming. As used herein, the term “television transmission” includes all of these modalities of television distribution. As used herein, “television” means the distribution of program content, either with or without commercials and includes both conventional television programs, as well as the distribution of video games.
Systems are known for determining what programs users are watching. For example, the IntoNow service records, on a cell phone, audio signals from television programs being watched, analyzes those signals, and uses that information to determine what programs viewers are watching. One problem with audio analysis is that it is subject to degradation from ambient noise. Of course, ambient noise in the viewing environment is common and, thus, audio based systems are subject to considerable limitations.
In accordance with some embodiments, a multimedia clip, such as a limited duration electronic representation of a video frame or clip, metadata or audio, may be grabbed from the actively tuned television transmission currently being watched by one or more viewers. A hand gesture may be recognized to select a currently played multimedia clip for searching. This multimedia clip may then be transmitted to a mobile device in one embodiment. The mobile device may then transmit the information to a server for searching. For example, image searching may ultimately be used to determine who the actors are in a video. Once the content is identified, then it is possible to provide the viewer with a variety of other services. These services can include the provision of additional content, including additional focused advertising content, social networking services, and program viewing recommendations.
Referring to
In one embodiment, the processor-based system 14 includes a multimedia grabber 16 that grabs an electronic representation of a video frame or clip (i.e. a series of frames), metadata or sound from the decoded television transmission currently tuned to by a receiver (that may be part of the system 14 in one embodiment). The processor-based system 14 may also include a wired or wireless interface 18 which allows the multimedia that has been grabbed to be transmitted to an external control device 24. This transmission may be over a wired connection, such as a Universal Serial Bus (USB) connection, widely available in television receivers and set top boxes, or over any available wireless transmission medium, including those using radio frequency signals and those using light signals. The metadata may be metadata about the content itself (e.g., rating information, plot, director name, year of release).
In one embodiment, non-decoded or raw electronic representation of video clips may be transferred to the control device 24. The video clips may be decoded locally at the control device 24 or remotely, for example, at a server 30.
Also coupled to the system 14 and/or the display 20 may be a video camera 17 to capture images of the viewer for detecting user gestural commands, such as hand gestures. A gestural command is any movement recognized, via image analysis, as a computer input.
The control device 24 may be a mobile device, including a cellular telephone, a laptop computer, a tablet computer, a mobile Internet device, or a remote control for a television receiver, to mention a few examples. The device 24 may also be non-mobile, such as a desk top computer or entertainment system. The device 24 and the system 14 may be part of a wireless home network in one embodiment. Generally, the device 24 has its own separate display so that it can display information independently of the television display screen. In embodiments where the device 24 does not include its own display, a display may be overlaid on the television display, for example, by a picture-in-picture display.
The control device 24, in one embodiment, may communicate with a cloud 28. In the case where the device 24 is a cellular telephone, for example, it may communicate with the cloud by cellular telephone signals 26, ultimately conveyed over the Internet. In other cases, the device 24 may communicate through hard wired connections, such as network connections, to the Internet. As still another example, the device 24 may communicate over a television transport medium. For example, in the case of a cable system, a device 24 may provide signals through the cable system to the cable head end or server 11. Of course, in some embodiments, this may consume some of the available transmission bandwidth. In some embodiments, the device 24 may not be a mobile device and may even be part of the processor-based system 14.
Referring to
The system interconnect may be coupled to transport processor 36, security processor 38, and a dual audio digital signal processor (DSP) 40. The digital signal processor 40 may be responsible for decoding the incoming video transmission. A general input/output (I/O) module 42 may, for example, be coupled to a wireless adaptor, such as a WiFi adaptor 18a. This will allow it to send signals to a wireless control device 24 (
In some embodiments, the processor-based system 14 may be programmed to output multimedia clips upon the satisfaction of a particular criteria. One such criteria is the detection of a user hand gesture. User hand gestures may be recorded by the camera 17 (
Referring to
Initially, a check at diamond 72 determines whether the grabber feature has been activated. The grabber device 16 (
While an embodiment is depicted in
When the control device 24 receives a multimedia clip from the system 14, as detected at diamond 56, in some embodiments, the control device 24 may send the annotated multimedia clip to the cloud 28 for analysis (block 58). Then the device 24 may display a user interface to aid the user in annotating the captured clip (block 57) now displayed on the device 24.
In some embodiments, the user may append annotations to focus the analysis of the clip, as indicated in block 57. An annotation may also include questions about the clip for distribution as an annotation with the clip over social networking tools. For example, a text block may be automatically displayed over the transferred video clip on the control device 24. The user can then insert text that may be used as keywords for Internet or database searches. Also, the user may select particular depicted objects for providing search focus. For example, if two people appear in the clip, one of them may be indicated. Then, in the text box, the user may enter “Who is this actress?”. The search is then focused on identifying the indicated person.
The person in the clip can be selected using a mouse cursor or a touch screen. Also, video analysis of the user's finger pointing at the screen may be used to identify the user's focus. Similarly, eye gaze detection can be used in the same way.
Of course, the multimedia clip can be sent over a network to any server for image searching and/or analysis in other embodiments. The multimedia clip can also be sent to the head end 11 for image, text, or audio analysis, as another example.
If an electronic representation of audio is captured, the captured audio may be converted to text, for example, in the control device 24, the system 14 or the cloud 28. Then the text can be searched to identify the television program.
Similarly, metadata may be analyzed to identify information to use in a text search to identify the program. In some embodiments, more than one of audio, metadata, video frames or clips, may be used as input for keyword Internet or database searches.
A transferred video clip may also be distributed to friends using social networking tools. Those friends may also provide input about the video clip, for example, answering questions, accompanying the clip as annotations, like, “Who is this actress?”.
An analysis engine then may perform a multimedia search to identify the television transmission being viewed or to obtain other information about the clip, including scene or actor/actress identification or program identification, as examples. This search may be a simple Internet or database search or it may be a more focused search.
For example, the transmission in block 58 may include the current time or video capture and location of the control device 24. This information may be used to focus the search using information about what programs are being broadcast or transmitted at particular times and in particular locations. For example, a database may be provided on a website that correlates television programs available in different locations at different times and this database may be image searched to find an image that matches a captured frame to identify the program.
The identification of the program may be done by using a visual or image search tool. The image frame or clip is matched to existing frames or clips within the image search database. In some cases, a series of matches may be identified in a search and, in such case, those matches may be sent back to the control device 24. When a check at diamond 60 determines that the search results have been received by the control device 24, the search results may be displayed for the user, as indicated at block 62. The control device 24 then receives the user selection of one of the search results that conforms to the information the user wanted, such as the correct program being viewed. Then, once the user selection has been received, as indicated in diamond 64, the selected search result may then forwarded to the cloud, as indicated in block 66. This allows the television program identification or other query to be used to provide other services for the viewer or for third parties.
Referring to
While an embodiment using a cloud is illustrated, of course, the same sequence could be implemented by any server, coupled over any suitable network, by the control device 24 itself, by the processor-based device 14, or by the head end 11 in other embodiments.
Initially, a check at diamond 82 of
For example, the task may be to provide information to a pre-selected group of friends for social networking purposes. For example, the user's friends on Facebook may automatically be sent a message indicating which program the user is watching at the current time. Those friends can then interact over Facebook with the viewer to chat about the television program using the control device 24, for example.
As other examples, the task may be to analyze demographic information about viewers and to provide head ends or advertisers information about the programs being watched by different users at different times. Still other alternatives include providing focused content to viewers watching particular programs. For example, the viewers may be provided information about similar programs coming up next. The viewers may be offered advertising information focused on what the viewer is currently watching. For example, if the ongoing television program highlights a particular automobile, the automobile manufacturer may provide additional advertising to provide viewers with more information about that vehicle that is currently being shown in the program. This information could be displayed as an overlay, in some cases, on the television screen, but may be advantageously displayed on a separate display associated with the control device 24, for example. In the case where the broadcast is an interactive game, information about the game progress can be transmitted to the user's social networking group. Similarly, advertising may be used and demographics may be collected in the same way.
In some embodiments, a plurality of users may be watching the same television program. In some households, a number of televisions may be available. Thus, many different users may wish to use the services described herein at the same time. To this end, the processor-based system 14 may maintain a table which identifies identifiers for the control devices 24, a television identifier and program information. This may allow users to move from room to room and still continue to receive the services described herein, with the processor-based system 14 simply adapting to different televisions, all of which receive their signal downstream of the processor-based 14, in such an embodiment.
In some embodiments, the table may be stored in the processor-based system 14 or may be uploaded to the head end 11 or, perhaps, even may be uploaded through the control device 24 to the cloud 28.
Thus, referring to
Initially, the system receives and stores an identifier for each of the control devices that provides commands to the system 14, as indicated in block 94. Then, the various televisions that are coupled through the system 14 may be identified and logged, as indicated in block 96. Finally, a table is setup that correlates control devices, channels, and television receivers (block 100). This allows multiple televisions to be used that are connected to the same control device in a seamless way so that viewers can move from room to room and continue to receive the services described herein. In addition, a number of viewers can view the same television and each can independently receive the services described herein.
References throughout this specification to “one embodiment” or “an embodiment” mean that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one implementation encompassed within the present invention. Thus, appearances of the phrase “one embodiment” or “in an embodiment” are not necessarily referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be instituted in other suitable forms other than the particular embodiment illustrated and all such forms may be encompassed within the claims of the present application.
While the present invention has been described with respect to a limited number of embodiments, those skilled in the art will appreciate numerous modifications and variations therefrom. It is intended that the appended claims cover all such modifications and variations as fall within the true spirit and scope of this present invention.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CN2011/001548 | 9/12/2011 | WO | 00 | 6/17/2013 |