1. Field of the Invention
This invention relates generally to personal streaming media broadcasters, and in particular to streaming media from a media source input to a client device over a network.
2. Background of the Invention
While people spend a great deal of time watching television programming and other forms of audio-visual (A/V) content, they are also spending an increasing amount of time interfacing with computing devices such as personal computers, personal digital assistants, mobile phones, dedicated multimedia devices, and other devices that, like the traditional television, include a display. These types of computing devices allow people to be increasingly mobile, but this mobility reduces the time people spend at home in front of their televisions. It would therefore be beneficial to enable people to enjoy their television programming and other forms of A/V content they now receive at home on these computing devices as well, regardless of location and without dependence on physical connections.
This ability would enable several desirable applications. For example, a user might want to access and control television and other regularly consumed A/V content from a personal computer (desktop as well as notebook computers) or other computing devices around the home via the user's local network in the home. Since cable, satellite, and other sources of television content typically enter the house at a few discrete locations, allowing access to the content over a home network gives the user more freedom to enjoy the content throughout the home. Another possible application would be to enable a user to access and control television and other A/V content from any number of remote networks where a broadband connection is available to the user (e.g., at an airport, at work, at school, in a hotel, in a cafe, at an acquaintance's house). Yet another application would be to enable a user to access and control television and other A/V content from a mobile phone or other computing devices that can be connected to a wide area network (e.g., GPRS, W-CDMA, CDMA-2000, 1XRTT, 1XEVDO, and the like). In various applications, users are likely to want to access their media content stored on devices, such as personal computers and other devices having storage, from remote networks. Nevertheless, network bandwidth and other limitations have made it difficult to provide an effective and enjoyable remote media experience for the user.
But traditional streaming media solutions do not enable these applications in any effective way; moreover, they suffer from technical limitations that would prevent their use in personal media broadcasting applications like those described above. Accordingly, it would be desirable to enable users to access their A/V content from any of a variety of remote locations inside or outside the home, as such content is currently available only from locations in the home that traditionally receive and play it (e.g., a television set).
A personal media broadcasting system enables video transmission over a computer network, allowing a user to view and control media sources over a computer network from a remote location. In one embodiment, the personal media broadcasting system includes a personal media broadcaster and a media player. The personal media broadcaster may be configured to receive as an input virtually any media source. The broadcaster digitizes and compresses the received media content (if necessary) and streams the media over a computer network to the media player. The media player may reside on any of a wide range of client devices for viewing the media. A user may send control commands (e.g., “channel up”) using the media player back over the network to be executed by the media source device, thus affecting the media stream received by the media player.
In one embodiment, a personal media broadcasting system includes a personal media broadcaster and media player client, which communicate with each other over a network. The personal media broadcaster can accept an A/V signal from one or more A/V source devices. From this A/V signal, the personal media broadcaster constructs a media stream that is suitable for transmission over a network to the media player client. The personal media broadcaster includes a network interface for transmitting the media stream to the media player client. As the media stream is received at the client, it can be viewed by a user using the media player client. In one embodiment, the personal media broadcaster is a dedicated appliance, not a general purpose computer. In this way, a general purpose computer need not be powered on and connected to the A/V source devices for the user to receive media content remotely. In another embodiment, the media player client can run on any general or multi-purpose device (such as a personal computer or cellular phone), beneficially avoiding the need for the user to carry special equipment to use the broadcasting system. The computer network over which the broadcaster and media player client communicate may comprise a wide area network, such as the Internet, allowing the user to receive media content from the home to anywhere in the world where a connection to the network is available.
To improve the user's experience, the person media broadcasting system may employ any of a number of techniques for buffering, transmitting, and viewing the content. In one embodiment, for example, the media stream is constructed by encoding the audio/visual signal using a bit rate selected at least in part based on an amount of free space remaining in an intermediate output buffer used to temporarily store the media stream before it is transmitted to the media player client. While various measures of occupancy of the intermediate output buffer can be used, one involves the use of multiple watermarks to measure the free space available in the buffer. The encoding bit rate can be decreased each time the free space dips below a watermark, while it can be increased when the free space rises above a watermark or if the free space never drops below the watermark in the first place. In this way, the encoding bit rate can be dynamically adjusted based on current performance of the system, and this dynamic adjustment can occur at the broadcaster without requiring feedback from the client.
In another embodiment, the media player client implements a fast-start mechanism by which the media player receives a media stream in real time from the personal media broadcaster and stores the media stream in a buffer. The media player client plays the received media stream from the buffer at a decreased playback speed without waiting for the buffer to fill to a predetermined threshold. While the playback speed is kept at a slower rate than the streaming media content is received, the buffer fills (albeit at a rate slower than the media stream is received). This allows the media player client to play a received media stream without waiting for its buffer to fills, which improves the user's experience dramatically in situations such as where the user is changing channels or operating a menu on the remote A/V source device. In addition, the media player client may further enhance this experience by switching into a control mode when the client receives a user command to control content in the media stream. When the client enters the control mode, the buffer is flushed and the received media stream is played. In this way, the content in the buffer need not be displayed on the client device, and the user can then almost immediately receive visual feedback confirming the user's command to control content in the media stream.
Embodiments of the invention allow for distribution of A/V content from a variety of A/V source devices over a computer network to a client device. As used herein, A/V content may include any type of media content, including audio or video content, or both. In one embodiment, a personal media broadcaster takes as an input an audio or video signal, digitizes and compresses the content (e.g., in Windows Media Video, MPEG-4, or H.264), and streams the content over an IP network (such as IP4 or IP6) to a client device for viewing and/or listening to the content. The personal broadcaster, which may be implemented as an embedded solution, may allow the user to issue control commands, such as “channel up” via an IR or serial command, back over the network to be executed by the original source device. Using various embodiments of the present invention, a user can connect to the personal broadcaster a cable or satellite set-top box, a personal video recorder, a DVD player, a DVD jukebox, a music server, a satellite radio receiver, a camcorder, a digital video recorder (DVR) or any other A/V source component. This way, the user can view and control the live output of these sources from any networked device. Various embodiments of the present invention may include different components, including a personal broadcaster and media player. The media player may be a software application that runs on a client device, which is configured to receive the media stream created by the personal broadcaster. Different client software applications can exist for different classes of client devices, such as personal computers and cell phones. The client may be able to connect to the personal broadcaster through a Web browser interface.
The system described herein may be applied in a number of applications or usage scenarios. For example, there are a variety of uses in the home over a wired or wireless home network for the system. In the home, users can view and control their A/V source components, such as a personal video recorder (PVR) or a cable set-top box, from any desktop PC, notebook PC, PDA, or other network devices. For example, this application allows a user to convert a wireless notebook PC situated in the kitchen into a wireless LCD TV. As another example of in-home use, young parents can pair a personal broadcaster with a simple analog camera and place them in the baby's room to provide a real time video and audio stream of their child displayed on a device in another location in the home.
Another usage scenario category relates to users who are away from home but have access to a broadband connection. This access can be in a variety of locations, such as at work, at school, at a friend's house, in a cafe, at the airport or in a plane, or in a hotel. From these locations, users can connect to their personal broadcaster over the public Internet and enjoy the same live television experience that they have at home. The present invention allows users that are away from home to access the full breadth of programming they are accustomed to, and even view content that they previously captured on their personal video recorder.
Yet another usage scenario category relates to individuals with mobile phones, communicators, or other wide area network devices. On emerging networks, the bit rate is now high enough for users to enjoy streaming video on their devices. This opens up new possibilities for the personal broadcaster to deliver a wealth of programming to the mobile user, all of which originates from the user's own home.
The media stream transmitted from the personal media broadcaster to the media player client networks can take advantage of a combination of any number of networking protocols, including HTTP over TCP/IP, as well as UDP, RTSP, RTP, RSVP, and the like. Because embodiments of the invention can accept, digitize, encode, and stream any analog A/V source, there are a vast number of uses for the personal broadcaster—from a security camera system to a method for extending a user's satellite radio to the user's cell phone.
System Architecture
The gateway no may be coupled to a local area network (LAN) 140 that couples several computing devices in a user's home. According to known techniques, any number of local clients 150 may be able to communicate with the home Internet gateway 110. In this way, created by the media broadcaster 100 may be routed to any of these local clients 150 by way of the local network 140, either through the gateway no or directly. The local area network 140 can be wired or wireless, as the present invention is not limited to any particular network technology or configuration. The local clients 140 can be any number of device types, including but not limited to desktop and notebook PCs, Web tablets, PDAs, embedded clients built expressly for the purposes of decoding the streams of the personal broadcaster, and other devices capable of receiving and/or playing a media stream over a network.
The media streams created by the personal broadcaster 100 may also be received by remote clients 170 from a remote network 160. The remote network 160 may comprise any suitable networking technology, including but not limited to wide area mobile networks (e.g., GPRS, EDGE, 1X-RTT, 1x-EvDO, and FOMA 2.5G and 3G cellular networks), WiFi and other public broadband access locations, WiMAX networks, other LANs (such as at work, school, or a friend's home), and direct connections to other Internet service providers. As with the local clients 150, the remote clients 170 may include any number of device types, but not limited to desktop and notebook PCs, Web tablets, PDAs, embedded clients built expressly for the purposes of decoding the streams of the personal broadcaster, and other devices capable of receiving and/or playing a media stream over a network.
In one embodiment, the local clients 150 and/or the remote clients 170 execute a client software application that includes a user interface for requesting content from the broadcaster 100 and for viewing that content. In another embodiment, the client functionality is provided by a Web site and is accessible by the local clients 150 and/or the remote clients 170 via a Web browser. When a remote client 170 wishes to connect to the stream of the personal broadcaster 100 using the client application or via a Web browser interface, it may specify the home IP address of the user to access and pull the media stream from the personal broadcaster. This action sends a request to the personal broadcaster, and the request travels across the public Internet, to the network of the user's Internet service provider (ISP), into the home via the telephony or cable infrastructure (or wirelessly in the case of the fixed wireless or satellite broadband ISP), to the home Internet gateway 110, and finally to the personal broadcaster 100.
In one embodiment, a central server 180 is coupled to the remote network 160 and provides one or more roles, including that of DNS server. Because most residential ISPs allocate dynamic IP addresses via DHCP, as opposed to providing static IP addresses, there is a need for a system that provides a consistent method for accessing the user's home network from remote networks. In the embodiment illustrated in
To make this process even easier for the user, so that the user need not manage a constantly changing IP address or enter a DNS name, much of the process for connecting to a personal broadcaster 100 can be abstracted from the user. For example, in one embodiment, the user need only to enter the name of the personal broadcaster, or select an icon representing the personal broadcaster, and then enter the corresponding password before being automatically directed to their personal broadcaster 100. This can be accomplished by tying a unique device name to the DNS name assigned to the user's dynamic IP address. The translation between the device user name and the DNS name can take place within the remote client 170 itself, or it can be accomplished through a directory maintained on the central server 180.
Connections from a local client 140 or a remote client 170 can be accomplished either by using a client application designed specifically for the purpose of accessing the personal broadcaster stream or via a traditional Web browser. The option of using a Web browser provides for wide range of client compatibility without the need for configuration, while the client application provides for an optimized experience. The client application or the Web interface may prompt the user for a password before allowing communication with the broadcaster 100, as a security measure. As an additional measure of security, the media stream can be encrypted.
In one embodiment of the invention, there is a limit of one connected client (applies equally to remote clients and local clients) per device. That is, only one client at a given point in time can be connected to and streaming from the personal broadcaster. Other variations of this embodiment can provide for multiple simultaneous sessions. Still other variations can allow for multiple simultaneous sessions from local clients, but maintain a single session limit for remote clients.
In
The personal broadcaster 100 connects to the home Internet gateway 110 from its Ethernet port 290 (using, e.g., a Cat5 cable), which connection can be direct or via an Ethernet wall jack located near the personal broadcaster 100 (which in turn connects to the home Internet gateway 110). In other embodiments, the connection between the personal broadcaster 100 and the home Internet gateway 110 is wireless, where the broadcaster 100 may include built-in wireless or power line networking capability.
In one embodiment, the A/V decoder 315 converts the analog video input into YUV video and applies various types of filters and color corrections to the signal. The A/V decoder 315 also extracts vertical blanking interval (VBI) data, such as close caption, tele-text, and copy protection bits. The A/V decoder 315 also decodes the tuner audio and converts it into stereo or mono digital audio, depending on the broadcast signal. The analog signal is further converted into a digital signal in the A/V decoder 315. The digital video and audio from the A/V decoder 315 is then sent to a processor 320 for further processing. The personal broadcaster 100 may include memory 330, such as flash memory or SDRAM, used by the processor 320 for performing its associated processing tasks. The memory 330 may also be used as a buffer for the outgoing media stream, as described herein for various embodiments.
In one embodiment, the processor 320 performs pre-processing on the digital audio and video before compression. The pre-processing can be performed based on the input type, compression properties, and target bit rate. After pre-processing, the processor compresses the audio and video to a desired bit rate using any suitable compression technique (such as WM9, MPEG-4, H.263, and H.264). The compressed audio and video are multiplexed into a single media stream together along with other user data such as close caption, tele-text, parental control, and Macrovision. In one embodiment, the processor 320 is capable of statically and/or dynamically adjusting the compression bit rate, frame rate, resolution, and filtering depending on a user request, the input content, available network bandwidth, or on any other data known to the processor 320. The compressed media stream is then converted into network packets for transmission over the local network 140 or a remote network 160 via the network interface 325. The network interface 325 may be a wireless or a wired interface, or may have provisions for both types. As mentioned above, the personal broadcaster 100 may also receive and process commands received from a client over the network interface 325. Some examples of these commands include selecting a particular channel, automatic scanning of channels, switching between RF input and base-band input, changing compression properties (compression type, bit rate, frame rate, resolution, and other properties), remotely controlling commands for the IR blaster, and any other command that a user may desire for viewing content from the A/V source device 120.
The broadcaster 100 may further comprise a controller interface 335 for interfacing with an output for controlling an A/V source device 120. As mentioned above, control of an A/V source device 120 may be performed using an RS-232 serial controller or an IR emitter 285. The controller interface 335 thus receives the appropriate output signals from the processor 320 and provides the corresponding interface for controlling an operation of an A/V source device 120.
While
Despite the specific network topology illustrated in
As shown in
When wishing to stream media stored on an on-network storage device 130 to a remote location, a challenge arises. The bit rate of the media clips present on the on-network storage device can be higher than the bit rate supported by the upstream link of the user's broadband service. For example, a video clip on an on-network storage device 130 can have a bit rate of 800 kbps, whereas very few broadband connections currently have an uplink speed equal to or greater than that. In such cases, the bit rate of the source media signal is reduced and its encoding format is possibility changed. The personal broadcaster 100 may perform this transrating and transcoding functionality. In this situation, the personal broadcaster 100 acts as a networked-attached transrater and transcoder. The broadcaster receives the media stream from the on-network storage device 130, transrates and possibility transcodes the media, and outputs a media stream with a sufficiently low bit rate so the media can be effectively streamed upstream from the user's broadband service. The method for determining the proper transrating or transcoding settings (e.g., the bit rate to which the source content is to be reduced) can be accomplished within the framework for determining the throughput currently supported between the personal broadcaster and the local or remote client, outlined below.
The use of the personal broadcaster 100 as an agent to transrate and/or transcode the material residing on one or more on-network storage devices 130 has the benefit of creating a system where only one device (the personal broadcaster) streams media upstream through the home Internet gateway 110 and out to the Internet 160. This is beneficial because the user does not have to make further configurations to the NAT or firewall of the home Internet gateway, which can include manually forwarding a port to allow direct access to each on-network storage device.
In addition to requiring further configuration, streaming content directly from the on-network storage device 130 can create security concerns for the user, especially if the on-network storage device is a PC. Because allowing a PC to stream directly upstream to the public Internet involves opening a port on the firewall/NAT that forwards to the PC, a user can be concerned that other personal or private information is at risk for being exposed. By relying on the personal broadcaster to be a gateway for streaming media to the public Internet, the PC or on-network storage device on which the media is stored need not have a port forwarded to it. In such an embodiment, the personal broadcaster is the only device for which NAT port configuration is required.
In addition, using the personal broadcaster instead of the on-network storage device to transcode and transrate content prevents the CPU on the on-network storage device 130 from being unnecessarily taxed. This is especially important if the on-network storage device 130 is a PC, because the increase in CPU utilization would detrimentally impact the performance of the PC for accomplishing other tasks.
Because the personal broadcaster 100 provides the user with access to the same A/V source devices 120 available at home, it makes sense to provide the user with an interface to the A/V source device 120 similar to the one used in the living room setting. Most often, this is a handheld remote control. In one embodiment, therefore, a “virtual” remote control is provided by the client application that includes either a generic image representing the A/V source device's remote control or an image or likeness of the actual A/V source device's remote control. Moreover, the client application may support a number of virtual remote controls, one customized for each A/V source device 120. Interfaces on the client application are thus selected by the user to resemble each particular A/V source device 120 found in the user's home. For example, if a TiVo personal video recorder is connected to the personal broadcaster, the user can decide to use the TiVo skin which modifies the virtual remote control on the client application to resemble the TiVo remote control. Remote control commands are mapped to the graphical image in such a way that a press of the button triggers the action suggested by the image of the button (e.g., pressing on CH+ button turns the channel from 3 to 4).
In one embodiment, the client application contains a database of remote control skins from which to choose. In this case, a central database maintained on the central server, is likely to update the client application upon configuration to ensure that the latest remote control skins are available to the users. In addition, third parties may be allowed to create and share images with commands mapped to particular regions of the image. In this case, a method for “plugging-in” the third party remote control skins is provided to the user. To allow third parties to create skins for the media player, an API is provided to allow access to some of the features and functionality within the media player client. Third party skins allow users to develop content as well as third party device manufacturers and service providers to make virtual remote controls that closely resemble the physical remote control associated with an A/V source device 120. The virtual remote on the client application can thus be made easier to use, since the user is used to its layout.
Operation of Personal Media Broadcasting System
As described above with reference to
Control of Audio/Visual Source Devices
As stated above, embodiments of the personal media broadcaster allow a user to control an A/V source device from clients connected to the remote or local networks. The client may allow for control of the user's specific model of A/V source device. Upon initial configuration of the personal broadcaster and a client, the user indicates which make and model of A/V source device the user would like to control (e.g., TiVo Personal Video Recorder Series 2). The configuration software on the client then identifies the group of IR or serial codes that correspond to the specific A/V source in question. This can be performed by searching a database that ships with the included software, an updated database residing on the central server, or a database that resides on the personal broadcaster.
In one embodiment of the invention, the IR codes is then stored on the client device. When a user wishes to invoke a code, the user selects the given command on the control panel or virtual remote control in the client application. The client application then sends the corresponding IR or serial command over the IP network. Once the IR or serial command reaches the personal broadcaster, the personal broadcaster processes the code and sends it out the IR or serial output, triggering the requested action in the A/V source device. The connection between personal broadcaster and the A/V source device can be an IR emitter, in the case that the A/V source device was to be controlled via IR, or an RS-232 port, in the case that the A/V source device was to be controlled by serial commands.
In another embodiment, the IR and serial codes are stored on the personal broadcaster rather than the clients. In this case, a client device that requests a given command sends a notation representing that command (e.g., “CH UP” if the user wants to change the channel from 4 to 5) rather than the IR or serial control code itself. When the notation representing the command reaches the personal broadcaster, the personal broadcaster performs a simple lookup, and outputs the appropriate IR or serial code. Certain commands or selections made on the client device can activate a series of commands, also known as a macro. For example, by clicking on the “CNN” button (which can possibly be represented by a logo of the network), the combination of commands that tune the A/V source device to CNN would be triggered. For example, if CNN was channel 202 on a user's DirecTV system, a press of the “CNN” button can trigger the following commands in succession: “2,” “0,” “2,” and “Enter.”
To make the set up of multiple client devices easier, profiles of the A/V source devices used can be stored in the personal broadcaster during the configuration of the initial client device. This enables easy configuration for subsequent client devices, as the personal broadcaster informs the subsequent client device which A/V source devices it is able to connect to, and which IR or serial codes it uses.
Adjustment of Encoder Settings Based on Throughput and Device Capability
Because the broadcaster enables access of a media stream by a variety of client device types connected to the local area network as well a various remote networks, the available data throughput present between the personal broadcaster on one end and the local clients and remote clients on another can vary considerably based on network topology. There is also likely to be considerable throughput variation in a given connection, due to competing traffic and general network congestion. In one embodiment of the invention, a method for optimizing the audio (e.g., bit rate and sampling rate) and video (e.g., bit rate, resolution, and frame rate) compression based on available network bandwidth and capabilities of client device is implemented.
Because various embodiments of the present invention encompass both the personal broadcaster and the client devices, and these elements can operate in a 1:1 relationship (i.e., each broadcaster may accept only one client connection at a time), the two components are able to act in concert to optimize the experience for the user. In one embodiment, the optimization process includes an initial optimization stage and an ongoing optimization stage.
In the initial optimization process, the client and personal broadcaster communicate to mutually establish the capabilities of the client device, as well as the throughput of the connection between the personal broadcaster and the client device. The client device first requests the personal broadcaster to send a set number of bits to the client (this can happen automatically at first connection, or can be manually ordered by the user to recalibrate initial optimization). Based on the time it takes for the client to receive those bits, the client has an idea of the actual data throughput between the personal broadcaster and the client. With this information in hand, the client instructs the personal broadcaster to begin streaming at a rate compatible with this throughput. This is not likely to be the full rate of throughput, but some rate less (for example, 80% of throughput), to allow for inevitable variation in network bandwidth. In choosing the proper resolution setting, the application residing on the client notes its current capabilities (i.e., resolution of its display) and pairs the appropriate bit rate setting with the appropriate resolution setting in its command to the personal broadcaster.
The client can learn of its capabilities in several ways. One can be related to the version of the application itself. For example, the application for a Pocket PC can know that the device best supports streams at or below a certain resolution and frame rate. Another way is to take inventory of system resources before it sends the request. For example, the client can identify its display resolution and incorporate this information in the streaming request to the personal broadcaster.
The initial optimization process represents a starting point that can very well provide for the proper encoder settings. However, the variability of network bandwidth over time calls for a system that is dynamic in nature and capable of real time changes to the encoder settings. To address this variability, one embodiment of the present invention implements a feedback loop between the client and personal broadcaster to maintain the proper encoder settings over time. This feedback loop can be implemented in a number of ways. In one embodiment the client gives notification to the personal broadcaster when it experiences frame drops. Alternatively, the client communicates, from time to time, the size of the buffer, or the total measured throughput. The client may communicate any or all of these statistics, or any other data that reflects on the need to adjust the settings.
Given this feedback from the client, the personal broadcaster alters the encoder settings. For example, the personal broadcaster can reduce the encoding bit rate from 350 kbps to 280 kbps if the size of the buffer becomes reduced or an unacceptable number of frame drops are observed. The time between measurements and feedback received from the client can vary, and the figure depends on a balance between reacting quickly to significant changes and overcorrecting based on temporary blips. Based on the feedback, the personal broadcaster can adjust the settings upward (e.g., increase encoding bit rate) as well as downward.
While one embodiment provides for a method of automatic adjustment of encoder settings, the user may also have the ability to set the encoder settings manually. These manually adjustable settings include frame rate, bit rate, resolution, “quality,” and time between key frames for video, bit rate and sampling rate for audio, as well as client side settings that can impact performance, such as buffer size and smoothing.
Adjustment of Encoder Settings Based on Programming Type
While the data throughput between the personal broadcaster and the client may be one important determinant of the proper encoding settings, the type of content being viewed may also be an important criterion. For example, fast motion video from a sports program requires a higher frame rate than a talk show, which features much less motion. Likewise, video with little movement can require comparatively lower bit rates or resolution than fast motion video to achieve an acceptable quality. This is because video with slower movement tends to be encoded much more efficiently. Accordingly, the encoder settings may be selected based on the type of programming being encoded.
There are various methods for determining the type of content being viewed, and hence the proper range of encoder settings. First, there can be settings that apply generally to all content on a given channel. By identifying the programming channel or network to which the personal broadcaster is tuned (e.g., HBO or NBC), the client can request suitable encoder settings. For example, there can be a rule that when the personal broadcaster is streaming content from ESPN, the frame rate is always set at 30 frames per second. Such rules can be stored on the client or on the central server, which informs the client of the proper settings for a requested channel. Moreover, the rules can be learned over time by an individual's own client, which observes the settings chosen by the user for certain network programming.
In a further refinement, the encoder settings may be customized based on the specific programming being watched. Because a major network typically has content that is both demanding, high-motion programming (such as a major sporting event) and easy, low-motion programming (such as a newscast), selecting encoder settings based solely on the channel may not be efficient for many channels. Accordingly, by cross referencing the current channel with the current time of day, the system can determine the program that is being viewed, and the encoder's settings can be selected based on the actual program being viewed.
In another embodiment, the system constantly monitors the A/V content being encoded. Metrics based on pre-selected criteria (e.g., amount of motion in video) are generated, and this information is used to assign or adjust the encoder settings to the personal broadcaster dynamically. Alternatively, the media player may receive from the user an indication of what kind of content is being watched (e.g., action, music, news, etc.), which is mapped to predefined profiles in the framework optimized for that kind of content.
Buffering and Control of Buffer Resources
In accordance with one embodiment of the invention, the personal media broadcaster implements a buffering scheme to manage its buffer resources without requiring feedback from a client device. As mentioned above, the broadcaster and the client may communicate using TCP as a transport protocol, where the broadcaster acts as a server. Beneficially, TCP is a reliable protocol and ensures that sent data always reaches its destination in the correct order. Parameters and/or behavior of the TCP stack on a server can be monitored to estimate network congestion and speed according to one or more known techniques.
In accordance with an embodiment of the invention, a large buffer is added between the encoder (which generates the data) and the TCP stack on the network interface (which transmits the data). This additional buffer layer added above the TCP stack helps to avoid loss of data due to network congestion and the variability of data rates. In one embodiment, with reference to
As the data for the media stream are being generated by the broadcaster, the intermediate buffer acts as a FIFO queue. When the available network bandwidth is more than the encoder's bandwidth, the broadcaster is able to send data as soon as it is generated. The intermediate buffer will begin to empty. When the available network bandwidth is less than the encoder's bandwidth, the broadcaster will generate the data faster than it can be transmitted. This will begin to fill the buffer. The buffer may therefore vary between being completely full and completely empty. To classify the occupancy of the buffer, a number of watermarks are defined to indicate the amount of free space left in the buffer. While the number of watermarks can vary, in one embodiment four watermarks are used—at the 90%, 75%, 50%, and 30% levels. As data are added to and taken from the buffer, the amount of data filling the buffer can change over time. When this level reaches one of the watermarks, various actions are taken depending on which watermark has been reached.
The amount of free space left in the intermediate buffer is observed for a period of time (e.g., one minute). If the amount of free space in the buffer remains above the 90% watermark during the last observation period, the encoder's output bit rate may be increased. Although any increase can be implemented based on the application, in one embodiment the increment is about 10% of the bit rate then being used.
Because network bandwidth varies over time, sudden drops in available bandwidth may take place frequently. In such cases, the TCP stack transmit rate will go down and the occupancy of the intermediate buffer will increase. If this occurs for a long enough period, the free space in the buffer will decrease so that the 90% watermark may be breached. In response, the broadcaster reduces the encoder's bit rate by a small percentage, for example about 15% of the bit rate then being used. If this network problem is a temporary one, the TCP stack will again be able to send the backlog of data in the buffer so that the amount of free space in the buffer will rises again above the 90% mark. The encoder's bit rate can then be increased.
On the other hand, if the network problem persists, the amount of free space will continue to decline. Over a period of time, the buffer will fill and the other watermarks will be breached. As each watermark is breached, the encoder's bit rate is further reduced. In one embodiment, these subsequent reductions may be larger (e.g., 33%, 50%, 50% for each watermark, respectively).
As described, the system intelligently exploits TCP stack behavior to estimate network status and reacts to provide optimal user experience in presence of bandwidth variations. This may offer improved performance as compared to using client-server interaction, which can be complicated, react more slowly, and may not make correct decisions if the encoder's output bit rate is expected to vary.
Fast-Start Streaming
Conventionally, when a streaming media player receives a command to play a media stream, the media player fills its audio/video buffers before starting the playback. A buffer of five seconds or longer is typically maintained to ensure smooth playback of the media, since the time taken to transfer the media stream over a network typically varies while the media stream is meant to be played at a constant rate. Once the required amount of data is accumulated to fill the buffer sufficiently, the media player starts playing the requested content at normal play speed (i.e., 1.0.times.). Disadvantageously, the user must wait the amount of time required to fill the buffer before viewing the requested content. This time may be very small in applications that send the stream faster than real time (e.g., a media stream from a storage device), but it is noticeable when the received media stream is at normal playback speed (i.e., 1.0.times.). In such a case, filling a five-second buffer would take five seconds (assuming no network communication issues). While this delay may be tolerated in some streaming media applications, it becomes unbearable where the media streams are changing, such as in a personal media broadcaster in which the viewer is changing channels. In such a case, the user would have to wait for the buffer to refill each time the channel (and thus the media stream) is changed.
To avoid this disadvantage, in one embodiment of the invention, the media player performs a fast start when a new media stream is selected. With the fast start, the media player client application plays the video immediately as it is received from the media broadcaster while still filling its buffer. In this way, the user does not have to wait for the media player's A/V buffers to fill up, e.g., upon channels changes, and the media player can still build its buffers to provide smooth playback. The media player is able to fill its buffers even though it plays the media stream immediately because the media player plays the A/V stream slightly slower than normal play speed. By playing the stream slightly slower than normal playback speed, the part of the received media stream that runs ahead of the stream played back is added to the buffer, resulting in accumulation of the A/V buffer—albeit more slowly. Because the user need not wait for content to be buffered and can immediately start watching the content, the experience is much more similar to that with a normal television.
In one example, the media player begins in fast-start mode playing back a received media stream immediately but at slightly slower speed than normal play speed, e.g., 0.85.times. speed. Since the broadcaster is streaming at normal speed, the media player slowly accumulates media stream data in its A/V buffer. For example, if the playback speed is 85% normal, the data accumulates in the A/V buffer at a rate of 15% (or 1.5 second for every 10 seconds of received streaming data). Once the media player's buffer is full or otherwise reaches an acceptable level, the media player begins to play the media stream at a normal rate, and the A/V buffer stops accumulating data.
When the play speed of stream is changed slightly, the change in the video stream is generally not noticeable. However, the change in audio becomes immediately perceptible. To take care of this, the media player may use time-stretching on the audio stream while maintaining the “pitch” in the audio stream. One software tool that can be used to time-stretch the audio stream is SoundTouch, an open-source audio processing library. In one embodiment, the playback speed is increased gradually from the low threshold (e.g., 0.85.times.) to the real time playback speed (e.g., 1.0.times.). The rate at which the playback speed is increased may be a function of the buffer level, so the user does not perceive any drastic change in the playback speed. The timestamps of the audio and video samples may be changed according to the existing playback speed to reduce jerks in the video stream. As a result, the change in the audio stream is also not perceptible, and the user is less likely to perceive the difference between the normal streaming speed and the initial fast-start playback speed.
Control Mode for Low Latency
One of the most used features in TV viewing is channel control, and users expect such control operations to take at most one second to execute. But streaming video over IP networks performs best when some delay is added between the server and the client. This delay is needed to fill the A/V buffers in the media player client. The two requirements of low delay for user interactions and smooth audio-video display are conflicting in nature. To deal with both of these requirements, in one embodiment of the invention, two modes of operations are introduced: normal mode and control mode.
In the normal mode, the system performs conventional audio-video streaming, wherein a buffer of five or more seconds is maintained by the media player to ensure smooth playback. In this mode, the media broadcaster may also start buffering data if the network bit rate drops below encoder bit rate (i.e., the encoder at the broadcaster runs ahead of the media stream transmission). The total delay between video input to the media broadcaster and its viewing on the media player is thus the sum of three parameters: the time taken to buffer on the media broadcaster, the network transmission time taken for the stream to move from the media broadcaster to the media player, and the time taken to buffer on the media player. The network transmission time taken for the stream to move from the media broadcaster to the media player cannot be directly controlled; however, both the media broadcaster and the media player can minimize delays caused by their buffers.
The system enters control mode when the user starts interacting with the media broadcaster by way of the media player. In one embodiment, user interaction is defined as when the user requests an IR command or tuner command to control the operation of the A/V source device through the media player user interface. In one embodiment, the system returns to normal mode from control mode after a predetermined amount of time has passed since the last action that could cause the system to start control mode. In another embodiment, the system returns to normal mode immediately after it performs the operations associated with going into control mode, as described below.
On entering control mode, the media player and the media broadcaster change their behavior relative to that in normal mode in a number of ways.
As mentioned, in the normal mode the media player performs normal A/V streaming wherein the media player reads the media stream from the network buffer, parses the stream, and fills its audio/video buffers. This buffer is maintained to ensure smooth playback. When going into control mode, the media player flushes the data that is present in the A/V buffers and the network buffer. In one embodiment, when going into control mode, the media player makes the source filter flush all the data buffered in the A/V buffers and also in all the filters downstream (decoders and renderers), which may be holding 2-3 seconds worth of content. SP then flushes all the data present in the network buffer. Thereafter, the media player sends a notification to the media broadcaster to go into control mode and waits for the next I-frame received from the media broadcaster.
After the flush operation on the source filter and the network buffer, discontinuities occur in the media stream. The filter has intelligence built in so that if there is any packet discontinuity in the media stream, the filter waits for the next I-frame. Any incoming data that is not an I-frame is discarded by the source filter until a valid I-frame is detected. When the media broadcaster goes into control mode, it sends an I-frame immediately. As soon as this I-frame is detected by the source filter, it is sent downstream for rendering. While the media player remains in control mode, no buffering occurs in the source filter; the samples are sent downstream for rendering as soon as they are read from the network buffer. In this manner, the media player reduces the latency on the client side.
In one embodiment, upon a change into control mode, the media broadcaster stops buffering data and flushes data currently contained in its buffers. The media broadcaster then immediately generates an I-frame (also known as key frame) to send to the media player. The I-frame allows the media player to reconstruct an image, whereas other types of frames that encode the frame based on previous frames would not allow the media player to reconstruct the frame due to the discontinuity in the media stream.
Given the dependency of modern audio-video encoding standards, however, it may not be desirable for the media broadcaster to stop buffering data completely. This is because without buffering there may be too much data loss, leading to an extremely poor user experience. A compromise can therefore be made to balance low delay and a reasonable user experience. Based on a set of empirical values for achieving a good balance, an amount of data is allowed to be buffered by the media broadcaster during control mode as a function of bit rate of content. For example, in control mode, buffered data can be reduced by flushing the data if it crosses a limit of about one second; however, other empirical values may be used for various applications.
In one embodiment, the media broadcaster keeps track of effective transmission bit rate, for example by periodically calculating the average transmission bit rate for the last few seconds. Based on the bit rate it calculates permissible buffer usage. If usage goes beyond the current limit then all data is removed and an I-frame is forced. This also means encoder output rate is more than network transmission rate; therefore, encoder output is reduced to half of observed transmission rate. Due to the resulting discontinuity in the media stream, an I-frame is also forced.
Too many user control commands in quick succession will force many I-frames, which will adversely affect the encoder. Therefore, all buffer flushes and forcing of I-frames are preferably timed and spaced out by at least one second.
When returning to normal mode from control mode, the source filter of the media player pauses the playback, buffers (e.g., up to five or more seconds) content in the A/V buffers, and then sends the data downstream for rendering.
To operate effectively in a low latency, low buffer environment, a command can be immediately sent to the personal broadcaster to reduce the bit rate when in control mode. Without any additional adjustments, this would result in a reduction in image quality; however, because the menu screens being navigated typically feature very little motion, the frame rate of the video can be significantly reduced. With less frames to process, the personal broadcaster can output a stream with image quality good enough to effectively read the on-screen text.
The low latency achieved by the dual mode operation of this embodiment is a very desirable characteristic for systems in which the users interacts with the media broadcaster through the media player. The addition of a control mode achieves this low latency, while the return to normal mode once user interaction has stopped achieves the smooth streaming desired for normal viewing. Moreover, in combination with the fast-start streaming described above, the control mode function provides a low latency control interaction experience for the user with a seamless transition from control mode to normal viewing.
Trick Play of Received IP Stream
In one embodiment, a user viewing a live stream being encoded in real time by the personal broadcaster and displayed on the media player can replay the last several seconds or minutes of content just viewed, or it can pause the live stream for resumption at some point in the near future. The personal broadcaster receives and processes the analog input signal; converts it into digital format; encodes the video in a compression algorithm such as MPEG-4, H.264, Windows Media Video Series 9, or another appropriate format; and then streams the encoded via stream over TCP/IP (or an alternative protocol, such as UDP, RTP, RTSP) to the media player.
To enable this functionality, the media player caches the last five minutes (or some other fixed period of time as chosen by user or limited by product manufacturer) of the audio and video content to a local storage medium while receiving a video stream. The media player may store the data on a local storage medium, such as a hard drive in the case of a PC, or some removable media, including but not limited to compact flash, smart media, a memory stick, or a micro drive.
When the user wishes to pause or replay the content, the user instructs the media player to do so by selecting an appropriate labeled button in the media player user interface. When the user instructs the media player to rewind, the media player accesses the content cached in its storage, allowing the user to scan through it, and play as desired. Once the user has “rewound” content, the user can then “fast forward” the viewing of the stream at faster than real-time speed until the current-most point in the video stream is being displayed. When the user instructs the media player to pause, the media player pauses the video stream being displayed but continues to receive the incoming stream, which it caches to its local storage medium. The media player continues to cache the stream as it is received, until the point when the maximum number of minutes (or size of data) allowed to be cached when in pause mode is reached. If the maximum number of minutes (or size of data) is reached, the media player resumes playback. Otherwise, the media player resumes playback when the user instructs the media player to do so.
Capturing, Editing, and Sending Video Clips from a Streaming Source
Embodiments of the invention also allow people to send video clips to friends and acquaintances. In accordance with one embodiment, the media player device continually captures and caches the last five minutes (or some other designated time period) of video and audio received. The media player automatically discards the earliest recorded content when the cached stream hits five minutes (or some other designated time period), so that the five minutes being cached is always the five most recent minutes of media streamed.
When a user sees something of interest that the user would like to capture and save or share, the user clicks on or selects a designated button or command using the media player's graphical interface. Once selected, the media player brings up a simple video editing interface, which enables the user to select the beginning and end of the clip that the user would like to capture. Once the bounds of the clip have been selected, the user selects a command to save the clip or send the clip to another person.
In response to a command to send the captured video clip, the media player calls the email client resident on the user's client device, creates a new message, and attaches the video clip to the message. The user can then select the intended recipients of the message and send the message. The address book databases can be those available from email clients such as Microsoft Outlook, Lotus Notes, and others, as well as any Internet based messaging services, such as MSN Messenger and AOL Instant Messenger.
Personal Video Recorder (PVR) Functionality for Mobile Devices
As users get more accustomed to video experiences on mobile devices, and mobile storage solutions (e.g., removable flash media and small hard disk drives) become more robust and inexpensive, the desire to add personal video recorder functionality to these devices will emerge. One embodiment of the present invention adds the functionality of a personal video recorder to a mobile phone, communicator, PDA or other device connecting to a Wide Area Network or other remote network outside a user's local area network.
One variation that accomplishes this can feature a scheduler as part of the remote client application which coordinates with an electronic programming guide. When the user launches the remote client application, he can search for and select the shows he would like to have recorded to his remote client. When the time for the scheduled recording occurs, the remote client application initiates a recording. The application can call the connection manager on the remote client, which in turn opens an Internet connection. The remote client application then connects to the personal broadcaster at the user's home.
Next, the remote client application can issue the right set of commands related to channel selection and encoder settings on the server, and begins to capture the incoming stream onto the local storage medium (e.g., hard disk drive or flash memory). This can happen in the background, so a user can be on phone call or listening to music (depending on other capabilities of the device) while the remote client is recording.
If the recording is interrupted because a network connection is lost, the live streaming content can be temporarily stored on the local (built-in or removable) storage that exists in one variation of the personal broadcaster. Once a connection between the personal broadcaster and the remote client is reestablished, the content resumes the stream from the personal broadcaster to the remote client, taking into consideration the point at which the previously interrupted transfer left off. In another variation, an on-network storage device, such as a PC, can be used to store content temporarily after a connection between the personal broadcaster and the remote client has been interrupted. In another embodiment, an entire show or media program is encoded and stored on the personal broadcaster (or in a storage device), and then transferred to the mobile device in the background. Beneficially, the transfer can be performed at a more efficient time for data transfer e.g., the file downloaded to a cellular phone overnight during non peak hours so that the content is available to the user for offline viewing the following day.
In the case where the original source of the media being streamed is an on-network storage device (using the personal broadcaster as a transcoder/transrater), transfers between the personal broadcaster and the remote client can be accomplished over time. That is, if a remote client loses network connectivity, the download pick ups and continues later at the proper spot in the content.
Another embodiment enabling a PVR on a remote device, a variation of the embodiment described above, allows the personal broadcaster to work with an on-network storage device (such as a PC or a NAS or SAN device) to provide for the user a PVR that can be accessed remotely. The personal broadcaster takes in the analog content, digitizes and encodes the content, and then streams the resulting media stream to the on-network storage device for storage. When a user wants to access the content from a remote client, the personal broadcaster acts as an arbiter between the remote client, which wants to view the content, and the on-network storage device. The personal broadcaster receives the stream from the on-network storage device and transcodes or transrates the content if necessary before repackaging and streaming it to the remote client.
Pairing Programming Being Viewed with Context-Specific Content/Advertisements
One embodiment of the invention provides the ability for the client to present the user with Web pages, scrolling text with news, or other information that varies based on the content the user is currently viewing. For example, if the user is currently watching a San Francisco Giants baseball game, the user can be presented with a “news ticker” that details other baseball scores around the league, or perhaps a Web page with statistics and facts about the San Francisco Giants. The content can be embedded in the interface of the client application itself or can be presented through the launching of another application, such as a Web browser. Similarly, the user can be presented with content and context-specific advertisements. An example in the context of the same baseball game would be an advertisement from the San Francisco Giants ticket office, which hopes to attract viewers to purchase tickets to a future game.
The broadcasting system can determine the content currently being viewed by the user in a number of ways. In one embodiment, the personal broadcaster or client sends information, including the current time, the channel lineup being used (e.g., location and cable TV provider), and the current channel being tuned to, to the central server. The central server then takes the information received and examines the electronic programming guide applicable to the user's service. From information it is determined which show the user is currently viewing. Alternatively, the client can perform the programming guide lookup itself and transmit this information to the central server.
After determining what show the user is currently viewing, overlaid information and/or advertisements can be transmitted from the central server or other server on the public Internet to the client media player. Alternatively, applications, such as a Web browser, can be launched, simply sending the user to a specified URL. The client can also contain the information necessary to queue any relevant information or advertisement for display to the user. For example, the client can have stored in its memory certain advertising banners or URLs of relevant Web sites that are queued depending on what type of content the user views. This method alleviates much of the need for transmission of content and or advertisements from a central server or other remote server.
Embodiments that include personal video recorder functionality (using local storage on the personal broadcaster or using storage of the on-network storage device) can provide for context-specific ads and information even when playing a previously recorded show. When recorded, the content is “tagged” with information indicating the programming being captured. On play back, this information can be used to prompt content specific information and ads.
Content-Specific Edge Preserving Pre-Filter
Low bit rate coding involving block based video coders produce strong Mocking artifacts. To reduce the severity of these artifacts, in one embodiment of the invention, pre-filtering is employed to simplify the image content before compression. The pre-filters often comprise low pass filters, which not only smooth Mocking artifacts but also undesirably reduce image details. Moreover, low pass filters are not adaptive to video content, which makes them ill-suited for perceptual video encoding. To avoid the limitations of previous pre-filters, one embodiment of media player performs a content-specific edge estimation algorithm on the media stream received from the media broadcaster. In accordance with this embodiment, an edge estimate gives the location of the details in the image that should be preserved while a motion estimate gives a classification of whether the content is in high, medium, or low motion.
In one embodiment, the media player applies a pre-filter on the received stream. The pre-filter comprises a low pass filter having pass band characteristics that are changed based on an edge estimate, a motion level estimate, and an encoding bit rate. Based on the motion level estimate and the encoding bit rate, a filter characteristic is defined for a frame. The filter characteristic is then fine tuned at the pixel level based on the edge estimate. Using this approach, a higher degree of smoothing is used for high motion contents and lower degree for low motion content, while leaving the details intact.
In one embodiment, two sets of adaptive low pass filters are defined. For low bit rate encoding, the following Gaussian filter of size 5.times.5 is used:
G(x,y)=½.pi.sigma.2exp(−(x2+y2)2.sigma.2).##EQU00001##
For high bit rate encoding, a low pass filter having an average mask of size 3.times.3 is used, as shown in the table below.
TABLE-US-000011111.psi.1111
The parameters .sigma. and .psi. are varied based on the edge estimate, motion level estimate, and the encoding bit rate. Increasing .sigma. or decreasing .psi. increases the smoothness of the video, while the opposite decreases the smoothness of the video. The edge estimate, is operated at pixel level and is obtained by taking the gradient at all eight directions. This is given by:
=.SIGMA.DELTA.f.sub.theta.,.theta.{0.degree.,45.degree.,90.degree.,135.degree.,180.degree.,225.degree.,270.degree.,315.degree.},
where f.sub.0.degree.={(x,y)−(x+1,y)}, f.sub.45.degree.={(x,y)−(x,y+1)}, . . . .
If is greater than a threshold T, no pre-filtering is performed on the current pixel, thereby preserving the image details. The motion level estimate gives a classification of whether the motion is high, medium, or low motion. The estimate is based on the bits required for the current frame and the recursive average of the bits per frame. If the current frame bits are greater than .alpha. (always >1) times the recursive average, the current video content is declared as high motion, and if the current frame bits are less than .beta.(<1) times the recursive average, the video content is classified as low motion. Otherwise, it is classified as medium motion. Based on the type of motion, the filter strength is varied. For high motion content, for example, a relatively heavy smoothing filter can be applied as compared to low motion content.
This pre-filtering method not only retains the image details but also is content specific. That is, the filters are adapted to the motion type: high, medium, or low. This pre-filtering method provides non-blocky, constant quality encoded video for all motion types. Unlike conventional pre-filters this reduces the frame drops without compromising quality.
Frame Rate Regulation and Quality Control for Encoder
Most video software encoders at low bit rates dynamically drop video frames to meet some specific target bit rate, often during high motion. This dropping of frames can lead to jerky video and fluctuating quality levels. As potentially large sections of frames are dropped, the motion estimation process becomes ineffective. To avoid this problem, one embodiment of the media broadcaster regulates frame rate using a multi-level approach designed to enhance the viewing experience of the user by sustaining both frame rate and quality at each level.
In one embodiment, the encoder in the media broadcaster estimates a sustainable frame rate based on a Sustainability measure, SM. Four levels of frame rate regulation are defined. Based on SM, the frame rate is selected and an appropriate quality level is defined. Each stage of the algorithm keeps the inter-frame distance constant, preserving temporal video quality. This results in better motion estimation and acceptable spatial video quality levels.
Frame rate regulation is performed in one embodiment according to the following algorithm. The target frame rate defined by the encoder is F.sub.0. A sustainability measure, SM, determines whether the frame rate can be sustained over the time interval T. SM can be defined as the ratio of the bit pool available to the estimated bits required for the next T seconds. The bit estimate is a product of frames per T seconds with the recursive average of the bits per frame. The SM may be calculated based on the “motion” estimate and the target bit rate. The “motion” estimate is based on the recursive average of the encoded bits per frame. In one embodiment, the frame rates at subsequent levels are: F.sub.1=F.sub.0/2; F.sub.2=F.sub.0/3; and F.sub.3=F.sub.0/5.
If F.sub.i denotes the assigned frame rate for level i, based on SM, a decision is made every T whether the encoder should continue in the same level or move to a level higher or a level lower. If SM is less than a (indicating that that the current frame rate, F.sub.i, cannot be sustained for the time interval T), the encoder moves to level i+1. If the encoder is already at the lowest frame rate, F.sub.3, it stays there. This operation is again performed after every T. The level thus goes down when bit rate is not sufficient enough to cater high motion content bit rate need.
While in level i, as long as i is not o, SM for level i−1 is checked after time interval of T. If the SM is greater than .beta.(>1), it indicates that the current content can be encoded at level i−1, so the encoder moves to level i−1. This method thus tries to maintain the inter-frame distances constant at each level thereby improving the overall video quality.
Dynamic Parameter Control for Video Encoder
In a real time video streaming environment, network characteristics change dynamically. To improve the use of the network resources, a video encoder should be able to adapt to these network changes; however, most of the standard video encoders do not support such adaptation. In accordance with one embodiment of the invention, a scheme for the Windows Media Video (WMV9) encoder is provided wherein the parameters can be changed dynamically during streaming.
The following parameters can be changed dynamically while streaming is in progress: bit rate, frame rate, video smoothness, and I-frame interval. The bit rate can adapt to the rate that can be supported by the network at a given time, and rate control buffer delay, bits per frame, and quantization step size vary according to changes in bit rate. Frame rate depends on the dynamically changed bit rate. At lower bit rates, high frame rates cannot be sustained, which creates the need in some circumstances for a dynamic frame rate. The video smoothness parameter indicates the encoded video quality. Lower bit rates supporting high values of video smoothness can cause jerky video, while low values of video smoothness at higher bit rates leads to under utilization of the available resources. Depending on the video smoothness parameter, the quantization upper bound and lower bound step values are changed, which affects the quality of the video. Because the bits required to encode I-frames are greater than bits required for P frames, I-frame intervals are larger for lower bit rates. I-frame intervals can be reduced at higher bit rates, where more bits can be afforded.
Accordingly, the encoder in the media broadcaster can dynamically adjust these parameters to adapt to the changing network characteristics and optimize use of system resources. This allows for improved performance than the standard WMV9 encoder, which does not dynamically change the above parameters in a real time streaming environment. Although described in the context of video encoding with Windows Media Video (WMV9), this method can be applied to other video compression formats, including MPEG-4, H.263, H.264, and any other compression formats that use the same or similar parameters mentioned herein.
Point and Click Interaction with Traditional CE Menus
Because the personal broadcaster digitizes, encodes and streams the analog output of the A/V source devices being used, the client presents the user with the full interface of the A/V source device. The default paradigm for navigating menus of the A/V source device that are rendered by the client is exactly the same as it would be for a user viewing the A/V source device in a more traditional fashion. That is, the mode of navigation typically involves “Up,” “Down,” “Right,” “Left,” and “Select” as its key components. However, alternative methods for interacting with menus and other lists are possible. For example, instead of pressing the “Down” command four times to highlight an item four spaces down from what is currently highlighted, an embodiment of the media player allows a user simply to point and click directly on the desired menu item. This speeds up the interaction and takes advantage of the rich human interface tools (e.g., keyboard and mouse) that are available in many computing devices but not generally used with televisions.
Hot Spot Finder
In one embodiment of the invention, the media player includes a directory of hotspots (wireless broadband networks available in public locations) stored on the client device and accessible when the client is not connected to a network. Such a directory allows the user more easily to find a location where the user can gain access to the personal media broadcaster.
Testing and Configuring a User's NAT
For users that have a home Internet gateway that includes a Network Address Translation (NAT), some configuration can be required to allow a user to access the personal broadcaster from a remote network. This is because incoming requests are rejected by many NATs unless the NAT has been explicitly instructed how to forward the incoming packets. There are many methods for solving this problem, some more desirable and automated than others. Some embodiments of the invention include a multi-step process for determining and implementing the possible solutions.
The system may first attempt to determine whether the home Internet gateway supports UPnP (Universal Plug 'N Play). If so, the personal broadcaster may be able to control the home Internet gateway using UPnP. The personal broadcaster can thus instruct the NAT to forward a specified port to the internal IP address dedicated to the personal broadcaster.
If the home Internet gateway does not support UPnP, the system may then attempt to determine the type of NAT in the home Internet gateway, and specifically whether the NAT is a full cone NAT. This detection may be performed by using the central server as a STUN server, which runs various tests to determine what the type of the NAT behind which the personal broadcaster sits. There are four basic types of NATs: full cone, restricted cone, port restricted cone, and symmetric. A full cone NAT is a NAT that allows a client behind it to receive messages from one external machine that are addressed to an IP address and port that the internal client used in sending a message to another external machine. If the personal broadcaster sits behind a full cone NAT, the following is possible: The personal broadcaster from time to time sends a message to the central server, and the central server makes a note of the IP address and port from which it was received. When a remote client wishes to connect to the personal broadcaster, it asks the central server for the address and port recently used by the personal broadcaster to send its message. The remote client can use the same IP address and port to locate and connect to the personal broadcaster. If the NAT is not a full cone NAT, another method is used.
A third possible method to use comprises a UDP “hole punching” technique. This technique, which is known to those of ordinary skill in the relevant art, would involve using the central sever as a way to “introduce” the remote client and the personal broadcaster. The method works for all NAT types except for symmetric NATs (so the STUN test would be useful for determining if this is a viable option) and uses the central server to cause both devices to send messages to one another simultaneously. Because both devices believe they are “initiating” the conversation, the return packets are permitted to flow through the NAT to the destination device.
If none of the above or other methods are viable solutions, the present invention can walk the user through the steps for manual configuration of the NAT on the home Internet gateway. To make this easy and integrated into the set up process for the personal broadcaster, the configuration screens for the home Internet gateway can be embedded in the window that houses the set up application. This provides the user with a greater sense of seamlessness.
On-Device EPG
One variation of the present invention features an electronic programming guide (EPG) that resides locally on a client. The EPG is configured at initial configuration of the personal broadcaster. The user is asked for a zip code and the service and package to which the user subscribes. Based on this information, the client application downloads an EPG that covers the next several days. From time to time, the EPG is updated via download from the central server or another server from a third party provider. The EPG can be stored locally on the client.
Beneficially, the EPG can be made interactive. Among the many features enabled by an on-device EPG, a user could search and sort programming content by a number of variables, and a single “click” on a channel can automatically tune the A/V source device and media player to the desired channel.
Community or “Buddy” List
One embodiment of the media player incorporates a “buddy list.” Using the buddy list, a user can connect to personal broadcasters that reside at different locations. For example, if Charlie declares Amy as a “buddy,” Charlie's personal broadcaster appears on Amy's buddy list. By choosing Charlie's personal broadcaster, Amy connects to Charlie's device. All of the necessary settings (e.g., IP address, port, password, and any other required settings) are automatically provided to the buddy.
Last Come, Last Served
Another embodiment of the invention allows only a single client to be connected to the personal broadcaster at any given time. The broadcaster may implement a number of priority schemes, one of which is last come last served. In this scheme, if a client A is connected and client B attempts to connect, priority is given to the client B. One embodiment can provide a message to client A informing the user of client A that the client is about to be disconnected. The user of client A may be provided the opportunity to override this rule and remain connected. This priority system is especially useful in the situation where a person logs in at home, then leaves the house without disconnecting and attempts to log in from a remote client.
Encryption and Security
Various known security mechanisms can be used in different embodiments of the present invention. Example of security mechanisms that can be used with the personal media broadcasting system described herein include, but are not limited to, password protection, communication over a secure link, encrypting the content sent over the remote network.
Blocking Out the Local Display
In another embodiment the media broadcaster includes A/V pass-throughs for all of the inputs, where there is an output that corresponds to each of the inputs. This saves the user from using multiple outputs on an A/V source device and may provide a complete method for prohibiting two sessions. Moreover, this embodiment can be used to prohibit the simultaneous occurrence of a client connection to a personal broadcaster and a local viewing of the A/V source which is plugged into the personal broadcaster. For example, if someone were watching a pay per view movie at home, and a user connected to the personal broadcaster tunes into the same movie, the personal broadcaster can disable its A/V outputs. Whether the A/V outputs are disabled can depend on the content. For example, in one embodiment only pay per view content triggers a blocking, whereas regular programming does not. This discrimination scheme can be integrated with an EPG, as discussed above.
The foregoing description of the embodiments of the invention has been presented for the purpose of illustration; it is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Persons skilled in the relevant art can appreciate that many modifications and variations are possible in light of the above teachings. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by the claims appended hereto.
This application is a continuation application of U.S. patent application Ser. No. 13/079,644 filed on Apr. 4, 2011, which is a continuation application of U.S. application Ser. No. 12/643,812, which was filed on Dec. 21, 2009. That application is a continuation of U.S. patent application Ser. No. 11/147,985, which was filed on Jun. 7, 2005 and claims the benefit of U.S. Provisional Application No. 60/577,833, filed Jun. 7, 2004, which is incorporated by reference in its entirety. That application is also related to U.S. Application entitled, “Personal Media Broadcasting System,” to Krikorian et al., filed Jun. 7, 2005, and U.S. Application entitled, “Personal Media Broadcasting System With Output Buffer,” to Krikorian et al., filed Jun. 7, 2005, each of which is incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3416043 | Jorgensen | Dec 1968 | A |
4254303 | Takizawa | Mar 1981 | A |
5132992 | Yurt et al. | Jul 1992 | A |
5161021 | Tsai | Nov 1992 | A |
5237648 | Mills et al. | Aug 1993 | A |
5293357 | Hallenbeck | Mar 1994 | A |
5321846 | Yokota et al. | Jun 1994 | A |
5377332 | Entwistle et al. | Dec 1994 | A |
5386493 | Degen et al. | Jan 1995 | A |
5434590 | Dinwiddie, Jr. et al. | Jul 1995 | A |
5493638 | Hooper et al. | Feb 1996 | A |
5537530 | Edgar et al. | Jul 1996 | A |
5602589 | Vishwanath et al. | Feb 1997 | A |
5661516 | Carles | Aug 1997 | A |
5666426 | Helms | Sep 1997 | A |
5682195 | Hendricks et al. | Oct 1997 | A |
5706290 | Shaw et al. | Jan 1998 | A |
5708961 | Hylton et al. | Jan 1998 | A |
5710605 | Nelson | Jan 1998 | A |
5722041 | Freadman | Feb 1998 | A |
5757416 | Birch et al. | May 1998 | A |
5774170 | Hite et al. | Jun 1998 | A |
5778077 | Davidson | Jul 1998 | A |
5794116 | Matsuda et al. | Aug 1998 | A |
5822537 | Katseff et al. | Oct 1998 | A |
5831664 | Wharton et al. | Nov 1998 | A |
5850482 | Meany et al. | Dec 1998 | A |
5852437 | Wugofski et al. | Dec 1998 | A |
5880721 | Yen | Mar 1999 | A |
5889506 | Lopresti et al. | Mar 1999 | A |
5898679 | Brederveld et al. | Apr 1999 | A |
5909518 | Chui | Jun 1999 | A |
5911582 | Redford et al. | Jun 1999 | A |
5922072 | Hutchinson et al. | Jul 1999 | A |
5936968 | Lyons | Aug 1999 | A |
5968132 | Tokunaga | Oct 1999 | A |
5969764 | Sun et al. | Oct 1999 | A |
5987501 | Hamilton et al. | Nov 1999 | A |
6002450 | Darbee et al. | Dec 1999 | A |
6008777 | Yiu | Dec 1999 | A |
6014694 | Aharoni et al. | Jan 2000 | A |
6020880 | Naimpally | Feb 2000 | A |
6023233 | Craven et al. | Feb 2000 | A |
6031940 | Chui et al. | Feb 2000 | A |
6036601 | Heckel | Mar 2000 | A |
6040829 | Croy et al. | Mar 2000 | A |
6043837 | Driscoll, Jr. et al. | Mar 2000 | A |
6044396 | Adams | Mar 2000 | A |
6049671 | Slivka et al. | Apr 2000 | A |
6075906 | Fenwick et al. | Jun 2000 | A |
6088777 | Sorber | Jul 2000 | A |
6097441 | Allport | Aug 2000 | A |
6104334 | Allport | Aug 2000 | A |
6108041 | Faroudja et al. | Aug 2000 | A |
6115420 | Wang | Sep 2000 | A |
6117126 | Appelbaum et al. | Sep 2000 | A |
6141059 | Boyce et al. | Oct 2000 | A |
6141447 | Linzer et al. | Oct 2000 | A |
6160544 | Hayashi et al. | Dec 2000 | A |
6201536 | Hendricks et al. | Mar 2001 | B1 |
6212282 | Mershon | Apr 2001 | B1 |
6222885 | Chaddha et al. | Apr 2001 | B1 |
6223211 | Hamilton et al. | Apr 2001 | B1 |
6240459 | Roberts et al. | May 2001 | B1 |
6240531 | Spilo et al. | May 2001 | B1 |
6243596 | Kikinis | Jun 2001 | B1 |
6256019 | Allport | Jul 2001 | B1 |
6263503 | Margulis | Jul 2001 | B1 |
6279029 | Sampat et al. | Aug 2001 | B1 |
6282714 | Ghori et al. | Aug 2001 | B1 |
6286142 | Ehreth | Sep 2001 | B1 |
6289485 | Shiomoto | Sep 2001 | B1 |
6310886 | Barton | Oct 2001 | B1 |
6340994 | Margulis et al. | Jan 2002 | B1 |
6353885 | Herzi et al. | Mar 2002 | B1 |
6356945 | Shaw et al. | Mar 2002 | B1 |
6357021 | Kitigawa et al. | Mar 2002 | B1 |
6370688 | Hejna, Jr. | Apr 2002 | B1 |
6389467 | Eyal | May 2002 | B1 |
6434113 | Gubbi | Aug 2002 | B1 |
6442067 | Chawala et al. | Aug 2002 | B1 |
6456340 | Margulis | Sep 2002 | B1 |
6466623 | Youn et al. | Oct 2002 | B1 |
6470378 | Tracton et al. | Oct 2002 | B1 |
6476826 | Plotkin et al. | Nov 2002 | B1 |
6487319 | Chai | Nov 2002 | B1 |
6493874 | Humpleman | Dec 2002 | B2 |
6496122 | Sampsell | Dec 2002 | B2 |
6505169 | Bhagavath et al. | Jan 2003 | B1 |
6510177 | De Bonet et al. | Jan 2003 | B1 |
6529506 | Yamamoto et al. | Mar 2003 | B1 |
6553147 | Chai et al. | Apr 2003 | B2 |
6557031 | Mimura et al. | Apr 2003 | B1 |
6564004 | Kadono | May 2003 | B1 |
6567984 | Allport | May 2003 | B1 |
6584201 | Konstantinou et al. | Jun 2003 | B1 |
6584559 | Huh et al. | Jun 2003 | B1 |
6597375 | Yawitz | Jul 2003 | B1 |
6598159 | McAlister et al. | Jul 2003 | B1 |
6600838 | Chui | Jul 2003 | B2 |
6609253 | Swix et al. | Aug 2003 | B1 |
6611530 | Apostolopoulos | Aug 2003 | B1 |
6628716 | Tan et al. | Sep 2003 | B1 |
6642939 | Vallone et al. | Nov 2003 | B1 |
6647015 | Malkemes et al. | Nov 2003 | B2 |
6658019 | Chen et al. | Dec 2003 | B1 |
6665751 | Chen et al. | Dec 2003 | B1 |
6665813 | Forsman et al. | Dec 2003 | B1 |
6697356 | Kretschmer et al. | Feb 2004 | B1 |
6701380 | Schneider et al. | Mar 2004 | B2 |
6704678 | Minke et al. | Mar 2004 | B2 |
6704847 | Six et al. | Mar 2004 | B1 |
6708231 | Kitagawa | Mar 2004 | B1 |
6718551 | Swix et al. | Apr 2004 | B1 |
6754266 | Bahl et al. | Jun 2004 | B2 |
6754439 | Hensley et al. | Jun 2004 | B1 |
6757851 | Park et al. | Jun 2004 | B1 |
6757906 | Look et al. | Jun 2004 | B1 |
6766376 | Price | Jul 2004 | B2 |
6768775 | Wen et al. | Jul 2004 | B1 |
6771828 | Malvar | Aug 2004 | B1 |
6774912 | Ahmed et al. | Aug 2004 | B1 |
6781601 | Cheung | Aug 2004 | B2 |
6785700 | Maasud et al. | Aug 2004 | B2 |
6788882 | Geer et al. | Sep 2004 | B1 |
6795638 | Skelley, Jr. | Sep 2004 | B1 |
6798838 | Ngo | Sep 2004 | B1 |
6806909 | Radha et al. | Oct 2004 | B1 |
6807308 | Chui et al. | Oct 2004 | B2 |
6816194 | Zhang et al. | Nov 2004 | B2 |
6816858 | Coden et al. | Nov 2004 | B1 |
6826242 | Ojard et al. | Nov 2004 | B2 |
6834123 | Acharya et al. | Dec 2004 | B2 |
6839079 | Barlow et al. | Jan 2005 | B2 |
6847468 | Ferriere | Jan 2005 | B2 |
6850571 | Tardif | Feb 2005 | B2 |
6850649 | Malvar | Feb 2005 | B1 |
6868083 | Apostolopoulos et al. | Mar 2005 | B2 |
6889385 | Rakib et al. | May 2005 | B1 |
6892359 | Nason et al. | May 2005 | B1 |
6898583 | Rising, III | May 2005 | B1 |
6907602 | Tsai et al. | Jun 2005 | B2 |
6910175 | Krishnamachari | Jun 2005 | B2 |
6927685 | Wathen | Aug 2005 | B2 |
6930661 | Uchida et al. | Aug 2005 | B2 |
6941575 | Allen | Sep 2005 | B2 |
6944880 | Allen | Sep 2005 | B1 |
6952595 | Ikedo et al. | Oct 2005 | B2 |
6970602 | Smith et al. | Nov 2005 | B1 |
6981050 | Tobias et al. | Dec 2005 | B1 |
7016337 | Wu et al. | Mar 2006 | B1 |
7020892 | Levesque et al. | Mar 2006 | B2 |
7032000 | Tripp | Apr 2006 | B2 |
7047305 | Brooks et al. | May 2006 | B1 |
7099951 | Laksono | Aug 2006 | B2 |
7103905 | Novak | Sep 2006 | B2 |
7110558 | Elliott | Sep 2006 | B1 |
7124366 | Foreman et al. | Oct 2006 | B2 |
7151575 | Landry et al. | Dec 2006 | B1 |
7155734 | Shimomura et al. | Dec 2006 | B1 |
7155735 | Ngo et al. | Dec 2006 | B1 |
7184433 | Oz | Feb 2007 | B1 |
7224323 | Uchida et al. | May 2007 | B2 |
7239800 | Bilbrey | Jul 2007 | B2 |
7344084 | Dacosta | Mar 2008 | B2 |
7382729 | Honda et al. | Jun 2008 | B2 |
7430686 | Wang et al. | Sep 2008 | B1 |
7464396 | Hejna, Jr. | Dec 2008 | B2 |
7502733 | Andrsen et al. | Mar 2009 | B2 |
7502818 | Kohno et al. | Mar 2009 | B2 |
7505480 | Zhang et al. | Mar 2009 | B1 |
7565681 | Ngo et al. | Jul 2009 | B2 |
7616955 | Kim | Nov 2009 | B2 |
7627227 | Evans et al. | Dec 2009 | B2 |
7634373 | Batterberry et al. | Dec 2009 | B2 |
7647614 | Krikorian et al. | Jan 2010 | B2 |
7676823 | Acharya et al. | Mar 2010 | B2 |
7716358 | Price | May 2010 | B2 |
7720432 | Colby et al. | May 2010 | B1 |
7725912 | Margulis | May 2010 | B2 |
7778372 | Takashima | Aug 2010 | B2 |
7788696 | Burges et al. | Aug 2010 | B2 |
7911946 | Poli et al. | Mar 2011 | B2 |
7913278 | Ellis et al. | Mar 2011 | B2 |
7921446 | Krikorian et al. | Apr 2011 | B2 |
7954056 | Graham | May 2011 | B2 |
7975062 | Krikorian et al. | Jul 2011 | B2 |
7995904 | Mirsky et al. | Aug 2011 | B2 |
8019202 | Takase et al. | Sep 2011 | B2 |
8019885 | Yu et al. | Sep 2011 | B2 |
8051454 | Krikorian et al. | Nov 2011 | B2 |
8060906 | Begeja et al. | Nov 2011 | B2 |
8099755 | Bajpai et al. | Jan 2012 | B2 |
8365236 | Krikorian et al. | Jan 2013 | B2 |
8621533 | Krikorian et al. | Dec 2013 | B2 |
8799969 | Krikorian et al. | Aug 2014 | B2 |
20010021998 | Margulis | Sep 2001 | A1 |
20020004839 | Wine et al. | Jan 2002 | A1 |
20020010925 | Kikinis | Jan 2002 | A1 |
20020012530 | Bruls | Jan 2002 | A1 |
20020019984 | Rakib | Feb 2002 | A1 |
20020031333 | Mano et al. | Mar 2002 | A1 |
20020032907 | Daniels | Mar 2002 | A1 |
20020046404 | Mizutani | Apr 2002 | A1 |
20020053053 | Nagai et al. | May 2002 | A1 |
20020080753 | Lee | Jun 2002 | A1 |
20020090029 | Kim | Jul 2002 | A1 |
20020105529 | Bowser et al. | Aug 2002 | A1 |
20020112247 | Horner et al. | Aug 2002 | A1 |
20020122137 | Chen et al. | Sep 2002 | A1 |
20020131497 | Jang | Sep 2002 | A1 |
20020138843 | Samaan et al. | Sep 2002 | A1 |
20020143973 | Price | Oct 2002 | A1 |
20020147634 | Jacoby et al. | Oct 2002 | A1 |
20020147687 | Breiter et al. | Oct 2002 | A1 |
20020151992 | Hoffberg et al. | Oct 2002 | A1 |
20020167458 | Baudisch et al. | Nov 2002 | A1 |
20020188818 | Nimura et al. | Dec 2002 | A1 |
20020191575 | Kalavade et al. | Dec 2002 | A1 |
20030001880 | Holtz et al. | Jan 2003 | A1 |
20030028643 | Jabri | Feb 2003 | A1 |
20030028873 | Lemmons | Feb 2003 | A1 |
20030030752 | Begeja et al. | Feb 2003 | A1 |
20030065915 | Yu et al. | Apr 2003 | A1 |
20030078973 | Przekop et al. | Apr 2003 | A1 |
20030088686 | Jennings | May 2003 | A1 |
20030093260 | Dagtas et al. | May 2003 | A1 |
20030095791 | Barton et al. | May 2003 | A1 |
20030110514 | West et al. | Jun 2003 | A1 |
20030115167 | Sharif et al. | Jun 2003 | A1 |
20030159143 | Chan | Aug 2003 | A1 |
20030187657 | Erhart et al. | Oct 2003 | A1 |
20030192054 | Birks et al. | Oct 2003 | A1 |
20030208612 | Harris et al. | Nov 2003 | A1 |
20030231621 | Gubbi et al. | Dec 2003 | A1 |
20030231655 | Kelton et al. | Dec 2003 | A1 |
20030231868 | Herley | Dec 2003 | A1 |
20030234803 | Toyama et al. | Dec 2003 | A1 |
20040003406 | Billmaier | Jan 2004 | A1 |
20040049791 | Shah et al. | Mar 2004 | A1 |
20040052216 | Roh | Mar 2004 | A1 |
20040068334 | Tsai et al. | Apr 2004 | A1 |
20040083301 | Murase et al. | Apr 2004 | A1 |
20040098748 | Bo et al. | May 2004 | A1 |
20040100486 | Flamini et al. | May 2004 | A1 |
20040101271 | Boston et al. | May 2004 | A1 |
20040103340 | Sundareson et al. | May 2004 | A1 |
20040139047 | Rechsteiner et al. | Jul 2004 | A1 |
20040162845 | Kim et al. | Aug 2004 | A1 |
20040162903 | Oh | Aug 2004 | A1 |
20040172410 | Shimojima et al. | Sep 2004 | A1 |
20040205830 | Kaneko | Oct 2004 | A1 |
20040212640 | Mann et al. | Oct 2004 | A1 |
20040215458 | Kobayashi et al. | Oct 2004 | A1 |
20040216173 | Horoszowski et al. | Oct 2004 | A1 |
20040236844 | Kocherlakota | Nov 2004 | A1 |
20040246936 | Perlman | Dec 2004 | A1 |
20040255249 | Chang et al. | Dec 2004 | A1 |
20050021398 | McCleskey et al. | Jan 2005 | A1 |
20050021830 | Urzaiz et al. | Jan 2005 | A1 |
20050027821 | Alexander et al. | Feb 2005 | A1 |
20050038981 | Connor et al. | Feb 2005 | A1 |
20050044058 | Matthews et al. | Feb 2005 | A1 |
20050050462 | Whittle et al. | Mar 2005 | A1 |
20050053356 | Mate et al. | Mar 2005 | A1 |
20050055595 | Frazer et al. | Mar 2005 | A1 |
20050060759 | Rowe et al. | Mar 2005 | A1 |
20050097542 | Lee | May 2005 | A1 |
20050114852 | Chen et al. | May 2005 | A1 |
20050132351 | Randall et al. | Jun 2005 | A1 |
20050138560 | Lee et al. | Jun 2005 | A1 |
20050172028 | Nilsson et al. | Aug 2005 | A1 |
20050198584 | Matthews et al. | Sep 2005 | A1 |
20050204046 | Watanabe | Sep 2005 | A1 |
20050216851 | Hull et al. | Sep 2005 | A1 |
20050223087 | Van Der Stok | Oct 2005 | A1 |
20050227621 | Katoh | Oct 2005 | A1 |
20050229118 | Chiu et al. | Oct 2005 | A1 |
20050246369 | Oreizy et al. | Nov 2005 | A1 |
20050251833 | Schedivy | Nov 2005 | A1 |
20050262534 | Bontempi et al. | Nov 2005 | A1 |
20050283791 | McCarthy et al. | Dec 2005 | A1 |
20050288999 | Lerner et al. | Dec 2005 | A1 |
20060011371 | Fahey | Jan 2006 | A1 |
20060031381 | Van Luijt et al. | Feb 2006 | A1 |
20060050970 | Gunatilake | Mar 2006 | A1 |
20060051055 | Ohkawa | Mar 2006 | A1 |
20060093318 | Cohen et al. | May 2006 | A1 |
20060095401 | Krikorian et al. | May 2006 | A1 |
20060095471 | Krikorian et al. | May 2006 | A1 |
20060095472 | Krikorian et al. | May 2006 | A1 |
20060095942 | Van Beek | May 2006 | A1 |
20060095943 | Demircin et al. | May 2006 | A1 |
20060101499 | Aravamudan et al. | May 2006 | A1 |
20060107226 | Matthews et al. | May 2006 | A1 |
20060117371 | Margulis | Jun 2006 | A1 |
20060146174 | Hagino | Jul 2006 | A1 |
20060156219 | Haot et al. | Jul 2006 | A1 |
20060174026 | Robinson et al. | Aug 2006 | A1 |
20060184980 | Cole | Aug 2006 | A1 |
20060206581 | Howarth et al. | Sep 2006 | A1 |
20060280157 | Karaoguz et al. | Dec 2006 | A1 |
20070003224 | Krikorian et al. | Jan 2007 | A1 |
20070005783 | Saint-Hillaire et al. | Jan 2007 | A1 |
20070022328 | Tarra et al. | Jan 2007 | A1 |
20070053446 | Spilo | Mar 2007 | A1 |
20070074115 | Patten et al. | Mar 2007 | A1 |
20070076604 | Litwack | Apr 2007 | A1 |
20070097257 | El-Maleh et al. | May 2007 | A1 |
20070127437 | Ozawa | Jun 2007 | A1 |
20070168543 | Krikorian et al. | Jul 2007 | A1 |
20070180485 | Dua | Aug 2007 | A1 |
20070198532 | Krikorian et al. | Aug 2007 | A1 |
20070234213 | Krikorian et al. | Oct 2007 | A1 |
20070260462 | Andersen et al. | Nov 2007 | A1 |
20070286596 | Lonn | Dec 2007 | A1 |
20070290876 | Sato et al. | Dec 2007 | A1 |
20080019276 | Takatsuji et al. | Jan 2008 | A1 |
20080037573 | Cohen | Feb 2008 | A1 |
20080059533 | Krikorian | Mar 2008 | A1 |
20080134267 | Moghe et al. | Jun 2008 | A1 |
20080175559 | Jung et al. | Jul 2008 | A1 |
20080195744 | Bowra et al. | Aug 2008 | A1 |
20080199150 | Candelore | Aug 2008 | A1 |
20080256485 | Krikorian | Oct 2008 | A1 |
20080294759 | Biswas et al. | Nov 2008 | A1 |
20080307456 | Beetcher et al. | Dec 2008 | A1 |
20080307462 | Beetcher et al. | Dec 2008 | A1 |
20080307463 | Beetcher et al. | Dec 2008 | A1 |
20090074380 | Boston et al. | Mar 2009 | A1 |
20090080448 | Tarra et al. | Mar 2009 | A1 |
20090102983 | Malone et al. | Apr 2009 | A1 |
20090103607 | Bajpai et al. | Apr 2009 | A1 |
20090157697 | Conway et al. | Jun 2009 | A1 |
20090177758 | Banger et al. | Jul 2009 | A1 |
20090199248 | Ngo et al. | Aug 2009 | A1 |
20090252219 | Chen et al. | Oct 2009 | A1 |
20090300205 | Jabri | Dec 2009 | A1 |
20100001960 | Williams | Jan 2010 | A1 |
20100005483 | Rao | Jan 2010 | A1 |
20100064055 | Krikorian et al. | Mar 2010 | A1 |
20100064332 | Krikorian et al. | Mar 2010 | A1 |
20100070925 | Einaudi et al. | Mar 2010 | A1 |
20100071076 | Gangotri et al. | Mar 2010 | A1 |
20100100915 | Krikorian et al. | Apr 2010 | A1 |
20100129057 | Kulkarni | May 2010 | A1 |
20100146527 | Craib et al. | Jun 2010 | A1 |
20100192007 | Tarra et al. | Jul 2010 | A1 |
20100192186 | Margulis | Jul 2010 | A1 |
20100192188 | Rao | Jul 2010 | A1 |
20100232437 | Bajpai et al. | Sep 2010 | A1 |
20100232438 | Bajpal et al. | Sep 2010 | A1 |
20100232439 | Asnis et al. | Sep 2010 | A1 |
20100268832 | Lucas et al. | Oct 2010 | A1 |
20110019839 | Nandury | Jan 2011 | A1 |
20110032986 | Banger et al. | Feb 2011 | A1 |
20110033168 | Iyer | Feb 2011 | A1 |
20110035462 | Akela | Feb 2011 | A1 |
20110035466 | Panigrahi | Feb 2011 | A1 |
20110035467 | Thiyagarajan et al. | Feb 2011 | A1 |
20110035668 | Thiyagarajan | Feb 2011 | A1 |
20110035669 | Shirali et al. | Feb 2011 | A1 |
20110035741 | Thiyagarajan | Feb 2011 | A1 |
20110035765 | Shirali | Feb 2011 | A1 |
20110051016 | Malode | Mar 2011 | A1 |
20110055864 | Shah et al. | Mar 2011 | A1 |
20110113354 | Thiyagarajan et al. | May 2011 | A1 |
20110119325 | Paul et al. | May 2011 | A1 |
20110150432 | Paul et al. | Jun 2011 | A1 |
20110153718 | Dham et al. | Jun 2011 | A1 |
20110153845 | Rao et al. | Jun 2011 | A1 |
20110158610 | Paul et al. | Jun 2011 | A1 |
20110191456 | Jain | Aug 2011 | A1 |
Number | Date | Country |
---|---|---|
2286641 | Oct 1998 | CA |
2324856 | May 2001 | CA |
1464685 | Dec 2003 | CN |
4407319 | Sep 1994 | DE |
0515101 | Nov 1992 | EP |
1443766 | Aug 2004 | EP |
2307151 | May 1997 | GB |
11289350 | Oct 1999 | JP |
2001054066 | Feb 2001 | JP |
2003046582 | Feb 2003 | JP |
2003114845 | Apr 2003 | JP |
2003179580 | Jun 2003 | JP |
2004007127 | Jan 2004 | JP |
2004072686 | Mar 2004 | JP |
2005032120 | Feb 2005 | JP |
20010211410 | Aug 2001 | KR |
19990082855 | Feb 2010 | KR |
200703018 | Jan 2007 | TW |
9837701 | Aug 1998 | WO |
0072596 | Nov 2000 | WO |
0133839 | May 2001 | WO |
0147248 | Jun 2001 | WO |
0193161 | Dec 2001 | WO |
02073443 | Sep 2002 | WO |
03052552 | Jun 2003 | WO |
2004032511 | Apr 2004 | WO |
2004045218 | May 2004 | WO |
2008024723 | Feb 2008 | WO |
Entry |
---|
USPTO, Non-Final Office Action, mailed May 15, 2014 for U.S. Appl. No. 13/194,783. |
USPTO, Notice of Allowance, mailed May 28, 2014 for U.S. Appl. No. 13/615,306. |
International Search Report and Written Opinion, PCT/US2005/020105, Feb. 15, 2007, 6 pages. |
International Search Report and Written Opinion for PCT/US2006/04382, mailed Apr. 27, 2007. |
Archive of “TV Brick Home Server,” www.tvbrick.com, [online] [Archived by http://archive.org on Jun. 3, 2004; Retrieved on Apr. 12, 2006] retrieved from the Internet <URL:http://web.archive.org/web/20041107111024/www.tvbrick.com/en/affiliate/tvbs/tvbrick/document18/print>. |
Faucon, B. “TV ‘Brick’ Opens up Copyright Can of Worms,” Financial Review, Jul. 1, 2003, [online [Retrieved on Apr. 12, 2006] Retrieved from the Internet, URL:http://afr.com/cgi-bin/newtextversions.pl?storyid+1056825330084&3ate+2003/07/01&pagetype+printer§ion+1053801318705&path+articies/2003/06/30/0156825330084.html.]. |
Balster, Eric J., “Video Compression and Rate Control Methods Based on the Wavelet Transform,” The Ohio State University 2004, pp. 1-24. |
Kulapala et al., “Comparison of Traffic and Quality Characteristics of Rate-Controlled Wavelet and DCT Video” Arizona State University, Oct. 11, 2004. |
Skodras et al., “JPEG2000: The Upcoming Still Image Compression Standard,” May 11, 2000, 14 pages. |
Taubman et al., “Embedded Block Coding in JPEG2000,” Feb. 23, 2001, pp. 1-8 of 36. |
Kessler, Gary C., An Overview of TCP/IP Protocols and the Internet; Jan. 16, 2007, retrieved from the Internet on Jun. 12, 2008 at http://www.garykessler.net/library/tcpip.html; originally submitted to the InterNIC and posted on their Gopher site on Aug. 5, 1994. |
Roe, Kevin, “Third-Party Observation Under EPC Article 115 on the Patentability of an Invention,” Dec. 21, 2007. |
Roe, Kevin, Third-Party Submission for Published Application Under CFR §1.99, Mar. 26, 2008. |
International Search Report for International Application No. PCT/US2007/063599, mailed Dec. 12, 2007. |
International Search Report for International Application No. PCT/US2007/076337, mailed Oct. 20, 2008. |
International Search Report and Written Opinion for International Application No. PCT/US2006/025911, mailed Jan. 3, 2007. |
European Patent Office “European Search Report,” mailed May 7, 2010 for Application No. 06786174.0. |
China State Intellectual Property Office “Office Action,” issued Apr. 13, 2010; Application No. 200580026825.X. |
Canadian Intellectual Property Office “Office Action” mailed Feb. 18, 2010 for Application No. 2569610. |
China State Intellectual Property Office “Office Action” issued Mar. 18, 2010, for Application No. 200680022520.6. |
China Patent Office “Office Action” mailed Feb. 23, 2011; Chinese Patent Appln. No. 200680022520.6. |
China Patent Office “Office Action” mailed Jan. 6, 2911; Chinese Patent Appln. No. 200810126554.0. |
Japan Patent Office “Notice of Grounds for Rejection (Office Action)” mailed Jun. 28, 2011; Japanese Patent Appln. No. P2008-507651. |
China State Intellectual Property Office “First Office Action,” issued Jul. 31, 2009, for Application No. 200580026825.X. |
USPTO, Non-Final Office Action, mailed Aug. 4, 2009; U.S. Appl. No. 11/734,277, filed Apr. 12, 2007. |
USPTO, Non-Final Office Action, mailed Jul. 31, 2009; U.S. Appl. No. 11/683,882, filed Mar. 8, 2007. |
USPTO, Non-Final Office Action, mailed Aug. 5, 2009; U.S. Appl. No. 11/147,663, filed Jun. 7, 2005. |
USPTO, Non-FinaI Office Action, maded Sep. 3, 2009; U.S. Appl. No. 11/620,711, filed Jan. 7, 2007. |
European Patent Office, International Searching Authority, “International Search Report,” for International Application No. PCT/US2009/049006, mailed Sep. 11, 2009. |
Australian Government “Office Action,” Australian Patent Application No. 2006240518, mailed Nov. 12, 2009. |
USPTO, Final Office Action, mailed Nov. 6, 2009; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Final Office Action mailed Nov. 12, 2009; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
Newton's Telecom Dictionary, 21st ed., Mar. 2005. |
European Patent Office, European Search Report, mailed Sep. 28, 2009 for European Application No. EP 06 78 6175. |
USPTO Final Office Action mailed Jan. 25, 2010; U.S. Appl. No. 11/734,277, filed Apr. 12, 2007. |
USPTO, Non-Final Office Action mailed Nov. 23, 2009; U.S. Appl. No. 11/683,862, filed Mar. 8, 2007. |
International Search Report for PCT/US2008/069914 mailed Dec. 19, 2008. |
Ditze M. et all “Resource Adaptation for Audio-Vsual Devices in the UPnP QoS Architecture,” Advanced Networking and Applications, 2006; AINA, 2006; 20% H International conference on Vienna, Austria Apr. 18-20, 2006. |
Joonbok Lee et al. “Compressed High Definition Television (HDTV) Over IPv6” Applicatons and the Intenet Workshops, 2006; Saint Workshops, 2006; International Symposium, Phoenix, AZ, USA, Jan. 23-27, 2006. |
Lowekamp, B. et al. “A Hierarchy of Network Performance Characteristics for Grid Applications and Services,” GGF Network Measurements Working Group, pp. 1-29, May 24, 2004. |
USPTO, Non-Final Office Action mailed Oct. 1, 2009; U.S. Appl. No. 11/778,287, filed Jul. 16, 2007. |
USPTO Final Office Action mailed Dec. 30, 2009; U.S. Appl. No. 11/147,664, filed Jun. 7, 2005. |
Meyer, Derrick “MyReplayTV™ Creates First-Ever Online Portal to Personal TII Service; Gives Viewers Whole New Way to Interact With Programming,” http://web.archive.org/web/20000815052751/http://www.myreplaytv.com/, Aug. 15, 2000. |
Sling Media “Sling Media Unveils Top-of-Line Slingbox PRO-HD” [online], Jan. 4, 2008, XP002560049; retrieved from the Internet: URL:www.slingmedia.com/get/pr-slingbox-pro-hd.html; retrieved on Oct. 12, 2009. |
PCT Partial International Search, PCT/US2009/054893, mailed Dec. 23, 2009. |
Srisuresh, P. et al. “Taditional IP Network Address Translator (Traditional NAT)” Network Working Goup, The Internet Society, Jan. 2001. |
China State Intellectual Property Office “First Office Action,” Jan. 8, 2010; Application No. 200810126554.0. |
Qiong, Liu et al. “Digital Rights Management for Content Distribution,” Proceedings of the Australasian Information Security Workshop Conference on ACSW Frontiers 2003, vol. 21, 2003, XP002571073, Adelaide, Australia, ISSN: 1445-1336, ISBN: 1-920682-00-7, sections 2 and 2.1.1. |
.Gurzhi, Alexander et al. “Systems and Methods for Emulation Network-Enabled Media Components,” U.S. Appl. No. 12/711,830, filed Feb. 24, 2010. |
Newton's Telcom Dictionary, 20th ed., Mar. 2004. |
“The Authoritative Dictionary of IEEE Standard Terms,” 7th ed. 2000. |
Lee, M. et al. “Video Frame Rate Control for Non-Guaranteed Network Services with Explicit Rate Feedback,” Globecom'00, 2000 IEEE Global Telecommunications conference, San Francisco, CA, Nov. 27-Dec. 1, 2000; [IEEE Global Telecommunications Conference], New York, NY; IEEE, US, vol. 1,Nov. 27, 2000, pp. 293-297, XP001195580; ISBN: 978-0-7803-6452-3, lines 15-20 of sec. II on p. 293, fig. 1. |
European Patent Office, International Searching Authority, “International Search Report and Written Opinion,” mailed Jun. 4, 2010 for International Application No. PCT/IN2009/000728, filed Dec. 18, 2009. |
USPTO Non-Final Office Action mailed Jun. 23, 2010; U.S. Appl. No. 11/933,969, filed Nov. 1, 2007. |
Korean Intellectual Property Office “Official Notice of Preliminary Rejection,” issued Jun. 18, 2010; Korean Patent Application No. 10-2008-7021254. |
USPTO Non-Final Office Action mailed Aug. 11, 2010; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO Final Office Action mailed Sep. 24, 2010; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
Canadian Patent Office “Office Action” mailed Mar. 17, 2011; Canadian Patent Appln. No. 2,606,235. |
USPTO “Final Office Action” mailed Mar. 10, 2011; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO Non-Final Office Action mailed Sep. 30, 2010; U.S. Appl. No. 12/758,196, filed Apr. 12, 2010. |
USPTO Non-Final Office Action mailed Oct. 4, 2010; U.S. Appl. No. 12/643,812, filed Dec. 21, 2009. |
Japan Patent Office “Notice of Grounds for Rejection (Office Action),” mailed May 25, 2010; Patent Application No. 2007-0268269. |
Japan Patent Office “Notice of Grounds for Rejection (Office Action),” mailed May 25, 2010; Patent Application No. 2007-527683. |
China State Intellectual Property Office “First Office Action of China State Intellectual Property Office,” issued Oct. 9, 2010; Application No. 200780030811.4. |
European Patent Office “EPO Communication” dated Nov. 29, 2010; Application No. 08 167 880.7-2202. |
Korean Intellectual Property Office “Official Notice of Preliminary Rejection,” issued Dec. 29, 2010; Patent Application No. 10-2009-7003607. |
China State Intellectual Property Office “Second Office Action of China State Intellectual Property Office,” issued Aug. 26, 2010; Application No. 200810126554.0. |
Sikora, Thomas “Trends and Perspectves in Image and Video Coding,” Proceedings of the IEEE, vol. 93 No. 1, pp. 6-17; Jan. 1, 2005. |
Vetro, Anthony et al. “Video Transcoding Architectures and Techniques: An Overview,” IEEE Signal Processing Magazine, vol. 20, No. 2, pp. 18-29, Mar. 1, 2003. |
De Lattre, Alexis et al., “VideoLAN Streaming Howto” Internet Citation, Jan. 22, 2005; http://www.videolan.org/doc/streaming-howto/en/. |
European Patent Office “Extednded European Search Report” dated Feb. 28, 2011; Appln. No. 06734554. |
European Patent Office, European Search Report for European Application No. EP 08 16 7880, mailed Mar. 4, 2009. |
Mythtv Wiki, “MythTV User Manual” [Oniine], Aug. 27, 2007, XP002515046; retrieved from the Internet: <URL: http://www.mythtv.org/wiki?title=User—Manual:Introduction&oldid=25549>. |
International Searching Authority, Written Opinion and International Search Report for International Application No. PCT/US2008/077733, mailed Mar. 18, 2009. |
International Searching Authority, Written Opinion and International Search Report for International Application No. PCT/US2008/087005, mailed Mar. 20, 2009. |
Watanabe Y. et al., “Multimedia Database System for TV Newscasts and Newspapers”; Lecture Notes in Computer Science, Springer Verlag, Berlin, Germany; vol. 1554, Nov. 1, 1998, pp. 208-220, XP002402824, ISSN: 0302-9743. |
Yasuhiko Watanabe et al., “Aligning Articles in TV Newscasts and Newspapers”; Proceedings of the International Conference on Computationallinguistics, XX, XX, Jan. 1, 1998, pp. 1381-1387, XP002402825. |
Sodergard C. et al., “Integrated Muitimedia Publishing: Combining TV and Newspaper Content on Personal Channels”; Computer Networks, Elsevier Science Publishers B.V., Amsterdam, Netherlands; vol. 31, No. 11-16, May 17, 1999, pp. 1111-1128, XP004304543, ISSN: 1389-1286. |
Ariki Y. et al., “Automatic Classification of TV News Articles Based on Telop Character Recognition”; Multimedia Computing and Systems, 1999; IEEE International Conference on Florence, Italy, Jun. 7-11, 1999, Los Alamitos, California, USA, IEEE Comput. Soc. US; vol. 2, Jun. 7, 1999, pp. 148-152, XP010519373, ISBN: 978-0-7695-0253-3; abstract, paragraph [03.1], paragraph [05.2], figures 1,2. |
USPTO, Non-Final Office Action mailed Dec. 17, 2004; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Final Office Action mailed Jul. 28, 2005; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Non-Final Office Action mailed Jan. 30, 2006; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Final Office Action mailed Aug. 10, 2006; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Non-Final Office Action mailed Jun. 19, 2007; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Non-Final Office Action mailed Apr. 16, 2008; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Final Office Action mailed Sep. 18, 2008; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Non-Final Office Action mailed Mar. 31, 2009; U.S. Appl. No. 09/809,868, filed Mar. 15, 2001. |
USPTO, Non-Final Office Action mailed May 1, 2008; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO, Final Office Action mailed Dec. 29, 2008; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO, Non-Final Office Action mailed Jun. 8, 2009; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO, Non-Final Office Action mailed Jun. 26, 2008; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
USPTO, Final Office Action mailed Oct. 21, 2008; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
USPTO, Non-Final Office Action mailed Mar. 25, 2009; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
USPTO, Non-Final Office Action mailed Aug. 7, 2008; U.S. Appl. No. 11/620,711, filed Jan. 7, 2007. |
USPTO, Final Office Action mailed Feb. 9, 2009; U.S. Appl. No. 11/620,711, filed Jan. 7, 2007. |
USPTO, Non-Final Office Action mailed Feb. 25, 2009; U.S. Appl. No. 11/683,862, filed Mar. 8, 2007. |
USPTO, Non-Final Office Action mailed Dec. 24, 2008; U.S. Appl. No. 11/147,985, filed Jun. 7, 2005. |
USPTO, Non-Final Office Action mailed Jun. 25, 2008; U.S. Appl. No, 11/428,254, filed Jun. 30, 2006. |
USPTO, Final Office Action mailed Feb. 6, 2009; U.S. Appl. No. 11/428,254, filed Jun. 30, 2006. |
USPTO, Non-Final Office Action mailed May 15, 2009; U.S. Appl. No. 11/147,664, filed Jun. 7, 2005. |
Sonic Blue “ReplayTV 5000 User's Guide,” 2002, entire document. |
Bluetooth-News; Main Future User Models Document Verification & Qualification: Bluetooth Technical Background, Apr. 21, 1999; pp. 1 of 7 and 2 of 7; http://www.bluetooth.com/v2/news/show.asp 1-2. |
Microsoft Corporation; Harman/Kardon “Master Your Universe” 1999. |
Matsushita Electric Corporation of America MicroCast : Wireless PC Multimedia Transceiver System, Nov. 1998. |
“Wireless Local Area Networks: Issues in Technology and Standards” Jan. 6, 1999. |
USPTO, Final Office Action mailed Jun. 25, 2009; U.S. Appl. No. 11/147,985, filed Jun. 7, 2005. |
International Search Report and Written Opinion for International Application No. PCT/US2006/025912, mailed Jul. 17, 2008. |
International Search Report for International Application No. PCT/US2008/059613, mailed Jul. 21, 2008. |
International Search Report and Written Opinion for International Application No. PCT/US2008/080910, mailed Feb. 16, 2009. |
Wikipedia “Slingbox” [Online], Oct. 21, 2007, XP002512399; retrieved from the Internet: <URL:http://en.wikipedia.org/w/index.php?title=Slingbox&oldid=166080570>; retrieved on Jan. 28, 2009. |
Wikipedia “LocationFree Player” [Online], Sep. 22, 2007, XP002512400; retrieved from the Internet: <URL: http://en.wikipedia.org/w/index.php?title=LocationFree—Player&oldid=159683564>; retrieved on Jan. 28, 2009. |
Capable Networks LLC “Keyspan Remote Control—Controlling Your Computer With a Remote” [Online], Feb. 21, 2006, XP002512495; retrieved from the Internet: <URL:http://www.slingcommunity.com/article/11791/Keyspan-Remote-Control---Controlling-Your-Computer-With-a-Remote/?highlight=remote+control>; retrieved on Jan. 28, 2009. |
Sling Media Inc. “Slingbox User Guide” [Online] 2006, XP002512553; retrieved from the Internet: <URL:http://www.slingmedia.hk/attach/en-US—Slingbox—User—Guide—v1.2.pdf>; retrieved on Jan. 29, 2009. |
Sony Corporation “LocationFree TV” [Online], 2004, SP002512410; retrieved from the Internet: <URL:http://www.docs.sony.com/release/LFX1—X5revision.pdf>; retrieved on Jan. 28, 2009 [note—document uploaded in two parts as file exceeds the 25MB size limit]. |
Sony Corporation “LocationFree Player Pak—LocationFree Base Station—LocationFree Player” [Online] 2005, XP002512401; retrieved from the Internet: <URL:http://www.docs.sony.com/release/LFPK1.pdf>; retrieved on Jan. 28, 2009. |
USPTO Final Office Action mailed Mar. 3, 2010; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO Final Office Action mailed Mar. 12, 2010; U.S. Appl. No. 11/620,711, filed Jan. 7, 2007. |
USPTO Non-Final Office Action mailed Mar. 19, 2010; U.S. Appl. No. 11/147,664, filed Jun. 7, 2005. |
USPTO Non-Final Office Action mailed Mar. 31, 2010; U.S. Appl. No. 11/620,707, filed Jan. 7, 2007. |
USPTO Non-final Office Action mailed Apr. 1, 2010; U.S. Appl. No. 12/237,103, filed Sep. 24, 2008. |
European Patent Office, International Authority, “International Search Report,” mailed Mar. 30, 2010; International Application PCT/US2009/068468 filed Dec. 27, 2009. |
USPTO “Non-Final Office Action” mailed Sep. 16, 2011; U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO “Non-Final Office Action” mailed Sep. 6, 2011; U.S. Appl. No. 12/347,465, fiied Dec. 31, 2008. |
USPTO “Notice of Allowance” mailed Sep. 22, 2011; U.S. Appl. No. 12/979,145, filed Dec. 27, 2010. |
USPTO “Notice of Allowance”, mailed Oct. 18, 2011; U.S. Appl. No. 12/339,878, filed Dec. 19, 2008. |
China State Intellectual Property Office, Decision of Rejection:, Issued Sep. 15, 2011; Chine Patent Appln. No. 200780030811.4. |
USPTO “Non-Final Office Action”, mailed Oct. 18, 2011; U.S. Appl. No. 12/538,659, filed Aug. 10, 2009. |
Canadian Intellectual Property Office, “Office Action” mailed Nov. 28, 2011; Canadian Appln. No. 2,606,235. |
USPTO, Non-Final Office Action, mailed Jan. 6, 2014 for U.S. Appl. No. 13/615,306. |
USPTO, Non-Final Office Action, mailed Jan. 6, 2014 for U.S. Appl. No. 12/827,964. |
Chinese Intellectual Property Office “Office Action” mailed Apr. 10, 2012 for Chinese Application 200810126554.0. |
Japan Patent Office “Office Action” mailed Feb. 28, 2012 for Japanese Application P2008-507651. |
Canadian Intellectual Property Office “Office Action” mailed on Feb. 16, 2012 for Canadian Application 2,660,350. |
Taiwan Intellectual Property Office “Office Action” mailed Feb. 23, 2012 for Taiwan Application 097137393. |
USPTO “Final Office Action” mailed Apr. 13, 2012; U.S. Appl. No. 12/538,659, filed Aug. 10, 2009. |
USPTO “Final Office Action” mailed Apr. 27, 2012; U.S. Appl. No. 12/347,465, filed Dec. 31, 2008. |
Japan Patent Office “Hearing” mailed Apr. 24, 2012 for Japanese Patent Appln. No. 2007-527683. |
Japan Patent Office “Hearing” mailed Apr. 24, 2012 for Japanese Patent Appln. No. 2007-268269. |
European Patent Office “Office Action” dated Aug. 7, 2012 for European Patent Appln. No. 06 734 554.6. |
China Patent Office “Office Action” issued Aug. 3, 2012 for Chinese Patent Appln. No. 200810161874.X. |
USPTO “Notice of Allowance” mailed Aug. 31, 2012 for U.S. Appl. No. 11/620,711, filed Jan. 7, 2007. |
M2 Presswire, Aug. 16, 2006, OTCPicks.com; Daily Market Movers Digest Midday Stock Alerts for Wednesday, Aug. 16th, AMRU, IPTM, BCLC, IHDR, EGIL © 1994-2006 M2 Communications Ltd. |
USPTO “Notice of Allowance” mailed Jun. 12, 2012 for U.S. Appl. No. 13/240,932, filed Sep. 22, 2011. |
USPTO “Final Office Action” mailed Jun. 6, 2012 for U.S. Appl. No. 12/827,964, filed Jun. 30, 2010. |
USPTO “Notice of Allowance” mailed Jul. 6, 2012 for U.S. Appl. No. 11/111,265, filed Apr. 21, 2005. |
USPTO “Non-Final Office Action” mailed Oct. 23, 2012 for U.S. Appl. No. 13/194,783, filed Jul. 29, 2011. |
USPTO “Notice of Allowance” mailed Sep. 25, 2012 for U.S. Appl. No. 13/240,932, filed Sep. 22, 2011. |
USPTO “Non-Final Office Action” maiied Jan. 10, 2012; U.S. Appl. No. 12/827,964, filed Jun. 30, 2010. |
Chinese Office Action, dated Dec. 31, 2011, for Chinese Patent Application No. 200810161874.X. |
European Office Action, dated Nov. 30, 2011, for European Patent Application No. 06 734 554.6-2223. |
USPTO “Non-Final Office Action” mailed Feb. 24, 2012; U.S. Appl. No. 13/240,932, filed Sep. 22, 2011. |
China Intellectual Property Office, “Second Office Action” mailed Feb. 15, 2012; Appln. No. CN200680022520.6. |
China State Intellectual Property Office, “Reexamination Office Action” mailed Mar. 5, 2012; Chinese Appln. No. 200580026825.X. |
Li Zhuo et al: “Adaptive forward error correction for streaming stored MPEG-4 FGS video over wireless channel”, IEEE 5th Workshop on Signal Processing Advances in Wireless Communications, 2004, Jan. 1, 2004, pp. 26-30, XP055091312, DOI: 10.1109/SPAWC.2004.1439196 ISBN: 978-0-78-038337-1. |
European Patent Office, Summons to attend oral proceedings pursuant to Rule 115(1) EPC, mailed Jan. 15, 2014 for European Patent Application No. 06734554.6. |
Japan Patent Office, Notice of Ground(s) for Rejection, mailed Feb. 18, 2014 for Japanese Patent Application No. P2008-507651. |
Office of the Controller-General of Patents, Designs and Trademarks, First Examination Report, mailed Mar. 21, 2014 for India Patent Application No. 3740/KOLNP/2006. |
USPTO, Non-Final Office Action, mailed Mar. 27, 2014 for U.S. Appl. No. 13/073,469. |
Japan Patent Office, Hearing, dated Aug. 13, 2013 for Japanese Patent Application No. P2008-507651. |
USPTO, Final Office Action, mailed Jun. 18, 2014 for U.S. Appl. No. 13/609,033. |
USPTO, Non-Final Office Action, mailed Jul. 16, 2014 for U.S. Appl. No. 13/194,834. |
China State Intellectual Property Office, Office Action, dated Sep. 29, 2013 for Chinese Patent Application No. 200810161874.X. |
China State Intellectual Property Office “Third Office Action” issued Dec. 5, 2012 for Chinese Patent Appln. No. 200680022520.6. |
European Patent Office “Extgended Search Report” dated Nov. 27, 2011 for European Patent Appln. No. 12187858.1. |
USPTO “Non-Final Office Action” mailed Dec. 21, 2012 for U.S. Appl. No. 12/538,659, filed Aug. 10, 2009. |
Japanese Patent Office “Decision of Rejection (Final Rejection)” dated Jan. 29, 2013 for Japanese Patent Appln. No. 2008-507651. |
China State Intellectual Property Office “Fourth Office Action” dated Mar. 5, 2013 for Chinese Patent Appln. No. 200810161874.X. |
USPTO “Non-Final Office Action” dated Mar. 5, 2013 for U.S. Appl. No. 13/107,341. |
European Patent Office Supplementary Search Report dated Mar. 6, 2013 for European Patent Appln. No. 05758509.3. |
Intellectual Property Office “Office Action” dated Feb. 25, 2013 for Taiwan Patent Appln. No. 098146025. |
European Patent Office, Extended Search Report mailed May 24, 2013 for European Patent Application No. 07814265.0. |
USPTO, Final Office Action mailed May 8, 2013 fo U.S. Appl. No. 13/194,783. |
USPTO, Office Action mailed May 8, 2013 for U.S. Appl. No. 13/194,834. |
USPTO, Office Action mailed Jun. 19, 2013 for U.S. Appl. No. 12/827,964. |
USPTO, Office Action mailed Jun. 20, 2013 for U.S. Appl. No. 13/615,306. |
USPTO, Notice of Allowance mailed May 10, 2013 for U.S. Appl. No. 12/538,659. |
China State Intellectual Property Office, Office Action, dated Jul. 1, 2013 for China Patent Appln. No. 200680022520.6. |
Eurpoean Patent Office, Examination Report, dated Sep. 6, 2013 for European Patent Application No. 12 187 858.1. |
European Patent Office, Examination Report, dated Jul. 25, 2013 for European Patent Application No. 05 758 509.3. |
China State Intellectual Property Office, First Office Action, dated Aug. 2, 2013 for Chinese Patent Application No. 200980151308.3. |
USPTO, Non-Final Office Action, dated Aug. 7, 2013 for U.S. Appl. No. 13/609,033. |
USPTO, Final Office Action, dated Sep. 19, 2013 for U.S. Appl. No. 13/107,341. |
Canadian Intellectual Property Office, Office Action, dated Sep. 3, 2013 for Canadian Patent Application No. 2,660,350. |
Office of the Controller-General of Patents, Designs and Trademarks, First Examination Report for Indian Patent Application No. 4190/KOLNP/2007 mailed Feb. 27, 2015. |
USPTO, Final Office Action, mailed Sep. 4, 2014 for U.S. Appl. No. 12/827,964. |
European Patent Office, Communication under Rule 71(3) EPC for European Patent Application No. 06 734 554.6 mailed Feb. 13, 2015. |
European Patent Office, Communication pursuant to Article 94(3) EPC for European Patent Application No. 05 758 509.3 mailed Mar. 3, 2015. |
USPTO, Office Action for U.S. Appl. No. 14/449,896 mailed Feb. 18, 2015. |
USPTO, Final Office Action for U.S. Appl. No. 13/194,834 mailed Mar. 3, 2015. |
USPTO, Utility U.S. Appl. No. 14/449,896, filed Aug. 1, 2014. |
Number | Date | Country | |
---|---|---|---|
20140115642 A1 | Apr 2014 | US |
Number | Date | Country | |
---|---|---|---|
60577833 | Jun 2004 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13079644 | Apr 2011 | US |
Child | 14144225 | US | |
Parent | 12643812 | Dec 2009 | US |
Child | 13079644 | US | |
Parent | 11147985 | Jun 2005 | US |
Child | 12643812 | US |