Audio, video, text, and images are often streamed to mobile devices. These devices may render the streamed content locally or on a remote display. The rendered content may be sent to multiple devices at various locations from a single content provider.
The same numbers are used throughout the disclosure and the figures to reference like components and features. Numbers in the 100 series refer to features originally found in
Streaming media includes various content forms that are delivered by a content provider, received by a device, and presented to a user. The device enables users to connect to a network, such as the Internet, and stream music, movies, television programming, images, and other media. In embodiments, streaming indicates that the content forms are continually delivered by the content provider, and the content forms are received by the device and presented to a user while being continually delivered by the content provider. While streaming continually delivers content to a device, streaming is distinguished from downloading where the data is transferred and stored for later use. As used herein, “stored for later” use may refer to storage other than the temporary storage that may occur with streaming, such as buffering, placing streaming data in a queue, and the like.
In embodiments, streaming video data includes content such as images, text, video, audio, and animations associated with a streaming video session. Streaming video data may also include live streaming data, where live content is delivered a network. A live stream, as used herein, refers to content that is captured in real time and subsequently streamed to a device as the content is continually captured at a source. Regardless of the type of streaming data, the streaming data is typically rendered without regard to environmental conditions. In some cases, the rendered data is of poor quality in view of the environmental conditions where the device that renders the data is located. Embodiments described herein enable remote adaptation of streaming data based on the environment at the receiver. In embodiments, the present techniques enable remote chromatic color adaptation of streaming video based on environmental light at the receiver.
Some embodiments may be implemented in one or a combination of hardware, firmware, and software. Further, some embodiments may also be implemented as instructions stored on a machine-readable medium, which may be read and executed by a computing platform to perform the operations described herein. A machine-readable medium may include any mechanism for storing or transmitting information in a form readable by a machine, e.g., a computer. For example, a machine-readable medium may include read only memory (ROM); random access memory (RAM); magnetic disk storage media; optical storage media; flash memory devices; or electrical, optical, acoustical or other form of propagated signals, e.g., carrier waves, infrared signals, digital signals, or the interfaces that transmit and/or receive signals, among others.
An embodiment is an implementation or example. Reference in the specification to “an embodiment,” “one embodiment,” “some embodiments,” “various embodiments,” or “other embodiments” means that a particular feature, structure, or characteristic described in connection with the embodiments is included in at least some embodiments, but not necessarily all embodiments, of the present techniques. The various appearances of “an embodiment,” “one embodiment,” or “some embodiments” are not necessarily all referring to the same embodiments. Elements or aspects from an embodiment can be combined with elements or aspects of another embodiment.
Not all components, features, structures, characteristics, etc. described and illustrated herein need be included in a particular embodiment or embodiments. If the specification states a component, feature, structure, or characteristic “may”, “might”, “can” or “could” be included, for example, that particular component, feature, structure, or characteristic is not required to be included. If the specification or claim refers to “a” or “an” element, that does not mean there is only one of the element. If the specification or claims refer to “an additional” element, that does not preclude there being more than one of the additional element.
It is to be noted that, although some embodiments have been described in reference to particular implementations, other implementations are possible according to some embodiments. Additionally, the arrangement and/or order of circuit elements or other features illustrated in the drawings and/or described herein need not be arranged in the particular way illustrated and described. Many other arrangements are possible according to some embodiments.
In each system shown in a figure, the elements in some cases may each have a same reference number or a different reference number to suggest that the elements represented could be different and/or similar. However, an element may be flexible enough to have different implementations and work with some or all of the systems shown or described herein. The various elements shown in the figures may be the same or different. Which one is referred to as a first element and which is called a second element is arbitrary.
The electronic device 100 also includes a graphics processing unit (GPU) 108. As shown, the CPU 102 can be coupled through the bus 106 to the GPU 108. The GPU 108 can be configured to perform any number of graphics operations within the electronic device 100. For example, the GPU 108 can be configured to render or manipulate graphics images, graphics frames, videos, streaming data, or the like, to be rendered or displayed to a user of the electronic device 100. In some embodiments, the GPU 108 includes a number of graphics engines, wherein each graphics engine is configured to perform specific graphics tasks, or to execute specific types of workloads.
The CPU 102 can be linked through the bus 106 to a display interface 110 configured to connect the electronic device 100 to one or more display devices 112A. The display devices 112 can include a display screen that is a built-in component of the electronic device 100. In embodiments, the display interface 110 is coupled with the display devices 112B via any networking technology such as cellular hardware 124, Wifi hardware 126, or Bluetooth Interface 128 across the network 132. The display devices 112B can also include a computer monitor, television, or projector, among others, that is externally connected to the electronic device 100.
The CPU 102 can also be connected through the bus 106 to an input/output (I/O) device interface 114 configured to connect the electronic device 100 to one or more I/O devices 116A. The I/O devices 116A can include, for example, a keyboard and a pointing device, wherein the pointing device can include a touchpad or a touchscreen, among others. The I/O devices 116A can be built-in components of the electronic device 100, or can be devices that are externally connected to the electronic device 100. Accordingly, in embodiments, the I/O device interface 114 is coupled with the I/O devices 116B via any networking technology such as cellular hardware 124, Wifi hardware 126, or Bluetooth Interface 128 across the network 132. The I/O devices 116B can also include any I/O device that is externally connected to the electronic device 100.
The electronic device 100 also includes an environmental condition capture unit 118. The environmental condition capture unit 118 is to capture data describing the environmental conditions surrounding the electronic device 100. The environmental capture unit 118 may include, for example, a plurality of sensors that are used to obtain environmental conditions. The sensors may include, a light sensor, a temperature sensor, a humidity sensor, a motion sensor, and the like.
In addition to sensors, an image capture device 120 may be used to obtain environmental information. The image capture device may be a camera or an image sensor. Images captured by the image capture device 120 can be analyzed to determine environmental conditions, such as lighting and color temperatures of the surrounding space. A microphone array 122 may be used to capture audio environmental information.
The storage device 124 is a physical memory such as a hard drive, an optical drive, a flash drive, an array of drives, or any combinations thereof. The storage device 124 can store user data, such as audio files, video files, audio/video files, and picture files, among others. The storage device 124 can also store programming code such as device drivers, software applications, operating systems, and the like. The programming code stored to the storage device 124 may be executed by the CPU 102, GPU 108, or any other processors that may be included in the electronic device 100.
The CPU 102 may be linked through the bus 106 to cellular hardware 126. The cellular hardware 126 may be any cellular technology, for example, the 4G standard (International Mobile Telecommunications-Advanced (IMT-Advanced) Standard promulgated by the International Telecommunications Union-Radio communication Sector (ITU-R)). In this manner, the electronic device 100 may access any network 132 without being tethered or paired to another device, where the cellular hardware 126 enables access to the network 132.
The CPU 102 may also be linked through the bus 106 to WiFi hardware 128. The WiFi hardware 128 is hardware according to WiFi standards (standards promulgated as Institute of Electrical and Electronics Engineers' (IEEE) 802.11 standards). The WiFi hardware 128 enables the electronic device 100 to connect to the Internet using the Transmission Control Protocol and the Internet Protocol (TCP/IP). Accordingly, the electronic device 100 can enable end-to-end connectivity with the Internet by addressing, routing, transmitting, and receiving data according to the TCP/IP protocol without the use of another device. Additionally, a Bluetooth Interface 130 may be coupled to the CPU 102 through the bus 106. The Bluetooth Interface 130 is an interface according to Bluetooth networks (based on the Bluetooth standard promulgated by the Bluetooth Special Interest Group). The Bluetooth Interface 130 enables the electronic device 100 to be paired with other Bluetooth enabled devices through a personal area network (PAN). Accordingly, the network 132 may be a PAN. Examples of Bluetooth enabled devices include a laptop computer, desktop computer, ultrabook, tablet computer, mobile device, or server, among others.
The network 132 may be used to obtain streaming data from a content provider 134. The content provider 134 may be any source that provides streaming data to the electronic device 100. The content provider 134 may be cloud based, and may include a server. Users of a mobile device, such as the electronic device 100, stream content to their respective mobile device that originates at the content provider 134. Frequently, users watch streaming data in extreme environments, whether the environments is indoors or outdoors. To facilitate rendering streaming data in any environment, the content provider 134 includes a compensation unit 136 that is to adjust various parameters of the streaming data based on the environmental information captured by the environmental condition capture mechanism of an electronic device 100 that is to receive streaming data. In this manner, streaming video content providers such as the content provider 134 (either video-on-demand (VOD) or user generated-content (UGC) or Multimedia Broadcast Multicast Service (eMBMS) content providers) do not serve the same video stream to all end-users, and the particular video stream is dependent upon the environment in which the end users are located.
In this manner, the chromatic content of served video renders appropriately to the eyes of an end-user watching streaming video outdoors under a bright sun, or to an end-user watching streaming video indoors under a blue-ish LED light. Without this chromatic compensation, overall colors rendered on a display in extreme light environments may look washed out, and end-user experience may be far from optimal. Additionally, the user may tend to not view streaming content without chromatic compensation from the content provider since the streaming content renders poorly in view of the environment of the user.
The block diagram of
In embodiments, light environment information captured at device/receiver location is used to control color chromaticity of video content streamed by a remote VOD/UGC/eMBMS server. The mobile device/receiver may capture local environment light, at location where video streaming services will be watched. The mobile device/receiver then shares captured local environment light with a remote VOD/UGC/eMBMS service provider. The network hosted VOD/UGC/eMBMS service provider then uses the environment light information sent by the device and improves end-user video playback experience on its local device/receiver by sending chromatic color compensated VOD/UGC/eMBMS streaming videos to the local device/receiver. The information about end-user light environment may be used to either generate on the fly chromatic compensated streaming videos (in case of retransmission of live events over VOD or eMBMS), or to select the best streaming video (or video segment) out of a library of chromatic compensated streaming videos stored at the content provider.
In particular, at block 212, input data is input to the color compensation unit. The input data may be an image stream captured by a camera in case of a live event, or resulting from the playback of a movie. At block 202, the RGB input data may be chroma subsampled by implementing less resolution for the chroma information than for luma information. This subsampling may be performed via a YUV family of color spaces, where the Y component determines the brightness of the color, referred to as luminance or luma. The U and V components determine the color itself, which is the chroma. For example, U may represent the blue-difference chroma component, and V may represent the red-difference chroma component. In embodiments, the chroma subsampling is a YUV4:2:0 subsampling ratio. The YUV family of color spaces describes how RGB information is encoded and decoded, and the sampling ratio describes how the data will be decoded by implementing less resolution for chroma information than for luma information, taking advantage of the human visual system's lower acuity for color differences than for luminance. The subsampled data may be placed in an RGB pixel buffer. A matrix 214 illustrates the pixel data after subsampling at the color space conversion 202.
At block 206, a chromatic color compensation is performed based on an external light environment parameter. The external light environment parameter may be obtained from the environmental information 204. In particular, the environmental information 204 may be used in conjunction with a chromatic color compensation look up table (LUT) 216. The chromatic color compensation LUT 216 may be used to convert a range of input data into another range of parameters to be applied to streaming data. The LUT may be configured as indicated at block 218 by determining the best compensation parameters based on the environmental information. In embodiments, the compensation parameters are to cancel any undesirable rendering parameters in favor of parameters that will render properly such that a user in a particular environment can view the video. The chromatic color compensation LUT 216 can then store those parameters to be applied when the environmental conditions match those in the chromatic color compensation LUT 216. In embodiments, the parameters are stored as a color compensation matrix.
For example, for each RGBin pixel in the image, a new RGBout pixel value is computed, using a chromatic color compensation matrix 220. The resulting compensated data is a function of the RGBin matrix 214, such that the RGBout matrix 222 is RGBout=(Rout, Gout, Bout) with Rout=f(Rin, Gin, Bin), Gout=f(Rin, Gin, Bin), Bout=f(Rin, Gin, Bin), where f is the compensation matrix 220 capturing RGB light environment information, and RGBin is the pixel data where the compensation will be applied. In embodiments, the color compensation matrix 220 contains a light environment value measured at the receiver by an RGB sensor or a camera.
In the color space conversion 208, the RGBout matrix 222 may be encoded using a chroma subsampling ratio of YUV 4:2:0. As discussed above, the chroma sampling ratio describes how the data will be decoded by implementing less resolution for chroma information than for luma information, taking advantage of the human visual system's lower acuity for color differences than for luminance. In embodiments, encoding the adapted video according to a chroma subsampling ratio enables a video stream that can be compressed in the same manner as the original, non-adapted video. Moreover, the present techniques do not increase the size of the adapted video, as typical encoding includes chroma subsampling that implements less resolution for chroma information than for luma information. In adapting the streaming data, the chroma information may be adapted according to the present techniques. The new RGBout pixel buffer that has been chroma subsampled at a ratio of YUV 4:2:0 is sent to the video encoder 210 for further processing and encoding for transmission. In embodiments, a video encoder 210 is a “genuine” video encoder that takes as an input a YUV420 (typically) video stream, and outputs a compressed bitstream.
The color compensation as described above can be applied to both live streaming data as well as to pre-encoded streaming data. In embodiments, to improve overall end user experience when streaming video to a device in varying lighting environments, the cloud server can use the color chromatic processing chain in at least two different ways for use with each of live streaming data and pre-encoded streaming data. In particular, the cloud server may either select a video stream with colors previously adapted for transmission or the cloud server can convert the colors in a live video stream for transmission.
The transmission of an already encoded video stream with colors previously adapted to the chromatic condition at the receiver or a color converted live stream results in an overall reduction in the volume of information that is transmitted to the receiver when compared to sending a non-color compensated stream to the receiver. The reduction in the volume of information is a result of sending only the relevant chromatic content to the receiver in the color compensated video. Relevant chromatic content is chromatic data that has been color compensated, and may also include the associate luma information. Chromatic content that is visible by an end-user eye under the current ambient light gets sent to the device instead of an entire range of chromatic content that is traditionally sent to the receiver. As such, the amount of information to encode by the video encoder at server side is reduced when compared to the amount of information that would have to be encoded without chromatic pre-filtering as described herein.
For example, if the receiver is in a red-ish environment, the server would broadcast a stream with reduced red content. Artificially boosted green and blue colors may be used to compensate for the red-ish environment at the receiver. This change in chroma values can result in less data overall that is transmitted to the receiver. In another example, a live stream is color converted prior to being encoded in order to reduce amount of red content by artificially boosting all non-red colors as in the prior example. The blue and green boosted YUV 4:2:0 color stream is sent to the encoder that will encode it and stream it to the receiver. In this manner, the receiver will receive a stream with less red and more blue and green content. When displayed in a red-ish environment, the viewing experience will be enhanced as the overall video content will not look washed out, as would be the case if the blue and green content had not been boosted.
At the server 302, a live VOD/eMBMS raw video stream is captured at block 312. Chromatic color compensation is applied to the raw image stream at block 314. In particular a correction faction is applied to the image stream as indicated at block 308B. The correction factor may be based on the environmental information captured by the receiver. In embodiments, the color compensation is performed as is described with respect to
At the server 402, a VOD/UGC raw image stream is captured and/or stored at block 412. Chromatic color compensation is applied to the raw image stream at block 414. A set of typical correction factors is applied to the image stream as indicated at block 408B. The correction factor is typical in that it is configured to apply to a representative range of the lighting environments that the receiver may be located. In embodiments, the color compensation is performed as is described with respect to
At block 504, the environmental information is sent to the content provider. In embodiments, the environmental information is sent to the content provider on a periodic basis. The environmental information may be sent to the network hosted server or the remote streaming video provider (VOD/UGC provider) using whatever IP connection is available (for example, WLAN, or cellular). At block 506, the content is adapted based on the environmental information. In particular, upon reception of environment information captured at device level, or upon reception of environment information captured at device level that shows significant degradation when compared to a previously received and stored message, the network hosted server will take the decision to adapt the chromatic content of its streaming video to be sent. In this manner, the transmitted content is suitable for the lighting environment at receiver location. For example, chromatic color content may be made to look more yellow (R+G) when receiver light environment is blue-ish (B), as is the case under some LED lighting. Since the content has been adapted for the environmental conditions, a power savings at the mobile device can occur since there is not additional processing performed at the mobile device. Moreover, adaptation can be done either on-the-fly, or based on library or pre-compensated video segments.
Adaptation of the video may occur at a network or cloud hosted server as described above. In embodiments, the video may also be adapted at the device according to the techniques described herein. In such a scenario, the environmental information is captured by the device, and applied to a streaming data received at the device. The device may include a buffer to store streaming data prior to adaptation at the device. In embodiments, the buffer at the device in a first in, first out buffer. Additionally, in embodiments, a cloud streaming server receives both the video streaming request and the environment information and uses this information to either encode video content adapted for the receiver on the fly, or fetches an appropriate chromatic color compensated video (or video segment) out of a repository, and then sends the video to the receiver using typical streaming protocols, such as Real-time Transport Protocol (RTP) over IP.
Thus, the present techniques use end-user reported information to control chromatic color adaptation of streaming video in order to optimize overall end user experience. In examples, the color content is adapted so a streaming video does not look ‘washed out’ when video playback is rendered on a device in a blue-ish light environment which is typical of some LED lighting. In such an example, the aim may be to boost Red and Green content over Blue to implement the compensation described herein. In another example, the color content is adapted so a streaming video does not look ‘washed out’ when video playback is rendered on a device in a red-ish light environment, which is typical of a sunset environment. In this example, the aim may be to boost Blue and Green content over Red content to implement the compensation described herein.
In a use case of advertisers or third party content providers, ensuring that the rendered video is optimally viewable in any environment enables a higher number of views among a group of electronic device users. In embodiments, cloud served video is compensated before it is streamed to the receiver, using light environment information captured at receiver location. In embodiments, the chromatic color compensation can be done on-the-fly for live video content (for example, a live retransmission of sport event over VOD or eMBMS) or offline for UGC content or movie VOD.
The media 600 may include modules 606-610 configured to perform the techniques described herein. For example, an information capture module 606 may be configured to capture environmental conditions at an electronic device. The information capture module may also be used to configure a look up table that includes a color compensation matrix. In embodiments, an end user sends a request for video streaming delivery to a video streaming server (VOD/UGC/eMBMS). In embodiments, the streaming server is a cloud-based server. A request for video streaming may be sent using transmission protocols such as Hypertext Transfer Protocol (HTTP) over Internet Protocol (IP). This request may result in the capture of environmental information. Accordingly, in embodiments, an end user device may send, information about the light environment where receiver is located on a periodic basis. The information about the light environment where receiver is located is typically also sent using HTTP over IP, and can be captured either by a camera sensor or by a simple RGB sensor (providing the environmental light level for each RGB component). In embodiments, using an RGB sensor has a significant advantage when compared to camera sensing, as it is also very low-power consuming and is a low-cost hardware component.
An adaptation module 608 may be configured to apply the environmental information to streaming data. In embodiments, a cloud streaming server receives both the video streaming request and the light environment information and uses this information to either encode video content adapted for the receiver on the fly, or fetches an appropriate chromatic color compensated video (or video segment) out of a repository, and then sends the video to the receiver using typical streaming protocols, such as Real-time Transport Protocol (RTP) over IP.
A streaming module 610 may be configured stream the adapted data. In embodiments, on a periodic basis, the cloud streaming server receives new light environment measures done at receiver location, and either adjusts its encoding parameters, or sends more appropriate video (or video segment) to the receiver. In some embodiments, the modules 606-610 may be modules of computer code configured to direct the operations of the processor 602.
The block diagram of
By hosting the chromatic video processing in the cloud or on the server side rather than on the device, several operator/video service provider benefits may be enabled. First, the technology reach may be increased because it does not require dedicated processing power/HW accelerator at device level. Additionally, chromatic color compensation can be done locally on the rendering device is the device includes the hardware or software used for color compensation. Color compensation by a content provider can allocate enough CPU processing power to run the processing in software.
In embodiments, local chromatic color compensation may be dependent on the end device being powerful enough, or the end device encompassing the right hardware. Running the chromatic compensation in the cloud/on a server means the solution can be deployed to a broader base of devices, even those who do neither have enough CPU horsepower nor the dedicated hardware acceleration to run the chromatic compensation locally.
Moreover, the present techniques increases the technology reach because the bandwidth required for server to device streaming is decreased. In embodiments, the RGB information about ambient light at the device level, sensed by the RGB sensor local to the device and periodically uploaded to the server, can be used by the server to pre-process video prior to the transmission, so only chromatic content that is visible by an end-user eye under the current ambient light gets sent to the device. The color chromatic processing done on the server can yield more optimized video streams as a result of a more efficient bit allocation during the encoding process. Moreover, color chromatic processing can yield streams encoded at a lower bitrate. Additionally, as limited chromatic video content is visible by the end user eye, there is no need to encode all chromatic content of source video at the same quality level; server-hosted video encoder can allocate less bits to encode chromatic content that is hardly visible by the end-user, and thus server-hosted encoder can yield smaller encoded video files or smaller encoded video segment when compared to video streams encoded without chromatic pre-processing.
Example 1 is an apparatus. The apparatus includes an environmental conditions capture mechanism to obtain environmental information from a receiver; a controller to control streaming data based on the environmental information, wherein the receiver is to receive streaming data that is adapted based on the environmental information and a volume of the streaming data is reduced as a result of adapting the streaming data.
Example 2 includes the apparatus of example 1, including or excluding optional features. In this example, the environmental information is ambient lighting information.
Example 3 includes the apparatus of any one of examples 1 to 2, including or excluding optional features. In this example, the environmental conditions capture mechanism is a plurality of sensors.
Example 4 includes the apparatus of any one of examples 1 to 3, including or excluding optional features. In this example, the environmental conditions capture mechanism is a camera, an RGB sensor, or any combination thereof.
Example 5 includes the apparatus of any one of examples 1 to 4, including or excluding optional features. In this example, control of the streaming data comprises manipulating the data based on environmental conditions.
Example 6 includes the apparatus of any one of examples 1 to 5, including or excluding optional features. In this example, reducing the volume of the streaming data is a result of sending only a relevant chromatic content to the receiver.
Example 7 includes the apparatus of any one of examples 1 to 6, including or excluding optional features. In this example, adapting the streaming data comprises compensating colors of the streaming data based on the environmental conditions.
Example 8 includes the apparatus of any one of examples 1 to 7, including or excluding optional features. In this example, adapting the streaming data comprises applying a color compensation matrix to streaming data in a pixel buffer.
Example 9 includes the apparatus of any one of examples 1 to 8, including or excluding optional features. In this example, the adapted streaming data is stored in a repository for later transmission to the receiver.
Example 10 includes the apparatus of any one of examples 1 to 9, including or excluding optional features. In this example, the streaming data is captured on the fly.
Example 11 includes the apparatus of any one of examples 1 to 10, including or excluding optional features. In this example, the streaming data is obtained from a data repository.
Example 12 includes the apparatus of any one of examples 1 to 11, including or excluding optional features. In this example, environmental information is ambient noise.
Example 13 includes the apparatus of any one of examples 1 to 12, including or excluding optional features. In this example, adapting the streaming data comprises compensating audio volume the streaming data based on the environmental information.
Example 14 is a method. The method includes obtaining environmental information from a receiver; obtaining streaming data; modifying the streaming data based on the environmental information; and transmitting the modified streaming data to the receiver, wherein a volume of the streaming data is reduced as a result of modifying the streaming data.
Example 15 includes the method of example 14, including or excluding optional features. In this example, the environmental information is ambient lighting information.
Example 16 includes the method of any one of examples 14 to 15, including or excluding optional features. In this example, the environmental information is obtained via a plurality of sensors.
Example 17 includes the method of any one of examples 14 to 16, including or excluding optional features. In this example, the environmental information is obtained via a camera, an RGB sensor, or any combination thereof.
Example 18 includes the method of any one of examples 14 to 17, including or excluding optional features. In this example, the streaming data is captured at a content provider.
Example 19 includes the method of any one of examples 14 to 18, including or excluding optional features. In this example, reducing the volume of the streaming data is a result of sending only a relevant chromatic content to the receiver.
Example 20 includes the method of any one of examples 14 to 19, including or excluding optional features. In this example, modifying the streaming data comprises compensating colors of the streaming data based on the environmental conditions.
Example 21 includes the method of any one of examples 14 to 20, including or excluding optional features. In this example, modifying the streaming data comprises applying a color compensation matrix to streaming data in a pixel buffer.
Example 22 includes the method of any one of examples 14 to 21, including or excluding optional features. In this example, the modified streaming data is stored in a repository for subsequent transmission to a receiver.
Example 23 includes the method of any one of examples 14 to 22, including or excluding optional features. In this example, the streaming data is captured on the fly, modified and transmitted to the receiver on the fly.
Example 24 is a system. The system includes a display; a radio; a memory that is to store instructions and that is communicatively coupled to the display; and a processor communicatively coupled to the radio and the memory, wherein when the processor is to execute the instructions, the processor is to: capturing environmental information; transmitting the environmental information to a content provider; obtaining receiving streaming data modified based on the environmental information from the content provider, wherein a volume of the streaming data is reduced as a result of modifying the streaming data; and rendering the modified streaming data at the receiver.
Example 25 includes the system of example 24, including or excluding optional features. In this example, the reduced volume of the streaming data is a result of sending only a relevant chromatic content to the receiver.
Example 26 includes the system of any one of examples 24 to 25, including or excluding optional features. In this example, the environmental information is ambient lighting information.
Example 27 includes the system of any one of examples 24 to 26, including or excluding optional features. In this example, the environmental conditions are captured via a plurality of sensors.
Example 28 includes the system of any one of examples 24 to 27, including or excluding optional features. In this example, the environmental conditions are captured via is a camera, an RGB sensor, or any combination thereof.
Example 29 includes the system of any one of examples 24 to 28, including or excluding optional features. In this example, modified streaming data is modified by manipulating the streaming data based on environmental conditions.
Example 30 includes the system of any one of examples 24 to 29, including or excluding optional features. In this example, modified streaming data is modified by compensating colors of the streaming data based on the environmental conditions.
Example 31 includes the system of any one of examples 24 to 30, including or excluding optional features. In this example, modified streaming data is modified by applying a color compensation matrix to streaming data in a pixel buffer.
Example 32 includes the system of any one of examples 24 to 31, including or excluding optional features. In this example, the modified streaming data is stored in a repository for later transmission to the receiver.
Example 33 includes the system of any one of examples 24 to 32, including or excluding optional features. In this example, the streaming data is captured on the fly.
Example 34 includes the system of any one of examples 24 to 33, including or excluding optional features. In this example, the streaming data is obtained from a data repository.
Example 35 includes the system of any one of examples 24 to 34, including or excluding optional features. In this example, environmental information is ambient noise.
Example 36 includes the system of any one of examples 24 to 35, including or excluding optional features. In this example, adapting the streaming data comprises compensating audio volume the streaming data based on the environmental information.
Example 37 is a computer-readable medium. The computer-readable medium includes instructions that direct the processor to obtain environmental information from a receiver; obtain streaming data; modify the streaming data based on the environmental information; and transmit the modified streaming data to the receiver, wherein a volume of the streaming data is reduced as a result of modifying the streaming data.
Example 38 includes the computer-readable medium of example 37, including or excluding optional features. In this example, the environmental information is ambient lighting information.
Example 39 includes the computer-readable medium of any one of examples 37 to 38, including or excluding optional features. In this example, the environmental information is obtained via a plurality of sensors.
Example 40 includes the computer-readable medium of any one of examples 37 to 39, including or excluding optional features. In this example, the environmental information is obtained via a camera, an RGB sensor, or any combination thereof.
Example 41 includes the computer-readable medium of any one of examples 37 to 40, including or excluding optional features. In this example, the streaming data is obtained captured at a content provider.
Example 42 includes the computer-readable medium of any one of examples 37 to 41, including or excluding optional features. In this example, reducing the volume of the streaming data is a result of sending only a relevant chromatic content to the receiver.
Example 43 includes the computer-readable medium of any one of examples 37 to 42, including or excluding optional features. In this example, modifying the streaming data comprises compensating colors of the streaming data based on the environmental conditions.
Example 44 includes the computer-readable medium of any one of examples 37 to 43, including or excluding optional features. In this example, modifying the streaming data comprises applying a color compensation matrix to streaming data in a pixel buffer.
Example 45 includes the computer-readable medium of any one of examples 37 to 44, including or excluding optional features. In this example, the modified streaming data is stored in a repository for subsequent transmission to a receiver.
Example 46 includes the computer-readable medium of any one of examples 37 to 45, including or excluding optional features. In this example, the streaming data is captured on the fly, modified and transmitted to the receiver on the fly.
Example 47 is an apparatus. The apparatus includes instructions that direct the processor to an environmental conditions capture mechanism to obtain environmental information from a receiver; a means to control streaming data based on the environmental information to the receiver, wherein the receiver is to receive streaming data that is adapted based on the environmental information, and a volume of the streaming data is reduced as a result of adapting the streaming data.
Example 48 includes the apparatus of example 47, including or excluding optional features. In this example, the environmental information is ambient lighting information.
Example 49 includes the apparatus of any one of examples 47 to 48, including or excluding optional features. In this example, the environmental conditions capture mechanism is a plurality of sensors.
Example 50 includes the apparatus of any one of examples 47 to 49, including or excluding optional features. In this example, the environmental conditions capture mechanism is a camera, an RGB sensor, or any combination thereof.
Example 51 includes the apparatus of any one of examples 47 to 50, including or excluding optional features. In this example, the means to control the streaming data manipulates the data based on environmental conditions.
Example 52 includes the apparatus of any one of examples 47 to 51, including or excluding optional features. In this example, reducing the volume of the streaming data is a result of sending only a relevant chromatic content to the receiver.
Example 53 includes the apparatus of any one of examples 47 to 52, including or excluding optional features. In this example, adapting the streaming data comprises compensating colors of the streaming data based on the environmental conditions.
Example 54 includes the apparatus of any one of examples 47 to 53, including or excluding optional features. In this example, adapting the streaming data comprises applying a color compensation matrix to streaming data in a pixel buffer.
Example 55 includes the apparatus of any one of examples 47 to 54, including or excluding optional features. In this example, the adapted streaming data is stored in a repository for later transmission to the receiver.
Example 56 includes the apparatus of any one of examples 47 to 55, including or excluding optional features. In this example, the streaming data is captured on the fly.
Example 57 includes the apparatus of any one of examples 47 to 56, including or excluding optional features. In this example, the streaming data is obtained from a data repository.
Example 58 includes the apparatus of any one of examples 47 to 57, including or excluding optional features. In this example, environmental information is ambient noise.
Example 59 includes the apparatus of any one of examples 47 to 58, including or excluding optional features. In this example, adapting the streaming data comprises compensating audio volume the streaming data based on the environmental information.
It is to be understood that specifics in the aforementioned examples may be used anywhere in one or more embodiments. For instance, all optional features of the computing device described above may also be implemented with respect to either of the methods or the computer-readable medium described herein. Furthermore, although flow diagrams and/or state diagrams may have been used herein to describe embodiments, the techniques are not limited to those diagrams or to corresponding descriptions herein. For example, flow need not move through each illustrated box or state or in exactly the same order as illustrated and described herein.
The present techniques are not restricted to the particular details listed herein. Indeed, those skilled in the art having the benefit of this disclosure will appreciate that many other variations from the foregoing description and drawings may be made within the scope of the present techniques. Accordingly, it is the following claims including any amendments thereto that define the scope of the present techniques.