This disclosure relates to data bus architecture and, more specifically, data bus architecture in a multi-resolution display, such as a foveated display.
This section is intended to introduce the reader to various aspects of art that may be related to various aspects of the present techniques, which are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present disclosure. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
Numerous electronic devices-including televisions, portable phones, computers, wearable devices, vehicle dashboards, virtual-reality glasses, and more-display images on an electronic display. To display an image, an electronic display may control light emission of its display pixels based at least in part on corresponding image data. In some scenarios, such as in virtual reality, mixed reality, and/or augmented reality, an image frame of the image data to be displayed may be blended from multiple sources. For example, graphics may be rendered in high definition and blended with a camera feed. Furthermore, the image data may be formatted in multiple resolutions, such as for a foveated display that displays multiple different resolutions of an image at different locations on the electronic display depending on a viewer's gaze or focal point on the display.
Foveated display architectures may use multiplexers to select which image data is routed to which columns of pixels of the electronic display. Depending on the number of columns of the electronic display, however, the multiplexers may consume a significant portion of the die area while also consuming a significant amount of energy. Moreover, sending image data across a pixel data bus to the source latches consumes energy, particularly as the number of columns of pixels of the electronic display increases.
This disclosure relates to implementing a multiplexer-free architecture for data bus in foveated displays.
Electronic displays may be found in numerous electronic devices, from mobile phones to computers, televisions, automobile dashboards, and augmented reality or virtual reality glasses, to name just a few. Electronic displays with self-emissive display pixels produce their own light. Self-emissive display pixels may include any suitable light-emissive elements, including light-emitting diodes (LEDs) such as organic light-emitting diodes (OLEDs) or micro-light-emitting diodes (μLEDs). By causing different display pixels to emit different amounts of light, individual display pixels of an electronic display may collectively produce images.
Foveated electronic displays efficiently present image data based on characteristics of human vision-namely, that the human eye only sees in full resolution at a narrow point of focus and at much lower resolution in peripheral vision. Rather than generate full resolution image data for the entire electronic display, foveated image data may be generated that only includes the full resolution where the viewer is focused. In this way, the foveated image data may be generated that only includes the full resolution where the viewer is focused. In this way, the foveated image data that is displayed on a foveated electronic display may take up less memory and less bandwidth, but may look the same to the viewer, since the human eye cannot tell that the periphery has a lower resolution.
On a foveated electronic display, foveated image data may include a variety of groupings of pixels in different resolutions for different parts of the display. For example, a foveated region of the display where the viewer's eye is focused may display full resolution image data (e.g., one image data pixel for a 1×1 block of display pixels), whereas other peripheral parts of the electronic display may display lower resolution image data (e.g., one image data pixel for a 2×2 block of display pixels, one image data pixel for a 4×4 block of display pixels, and so on). Since the foveated region changes based on the movement of the viewer's eye, different areas of the electronic display present different resolutions at different times. As such, different parts of the foveated image data are routed to different pixels of the electronic display. One way of routing data is to use multiplexers to select which image data is routed to which source latches of columns of pixels of the electronic display. Depending on the number of columns of the electronic display, however, the multiplexers may consume a significant portion of the die area, leaves less area for the display active area meaning FOV (field of view) decreases while also consuming a significant amount of energy. Moreover, sending image data across a pixel data bus to the source latches consumes energy, particularly as the number of columns of pixels of the electronic display increases. Accordingly, in an embodiment it may be beneficial to reduce or eliminate multiplexers from a data bus and source latch architecture. Additionally, it may be beneficial to reduce the amount of energy consumed by the pixel data bus by gating slices of the pixel data bus to correspond to which source latches are being loaded.
Various aspects of this disclosure may be better understood upon reading the following detailed description and upon reference to the drawings in which:
When introducing elements of various embodiments of the present disclosure, the articles “a,” “an,” and “the” are intended to mean that there are one or more of the elements. The terms “including” and “having” are intended to be inclusive and mean that there may be additional elements other than the listed elements. Additionally, it should be understood that references to “some embodiments,” “embodiments,” “one embodiment,” or “an embodiment” of the present disclosure are not intended to be interpreted as excluding the existence of additional embodiments that also incorporate the recited features. Furthermore, the phrase A “based on” B is intended to mean that A is at least partially based on B. Moreover, the term “or” is intended to be inclusive (e.g., logical OR) and not exclusive (e.g., logical XOR). In other words, the phrase A “or” B is intended to mean A, B, or both A and B.
Foveated electronic displays efficiently present image data based on characteristics of human vision-namely, that the human eye only sees in full resolution at a narrow point of focus and at much lower resolution in peripheral vision. Rather than generate full resolution image data for the entire electronic display, foveated image data may be generated that only includes the full resolution where the viewer is focused. In this way, the foveated image data may be generated that only includes the full resolution where the viewer is focused. In this way, the foveated image data that is displayed on a foveated electronic display may take up less memory and less bandwidth, but may look the same to the viewer, since the human eye cannot tell that the periphery has a lower resolution.
On a foveated electronic display, foveated image data may include a variety of groupings of pixels in different resolutions for different parts of the display. For example, a foveated region of the display where the viewer's eye is focused may display full resolution image data (e.g., one image data pixel for a 1×1 block of display pixels), whereas other peripheral parts of the electronic display may display lower resolution image data (e.g., one image data pixel for a 2×2 block of display pixels, one image data pixel for a 4×4 block of display pixels, and so on). Since the foveated region changes based on the movement of the viewer's eye, different areas of the electronic display present different resolutions at different times. As such, different parts of the foveated image data are routed to different pixels of the electronic display. One way of routing data is to use multiplexers to select which image data is routed to which source latches of columns of pixels of the electronic display. Depending on the number of columns of the electronic display, however, the multiplexers may consume a significant portion of the die area while also consuming a significant amount of energy. Moreover, sending image data across a pixel data bus to the source latches consumes energy, particularly as the number of columns of pixels of the electronic display increases. The die area consumed by multiplexers will reduce the FOV (Field of View) can achieved with the same die size. FOV is one of the most critical part of the user experience especially for VR, MR, AR devices.
In an embodiment, instead of using multiplexers to route foveated image data in the electronic display, groups of source latches of the electronic display may be hardwired to respective wires of a pixel data bus. For example, a first group of four source latches may be connected to a first wire of the pixel data bus, a second group of four source latches may be connected to a second wire of the pixel data bus, and so on. The source latches of each group may be enabled or disabled individually (e.g., by control circuitry, by a state machine, and so on). Thus, image data provided on the first wire may be stored in selected source latches of the first group based on which of the source latches are enabled at a given time. For lower-resolution groupings (e.g., all four source latches receive the same single image data pixel), all of the source latches of a group may be enabled at once while one image data pixel is sent across a corresponding wire of the pixel data bus. For higher-resolution groupings, (e.g., source latch receives a different image data pixel), time multiplexing across the wire of the pixel data bus may be used. For example, at a first time, a first image data pixel may be sent across a wire of the pixel data bus while the first source latch of a group of source latches is enabled; at a second time, a second image data pixel may be sent across the wire of the pixel data bus while a second source latch of the group of source latches is enabled; and so forth. In this way, foveated image data may be effectively routed to the proper source latches without multiplexers.
In another embodiment, to reduce the amount of energy consumed by the pixel data bus, slices of the pixel data bus may be gated to correspond to which source latches are being loaded. For instance, a first set of source latches corresponding to a first slice of the pixel data bus may be loaded with data while downstream slices of the pixel data bus may be gated to save energy. A token signal passed along the pixel data bus may un-gate the slices over time as image data is passed along to further downstream slices. Thus, fewer slices of the pixel data bus may be active and consuming dynamic power at any point in time. To reduce the peak energy consumed by the pixel data bus, the pixel data bus may be divided into two parts that are loaded from opposite sides. Thus, the total number of gated slices may remain stable throughout the loading process.
With the foregoing in mind,
The electronic device 10 may include one or more electronic displays 12, input devices 14, an eye tracker 15, input/output (I/O) ports 16, a processor core complex 18 having one or more processors or processor cores, local memory 20, a main memory storage device 22, a network interface 24, a power source 26, and image processing circuitry 28. The various components described in
The processor core complex 18 is operably coupled with local memory 20 and the main memory storage device 22. Thus, the processor core complex 18 may execute instructions stored in local memory 20 or the main memory storage device 22 to perform operations, such as generating or transmitting image data to display on the electronic display 12. As such, the processor core complex 18 may include one or more general purpose microprocessors such as reduced instruction set computing (RISC) processors, one or more application specific integrated circuits (ASICs), one or more field programmable gate arrays (FPGAs), or any combination thereof.
In addition to program instructions, the local memory 20 or the main memory storage device 22 may store data to be processed by the processor core complex 18. Thus, the local memory 20 and/or the main memory storage device 22 may include one or more tangible, non-transitory, computer-readable media. For example, the local memory 20 may include random access memory (RAM) and the main memory storage device 22 may include read-only memory (ROM), rewritable non-volatile memory such as flash memory, hard drives, optical discs, or the like.
The network interface 24 may communicate data with another electronic device or a network. For example, the network interface 24 (e.g., a radio frequency system) may enable the electronic device 10 to communicatively couple to a personal area network (PAN), such as a Bluetooth network, a local area network (LAN), such as an 802.11x Wi-Fi network, or a wide area network (WAN), such as a 4G, Long-Term Evolution (LTE), or 5G cellular network.
The power source 26 may provide electrical power to operate the processor core complex 18 and/or other components in the electronic device 10. Thus, the power source 26 may include any suitable source of energy, such as a rechargeable lithium polymer (Li-poly) battery and/or an alternating current (AC) power converter.
The I/O ports 16 may enable the electronic device 10 to interface with various other electronic devices. For example, when a portable storage device is connected, the I/O port 16 may enable the processor core complex 18 to communicate data with the portable storage device. Moreover, the input devices 14 may enable a user to interact with the electronic device 10. For example, the input devices 14 may include buttons, keyboards, mice, trackpads, and the like. Additionally or alternatively, the electronic display 12 may include touch sensing components that enable user inputs to the electronic device 10 by detecting occurrence and/or position of an object touching its screen (e.g., surface of the electronic display 12).
Additionally, the electronic display 12 may be a display panel with one or more display pixels. For example, the electronic display 12 may include a self-emissive pixel array having an array of one or more of self-emissive pixels or liquid crystal pixels. The electronic display 12 may include any suitable circuitry (e.g., display driver circuitry) to drive the self-emissive pixels, including for example row driver and/or column drivers (e.g., display drivers). Each of the self-emissive pixels may include any suitable light emitting element, such as an LED (e.g., an OLED or a micro-LED). However, any other suitable type of pixel, including non-self-emissive pixels (e.g., liquid crystal as used in liquid crystal displays (LCDs), digital micromirror devices (DMD) used in DMD displays) may also be used. The electronic display 12 may control light emission from the display pixels to present visual representations of information, such as a graphical user interface (GUI) of an operating system, an application interface, a still image, or video content, by displaying frames of image data. To display images, the electronic display 12 may include display pixels implemented on the display panel. The display pixels may represent sub-pixels that each control a luminance value of one color component (e.g., red, green, or blue for an RGB pixel arrangement or red, green, blue, or white for an RGBW arrangement). As used herein, a display pixel may refer to a collection of sub-pixels (e.g., red, green, and blue subpixels) or may refer to a single sub-pixel.
The eye tracker 15 may measure positions and movement of one or both eyes of someone viewing the electronic display 12 of the electronic device 10. For instance, the eye tracker 15 may include a camera that can record the movement of a viewer's eyes as the viewer looks at the electronic display 12. However, several different practices may be employed to track a viewer's eye movements. For example, different types of infrared/near infrared eye tracking techniques such as bright-pupil tracking and dark-pupil tracking may be used. In both of these types of eye tracking, infrared or near infrared light is reflected off of one or both of the eyes of the viewer to create corneal reflections. A vector between the center of the pupil of the eye and the corneal reflections may be used to determine a point on the electronic display 12 at which the viewer is looking. The processor core complex 18 may use the gaze angle(s) of the eyes of the viewer when generating/processing image data for display on the electronic display 12.
As described above, the electronic display 12 may display an image by controlling the luminance output (e.g., light emission) of the sub-pixels based on corresponding image data. In some embodiments, pixel or image data may be generated by an image source, such as the processor core complex 18, a graphics processing unit (GPU), or an image sensor (e.g., camera). Additionally, in some embodiments, image data may be received from another electronic device 10, for example, via the network interface 24 and/or an I/O port 16. Moreover, in some embodiments, the electronic device 10 may include multiple electronic displays 12 and/or may perform image processing (e.g., via the image processing circuitry 28) for one or more external electronic displays 12, such as connected via the network interface 24 and/or the I/O ports 16.
The electronic device 10 may be any suitable electronic device. To help illustrate, one example of a suitable electronic device 10, specifically a handheld device 10A, is shown in
The handheld device 10A may include an enclosure 30 (e.g., housing) to, for example, protect interior components from physical damage and/or shield them from electromagnetic interference. The enclosure 30 may surround, at least partially, the electronic display 12. In the depicted embodiment, the electronic display 12 is displaying a graphical user interface (GUI) 32 having an array of icons 34. By way of example, when an icon 34 is selected either by an input device 14 or a touch-sensing component of the electronic display 12, an application program may launch.
Input devices 14 may be accessed through openings in the enclosure 30. Moreover, the input devices 14 may enable a user to interact with the handheld device 10A. For example, the input devices 14 may enable the user to activate or deactivate the handheld device 10A, navigate a user interface to a home screen, navigate a user interface to a user-configurable application screen, activate a voice-recognition feature, provide volume control, and/or toggle between vibrate and ring modes. Moreover, the I/O ports 16 may also open through the enclosure 30. Additionally, the electronic device may include one or more cameras 36 to capture pictures or video. In some embodiments, a camera 36 may be used in conjunction with a virtual reality or augmented reality visualization on the electronic display 12.
Another example of a suitable electronic device 10, specifically a tablet device 10B, is shown in
Turning to
As described above, the electronic display 12 may display images based at least in part on image data. Before being used to display a corresponding image on the electronic display 12, the image data may be processed, for example, via the image processing circuitry 28. In general, the image processing circuitry 28 may process the image data for display on one or more electronic displays 12. For example, the image processing circuitry 28 may include a display pipeline, memory-to-memory scaler and rotator (MSR) circuitry, warp compensation circuitry, or additional hardware or software means for processing image data. The image data may be processed by the image processing circuitry 28 to reduce or eliminate image artifacts, compensate for one or more different software or hardware related effects, and/or format the image data for display on one or more electronic displays 12. As should be appreciated, the present techniques may be implemented in standalone circuitry, software, and/or firmware, and may be considered a part of, separate from, and/or parallel with a display pipeline or MSR circuitry.
To help illustrate, a portion of the electronic device 10, including image processing circuitry 28, is shown in
The electronic device 10 may also include an image data source 38, a display panel 40, and/or a controller 42 in communication with the image processing circuitry 28. In some embodiments, the display panel 40 of the electronic display 12 may be a self-emissive display panel (e.g., OLED, LED, μLED, μOLED), transmissive display panel (e.g., a liquid crystal display (LCD)), a reflective technology display panel (e.g., DMD display), or any other suitable type of display panel 40. In some embodiments, the controller 42 may control operation of the image processing circuitry 28, the image data source 38, and/or the display panel 40. The controller 42 may include a controller processor 44 and/or controller memory 46. The controller processor 44 may be any suitable microprocessor, such as a general-purpose microprocessor such as a reduced instruction set computing (RISC) processor, an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), or any combination thereof. In some embodiments, the controller processor 44 may be included in the processor core complex 18, the image processing circuitry 28, a timing controller in the electronic display 12, a separate processing module, or any combination thereof and execute instructions stored in the controller memory 46. Additionally, in some embodiments, the controller memory 46 may be included in the local memory 20, the main memory storage device 22, a separate tangible, non-transitory, computer-readable medium, or any combination thereof.
The image processing circuitry 28 may receive source image data 48 corresponding to a desired image to be displayed on the electronic display 12 from the image data source 38. The source image data 48 may indicate target characteristics (e.g., pixel data) corresponding to the desired image using any suitable source format, such as an RGB format, an αRGB format, a YCbCr format, and/or the like. Moreover, the source image data may be fixed or floating point and be of any suitable bit-depth. Furthermore, the source image data 48 may reside in a linear color space, a gamma-corrected color space, or any other suitable color space. As used herein, pixels or pixel data may refer to a grouping of sub-pixels (e.g., individual color component pixels such as red, green, and blue) or the sub-pixels themselves.
As described above, the image processing circuitry 28 may operate to process source image data 48 received from the image data source 38. The image data source 38 may include captured images (e.g., from one or more cameras 36), images stored in memory, graphics generated by the processor core complex 18, or a combination thereof. Additionally, the image processing circuitry 28 may include one or more image data processing blocks 50 (e.g., circuitry, modules, or processing stages) such as an enhancement block 52. As should be appreciated, multiple other processing blocks 54 may also be incorporated into the image processing circuitry 28, such as a pixel contrast control (PCC) block, a burn-in compensation (BIC)/burn-in statistics (BIS) block, a color management block, a dither block, a blend block, a warp block, a scaling/rotation block, etc. before and/or after the enhancement block 52. The image data processing blocks 50 may receive and process source image data 48 and output display image data 56 in a format (e.g., digital format, image space, and/or resolution) interpretable by the display panel 40. For example, in the case of a foveated display (e.g., an electronic display 12 outputting multi-resolution image data), the image processing blocks 50 may output display image data 56 in the multi-resolution format.
Furthermore, the functions (e.g., operations) performed by the image processing circuitry 28 may be divided between various image data processing blocks 50, and, while the term “block” and/or “sub-block” is used herein, there may or may not be a logical or physical separation between the image data processing blocks 50 and/or sub-blocks thereof. After processing, the image processing circuitry 28 may output the display image data 56 to the display panel 40. Based at least in part on the display image data 56, the display panel 40 may apply electrical signals to the display pixels of the electronic display 12 to output desired luminances corresponding to the image.
As discussed herein, in some scenarios, the display image data 56 may be output from the image processing circuitry 28 in a multi-resolution format to an electronic display 12 to be displayed in multiple resolutions. As should be appreciated, the boundaries of the regions of the multi-resolution format may be fixed or adjustable and may be based on the specifications of the electronic display 12 that receives the display image data 56 and/or based on a viewer's focal point, which may change on each image frame. To help illustrate,
In the depicted example, the foveated display 58 is divided into a set of 5×5 adjustable regions 60 according to their associated pixel groupings 62. In other words, five columns (e.g., L4, L2, C, R2, and R4) and five rows (e.g., T4, T2, M, B2, and B4) may define the adjustable regions 60. The center middle (C, M) adjustable region coincides with the focal point 64 of the viewer's gaze and may utilize the native resolution of the display panel 40 (e.g., 1×1 pixel grouping 62). Adjustable regions 60 in columns to the right of center (C), such as R2 and R4, have a reduced content resolution in the horizontal direction by a factor of two and four, respectively. Similarly, adjustable regions 60 in columns to the left of center, such as L2 and L4, have a reduced content resolution in the horizontal direction by a factor of two and four, respectively. Moreover, rows on top of the middle (M), such as T2 and T4, have a reduced content resolution in the vertical direction by a factor of two and four, respectively. Similarly, rows below the middle (M), such as B2 and B4, have a reduced content resolution in the vertical direction by a factor of two and four, respectively. As such, depending on the adjustable region 60, the content resolution may vary horizontally and/or vertically.
The pixel groupings 62 may be indicative of the set of display pixels that utilize the same image data in the reduced content resolutions. For example, while the adjustable region 60 at the focal point 64 may be populated by 1×1 pixel groupings 62, the adjustable region 60 in column LA and row M may be populated by 4×1 pixel groupings 62 such that individual pixel values, processed as corresponding to individual pixel locations in the reduced content resolution, are each sent to sets of four horizontal pixels of the display panel 40. Similarly, the adjustable region 60 in column LA and row T4 may be populated by 4×4 pixel groupings 62 such that pixel values are updated sixteen pixels at a time. As should be appreciated, while discussed herein as having reduced content resolutions by factors of two and four, any suitable content resolution or pixel groupings 62 may be used depending on implementation. Furthermore, while discussed herein as utilizing a 5×5 set of adjustable regions 60, any number of columns and rows may be utilized with additional or fewer content resolutions depending on implementation.
As the focal point 64 moves the boundaries 66 of the adjustable regions 60, and the sizes thereof, may also move. For example, if the focal point 64 were to be on the far upper right of the foveated display 58, the center middle (C, M) adjustable region 60, coinciding with the focal point 64, may be set to the far upper right of the foveated display 58. In such a scenario, the T2 and T4 rows and the R2 and R4 columns may have heights and widths of zero, respectively, and the remaining rows and columns may be expanded to encompass the foveated display 58. As such, the boundaries 66 of the adjustable regions 60 may be adjusted based on the focal point 64 to define the pixel groupings 62 for different portions of the foveated display 58.
As discussed herein, the pixel groupings 62 are blocks of pixels that receive the same image data as if the block of pixels was a single pixel in the reduced content resolution of the associated adjustable region 60. To track the pixel groupings 62, an anchor pixel 65 may be assigned for each pixel grouping 62 to denote a single pixel location that corresponds to the pixel grouping 62. For example, the anchor pixel 65 may be the top left pixel in each pixel grouping 62. The anchor pixels 65 of adjacent pixel groupings 62 within the same adjustable region 60 may be separated by the size of the pixel groupings 62 in the appropriate direction. Furthermore, in some scenarios, pixel groupings 62 may cross one or more boundaries 66. For example, an anchor pixel 65 may be in one adjustable region 60, but the remaining pixels of the pixel grouping 62 may extend into another adjustable region 60. As such, in some embodiments, an offset 67 may be set for each column and/or row to define a starting position for anchor pixels 65 of the pixel groupings 62 of the associated adjustable region 60 relative to the boundary 66 that marks the beginning (e.g., left or top side) of the adjustable region 60. For example, an anchor pixel 65 at a boundary 66 (e.g., corresponding to a pixel grouping 62 that abuts the left and/or upper boundary 66 of an adjustable region 60) may have an offset 67 of zero, while an anchor pixel 65 that is one pixel removed from the boundary 66 (e.g., one pixel to the right of or below the boundary 66) may have an offset 67 of one in the corresponding direction. As should be appreciated, while the top left pixel is exampled herein as an anchor pixel 65 and the top and left boundaries 66 are defined as the starting boundaries (e.g., in accordance with raster scan), any pixel location of the pixel grouping 62 may be used as the representative pixel location and any suitable directions may be used for boundaries 66, depending on implementation (e.g., read order).
As yet another example, a group of four 1× foveated input pixels 156 may each be provided to individual respective registers 150, such that the group of four 1× foveated input pixels 156 may provide image data to a total of four registers 150. Because each of the four 1× foveated input pixels 156 provide image data to a single register 150, the display resolution (e.g., foveation resolution) associated with each foveated input pixel of the group of four 1× foveated input pixels 156 may be twice as high as the display resolution associated with the group of four 2× foveated input pixels 154, and four times greater than the display resolution associated with the group of four 4× foveated input pixels 152.
To provide a multiplexer-free architecture, a foveation boundary may be aligned with a particular pixel group. That is, a group of pixels (e.g., a slice of the electronic display 12) may all have a constant foveation ratio. For example, if a slice includes 2× foveation pixels (e.g., foveation pixels with a 2× resolution), all foveation pixels in that slice may be 2×. If a 4× foveation pixel is included in a 2× slice, the 4× foveation pixel may be converted to two 2× foveation pixels to ensure that all foveation pixels in the given slice have a constant foveation ratio.
The timing controller 102 may adjust the foveated input image data 104 to the adjusted foveated image data 106 to provide the constant foveation ratio for the slice of the electronic display 12, as will be described in greater detail below.
It should be noted that a slice may be of any size or length, and the adjusted foveated image data may be converted to any appropriate length.
As previously mentioned, since a foveated region changes based on the movement of the viewer's eye, different areas of the electronic display 12 present different resolutions (e.g., 1×, 2×, 4×) at different times. As such, different parts of the foveated image data are routed to different pixels of the electronic display. One way of routing data is to use multiplexers to select which image data is routed to which source latches of columns of pixels of the electronic display 12. Depending on the number of columns of the electronic display 12, however, the multiplexers may consume a significant portion of the die area while also consuming a significant amount of energy. Moreover, sending image data across the data bus 112 to the source latches 110 consumes energy, particularly as the number of columns of pixels of the electronic display 12 increases. Instead of using multiplexers to route foveated image data in the electronic display 12, groups of registers 150 of the source latches 110 of the electronic display 12 may be hardwired to respective wires of the data bus 112.
With this in mind,
As will be discussed in greater detail below, all four registers 150A, 150B, 150C, and 150D (or 254A, 254B, 254C, and 254D) may be turned on, such that all four pixels of the foveated image data 106 may be provided to four display pixels associated with the four registers 150A, 150B, 150C, and 150D (or the display pixels associated with the registers 254A, 25B, 254C, and 254D). However, in some instances, one or more of the enable signals may be low, the registers corresponding to the low enable signal may not receive the adjusted foveated image data 106, and thus only a portion of the display pixels coupled to the registers 150 (or 254) may receive the adjusted foveated image data 106.
In a second clock cycle, the enable signal 252 may enable the registers 150C and 150D (and not the registers 150A and 150B) such that the registers 150C and 150D may receive the adjusted foveated image data 106 via the wire 250, and the enable signal 258 may enable the registers 254C and 254D (and not the registers 254A and 254B) such that the registers 254C and 254D receive the adjusted foveated image data via the wire 256. In this manner, 2× foveated image data may be effectively routed to source latches 110 and associated display pixels without the use of multiplexers.
As the adjusted foveated image data 106 includes 1× pixels, the wires 250 and 256 may carry the adjusted foveated image data 106 to one register 150, 254 per clock cycle. That is, in a first clock cycle, the enable signal 252 may activate the register 150A such that the register 150A may receive the adjusted foveated image data 106 via the wire 250 of the data bus 112, while the enable signal 258 may activate the register 254A such that the register 254A may receive the foveated image data 106 from the wire 256 of the data bus 112. In a second clock cycle, the enable signal 252 may activate the register 150B such that the register 150B may receive the adjusted foveated image data 106 via the wire 250 of the data bus 112, while the enable signal 258 may activate the register 254B such that the register 254B may receive the foveated image data 106 from the wire 256 of the data bus 112.
In a third clock cycle, the enable signal 252 may activate the register 150C such that the register 150C may receive the adjusted foveated image data 106 via the wire 250 of the data bus 112, while the enable signal 258 may activate the register 254C such that the register 254C may receive the foveated image data 106 from the wire 256 of the data bus 112. In a fourth clock cycle, the enable signal 252 may activate the register 150D such that the register 150D may receive the adjusted foveated image data 106 via the wire 250 of the data bus 112, while the enable signal 258 may activate the register 254D such that the register 254D may receive the adjusted foveated image data 106 from the wire 256 of the data bus 112. In this manner, 1× foveated image data may be effectively routed to source latches 110 and associated display pixels without the use of multiplexers.
As previously mentioned, sending image data across the data bus 112 to the source latches 110 consumes energy, particularly as the number of columns of pixels of the electronic display 12 increases. To reduce the amount of energy consumed by the data bus 112, slices of the data bus 112 may be gated to correspond to which source latches 110 are being loaded. For instance, a first set of source latches 110 corresponding to a first slice of the data bus 112 may be loaded with data while downstream slices of the data bus 112 may be gated to save energy. A token signal passed along the pixel data bus may un-gate the slices over time as image data is passed along to further downstream slices. Thus, fewer slices of the data bus 112 may be active and consuming dynamic power at any point in time.
With this in mind,
As the data bus 300 initializes by opening all of the data bus slices, the data bus consumes full power from initialization to completion. However, as may be observed from the data bus 304A and 304B (collectively, the data bus 304) of
Furthermore to reduce the peak energy consumed by the pixel data bus, the pixel data bus may be divided into two parts that are loaded from opposite sides.
As may be appreciated from the graphs 358 and 360, the peak power consumption of the data bus 304 may be greater than the peak power consumption of the data bus 350 (although the total power consumption may be the same), as the forward loading of the portion 352 and the reverse loading of the portion 354 have opposite peak power consumption. In this manner, the total number of gated slices may remain stable throughout the loading process and the overall power consumption and the peak power consumption of the data bus 112 may be reduced. It should be noted that the data bus 300 may include a number of flip-flops. A flip-flop may be disposed at the beginning of the portion 352 or the portion 354 or between any two slices 302 of the portion 352 or the portion 354.
The specific embodiments described above have been shown by way of example, and it should be understood that these embodiments may be susceptible to various modifications and alternative forms. It should be further understood that the claims are not intended to be limited to the particular forms disclosed, but rather to cover all modifications, equivalents, and alternatives falling within the spirit and scope of this disclosure.
The techniques presented and claimed herein are referenced and applied to material objects and concrete examples of a practical nature that demonstrably improve the present technical field and, as such, are not abstract, intangible or purely theoretical. Further, if any claims appended to the end of this specification contain one or more elements designated as “means for [perform] ing [a function] . . . ” or “step for [perform] ing [a function] . . . ,” it is intended that such elements are to be interpreted under 35 U.S.C. 112 (f). However, for any claims containing elements designated in any other manner, it is intended that such elements are not to be interpreted under 35 U.S.C. 112 (f).
It is well understood that the use of personally identifiable information should follow privacy policies and practices that are generally recognized as meeting or exceeding industry or governmental requirements for maintaining the privacy of users. In particular, personally identifiable information data should be managed and handled so as to minimize risks of unintentional or unauthorized access or use, and the nature of authorized use should be clearly indicated to users.