This disclosure relates to compensation of non-uniform properties of pixels.
Electronic displays are found in numerous electronic devices, from mobile phones to computers, televisions, automobile dashboards, and many more. Individual pixels of the electronic display may collectively produce images. Sometimes the different pixels emit light in such a way that creates perceivable non-uniform presentation of an image between portions of the display.
A summary of certain embodiments disclosed herein is set forth below. It should be understood that these aspects are presented merely to provide the reader with a brief summary of these certain embodiments and that these aspects are not intended to limit the scope of this disclosure. Indeed, this disclosure may encompass a variety of aspects that may not be set forth below.
This disclosure relates to compensating for non-uniform properties of pixels of an electronic display using a function derived in part by measuring light emitted by a pixel. Electronic displays are found in numerous electronic devices, from mobile phones to computers, televisions, automobile dashboards, and many more. Individual pixels of the electronic display may collectively produce images by permitting different amounts of light to be emitted from each pixel. This may occur by self-emission as in the case of light-emitting diodes (LEDs), such as organic light-emitting diodes (OLEDs), or by selectively providing light from another light source as in the case of a digital micromirror device (DMD) or liquid crystal display (LCD). These electronic displays sometimes do not emit light equally between pixels or groups of pixels of the electronic display. This may be due at least in part to non-uniform properties associated with the pixels caused by differences in component age, operating temperatures, material properties of pixel components, and the like. The non-uniform properties between pixels and/or portions of the electronic display may manifest as visual artifacts since different pixels and/or portions of the electronic display emit visibly different (e.g., perceivable by a user) amounts of light.
Systems and methods that compensate for non-uniform properties between pixels or groups of pixels of an electronic display may substantially improve the visual appearance of an electronic display by reducing perceivable visual artifacts. The systems to perform the compensation may be external to an electronic display and/or an active area of the electronic display, in which case they may be understood to provide a form of external compensation, or the systems to perform the compensation may be located within the electronic display (e.g., in a display driver integrated circuit). The compensation may take place in a digital domain or an analog domain, the net result of the compensation producing a compensated data signal (e.g., programming voltage, programming current) transmitted to each pixel of the electronic display before the data signal is used to cause the pixel to emit light. Because the compensated data signal has been compensated to account for the non-uniform properties of the pixels, the images resulting from compensated data signals to the pixels may have substantially reduced visual artifacts. Visual artifacts due to non-uniform properties of the pixels may be reduced or eliminated.
Indeed, this disclosure describes compensation techniques that use a per-pixel function to leverage a relatively small number of variables to predict a brightness-to-data relationship. In this disclosure, the brightness-to-data relationship is generally referred to a brightness-to-voltage (Lv-V) relationship, which is the case when the data signal is a voltage signal. However, the brightness-to-data relationship may also be used when the data signal represents a current (e.g., a brightness-to-current relationship (Lv-I)) or a power (e.g., a brightness-to-power relationship (Lv-W)). It should be appreciated that further references to brightness-to-voltage (Lv-V) are intended to also apply to any suitable brightness-to-data relationship, such as a brightness-to-current relationship (Lv-I), brightness-to-power relationship (Lv-W), or the like. The predicted brightness-to-data relationship may be expressed as a curve, which may facilitate determining the appropriate data signal to transmit to the pixel to cause emission at a target brightness level of light. In addition, some examples may include a regional or global adjustment to further correct non-uniformities of the electronic display.
A controller may apply the brightness-to-data relationship of a pixel or group of pixels to improve perceivable visual appearances of the electronic display by changing a data signal used to drive that pixel or by changing the data signals used to drive that group of pixels. The brightness-to-data relationship may reduce or eliminate perceivable non-uniformity between pixels or groups of pixels.
Various aspects of this disclosure may be better understood upon reading the following detailed description and upon reference to the drawings in which:
One or more specific embodiments are described below. In an effort to provide a concise description of these embodiments, not all features of an actual implementation are described in the specification. It should be appreciated that in the development of any such actual implementation, as in any engineering or design project, numerous implementation-specific decisions must be made to achieve the developers' specific goals, such as compliance with system-related and business-related constraints, which may vary from one implementation to another. Moreover, it should be appreciated that such a development effort might be complex and time consuming, but would nevertheless be a routine undertaking of design, fabrication, and manufacture for those of ordinary skill having the benefit of this disclosure.
When introducing elements of various embodiments of the present disclosure, the articles “a,” “an,” and “the” are intended to mean that there are one or more of the elements. The terms “comprising,” “including,” and “having” are intended to be inclusive and mean that there may be additional elements other than the listed elements. Additionally, it should be understood that references to “one embodiment” or “an embodiment” of the present disclosure are not intended to be interpreted as excluding the existence of additional embodiments that also incorporate the recited features.
Embodiments of the present disclosure relate to systems and methods that compensate non-uniform properties between pixels of an electronic display to improve perceived appearances of visual artifacts. Electronic displays may include light-modulating pixels, which may be light-emitting in the case of light-emitting diode (LEDs), such as organic light-emitting diodes (OLEDs), but may selectively provide light from another light source as in the case of a digital micromirror device (DMD) or liquid crystal display (LCD). While this disclosure generally refers to self-emissive displays, it should be appreciated that the systems and methods of this disclosure may also apply to other forms of electronic displays that have non-uniform properties of pixels causing varying brightness versus voltage relationships (Lv-V curves), and should not be limited to self-emissive displays. When the electronic display is a self-emissive display, an OLED represents one type of LED that may be found in a self-emissive pixel, but other types of LEDs may also be used.
The systems and methods of this disclosure may compensate for non-uniform properties between pixels. This may improve the visual appearance of images on the electronic display. The systems and methods may also improve a response by the electronic display to changes in operating conditions, such as temperature, by enabling a controller to accurately predict performance of individual pixels of the electronic display without tracking and recording numerous data points of pixel behavior to determine Lv-V curves. Instead, a controller may store a few variables, or extracted parameters, for each pixel or group of pixels that, when used in a function (e.g., per-pixel function or per-region function), may generally produce the Lv-V curve of each respective pixel. This reduces a reliance on large numbers of stored data points for all of the pixels of the electronic display, saving memory and/or computing or processing resources. In addition to the controller using a relatively small number of per-pixel or per-region variables, some embodiments may include a further compensation may be applied on a regional or global basis. By at least using the per-pixel function, the Lv-V curves for each pixel in the electronic display may be estimated without relying on large amounts of stored test data. Using the estimated Lv-V curves defined by the per-pixel function, image data that is to be displayed on the electronic display may be compensated before it is programmed into each pixel. The resulting images may have reduced or eliminated visual artifacts due to Lv-V non-uniformities among the pixels.
Furthermore, in some examples, a map used to generate each per-pixel function may be created at a particular brightness level of the display. For example, the map may be generated during manufacturing of the electronic device as part of a display calibration operation and may include data corresponding to one or more captured images. To generate the map, image capturing devices may capture an image of the display at a particular brightness level. In some cases, the per-pixel functions that result from the generated map may be optimally applied at the particular brightness level and less optimally applied at a brightness level not within a range of deviation from the particular brightness level or not at the particular brightness level. As will be appreciated, generating several maps at different brightness levels during calibration and selecting which map to reference to obtain relevant per-pixel functions may improve compensation operations of the electronic device. For example, a particular map may be selected from a group of maps in response to real-time operating conditions of the display (e.g., an input brightness value), and be used to derive per-pixel functions associated with the real-time operating condition. Improvements to compensation operations may improve an appearance of the display, such as by making the display appear relatively more uniform.
A general description of suitable electronic devices that may include a self-emissive display, such as a LED (e.g., an OLED) display, and corresponding circuitry of this disclosure are provided.
The processing core complex 12 of the electronic device 10 may perform various data processing operations, including generating and/or processing image data for presentation on the display 18, in combination with the storage device 14. For example, instructions that are executed by the processing core complex 12 may be stored on the storage device 14. The storage device 14 may be volatile and/or non-volatile memory. By way of example, the storage device 14 may include random-access memory, read-only memory, flash memory, a hard drive, and so forth.
The electronic device 10 may use the communication interface(s) 16 to communicate with various other electronic devices or elements. The communication interface(s) 16 may include input/output (I/O) interfaces and/or network interfaces. Such network interfaces may include those for a personal area network (PAN) such as Bluetooth, a local area network (LAN) or wireless local area network (WLAN) such as Wi-Fi, and/or for a wide area network (WAN) such as a cellular network.
Using pixels containing LEDs (e.g., OLEDs), the display 18 may show images generated by the processing core complex 12. The display 18 may include touchscreen functionality for users to interact with a user interface appearing on the display 18. Input structures 20 may also enable a user to interact with the electronic device 10. In some examples, the input structures 20 may represent hardware buttons, which may include volume buttons or a hardware keypad. The power supply 22 may include any suitable source of power for the electronic device 10. This may include a battery within the electronic device 10 and/or a power conversion device to accept alternating current (AC) power from a power outlet.
As may be appreciated, the electronic device 10 may take a number of different forms. As shown in
The electronic device 10 may also take the form of a tablet device 40, as is shown in
A computer 48 represents another form that the electronic device 10 may take, as shown in
As shown in
The scan lines S0, S1, . . . , and Sm and driving lines D0, D1, . . . , and Dm may connect the power driver 86A to the pixel 82. The pixel 82 may receive on/off instructions through the scan lines S0, S1, . . . , and Sm and may receive programming voltages corresponding to data voltages transmitted from the driving lines D0, D1, . . . , and Dm. The programming voltages may be transmitted to each of the pixel 82 to emit light according to instructions from the image driver 86B through driving lines M0, M1, . . . , and Mn. Both the power driver 86A and the image driver 86B may transmit voltage signals as programmed voltages (e.g., programming voltages) through respective driving lines to operate each pixel 82 at a state determined by the controller 84 to emit light. Each driver may supply voltage signals at a duty cycle and/or amplitude sufficient to operate each pixel 82.
The intensities of each pixel 82 may be defined by corresponding image data that defines particular gray levels for each of the pixels 82 to emit light. A gray level indicates a value between a minimum and a maximum range, for example, 0 to 255, corresponding to a minimum and maximum range of light emission. Causing the pixels 82 to emit light according to the different gray levels causes an image to appear on the display 18. In this way, a first brightness level of light (e.g., at a first luminosity and defined by a gray level) may emit from a pixel 82 in response to a first value of the image data and the pixel 82 may emit at a second brightness level of light (e.g., at a first luminosity) in response to a second value of the image data. Thus, image data may facilitate creating a perceivable image output by indicating light intensities to be generated via a programmed data signal to be applied to individual pixels 82.
The controller 84 may retrieve image data stored in the storage device 14 indicative of various light intensities. In some examples, the processing core complex 12 may provide image data directly to the controller 84. The controller 84 may control the pixel 82 by using control signals to control elements of the pixel 82. The pixel 82 may include any suitable controllable element, such as a transistor, one example of which is a metal-oxide-semiconductor field-effect transistor (MOSFET). However, any other suitable type of controllable elements, including thin film transistors (TFTs), p-type and/or n-type MOSFETs, and other transistor types, may also be used.
The programming voltage is applied to a transistor 93, causing a driving current to be transmitted through the transistor 93 onto the LED 92 based on the Lv-V curve characteristics of the transistor 93 and/or the LED 92. The transistor 93 may be any suitable transistor, such as in one example, an oxide thin film transistor (TFT). In this way, the light emitted from the LED 92 may be selectively controlled. When the Lv-V curve characteristics differ between two pixels 82, perceived brightness of different pixels 82 may appear non-uniform—meaning that one pixel 82 may appear as brighter than a different pixel 82 even when both are programmed by the same programming voltage. The controller 84 or the processing core complex 12 may compensate for these non-uniformities if the controller 84 or the processing core complex 12 are able to accurately predict the Lv-V behavior of the pixel 82. If the controller 84 or the processing core complex 12 are able to make the prediction, the controller 84 or the processing core complex 12 may determine what programming voltage to apply to the pixel 82 to compensate for differences in the brightness levels of light emitted between pixels 82.
Also depicted in
To help illustrate non-uniform Lv-V curves,
During operation, a programming voltage is transmitted to a pixel 82 in response to image data to cause the pixel 82 to emit light at a brightness level to suitably display an image. This programming voltage is transmitted to pixels 82 to cause an expected response (e.g., a first programming voltage level is used specifically to cause a first brightness level to display an image). The expected response of the pixels 82 to a first voltage (V1) level 106 is a first brightness (Lv1) level 108, however, both responses from the first pixel 82 and the second pixel 82 deviate from that expected response (e.g., line 104). As illustrated on the graph, the first pixel 82 indicated by the line 100 responds by emitting a brightness level corresponding to brightness level 110 while the second pixel 82 indicated by the line 102 responds by emitting a brightness level 112. Both the brightness level 110 and the brightness level 112 deviate from the target brightness level of 108. This deviation between the Lv-V curves may affect the whole relationship, including the responses to a second voltage (V2) level 114 as illustrated on the graph. It should be noted that, in some cases, the pixel non-uniformity caused at least in part by the Lv-V curves is worse at lower programming voltages than higher programming voltages (e.g., net disparity 118 at a lower voltage is greater than net disparity 120 at a higher voltage).
To correct for these non-uniformities, such as the differences between the portion 132 and the portion 134, a fixed correction may be used.
To improve the fixed correction techniques at lower brightness levels (e.g., to eliminate the net disparity 118 in addition to maintaining the eliminated net disparity 120), the controller 84 may use dynamic correction techniques, including applying a per-pixel function to determine a suitable correction a programming voltage.
The effect of basing the compensation at least in part on the per-pixel function is depicted through the difference in compensations used on the Lv-V curves. For example, to cause the first pixel 82 to emit light at a brightness level 184, the programming voltage is changed an amount 186 from a first voltage level 188 to a second voltage level 190, while to cause the first pixel 82 to emit light at a brightness level 192, the programming voltage is changed by an amount 194 from a voltage level 196 to a voltage level 198, where the amount 194 may be different from the amount 186 (based on the per-pixel function for the first pixel 82). In this way, the amount 194 and the amount 186 may be different from the corresponding compensation amounts used for the second pixel 82. For example, a compensation amount 194 differs from the corresponding compensation amount 186 used to correct pixel non-uniformities of the first pixel 82.
Thus, as shown in
To help explain the per-pixel function,
At block 202, the controller 84 receives one or more captured images of a display 18 panel. These images may be captured during a calibration and/or testing period, where test image data is used to determine what per-pixel compensations to apply to each pixel 82 of the display 18 being tested. Programming voltages based on the test image data may be used to drive the pixels 82 to display a test image corresponding to the test image data. After the pixels 82 begin to display the test image, an external image capture device, or other suitable method of capturing images, may be used to capture one or more images of the display 18 panel. The one or more images of the display 18 panel may capture an indication of how bright the different portions of the display 18 panel are or communicate relative brightness levels of light emitted by pixels 82 of the display 18 panel in response to the test image data.
After receiving the one or more images, at block 204, the controller 84 may process the one or more images to extract per-pixel Lv-V data. As described above, the received images indicate relative light intensity or brightness between pixels 82 and/or between regions of the display 18 panel. The controller 84 may process the received images to determine the response of the pixel 82 to the test data. In this way, the controller 84 processes the received images to determine (e.g., measure, calculate) the brightness of the light emitted from the respective pixels 82 in response to the test data. The per-pixel Lv-V data determined by the controller 84 includes the known programming voltages (e.g., based on the test image data) and the determined brightness of light emitted.
At block 206, the controller 84 fits a per-pixel function to the per-pixel Lv-V data. The controller 84 may perform this curve-fitting in any suitable matter using any suitable function. A suitable function indicates a relationship between a programming voltage used to drive each pixel 82 and the light emitted from the pixel 82 in response to the programming voltage. The per-pixel function may be, for example, a linear regression, a power law model (e.g., current or brightness equals power multiplied by a voltage difference exponentially raised by an exponent constant representative of the slope between voltages), an exponential model, or the like. The relationship defined by the per-pixel function may be specific to a pixel 82, to a display 18, to regions of the display 18, or the like. In this way, one per-pixel function may be used for determining extracted parameters to define an Lv-V curve for a first pixel 82 while a different per-pixel function may be used for determining extracted parameters to define an Lv-V curve for a second pixel 82.
After fitting the per-pixel function to the per-pixel Lv-V data, at block 208, the controller 84 generates extracted parameters from the per-pixel function and saves the extracted parameters. In this way, the per-pixel function may represent a curve that is fitted to several data points gathered as the per-pixel Lv-V data but may be defined through a few key variables that represent the extracted parameters. Examples of the extracted parameters may include an amplitude, a rate of growth (e.g., expansion), slopes, constants included in a per-pixel function, or the like, where an extracted parameter is any suitable variable used to defined a fitted curve. The extracted parameters are extracted and saved for each pixel 82. These values may be stored in one or more look-up tables to be referenced by the controller 84 to determine the response of a respective pixel to a particular programming voltage. Fitting the per-pixel function to a dataset including the known programming voltages and/or the determined brightness of light emitted enables the per-pixel function to predict an overall input/output relationship for the pixel 82 based on extracted parameters associated with the fitted per-pixel function without having to store each individual data point of the input/output relationship.
To better explain how the controller 84 may compensate for Lv-V non-uniformity among pixels 82,
In general, the controller 84 may apply the target brightness level 230 to a per-pixel function 232 that receives the target brightness level 230 and one or more extracted parameters 234 (e.g., variables based on the pixel 82). As described above, the per-pixel function 232 may be any suitable function that generally describes the Lv-V characteristics of each respective pixel 82. The extracted parameters 234 may be values stored in memory (e.g., in one or several look-up tables). When used in the function, the extracted parameters 234 permit the per-pixel function 232 to produce a first form of compensation for pixel values by, for example, translating the target brightness level to a corresponding programming voltage. This is shown in
As mentioned above, this first per-pixel function 232 may not always, on its own, provide a complete compensation. Indeed, the per-pixel function 232 may produce an approximation of the Lv-V curve of the pixel 82 based on the extracted parameters 234. Thus, rather than define the Lv-V curve of the pixel 82 using numerous measured data points, the Lv-V curve of the pixel 82 may be approximated using some limited number of variables (e.g., extracted parameters 234) that may generally define the Lv-V curve. The extracted parameters 234 may be determined based on measurements of the pixels 82 during manufacturing or based on measurements that are sensed using any suitable sensing circuitry in the display 18 to identify the Lv-V characteristics of each pixel 82.
Since the per-pixel function 232 provides an approximation of an actual Lv-V curve of a pixel 82, the resulting compensated programming voltage 236 (based on the target brightness level) may be further compensated in some examples (but not depicted). The compensated programming voltage 236 is used to program the pixels 82. Any additional compensations may be applied to the compensated programming voltage before being applied to the pixel 82.
At block 262, the controller 84 determines a target brightness level 230 for a pixel 82 to emit light at based on image data. The target brightness level 230 corresponds to a gray level associated with a portion of the image data assigned to the pixel 82. The controller 84 uses the target brightness level 230 to determine a compensated programming voltage 236 to use to drive the pixel 82. A proportion associating the gray level indicated by the image data to a target brightness level, or any suitable function, may be used in determining the target brightness level 230.
At block 264, the controller 84 applies the per-pixel function 232 to the target brightness level 230 for the pixel 82 to determine a compensated programming voltage 236. The controller 84 determines a compensated programming voltage 236 for the pixel 82 based on the target brightness level 230 and based on the extracted parameters 234. The extracted parameters 234 are used to predict the particular response of the pixel 82 to the various programming voltages that may be applied (e.g., the per-pixel function 232 for that pixel 82). Thus, based the per-pixel function, the controller 84 determines the programming voltage 236 to apply to cause the pixel 82 to emit at the target brightness level 230, or a compensation to make to a programming voltage to be transmitted to the pixel 82 (e.g., such as in cases where each pixel 82 to emit at the target brightness level 230 receives the same programming voltage that is later changed before being used to drive a pixel 82 based on the per-pixel function 232 for the pixel 82). It should be noted that although described as a programming voltage, the compensated programming voltage 236 may be any suitable data signal used to change a brightness of light emitted from the pixel 82 in response to image data. For example, the controller 84 may determine and/or generate a control signal used to change a data signal, such as a programming voltage, to generate a compensated data signal, such as the compensated programming voltage 236.
Using the compensated programming voltage 236, at block 268, the controller 84 may transmit the compensated programming voltage 236 to the pixel 82 by operating a driver 86 of the display 18 to output the compensated programming voltage 236 level to the pixel 82. The compensated programming voltage 236 causes the pixel 82 to emit light at the target brightness level 230. Thus, through the controller 84 transmitting the compensated programming voltage 236 to the pixel, visual artifacts of the display 18 are reduced via correction and compensation for non-uniform properties between pixels 82.
In some examples, a technique using a combination of a fixed correction and a dynamic correction may be applied by the controller 84 to compensate for non-uniform properties of pixels 82.
In addition to determining the per-pixel function 232 and extracted parameters 234 (e.g., via the process 200), the controller 84 receives one or more images at the block 202. The number of images received by the controller 84 may correspond to a number of missing variables of the per-pixel function, such that the images may facilitate a creation of a system of equations to determine one or more unknown variables. For example, three images may be captured and transmitted to the controller to be used to determine three unknown variables. These captured images may represent different outputs to different test data. In this way, a first test programming voltage may be used to generate a first captured image and a second test programming voltage may be used to generate a second captured image, where both the first captured image and the second captured image may be used to determine the extracted parameters. In some examples, the one or more unknown variables correspond to the extracted parameters 234.
Keeping the foregoing in mind, a map may result from the above-described image captures.
To do this, one or more input gray domain programming voltages are converted into voltage domain programming voltages via gray domain to voltage domain conversion operations 308. While at least one programming voltage is in the voltage domain, the processing core complex 12 and/or the controller 84 may reference a voltage map generated during manufacturing of the electronic device 10 (e.g., ΔV map generation operations 310) to determine the per-pixel function 232 applicable to the programming voltage. The per-pixel function 232 is applied to the programming voltage in the voltage domain via summation block 312, and the output is converted back into the gray domain via a voltage domain to gray domain conversion operation 314 for use in additional preparatory operations before being used as the compensated programming voltage 236. For ease of discussion herein, it should be understood that the processing core complex 12 and/or the controller 84 may perform the described operations even if the controller 84 is referred to as performing the operation.
The per-pixel function 232 may be derived from an image captured of the display 18 while operated at a particular input brightness value. In this way, the image captured during the image capture operations 300 may be used to generate one or more maps (e.g., during electronic device 10 manufacturing and/or calibration). For example, image data of the image captured via the image capture operations 300 may be used to generate a change in brightness map (e.g., ΔLv map via ΔLv map generation operations 316) and to generate, from the ΔLv map, a change in voltage map (ΔV map). In some cases, the per-pixel functions 232 based on the ΔV map may be relatively less accurate as the input brightness value of the display 18 at a time of compensation (e.g., during operation rather than manufacturing and/or calibration) deviates from the input brightness value of the display 18 at a time of the image capture operations 300 (e.g., during manufacturing and/or calibration).
An example of this deviation is shown in
Generating several maps at different brightness levels during image capture operations 300 and map generation operations 310, 316, and later selecting a specific map based on real-time operating conditions may improve compensation operations. For example, a map may be selected in response to an input brightness value and be used to derive a per-pixel function 232 associated with a particular pixel 82 and associated with the real-time operating condition.
To help explain further,
In this way, operational ranges 354 (354A, 354B, 354C) may be defined for a particular display 18. Each of the operational ranges 354 may correspond to one or more original image captures and a map. For example, the operational range 354A corresponds to a map that results from images captured at a luminance of capture equal to 0.1 nit while the operational range 354B corresponds to a map that results from images captured at a luminance of capture equal to 0.6 nit. Based on the input brightness value, a different operational range of the operational ranges 354 is selected as a way to select the map for uniformity compensation operations 304. In this way, if the input brightness value is less than 5 nits, the selected operational range is operational range 354A, and thus the map corresponding to operational range 354A may be applied as part of the compensation, while if the input brightness value exceeds 15 nits, the selected operational range is operational range 354C which leads to applying the map corresponding to the operational range 354C. Many different methods may be used to determine a suitable number and respective sizes of operational ranges 354. For example, as shown in
This selection process is described via
During manufacturing of the electronic device 10, when calibration operations are performed, multiple image capture operations 300 may be performed at different brightness levels (e.g., different luminance of capture levels). Any suitable number of brightness levels and image capture operations 300 may be performed. The image capture operations may be performed as part of map generation operations 372A used to generate one or more ΔV maps 374 in response to image captures performed at the different brightness levels. Each image capture operation 300, and resulting ΔV map 374, may correspond to one of the operational ranges 354 described in
This process is additionally depicted in
The input brightness value 368 may be a global brightness value. For example, the input brightness value 368 may correspond or be the brightness level of the display 18, and thus may change in response to ambient lighting conditions of the electronic device 10. In some examples, the input brightness value 368 may be a value derived or generated based on a histogram of an image to be displayed, a histogram of an image that is currently displayed, and/or a histogram of an image previously displayed. Furthermore, in some examples, the input brightness value 368 may correspond to a regional brightness, such as a brightness of a subset of pixels 82 of the display 18 or a brightness of an image to be presented via a subset of pixels 82 of the display. The input brightness value 368 may also be determined on a per-pixel basis, such as associated with a brightness that the pixel 82 is to emit light.
To help visualize further,
In some cases, the input brightness value 368 may be of a value that is between defined brightness values corresponding to the various ΔV maps 374. The map selection operation 366 may thus include performing an interpolation between two ΔV maps 374 (e.g., two of the ΔV maps 374 that correspond to the defined brightness values adjacent or close to the input brightness value 368). In this way, when the input brightness value 368 is between calibrated control points (e.g., brightness values corresponding to each of the ΔV maps 374), a new map may be dynamically generated that corresponds to the input brightness value 368. Linear interpolation may be used to generate a map that corresponds to the input brightness value 368 that falls between defined brightness values of ΔV maps 374.
At block 412, the controller 84 may receive one or more captured images of a panel of a display 18. These images may be captured during a calibration and/or testing period, where test image data is used to determine what per-pixel compensations to apply to each pixel 82 of the display 18 being tested. Programming voltages based on the test image data may be used to drive the pixels 82 to display a test image corresponding to the test image data. Furthermore, test image data may also include varying of an input brightness value to determine how the pixels 82 behave in response to varying input brightness values. After the pixels 82 begin to display the test image, an external image capture device, or other suitable method of capturing images, may be used to capture one or more images of the panel of the display 18. The one or more images of the panel of the display 18 may capture an indication of how bright the different portions of the panel of the display 18 are or communicate relative brightness levels of light emitted by pixels 82 of the panel of the display 18 in response to the test image data. These captured images and associated input brightness values, such as a global brightness value of the display 18 at the time of capture, are recorded and stored into memory (e.g., storage devices 14). These captured images and associated input brightness values may be used to define the different operational ranges 354.
After receiving the one or more images, at block 414, the controller 84 may process the one or more images to extract per-pixel Lv-V data for each captured image corresponding to the differing operational ranges 354. As described above, the received images indicate relative light intensity or brightness between pixels 82 and/or between regions of the display 18 panel. The controller 84 may process the received images to determine the response of the pixel 82 to the test data that is the same but applied at different input brightness values. In this way, the controller 84 processes the received images to determine (e.g., measure, calculate) the brightness of the light emitted from the respective pixels 82 in response to the test data. The per-pixel Lv-V data determined by the controller 84 includes the known programming voltages (e.g., based on the test image data) and the determined brightness of light emitted.
At block 416, the controller 84 may fit a per-pixel function to the per-pixel Lv-V data on a per optional range basis. The controller 84 may perform this curve-fitting in any suitable matter using any suitable function. A suitable function indicates a relationship between a programming voltage used to drive each pixel 82 and the light emitted from the pixel 82 in response to the programming voltage. The per-pixel function may be, for example, a linear regression, a power law model (e.g., current or brightness equals power multiplied by a voltage difference exponentially raised by an exponent constant representative of the slope between voltages), an exponential model, or the like. The relationship defined by the per-pixel function may be specific to a pixel 82, to a display 18, to regions of the display 18, or the like, at a specific input brightness value. In this way, one per-pixel function may generate one set of extracted parameters 234 to define an Lv-V curve for a first pixel 82 at a first input brightness value 368 while a different per-pixel function may generate a second set of extracted parameters 234 to define an Lv-V curve for a second pixel 82 at a same or different input brightness value 368.
After fitting the per-pixel function 232 to the per-pixel Lv-V data, at block 418, the controller 84 may generate extracted parameters 234 from the per-pixel function and may save the extracted parameters 234 at block 420. In this way, the per-pixel function may represent a curve that is fitted to several data points gathered as the per-pixel Lv-V data but may be defined through a few key variables that represent the extracted parameters 234. Examples of the extracted parameters 234 may include an amplitude, a rate of growth (e.g., expansion), slopes, constants included in a per-pixel function, or the like, where an extracted parameter 234 is any suitable variable used to at least partially define a fitted curve. The extracted parameters 234 are extracted and saved for each pixel 82 and for each of the operational ranges 354. These values may be stored in one or more look-up tables to be referenced by the controller 84 to determine the response of a respective pixel 82 to a particular programming voltage at a particular input brightness value 368. Fitting the per-pixel function to a dataset including the known programming voltages and/or the determined brightness of light emitted enables the per-pixel function to predict an overall input/output relationship for the pixel 82 based on extracted parameters 234 associated with the fitted per-pixel function without having to store each individual data point of the input/output relationship.
At block 434, the controller 84 may determine a target brightness level 230 for a pixel 82 to emit light at based on image data and/or based on the input brightness value 368. The target brightness level 230 corresponds to a gray level associated with a portion of the image data assigned to the pixel 82. The controller 84 uses the target brightness level 230 to determine a compensated programming voltage 236 to use to drive the pixel 82. A proportion associating the gray level indicated by the image data to a target brightness level, or any suitable function, may be used in determining the target brightness level 230.
At block 436, the controller 84 determines and applies the per-pixel function 232 based on the input brightness value 368 to the target brightness level 230 for the pixel 82 to determine a compensated programming voltage 236. The controller 84 determines a compensated programming voltage 236 for the pixel 82 based on the target brightness level 230, based on the extracted parameters 234, and based on the input brightness value 368 defining from which of the operational ranges 354 to source the extracted parameters 234. The extracted parameters 234 are used to predict the particular response of the pixel 82 to the various programming voltages that may be applied (e.g., the per-pixel function 232 for that pixel 82). Thus, based on the per-pixel function 232, the controller 84 determines the programming voltage 236 to apply to cause the pixel 82 to emit at the target brightness level 230, or a compensation to make to a programming voltage to be transmitted to the pixel 82 (e.g., such as in cases where each pixel 82 to emit at the target brightness level 230 receives the same programming voltage that is later changed before being used to drive a pixel 82 based on the per-pixel function 232 for the pixel 82). It should be noted that although described as a programming voltage, the compensated programming voltage 236 may be any suitable data signal used to change a brightness of light emitted from the pixel 82 in response to image data. For example, the controller 84 may determine and/or generate a control signal used to change a data signal, such as a programming voltage, to generate a compensated data signal, such as the compensated programming voltage 236.
Using the compensated programming voltage 236, at block 438, the controller 84 may transmit the compensated programming voltage 236 to the pixel 82 by operating a driver 86 of the display 18 to output the compensated programming voltage 236 level to the pixel 82. The compensated programming voltage 236 causes the pixel 82 to emit light at the target brightness level 230. Thus, through the controller 84 transmitting the compensated programming voltage 236 to the pixel, visual artifacts of the display 18 are reduced via correction and compensation for non-uniform properties between pixels 82.
Keeping the foregoing in mind, it is noted that the ΔV map 374 may be updated at various times during operation of the electronic device 10. For example, the ΔV map 374 may be updated during the processing of each image frame, or in response to a change in the input brightness value 368. Furthermore, in some embodiments, the ΔV map 374 is updated one or more frames after the input brightness value 368 was determined for a particular image frame. Other parameters than the input brightness value 368 may be used to select the ΔV map 374. For example, parameters like temperature and/or historic image data may be used in combination with or instead of the input brightness value 368. The input brightness value 368 may be determined independently of an image frame presented or to be presented via the display 18. For example, the input brightness value 368 may be an amount determined in response to a sensed amount of ambient light.
The foregoing descriptions relate to determining per-pixel (or per-regional) compensations based at least in part on per-pixel (or per-regional) functions generated using images captured of the display 18 during manufacturing. In some cases, an amount of data used to store per-pixel (or per-regional) functions may be reduced by instead storing anchor points of the per-pixel (or per-regional) function. An anchor point may define a compensation in terms of a voltage change, ΔV, to apply to a programming voltage (e.g., data voltage). In this way, the anchor point may represent a known compensation that is able to be used to derive other, unknown, or undefined adjustments to perform to input data voltages that do not correspond to an anchor point. For example, when a data voltage does not correspond to an anchor point, performing an interpolation on nearby anchor points may help to derive the ΔV compensation for the data voltage.
Keeping this in mind,
The anchor points 502 may define one or more compensations for a pixel 82. In this way, for the pixel 82, when the input image data 504 equals any of one or more defined input image data values, the anchor points 502 specify what compensation to apply to the input image data 504. However, when the input image data 504 does not match one of the defined input image data values, the anchor points 502 may be interpolated to derive a compensation (e.g., estimated compensation) to apply to the input image data 504 based on the other defined compensations. Thus, the anchor points 502 may provide a structured method in which to estimate a suitable and/or reasonable compensation to apply to a known input image data 504 when the input image data 504 is not specifically defined via the anchor point 502.
Similar to previous discussions associated with map selection, anchor points 502 of the pixel 82 may also change or be relatively less suitable at different brightness values. To counteract this, multiple ΔV maps of anchor points 502 corresponding to different brightness values may be defined during manufacturing.
To help elaborate,
During manufacturing of the electronic device 10, when calibration operations are performed, multiple image capture operations 300 may be performed at different brightness levels (e.g., different luminance of capture levels). Any suitable number of brightness levels and image capture operations 300 may be performed and one or more of the resulting images may be used to generate ΔV maps of anchor points 502 via a map generation of anchor points operation (shown as ΔV maps 506). Each image capture operation 300, and resulting ΔV map 506, may correspond to one of the operational ranges 354 described in
In this way, the controller 84 may receive image data 504 for presentation and perform initial data processing operations 302 on the image data 504. Using processed image data output from the initial data processing operations 302, the controller 84 may perform uniformity compensation operations 304 that include a grey domain to voltage domain conversion 308 and adding resulting voltage domain image data to a determined compensation value to generate the compensated programming voltage. The determined compensation value may be an output from the map selection and ΔV determination operation 366 since the controller 84 may use the selected ΔV map 506 during interpolation to determine an amount to compensate. The determined compensated value may be an analog offset voltage determined via map selection and anchor point interpolation operations. The determined compensation value may be summed with the voltage domain image data output from the grey domain to voltage domain conversion operations 308 to obtain the compensated programming voltage 236. The compensated programming voltage 236 may be converted back into the grey domain via voltage domain to grey domain conversion operations 314, and may be further processed via gamma processing 306 before being transmitted to one or more pixels 82 of the active area 83.
Each map 506 may include one or more anchor points used to describe a brightness and voltage relationship corresponding to a pixel 82, or a region of pixels 82. To help elaborate,
Explaining the interpolation further,
In some examples, a compensation technique using a combination of a fixed correction and a dynamic correction may be applied by the controller 84.
The controller 84 may use a combination of techniques to determine a programming voltage based on a threshold. The controller 84 may apply a certain technique for input image data 504 below a threshold, such as a threshold value of image data 504D (e.g., line 540 but corresponding to a brightness threshold value 542), and may apply a different technique for input image data 504 above or at the threshold. For example, the controller 84 may apply a fixed offset of ΔV 512E for input image data 504 at or above the threshold (e.g., line 540) but may use dynamic correction techniques for input image data 504 less than the threshold (e.g., line 540). It should be understood that the threshold (e.g., line 540) may equal any suitable input image data value and correspond to any suitable brightness level and/or that any number of thresholds may be used to control the compensation technique used for various target brightness levels. Using a combination of techniques may lessen processing resources while maximizing benefits from using the anchor points 502 for input image data 504 less than the threshold (e.g., line 540) and minimizing processing resources dedicated to input image data 504 above or at the threshold (e.g., line 540) where a fixed correction may be a suitable form of correction to apply.
In yet another case,
A low clipping threshold 560 and a high clipping threshold 562 may be used to define an input image data range via the anchor points 502. In this way, when input image data 504 is received that is greater than the high clipping threshold 562, the input image data 504 is compensated to a clipped voltage 564A, regardless of the amount by which the input image data 504 is greater than or equal to the high clipping threshold 562 (e.g., a value greater than the high clipping threshold 562 is adjusted to equal a uniform compensated programming voltage). Furthermore, when the input image data 504 is less than or equal to the low clipping threshold 560, the input image data is compensated to a clipped voltage 564B, regardless of the amount by which the input image data 504 is less than the low clipping threshold 560 (e.g., a value less than the low clipping threshold 560 is adjusted to equal a uniform compensated programming voltage). For example, input image data 504C and the input image data 504D may both exceed the high clipping threshold 562 and thus are both respectively compensated to (e.g., clipped to) the clipped voltage 564B. As a second example, input image data 504E may be clipped to clipped voltage 564C since the input image data 504E is less than the low clipping threshold 560. Clipping may be permitted because differences in light emission at either relatively low or relatively high nits is unperceivable to a viewer, thus reducing an amount of processing resources used during the compensation.
To help explain these operations described,
At block 572, the controller 84 may receive one or more captured images of a panel of a display 18. These images may be captured during a calibration and/or testing period, where test image data is used to determine compensations to apply to the display 18 being tested. Operations performed at block 572 may be similar to operations of block 412 of
After receiving the one or more images, at block 574, the controller 84 may process the one or more images to extract Lv-V data for each captured image corresponding to differing operational ranges 354. The controller 84 may process the received images to determine the response of one or more of the pixels 82 to the test data that is the same but applied at different input brightness values. Operations performed at block 574 may be similar to operations of block 414 of
At block 576, the controller 84 may use the Lv-V data for each captured image to determine anchor points for each of the one or more pixels 82 characterized by the Lv-V data. In this way, the controller 84 may determine anchor points 502 for regional groupings of pixels 82, for each pixel 82 of the active area 83, or for any suitable combination of pixels 82 of the display 18. The data stored as the anchor point 502 may be an input image data value and a corresponding adjustment for the input image data value. Using a per-pixel defined anchor point 502 as an example, the controller 84 may determine the adjustment to be stored as part of the anchor point 502 by comparing an identified behavior of a target pixel 82 to a desired behavior for each pixel 82 (e.g., a uniform behavior). In this way, the controller 84 may compare a brightness at which a target pixel 82 is to emit in response to the input image data value to an average emission behavior of the display 18 at the input image data value. The difference between the brightness level of the target pixel 82 and the desired brightness level may be correlated into a desired input image data value using the Lv-V data. A difference between the desired input image data value and the input image data value used to determine the desired input image data value may be used as an offset value to apply to input image data values received at a later time during actual operation of the electronic device 10. The controller 84 may determine offset values and/or adjustments to be stored with the input image data value as the anchor point using any suitable method.
Once the anchor points 502 are determined for each input brightness value (e.g., each captured image), at block 578, the controller may store the anchor points 502 in memory. The controller 84 may store the anchor points 502 as part of ΔV maps 504, organized by input brightness values. The ΔV maps 504 may each include a parameter, such as a variable stored in a field, that specifies which input brightness level each ΔV map 504 corresponds to. The controller 84 may reference these fields when performing map selection operations 366 of
At block 592, the controller 84 may receive input image data 504 to be used to generate driving signals to drive the pixel 82. The input image data 504 may correspond to target brightness level 230 or a target gray level associated with a portion of the image data assigned to the pixel 82.
At block 594, the controller 84 determines whether to clip the input image data 504 to a particular value. To do so, the controller 84 may compare a value of the input image data 504 to one or more clipping thresholds (e.g., low clipping threshold 560, high clipping threshold 562). In response to determining that the value of the input image data 504 is either greater than a high clipping threshold 562 or less than a low clipping threshold 560, the controller may, at block 596, clip the input image data 504 to generate the compensated programming voltage 236. As described with regard to
However, in response to determining that the value of the input image data 504 is not to be adjusted via clipping, the controller 84 may, at block 600, determine and apply an offset to the input image data 504 as specified by one or more anchor points 502. Each anchor point 502 may correspond to an input image data value and an offset value (e.g., ΔV 512), such that when the controller 84 receives input image data 504 equal to the input image data value, the controller 84 applies the offset value without performing additional operations related to determining a compensation. Sometimes two anchor points 502 may be used to determine an offset value if the input image data 504 does not equal one of the input image data values stored as anchor points 502. In these cases, the controller 84 may interpolate the two anchor points 502 to determine an offset valued between offset values associated with the two anchor points 502.
At block 600, the controller 84 may transmit the generated compensated programming voltage 236 to pixel 82. Driving the pixel 82 with the compensated programming voltage 236 may cause the pixel 82 to emit a uniform brightness level relative to other pixels 82 also emitting according to a same target brightness level (e.g., target grey level defined by image data).
As described above, the controller 84 may apply regionally-specific compensations.
Thus, the technical effects of the present disclosure include improving controllers of electronic displays to compensate for non-uniform properties between one or more pixels or groups of pixels, for example, by applying a per-pixel function to programming data signals used in driving a pixel to emit light. These techniques describe selectively generating a compensated data signal (e.g., programming voltage, programming current, programming power) based on a per-pixel function, where the compensated data signal is used to drive a pixel to emit light at a particular brightness level to account for specific properties of that pixel that are different from other pixels. These techniques may be further improved by generating compensated data signals with consideration for an input brightness value. By selecting a map based on the input brightness value, non-uniform properties of the display that manifest as visual artifacts may be reduced or mitigated. Different maps may be generated at a time of calibration and/or manufacturing by repeating, at different brightness values, generation of extracted parameters for multiple image captures as a way to gather information about how each pixel behaves when driven to present at different brightness values in addition to different image data. Maps may be generated to include per-pixel functions and/or to include anchor points. Furthermore, using anchor points to provide a compensated data signal may decrease an amount of time for compensation operations and/or may reduce an amount of memory used to store information used in the compensation.
The specific embodiments described above have been shown by way of example, and it should be understood that these embodiments may be susceptible to various modifications and alternative forms. It should be further understood that the claims are not intended to be limited to the particular forms disclosed, but rather to cover all modifications, equivalents, and alternatives falling within the spirit and scope of this disclosure.
The techniques presented and claimed herein are referenced and applied to material objects and concrete examples of a practical nature that demonstrably improve the present technical field and, as such, are not abstract, intangible or purely theoretical. Further, if any claims appended to the end of this specification contain one or more elements designated as “means for [perform]ing [a function] . . . ” or “step for [perform]ing [a function] . . . ”, it is intended that such elements are to be interpreted under 35 U.S.C. 112(f). However, for any claims containing elements designated in any other manner, it is intended that such elements are not to be interpreted under 35 U.S.C. 112(f).
This application claims priority to U.S. Provisional Patent Application No. 62/728,648, entitled “Dynamic Uniformity Compensation for Electronic Display,” filed Sep. 7, 2018, and is related to U.S. patent application Ser. No. 16/563,622, filed Sep. 6, 2019, entitled “Dynamic Uniformity Compensation for Electronic Display,” both of which are incorporated herein by reference in their entireties for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
9812071 | Matsui | Nov 2017 | B2 |
20060262147 | Kimpe | Nov 2006 | A1 |
20100123699 | Leon | May 2010 | A1 |
20140375704 | Bi | Dec 2014 | A1 |
20170122725 | Yeoh | May 2017 | A1 |
20180246375 | Zhang | Aug 2018 | A1 |
20180366074 | Choi | Dec 2018 | A1 |
20190066555 | Gu | Feb 2019 | A1 |
20190340980 | Yum | Nov 2019 | A1 |
20200058252 | Kim | Feb 2020 | A1 |
20200184912 | Wang | Jun 2020 | A1 |
Number | Date | Country | |
---|---|---|---|
62728648 | Sep 2018 | US |