Many electronic display devices, such as monitors, televisions, and phones, are 8-bit depth displays that are capable of displaying combinations of 256 different intensities of each of red, green, and blue (RGB) pixel data. Although these different combinations result in a color palette of more than 16.7 million available colors, the human eye is still able to detect color bands and other transition areas between different colors. These color banding effects can be prevented by increasing the color bit depth to 10 bits which is capable of supporting 1024 different intensities of each of red, green, and blue pixel data. However, since many display device devices only support an 8-bit depth, a 10-bit RGB input signal must be converted to an 8-bit signal to be displayed on a display device.
Many different dither methods, such as ordered dither, random dither, error diffusion, and so on, have been used to convert a 10-bit RGB input signal to 8-bit RGB data to reduce banding effects in the 8-bit RGB output. However, these dither methods have been applied at the display end, only after image data encoded at 10 bits has been received and decoded at 10 bits. These dither methods have not been applied to dithering 10-bit YCbCr to 8-bit YCbCr data before the 8-bit data is encoded and transmitted to a receiver for display on an 8-bit RGB display device.
One of the reasons that these dither methods have not been applied to dithering 10-bit YCbCr data before transmission is that many of the international standards that define YCbCr to RGB conversion cause loss of quantization levels in the output, even in the input and the output signals have the same bit depth. This may occur because the conversion calculation may map multiple input quantization levels into a same output level. The loss of these quantization levels during conversion from YCbCr to RGB negates the effects of applying a dither when converting a 10-bit signal to an 8-bit signal. As a result, the output images may contain banding artifacts.
There is a need to generate display images that do not contain banding artifacts when applying a dither during a bit reduction process as part of a conversion from YCbCr to RGB color space.
In an embodiment of the invention, YCbCr pixel data may be dithered and converted into 8-bit RGB data, which may be shown on a 8-bit display free of banding artifacts. Some methods and image processors generate display data that is free of banding artifacts by applying a stronger dither having a same mean with a larger variance to image data before conversion to RGB data. Other methods and image processors calculate a quantization or encoding error for a given pixel block and diffuse the calculated error among one or more neighboring pixel blocks. These options are discussed in more detail below.
Dither in Excess of Truncated Pixels
Prior random dither methods added dither noise to each of the three color channels before dropping bits in a quantization level reduction module. The dither noise that was added corresponded to the number of noise levels that the dropped bits could generate. For example, when the quantization level reduction is from 10 bits to 8 bits, the dither noise contains four possible digits: 0, 1, 2, and 3 with equal probabilities.
These random dither methods did not account for the further loss of quantization level loss when converting from YCbCr color space to RGB color space, even if the bit depth remained the same in both color spaces. Thus, the past random dither methods would have dithered 8-bit YCbCr data without banding, but the final 8-bit RGB output would include banding because of the loss of quantization level during the color space conversion.
To compensate for the loss of quantization levels during the color space conversion process, embodiments of the present invention may apply a dither noise to each of the three color channels that exceeds the number of levels corresponding to the dropped bits. For example, when the net quantization level is being reduced by two bits, such as by dropping two bits to get from a 10-bit input to an 8-bit signal, the applied dither noise may contain eight possible digits: −2, −1, 0, 1, 2, 3, 4, and 5, instead of the four digits in the past methods which were limited to a range having 2n values, where n is the number of bits being dropped. Other quantities of additional digits may be added in other embodiments.
These additional digits may be selected so that the mean of the new noise is the same as the mean in past random dither methods, while the variance of the new noise is increased. In some instances, each of digits may have an equal probably of being selected. The possibility for strong dither noise, given the greater noise variance makes it less likely that data at different input quantization levels will map to the output level during the conversion process.
The adder 110 may include functionality for adding a dither noise component to YCbCr image data. In this example, the YCbCr image data is shown as being 10-bits, but in different embodiments, other bit lengths may be used. Once the dither has been added to the YCbCr image data, a quantizer 120 may reduce the number of quantization levels of the YCbCr data. In some instances, this reduction may occur through decimation, but in other embodiments, different reduction techniques may be used. In this example, the 10-bit YCbCr image data is shown as being reduced to 8-bit YCbCr data to be outputted on an 8-bit display, but in different embodiments other bit lengths may be used.
An encoder 130, which may be an 8-bit encoder if the YCbCr data is 8 bits, may then encode the YCbCr data for transmission to the display. A decoder 140 may decode the received transmitted data. A converter 150 may converted the decoded 8-bit YCbCr data to 8-bit RGB data for display on an 8-bit display device.
In box 202, at least (2n+1) dither values may be selected using a processing device. The selected dither values may be chosen so that they have a mean equal to that of the dither values associated with the n truncated bits and a variance greater than that of the dither values associated with the n truncated bits. In some instances, one or more of the dropped bit values may be included in the set of at least (2n+1) dither values selected in box 202. In some instances, the dropped bit values may be a subset of the values included in the set of at least (2n+1) dither values selected in box 202. The dither values associated with the n truncated bits may include a quantity of 2n or fewer dither values.
In some instances, the selected dither values in box 202 may include a set of 2n+1 dither values, so that if 2 bits are dropped, eight dither values may be selected in box 202, whereas in the prior art only four dither values were selected. The eight selected dither values may include −2, −1, 0, 1, 2, 3, 4, and 5 and the set of four prior art dither values may include values 0, 1, 2, and 3.
In box 203, at least one of the selected dither values may be applied to the image data before reducing the quantization levels of the image data using the processing device. The selected dither values may be scaled before being applied to the image data.
Quantization Error Diffusion
Another option for avoiding banding is to diffuse a quantization error among its neighboring pixels. This may be accomplished by calculating a quantization error that is caused by dropping bits during a requantization operation, such as when converting 10 bit image data to an 8 bit format used by a display, and then diffusing the quantization error into neighboring pixels. The quantization error may be diffused by scaling the error by a factor and then adding the scaled amount to the pixel values of respective neighboring pixels.
The quantization error may be calculated in the RGB space rather than in the YCbCr color space. The calculation may be performed in the RGB space to avoid banding artifacts in the RGB space. A quantization error calculated in the YCbCr space may be unable to prevent banding artifacts caused by the color space conversion.
The quantizer 320 may reduce a quantization level of YCbCr pixel data. The error calculation unit 370 may calculate RGB pixel values of the YCbCr pixel data before and after the quantization level is reduced, calculate the quantization error in a RGB color space from a difference between the before and the after RGB pixel values, and convert the RGB quantization error to a YCbCr color space. The diffusion unit 380 may incorporate the converted YCbCr quantization error in at least one neighboring pixel block. The converter 330 may convert pixel data between YCbCr and RGB color spaces. The second adder 360 may calculate the difference between the before and the after RGB pixel values.
The encoder 340 may encode an original pixel block. The decoder 350 may generate a reconstructed pixel block from the encoded pixel block data. The processing device 365 and/or adder 360 may calculate a difference between values of the original pixel block and the reconstructed pixel block and then applying an error function to the difference to calculate an error statistic. The diffusion unit 380 may incorporate the error statistic in at least one value of at least one neighboring pixel block to the original pixel block.
In an exemplary method, the quantization error may be first calculated in the RGB color space. The quantization error of R, G, and B channels may then be converted to the error of Y, Cb, and Cr channels by the color space conversion. Finally, the errors of Y, Cb, and Cr may be diffused to the Y, Cb, and Cr of neighboring pixels.
In box 420, RGB pixel values of the YCbCr pixel data may be calculated before and after reducing the quantization level using a processing device. For example, an 8-bit RGB (RGB—8 bit) value may be calculated from YCbCr—8 bit using the following equation (1) shown below, where M3×3 and N3×1 are respect user selected 3×3 and 3×1 matrices, some of which may be selected from a set of international standards. Similarly, a floating point RGB value (RGB_floating) may be calculated by applying equation (1) to the original data YCbCr—10 bit.
In box 430, a quantization error may be calculated in a RGB color space from a difference between the before and the after RGB pixel values. For example, a quantization error Error_RGB=RGB_floating−RGB—8 bit may be calculated in the RGB color space from the results in box 420.
In box 440, the RGB quantization error may be converted to a YCbCr color space using a processing device. For example, the quantization error Error_RGB may be converted back to the YCbCr color space (Error_YCbCr) using the following equation (2) shown below:
In box 450, the converted YCbCr quantization error may be incorporated in at least one neighboring pixel block.
Encoding Error Diffusion
As discussed previously, the lossy encoding process may reduce quantization levels in image areas with smooth color transition gradients. Error diffusion may be applied in the encoding loop in order to distribute a reconstruction error into one or more neighboring areas.
In box 520, values of the original pixel block and the reconstructed pixel block may be compared.
In box 530, an error function may be applied to a difference between the compared values of the original pixel block and the reconstructed pixel block to calculate an error statistic for block i (E_i) using the processing device. For example, error statistic E_i may be computed from the coding noise by applying an error function ƒ(orig_block_i−rec_block_i) to the difference between the original values block and the reconstructed block for the respective block:
E—i=ƒ(orig_block—i−rec_block—i) (3)
In box 540, the error statistic may be incorporated in at least one value of at least one neighboring pixel block to the original pixel block. A neighboring pixel block may include any pixel block within a predetermined vicinity of the original pixel block. For example, the error may be distributed into one or more subsequent neighboring blocks (block_j) according to the function:
block—j=block—j+wi,j·g(E—i) (4)
The function g(E_i) may generate a compensating signal such that ƒ(g(E_i))≈E_i. For example, in an embodiment E_i may be an average and function ƒ( ) may compute a mean or average. In this embodiment, function g( ) may simply generate a block with identical values. In another embodiment E_i may be a transform coefficient and function ƒ( ) may compute a specific transform coefficient. In this embodiment, function g( ) may compute the corresponding inverse transform. In yet another embodiment E_i may be the n-th moment, and function ƒ( ) may compute the moment. In this embodiment, function g( ) may be an analytical generating function. The above algorithms for functions ƒ( ) and g( ) may also be applied in instances involving multiple statistics, such as when E_i is a vector of multiple statistics.
The diffusion coefficients wi,j may determine the distribution of error E_i to each neighboring block_j. In one embodiment, coefficients wi,j may be a fixed set of numbers. Coefficients wi,j may also vary depending on the sizes of block_i and/or block_j. In other instances, coefficients wi,j may vary depending on the spatial connectivity between block_i and block_j,
Coefficients wi,j may also vary based on a detection map indicating whether a block_i and/or block_j are part of an area subject to banding. If the two blocks are not in a similar area subject to banding, the coefficients wi,j for those blocks may be set to smaller values or zeroed out.
Coefficients wi,j may also be determined depending on the amount of texture and/or perceptual masking in the neighborhood or vicinity of block_i and/or block_j. If the neighborhood, is highly textured and/or has a high masking value, the coefficients wi,j may be set to smaller values or zeroed out. A perceptual mask may indicate how easily a loss of signal content may be observed when viewing respective blocks of an image.
Coefficients wi,j may be determined based on a relationship between original block values orig_block_i and orig_block_j. For example, coefficients wi,j may be lowered when the mean of the two blocks are far apart.
In some instances, the sum of all coefficients wi,j for a given block_i may equal one but need not equal one. In some instances, factors other than one may be used. Each of the blocks, such as block_i and block_j, may be defined as a coding unit, a prediction unit or a transform unit. In different instances, different criteria or considerations may be used to determine how the error will be diffused.
For example, in an embodiment the diffusion may be associated with a block size selection, where the amount of diffusion as well as the block size used are controlled by spatial gradients or detection maps indicating whether a particular block is part of a banding area. An example of this is shown in
In another embodiment the diffusion may only be carried out on neighboring transform units with small transform sizes, or on transform units within a unit, such as a coding or prediction unit.
In each of these instances, an error may be diffused among neighboring coding blocks. However, the same diffusion principle may also be applied within a given coding block. Several exemplary embodiments are described below for diffusion within a transform block. For example, after transform encoding and decoding, a reconstruction error of a given pixel may be diffused to neighboring pixels within the same transform unit. The diffusion process and encoding process may be iterative so that a diffusion also modifies the original signal before further encoding. Diffusion may also be carried out in the transform domain, where a quantization error may be assigned transform coefficient that are diffused to other transform coefficients.
The foregoing description has been presented for purposes of illustration and description. It is not exhaustive and does not limit embodiments of the invention to the precise forms disclosed. Modifications and variations are possible in light of the above teachings or may be acquired from the practicing embodiments consistent with the invention. For example, some of the described embodiments may refers to converting 10-bit YCbCr image data to 8-bit RGB image data, however other embodiments may convert different types of image data between different bit sizes.
The present application claims the benefit of U.S. Provisional Application Ser. No. 61/677,387 filed Jul. 30, 2012, entitled “ERROR DIFFUSION WITH COLOR CONVERSION AND ENCODING.” The aforementioned application is incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
6441867 | Daly | Aug 2002 | B1 |
6654887 | Rhoads | Nov 2003 | B2 |
7411707 | Ikeda | Aug 2008 | B2 |
20110164829 | Moribe | Jul 2011 | A1 |
20110285714 | Swic et al. | Nov 2011 | A1 |
20120050815 | Kodama et al. | Mar 2012 | A1 |
Number | Date | Country | |
---|---|---|---|
20140029846 A1 | Jan 2014 | US |
Number | Date | Country | |
---|---|---|---|
61677387 | Jul 2012 | US |