This disclosure generally relates to artificial reality, such as virtual reality and augmented reality.
Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured content (e.g., real-world photographs). The artificial reality content may include video, audio, haptic feedback, or some combination thereof, and any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Artificial reality may be associated with applications, products, accessories, services, or some combination thereof, that are, e.g., used to create content in an artificial reality and/or used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head-mounted display (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
Particular embodiments described herein relate to a method of using multiple dithering masks to generate spatio-temporal subframe images with less gray level bits (or color depth) to represent a target image with more gray level bits, without using an error buffer. The temporal subframe images may have smooth dithering pattern transition between grayscale levels and minimal temporal change among the subframes. For a target region (e.g., a tile region) of the target image, the system may generate a dithering mask for each subframe image. Each dithering mask may include a dot pattern with blue-noise distribution and satisfy spatial-stacking constraints. The dot pattern may include a number of stacked dot patterns with each dot pattern having a dot density corresponding to a grayscale level within the quantization range (e.g., 0-255 grayscale levels for 8-bit display). All dot patterns may be chosen to have blue-noise properties and may have a spatial stacking property according to which the dot pattern for grayscale level N+1 may include the dot pattern for all lower grayscale levels of 0 to N. Each dot in the dithering mask may correspond to a threshold value which equals to the lowest grayscale level for turning on that that dot (i.e., the lowest grayscale level with the corresponding dot pattern that includes that dot).
In particular embodiments, for representing a target grayscale value g (e.g., an average grayscale value of a target tile region), the dot patterns corresponding to all lower grayscale levels may be spatially stacked to represent the target grayscale level up to a distribution limit gz, (e.g., the dot pattern of the target grayscale level may include all dots of lower grayscale levels). The distribution limit gL, (e.g., 0.25) may be determined by dividing the maximum grayscale level (e.g., 1) by the number of subframes (e.g., 4 subframes). Under the condition of g<gL, the dot pattern of each dithering mask of each subframe may include a subset dots with have no overlapping dots with any other subframes. For presenting grayscale higher than the distribution limit gz, (e.g., g>0.25), additional dots could be incrementally added and turned on. To ensure temporal consistency, the dots that are incrementally added may be selected from the dots that are included in one or more dithering masks of the other subframes. For example, for quantizing grayscale between the distribution limit gz, to two times of distribution limit 2gL (e.g., 0.25<g<0.5), the dots added to the first subframe (which is at grayscale 0.25) may be incrementally selected from the dots included in the dithering mask of the second subframe. As another example, for quantizing grayscale in the range of two times of distribution limit 2gL to three times of distribution limit 3gL (e.g., 0.5<g<0.75), the dots to be turned on may include dots of the first and second subframe dithering masks (e.g., which are both at grayscale 0.25). The dots that are incrementally added may be selected from the dots included in the dithering mask of third subframe. The dither masks for generating the subframe images may be pre-determined and may be available for use when needed by the process for generating the subframe images. Therefore, all subframe images may be generated at essential the same time (or generated parallelly) and the quantization errors may be dithered in the temporal domain to other subframes during the subframe image generating process. Therefore, the system may not need to store the quantization error for temporal dithering process to other subframes. As a result, using the dithering masks generated following these principles, the temporal subframe images may be generated without using an error buffer, and therefore reduce the memory usage related to the subframe image generating processes. The subframe images may have smooth dither pattern transition between grayscales and minimal temporal change among the subframes.
In particular embodiments, the multiple masks used to generate the subframes may be generated from a single seed mask stored in the computer storage. The system may store the single seed mask instead of the multiple dithering masks to reduce the storage memory usage related to the subframe generating process. For an arbitrary number of subframes N, the mask for the n-th subframe may be generated by cyclically permuting the seed mask. For a target grayscale level g, the system may determine an offsetting coefficient kn based on a remainder of (n−1)·g divided by gmax, which is the maximum grayscale level. Then, the system may determine the threshold values of a subsequent subframe mask based on a remainder of (t1-kn) divided by gmax, where t1 is a threshold value for a dot in the first mark. As an example, for a target grayscale of 0.25 in a grayscale range of [0, 1] and 4 subframes, the first, second, third and fourth subframe masks may include the dots that have the threshold values within the ranges of [0, 0.25], [0.25, 0.5], [0.5, 0.75] and [0.75, 1], respectively. The threshold values of the first, second, third and fourth subframe dithering masks may be determined by mod(t1−0), 1), mod(t1−0.25, 1), mod(t1−0.5, 1) and mod(t1−0.75, 1), respectively. As another example, for a target grayscale of 0.6 within a grayscale range of [0, 1] and 4 subframes, the first, second, third and fourth subframe masks may include the dots that have the threshold values within the ranges of [0, 0.6], [0.2, 0.8], [0.4, 1] and [0.2, 0.8], respectively. The threshold values of the first, second, third and fourth subframe masks may be determined by mod(t1−0), 1), mod(t1−0.2, 1), mod(t1−0.3, 1) and mod(t1−0.2, 1), respectively.
The embodiments disclosed herein are only examples, and the scope of this disclosure is not limited to them. Particular embodiments may include all, some, or none of the components, elements, features, functions, operations, or steps of the embodiments disclosed above. Embodiments according to the invention are in particular disclosed in the attached claims directed to a method, a storage medium, a system and a computer program product, wherein any feature mentioned in one claim category, e.g. method, can be claimed in another claim category, e.g. system, as well. The dependencies or references back in the attached claims are chosen for formal reasons only. However, any subject matter resulting from a deliberate reference back to any previous claims (in particular multiple dependencies) can be claimed as well, so that any combination of claims and the features thereof are disclosed and can be claimed regardless of the dependencies chosen in the attached claims. The subj ect-matter which can be claimed comprises not only the combinations of features as set out in the attached claims but also any other combination of features in the claims, wherein each feature mentioned in the claims can be combined with any other feature or combination of other features in the claims. Furthermore, any of the embodiments and features described or depicted herein can be claimed in a separate claim and/or in any combination with any embodiment or feature described or depicted herein or with any of the features of the attached claims.
In an embodiment, a method may comprise, by a computing system:
The dots of each mask may be associated with a dot pattern, the dot pattern may comprise a plurality of stacked dot patterns, and each of the plurality of stacked dot patterns may satisfy a spatio stacking constraint by comprising all dot patterns corresponding to all lower grayscale levels.
Each dot of the dot pattern may be associated with a threshold value, and the threshold value may correspond to a lowest grayscale level which has a corresponding dot pattern that comprises that dot.
Each mask may have threshold values corresponding to all grayscale levels of a quantization grayscale range corresponding to the second number of bits per color.
The plurality of stacked dot patterns may correspond to all grayscale levels of the quantization grayscale range.
The dots in the dot pattern of each mask may have a blue-noise property.
A sum of the dot patterns of the masks may have a blue-noise property.
The plurality of images may be generated by satisfying a temporal stacking constraint, and the temporal stacking constraint may allow the plurality of images to have the luminosity within a threshold range.
The display may have the second number of bits per color.
In an embodiment, the masks may be available at a same time for a process of generating the plurality of images, and a method may comprise:
In an embodiment, a method may comprise:
The quantization grayscale range may have a plurality of evenly-placed grayscale levels.
The quantization grayscale range may have a plurality of unevenly-placed grayscale levels.
In an embodiment, a method may comprise:
When a target grayscale value associated with the target image is smaller than the grayscale limit, corresponding regions of the plurality of images may comprise non-overlapping sets of pixels from each other.
When a target grayscale value associated with the target image is greater than the grayscale limit, corresponding regions of the plurality of images may comprise overlapping sets of pixels, and wherein the overlapping sets of pixels are determined by incrementally selecting dots from at least another mask of the masks.
An average grayscale value of a target region of the target image may be used as a target grayscale value, and each of the plurality of masks may have a same size to the target region of the target image.
The plurality of images may be generated by repeatedly applying a corresponding mask to the target image.
In an embodiment, one or more computer-readable non-transitory storage media may embody software that is operable when executed to:
In an embodiment, a system may comprise: one or more non-transitory computer-readable storage media embodying instructions; and one or more processors coupled to the storage media and operable to execute the instructions to:
The number of available bits in a display may limit the display's color depth or gray scale level. Displays with limited color depth or gray scale level may use spatio dithering to generate the illusion of increased color depth or gray scale level, for example, by spreading quantization errors to neighboring pixels. To further increase the color depth or gray scale level, displays may generate a series of temporal subframe images with less gray level bits to give the illusion of a target image which has more gray level bits. Each subframe image may be generated using dithering techniques (e.g., spatio-temporal dithering methods). However, these dithering techniques may need an error buffer to provide temporal feedback, and therefore use more memory space.
To reduce the memory usage related to processes of generating subframe images, particular embodiments of the system may use a number of dithering masks to generate a series of subframe images with even luminance distribution across all subframe images to represent a target image. For generating N subframe images, the system may generate a dithering mask for each subframe image. Each dithering mask may include a number of dot patterns with each dot pattern having a dot density corresponding to a grayscale level within the quantization range (e.g., 0-255 grayscale levels for 8-bit display). The dot patterns may be generated based on blue-noise distribution and satisfy spatial stacking property. For example, the dot pattern for grayscale level N may include the dot patterns for all lower grayscale levels form 0 to N. The dithering mask may include the dot patterns corresponding to all grayscale levels of the quantization range. Each dot in the dithering mask may correspond to a threshold value which equals to the lowest grayscale level allowing that dot to be included in a dot pattern. The system may generate the subframe images based on the dithering masks without using an error buffer.
Particular embodiments of the system improve the efficiency of AR/VR display by reducing the memory usage related to generating the temporal subframe images without using an error buffer. Particular embodiments of the system provide better image quality and improve user experience for AR/VR display by using multiple subframe images with less color depth to represent an image with greater color depth. Particular embodiments of the system generate subframe images with more even luminance distribution across the subframe images for representing the target image and eliminate the temporal artifacts such as flashes or uneven luminance over time in AR/VR display when the user's eyes and head positions change between the subframe images. Particular embodiments of the system allow AR/VR display system to reduce the space and complexity of pixel circuits by having less gray level bits, and therefore miniaturize the size of the display system. Particular embodiments of the system make it possible for AR/VR displays to operate in monochrome mode with digital pixel circuits and eliminating analog pixel circuits for full RGB operations.
In particular embodiments, the display engine 130 may include a controller block (not shown). The control block may receive data and control packages such as position data and surface information from controllers external to the display engine 130 though one or more data buses. For example, the control block may receive input stream data from a body wearable computing system. The input data stream may include a series of mainframe images generated at a mainframe rate of 30-90 Hz. The input stream data including the mainframe images may be converted to the required format and stored into the texture memory 132. In particular embodiments, the control block may receive input from the body wearable computing system and initialize the graphic pipelines in the display engine to prepare and finalize the image data for rendering on the display. The data and control packets may include information related to, for example, one or more surfaces including texel data, position data, and additional rendering instructions. The control block may distribute data as needed to one or more other blocks of the display engine 130. The control block may initiate the graphic pipelines for processing one or more frames to be displayed. In particular embodiments, the graphic pipelines for the two eye display systems may each include a control block or share the same control block.
In particular embodiments, the transform block 133 may determine initial visibility information for surfaces to be displayed in the artificial reality scene. In general, the transform block 133 may cast rays from pixel locations on the screen and produce filter commands (e.g., filtering based on bilinear or other types of interpolation techniques) to send to the pixel block 134. The transform block 133 may perform ray casting from the current viewpoint of the user (e.g., determined using the headset's inertial measurement units, eye tracking sensors, and/or any suitable tracking/localization algorithms, such as simultaneous localization and mapping (SLAM)) into the artificial scene where surfaces are positioned and may produce tile/surface pairs 144 to send to the pixel block 134. In particular embodiments, the transform block 133 may include a four-stage pipeline as follows. A ray caster may issue ray bundles corresponding to arrays of one or more aligned pixels, referred to as tiles (e.g., each tile may include 16x 16 aligned pixels). The ray bundles may be warped, before entering the artificial reality scene, according to one or more distortion meshes. The distortion meshes may be configured to correct geometric distortion effects stemming from, at least, the eye display systems the headset system. The transform block 133 may determine whether each ray bundle intersects with surfaces in the scene by comparing a bounding box of each tile to bounding boxes for the surfaces. If a ray bundle does not intersect with an object, it may be discarded. After the tile-surface intersections are detected, the corresponding tile/surface pairs may be passed to the pixel block 134.
In particular embodiments, the pixel block 134 may determine color values or grayscale values for the pixels based on the tile-surface pairs. The color values for each pixel may be sampled from the texel data of surfaces received and stored in texture memory 132. The pixel block 134 may receive tile-surface pairs from the transform block 133 and may schedule bilinear filtering using one or more filer blocks. For each tile-surface pair, the pixel block 134 may sample color information for the pixels within the tile using color values corresponding to where the projected tile intersects the surface. The pixel block 134 may determine pixel values based on the retrieved texels (e.g., using bilinear interpolation). In particular embodiments, the pixel block 134 may process the red, green, and blue color components separately for each pixel. In particular embodiments, the display may include two pixel blocks for the two eye display systems. The two pixel blocks of the two eye display systems may work independently and in parallel with each other. The pixel block 134 may then output its color determinations (e.g., pixels 138) to the display block 135. In particular embodiments, the pixel block 134 may composite two or more surfaces into one surface to when the two or more surfaces have overlapping areas. A composed surface may need less computational resources (e.g., computational units, memory, power, etc.) for the resampling process.
In particular embodiments, the display block 135 may receive pixel color values from the pixel block 134, covert the format of the data to be more suitable for the scanline output of the display, apply one or more brightness corrections to the pixel color values, and prepare the pixel color values for output to the display. In particular embodiments, the display block 135 may each include a row buffer and may process and store the pixel data received from the pixel block 134. The pixel data may be organized in quads (e.g., 2×2 pixels per quad) and tiles (e.g., 16×16 pixels per tile). The display block 135 may convert tile-order pixel color values generated by the pixel block 134 into scanline or row-order data, which may be required by the physical displays. The brightness corrections may include any required brightness correction, gamma mapping, and dithering. The display block 135 may output the corrected pixel color values directly to the driver of the physical display (e.g., pupil display) or may output the pixel values to a block external to the display engine 130 in a variety of formats. For example, the eye display systems of the headset system may include additional hardware or software to further customize backend color processing, to support a wider interface to the display, or to optimize display speed or fidelity.
In particular embodiments, the dithering methods and processes (e.g., spatial dithering method, temporal dithering methods, and spatio-temporal methods) as described in this disclosure may be embodied or implemented in the display block 135 of the display engine 130. In particular embodiments, the display block 135 may include a model-based dithering algorithm or a dithering model for each color channel and send the dithered results of the respective color channels to the respective display driver ICs (e.g., 142A, 142B, 142C) of display system 140. In particular embodiments, before sending the pixel values to the respective display driver ICs (e.g., 142A, 142B, 142C), the display block 135 may further include one or more algorithms for correcting, for example, pixel non-uniformity, LED non-ideality, waveguide non-uniformity, display defects (e.g., dead pixels), etc.
In particular embodiments, graphics applications (e.g., games, maps, content-providing apps, etc.) may build a scene graph, which is used together with a given view position and point in time to generate primitives to render on a GPU or display engine. The scene graph may define the logical and/or spatial relationship between objects in the scene. In particular embodiments, the display engine 130 may also generate and store a scene graph that is a simplified form of the full application scene graph. The simplified scene graph may be used to specify the logical and/or spatial relationships between surfaces (e.g., the primitives rendered by the display engine 130, such as quadrilaterals or contours, defined in 3D space, that have corresponding textures generated based on the mainframe rendered by the application). Storing a scene graph allows the display engine 130 to render the scene to multiple display frames and to adjust each element in the scene graph for the current viewpoint (e.g., head position), the current object positions (e.g., they could be moving relative to each other) and other factors that change per display frame. In addition, based on the scene graph, the display engine 130 may also adjust for the geometric and color distortion introduced by the display subsystem and then composite the objects together to generate a frame. Storing a scene graph allows the display engine 130 to approximate the result of doing a full render at the desired high frame rate, while actually running the GPU or display engine 130 at a significantly lower rate.
In particular embodiments, the graphic pipeline 100D may include a resampling step 153, where the display engine 130 may determine the color values from the tile-surfaces pairs to produce pixel color values. The resampling step 153 may be performed by the pixel block 134 in
In particular embodiments, the graphic pipeline 100D may include a bend step 154, a correction and dithering step 155, a serialization step 156, etc. In particular embodiments, the bend step, correction and dithering step, and serialization steps of 154, 155, and 156 may be performed by the display block (e.g., 135 in
In particular embodiments, the optics system 214 may include a light combining assembly, a light conditioning assembly, a scanning mirror assembly, etc. The light source assembly 210 may generate and output an image light 219 to a coupling element 218 of the output waveguide 204. The output waveguide 204 may be an optical waveguide that could output image light to the user eye 202. The output waveguide 204 may receive the image light 219 at one or more coupling elements 218 and guide the received image light to one or more decoupling elements 206. The coupling element 218 may be, for example, but is not limited to, a diffraction grating, a holographic grating, any other suitable elements that can couple the image light 219 into the output waveguide 204, or a combination thereof. As an example and not by way of limitation, if the coupling element 350 is a diffraction grating, the pitch of the diffraction grating may be chosen to allow the total internal reflection to occur and the image light 219 to propagate internally toward the decoupling element 206. The pitch of the diffraction grating may be in the range of 300 nm to 600 nm. The decoupling element 206 may decouple the total internally reflected image light from the output waveguide 204. The decoupling element 206 may be, for example, but is not limited to, a diffraction grating, a holographic grating, any other suitable element that can decouple image light out of the output waveguide 204, or a combination thereof. As an example and not by way of limitation, if the decoupling element 206 is a diffraction grating, the pitch of the diffraction grating may be chosen to cause incident image light to exit the output waveguide 204. The orientation and position of the image light exiting from the output waveguide 204 may be controlled by changing the orientation and position of the image light 219 entering the coupling element 218. The pitch of the diffraction grating may be in the range of 300 nm to 600 nm.
In particular embodiments, the output waveguide 204 may be composed of one or more materials that can facilitate total internal reflection of the image light 219. The output waveguide 204 may be composed of one or more materials including, for example, but not limited to, silicon, plastic, glass, polymers, or some combination thereof. The output waveguide 204 may have a relatively small form factor. As an example and not by way of limitation, the output waveguide 204 may be approximately 50 mm wide along X-dimension, 30 mm long along Y-dimension and 0.5-1 mm thick along Z-dimension. The controller 216 may control the scanning operations of the light source assembly 210. The controller 216 may determine scanning instructions for the light source assembly 210 based at least on the one or more display instructions for rendering one or more images. The display instructions may include an image file (e.g., bitmap) and may be received from, for example, a console or computer of the AR/VR system. Scanning instructions may be used by the light source assembly 210 to generate image light 219. The scanning instructions may include, for example, but are not limited to, an image light source type (e.g., monochromatic source, polychromatic source), a scanning rate, a scanning apparatus orientation, one or more illumination parameters, or some combination thereof. The controller 216 may include a combination of hardware, software, firmware, or any suitable components supporting the functionality of the controller 216.
In particular embodiments, the image field 227 may receive the light 226A-B as the mirror 224 rotates about the axis 225 to project the light 226A-B in different directions. For example, the image field 227 may correspond to a portion of the coupling element 218 or a portion of the decoupling element 206 in
In particular embodiments, the light emitters 222 may illuminate a portion of the image field 227 (e.g., a particular subset of multiple pixel locations 229 on the image field 227) with a particular rotation angle of the mirror 224. In particular embodiment, the light emitters 222 may be arranged and spaced such that a light beam from each of the light emitters 222 is projected on a corresponding pixel location 229. In particular embodiments, the light emitters 222 may include a number of light-emitting elements (e.g., micro-LEDs) to allow the light beams from a subset of the light emitters 222 to be projected to a same pixel location 229. In other words, a subset of multiple light emitters 222 may collectively illuminate a single pixel location 229 at a time. As an example and not by way of limitation, a group of light emitter including eight light-emitting elements may be arranged in a line to illuminate a single pixel location 229 with the mirror 224 at a given orientation angle.
In particular embodiments, the number of rows and columns of light emitters 222 of the light source 220 may or may not be the same as the number of rows and columns of the pixel locations 229 in the image field 227. In particular embodiments, the number of light emitters 222 in a row may be equal to the number of pixel locations 229 in a row of the image field 227 while the light emitters 222 may have fewer columns than the number of pixel locations 229 of the image field 227. In particular embodiments, the light source 220 may have the same number of columns of light emitters 222 as the number of columns of pixel locations 229 in the image field 227 but fewer rows. As an example and not by way of limitation, the light source 220 may have about 1280 columns of light emitters 222 which may be the same as the number of columns of pixel locations 229 of the image field 227, but only a handful rows of light emitters 222. The light source 220 may have a first length L1 measured from the first row to the last row of light emitters 222. The image field 530 may have a second length L2, measured from the first row (e.g., Row 1) to the last row (e.g., Row P) of the image field 227. The L2 may be greater than L1 (e.g., L2 is 50 to 10,000 times greater than L1).
In particular embodiments, the number of rows of pixel locations 229 may be larger than the number of rows of light emitters 222. The display device 200B may use the mirror 224 to project the light 223 to different rows of pixels at different time. As the mirror 520 rotates and the light 223 scans through the image field 227, an image may be formed on the image field 227. In some embodiments, the light source 220 may also has a smaller number of columns than the image field 227. The mirror 224 may rotate in two dimensions to fill the image field 227 with light, for example, using a raster-type scanning process to scan down the rows then moving to new columns in the image field 227. A complete cycle of rotation of the mirror 224 may be referred to as a scanning period which may be a predetermined cycle time during which the entire image field 227 is completely scanned. The scanning of the image field 227 may be determined and controlled by the mirror 224 with the light generation of the display device 200B being synchronized with the rotation of the mirror 224. As an example and not by way of limitation, the mirror 224 may start at an initial position projecting light to Row 1 of the image field 227, and rotate to the last position that projects light to Row P of the image field 227, and then rotate back to the initial position during one scanning period. An image (e.g., a frame) may be formed on the image field 227 per scanning period. The frame rate of the display device 200B may correspond to the number of scanning periods in a second. As the mirror 224 rotates, the light may scan through the image field to form images. The actual color value and light intensity or brightness of a given pixel location 229 may be a temporal sum of the color various light beams illuminating the pixel location during the scanning period. After completing a scanning period, the mirror 224 may revert back to the initial position to project light to the first few rows of the image field 227 with a new set of driving signals being fed to the light emitters 222. The same process may be repeated as the mirror 224 rotates in cycles to allow different frames of images to be formed in the scanning field 227.
The coupling area 330 may include coupling elements (e.g., 334A, 334B, 334C) configured and dimensioned to couple light of predetermined wavelengths (e.g., red, green, blue). When a white light emitter array is included in the projector device 350, the portion of the white light that falls in the predetermined wavelengths may be coupled by each of the coupling elements 334A-C. In particular embodiments, the coupling elements 334A-B may be gratings (e.g., Bragg gratings) dimensioned to couple a predetermined wavelength of light. In particular embodiments, the gratings of each coupling element may exhibit a separation distance between gratings associated with the predetermined wavelength of light and each coupling element may have different grating separation distances. Accordingly, each coupling element (e.g., 334A-C) may couple a limited portion of the white light from the white light emitter array of the projector device 350 if white light emitter array is included in the projector device 350. In particular embodiments, each coupling element (e.g., 334A-C) may have the same grating separation distance. In particular embodiments, the coupling elements 334A-C may be or include a multiplexed coupler.
As illustrated in
In particular embodiments, the AR/VR system may use scanning waveguide displays or 2D micro-LED displays for displaying AR/VR content to users. In order to miniaturize the AR/VR system, the display system may need to miniaturize the space for pixel circuits and may have limited number of available bits for the display. The number of available bits in a display may limit the display's color depth or gray scale level, and consequently limit the quality of the displayed images. Furthermore, the waveguide displays used for AR/VR systems may have nonuniformity problem cross all display pixels. The compensation operations for pixel nonuniformity may result in loss on image grayscale and further reduce the quality of the displayed images. For example, a waveguide display with 8-bit pixels (i.e., 256 gray level) may equivalently have 6-bit pixels (i.e., 64 gray level) after compensation of the nonuniformity (e.g., 8:1 waveguide nonuniformity, 0.1% dead micro-LED pixel, and 20% micro-LED intensity nonuniformity).
To improve the displayed image quality, displays with limited color depth or gray scale level may use spatio dithering to spread quantization errors to neighboring pixels and generate the illusion of increased color depth or gray scale level. To further increase the color depth or gray scale level, displays may generate a series of temporal subframe images with less gray level bits to give the illusion of a target image which has more gray level bits. Each subframe image may be dithered using spatio dithering techniques within that subframe image. The average of the series of subframe image may correspond to the image as perceived by the viewer. For example, for display an image with 8-bit pixels (i.e., 256 gray level), the system may use four subframe images each having 6-bit pixels (i.e., 64 gray level) to represent the 8-bit target image. As another example, an image with 8-bit pixels (i.e., 256 gray level) may be represented by 16 subframe images each having 4-bit pixels (i.e., 16 gray level). This would allow the display system to render images of more gray level (e.g., 8-bit pixels) with pixel circuits and supporting hardware for less gray level (e.g., 6-bit pixels or 4-bit pixels), and therefore reduce the space and size of the display system.
However, using this segmented quantization and spatio dithering method, even though the average luminance of the all subframe images over time is approximately equal to the target image, the subframes 400B-D may have very different luminance, as illustrated in
To solve the problem of un-even luminance of subframe images, particular embodiments of the system may use a spatio-temporal dithering method to generate a series of subframe images for representing a target image with more even luminance distribution across all subframe images. The spatio-temporal dithering method may dither quantization errors both spatially to neighboring pixels of the same subframe image and temporally to the corresponding pixel of next subframe image of the series of subframe images. The temporally dithered quantization error of a pixel of a subframe image may be dithered to the corresponding pixel in the next subframe image of the series of subframe images in the time domain. The system may generate each subframe image using spatial-temporal dithering methods. However, these dithering methods may need an error buffer to provide temporal feedback, and therefore use more memory. To reduce the memory usage related to processes of generating subframe images, particular embodiments of the system may use a number of dithering masks to generate the series of subframe images for representing a target image with even luminance distribution across all subframe images. The system may generate the subframe images using the corresponding dithering masks by comparing target grayscale values the threshold values of the corresponding dithering masks and dithering the quantization errors to other subframes without using an error buffer, as will be described in detail in later sections of this disclosure.
In particular embodiments, each dot in the dithering mask may correspond to a threshold value which equals to the lowest grayscale level allowing that dot to be turned on (i.e., the lowest grayscale level whose corresponding dot pattern includes that dot). From the lowest grayscale level to the highest grayscale level, once a dot is turned on (i.e., being included in a dot pattern of a grayscale level), the dot may stay the turn-on state for all higher grayscale levels (i.e., being included in the dot patterns of all higher grayscale levels). The spatio stacking properties of the dot patterns may allow all dot patterns to be encoded into one dithering mask. In particular embodiments, the dithering mask (e.g., 500A in
In particular embodiments, for the quantization process, the system may compare a target grayscale value g to the threshold values associated with the dots in the dithering mask and determine the quantized grayscale value. For example, the system may select the closest threshold value (i.e., the closest grayscale level within the quantization range) in the dithering mask as the quantized grayscale value for the target grayscale value. Then, the system may determine the quantization error by comparing the quantized grayscale value with the target scale value. The system may dither the quantization error spatially to the nearby pixels or regions (e.g., tile regions) of the same subframe or/and temporally to the corresponding pixels or regions of other temporal subframes (e.g., corresponding pixels or region of the next subframe image). The system may determine display grayscale value based on the quantized grayscale value of the target grayscale value and the dithered quantization error to the pixels of the corresponding tile region the target grayscale value (e.g., dithered from the former subframe image or from neighboring tile region of the same subframe image). The system may a dot pattern corresponding to a grayscale level that is most close to the display grayscale value and use the selected dot pattern to represent the target grayscale value.
In particular embodiments, the process for dithering the quantization error may need an error buffer for propagating the quantization errors to other subframes. For example, the series of subframe images may be generated in a sequential order (e.g., from 1 to N). The quantization errors of the subframe n may be stored in an error buffer or a frame buffer (e.g., the same size as the subframe image) and may be dithered to the subframe n+1 during the generating process of the subframe n+1. It is notable that particular embodiments of the mask-based dithering method, as described in this disclosure, may not need to use the error buffer for prorogating the quantization errors to other subframes. Instead, for generating N number of subframe images, the system may generate N number of dithering masks. The dithering masks may be pre-determined or pre-generated (e.g., during an offline process) before the process for generating the subframe images. All N number of dithering masks may be available for being used to generate the subframe images at the same time. The system may use the N number of dithering masks to generate the N number of subframe images parallelly or/and essentially at the same time. The quantization errors of a subframe can be dithered to next subframe during the parallel subframe generating processes since they could be parallel or essentially at the same time. The temporal dithering process for dithering to next subframe and the spatial dithering process for dithering to neighboring pixels may be performed at essentially the same time during the subframe generating process. The system may not need to store the quantization errors in the error buffer, and therefore may reduce the memory and power usage during the subframe generating process. For example, the error buffer-based method may need an error buffer with a memory size of 6.6 Mbytes for the temporal dithering process for subframe images with 2560×1792 resolution, 4-bit grayscale level, and 3 color channels of RGB. The mask-based method may eliminate the needs for the error buffer or frame buffer, and therefore have less memory usage for the dithering process. Although the mask-based method may need N coupled dithering masks, the same dithering masks may be used for the three color channels of R, G and B. For example, for the subframe number N=16, the system may need 130 Kbytes memory for storing 16 subframe masks with each mask having 128×128 resolution and 4-bit grayscale levels (i.e., 130 Kbytes=128×128×4 bits×16 subframes), which is muss less than the error buffer size of 6.6 Mbytes. As another example, for the subframe number N=32, the system may need 260 Kbytes memory for storing 32 subframe masks with each mask having 128×128 resolution and 4-bit grayscale levels (i.e., 260 Kbytes=128×128×4 bits×32 subframes), which is muss less than the error buffer size of 6.6 Mbytes. Therefore, the mask-based dithering method may reduce the memory usage of the dithering process by 5-10 times comparing to the error buffer-based methods. In addition, the error buffer or frame buffer may need to be both writable and readable memory for the dithering operation to store and access the quantization errors during the dithering process. However, the mask-based dithering method may have fixed threshold values once after they are firstly determined and may store the masks (with fixed threshold values) in read-only memory units, which further improve the efficiency of the memory usage and reduce the power usage.
In particular embodiments, for generating N subframe images, the system may generate N coupled blue noise dithering masks that simultaneously satisfy both spatial and temporal stacking constraints and use the dithering masks to generate the subframe images. By using this approach, the system may generate subframes with high quality dithering and reduce the temporal integrated noise of the rendered images by 1/N2. This approach may have two attributes of spatial stacking property and temporal stacking property, simultaneously. The spatial stacking property, according to which the dot pattern for grayscale N may include all the dot patterns from grayscales 0 to N−1, may allow the generated subframes to have smooth dither pattern transition between grayscale levels. The temporal stacking property, according to which grayscale N rendered by the first subframe and grayscale N+1 rendered by the combination of the first subframe and second subframe may obey the stacking property, may allow the subframes to have minimal temporal change from one subframe to another subframe. All these advantages could be achieved without using an error buffer. Alternatively, the system may use N independently generated dithering masks for generating the N subframes with no temporal stacking property. However, the temporal integrated noise may only be reduced by 1/N by using the N independently generated dithering masks instead of using the N coupled blue noise dithering masks.
In particular embodiments, the system may use N dithering masks to generate N subframes for representing a target image (N can be any integer number). For determining a dot pattern for a target grayscale value (e.g., an average grayscale value a target tile region), the dot patterns corresponding to all lower grayscale levels may be spatially stacked to represent the target grayscale value. In particular embodiments, when the target grayscale value is below or equal to a grayscale limit gL (i.e., gL=gmax/N, where gmax is the maximum grayscale level and N is the number of subframes), the system may generate N dithering masks which have no overlapping dots between any two dithering masks. In other words, each dithering mask may include a different set of dots from any other dithering masks and the dots in all dithering masks when stacked together may correspond to all the pixels of the target image. Each dithering mask may include a dot pattern corresponding to the grayscale limit gL. The system may consequently use the N non-overlapping dithering masks to generate N subframes which have no overlapping pixels between any two subframes. Therefore, the subframes may have a temporal stacking property which allows that all the pixels of the N subframes once stacked together to correspond to all the pixels in the target image. As a result, the subframes may have more even luminance and more uniform display results for representing the target image.
In particular embodiments, the system may determine a grayscale limit gL for non-overlapping dithering masks by dividing the maximum grayscale level (e.g., 1) within a grayscale range (e.g., [0, 1]) by the number of subframes N (i.e., gL=gmax/N, where gmax is the maximum grayscale level). For example, with the maximum grayscale level of 1 and the number of subframes of 4, the grayscale limit for non-overlapping dithering masks may be determined to 0.25. As another example, with the maximum grayscale level of 1 and the number of subframe of 10, the grayscale limit for non-overlapping dithering mask may be determined to be 0.1. As another example, with the maximum grayscale level of 1 and the number of subframe of 16, the grayscale limit for non-overlapping dithering mask may be determined to be 1/16.
G1(G2∪G3∪G4) (1)
G2(G1∪G3∪G4) (2)
G3(G1∪G2∪G4) (3)
G4(G1∪G2∪G3) (4)
(G1∪G2∪G3∪G4)=1 (5)
where G1, G2, G3, and G4are the dot set of the first, second, third and fourth dithering mask, respectively. In other words, each dithering mask may have no overlapping dots with any other dithering masks and the combination of all dithering mask may correspond to all the pixels of the target image region.
In particular embodiments, the dot patterns of the dithering masks with blue-noise properties may be generated using a simulated annealing algorithm. When the target grayscale value is smaller than the grayscale limit of 0.25, the four subframe images may be generated by applying the four dithering masks as shown in
As an example and not by way of limitation, for representing a grayscale value gin the range of [0.25, 0.5] in a first subframe image, the system may determine a dot pattern that includes all the dots from the first dithering mask 700A (which corresponds to 0.25) and a subset of dots from the second dithering mask 700B. The subset of dots selected from the second dithering mask 700B may be stacked together to the dot pattern of the first dithering mask to make up the difference portion of the grayscale value to the grayscale limit (i.e., g−0.25). Because the dot pattern of the first dithering mask 700A and the second dithering mask 700B have no overlapping dots, the subset of dots selected from the second dithering mask 700B may be stacked to the dot pattern of the first dithering mask 700A without violating the spatio stacking constraints. As another example, for representing a target grayscale value g in the range of [0.25, 0.5] in the second subframe image, the system may determine a dot pattern that includes all the dots from the second dithering mask 700B and a subset of dots from the third dithering mask 700C. The subset of dots selected from the third dithering mask 700B may be stacked together to the dot pattern of the second dithering mask 700B to make up the difference portion of the target grayscale value to the grayscale limit (i.e., g−0.25). Because the dot pattern of the second dithering mask 700B and the third dithering mask 700C have no overlapping dots, the subset of dots selected from the third dithering mask 700C may be stacked to the dot pattern of the first dithering mask 700B without violating the spatio stacking constraints.
As another example, for representing a target grayscale value g in the range of [0.25, 0.5] in the third subframe image, the system may determine a dot pattern that includes all the dots from the third dithering mask 700C and a subset of dots from the fourth dithering mask 700D. The subset of dots selected from the fourth dithering mask 700D may be stacked together to the dot pattern of the third dithering mask 700C to make up the difference portion of the target grayscale value to the grayscale limit (i.e., g−0.25). Because the dot pattern of the third dithering mask 700C and the fourth dithering mask 700D have no overlapping dots, the subset of dots selected from the fourth dithering mask 700D may be stacked to the dot pattern of the third dithering mask 700C without violating the spatio stacking constraints. As another example, for representing a target grayscale value g in the range of [0.25, 0.5] in a fourth subframe image, the system may determine a dot pattern that includes all the dots from the fourth dithering mask 700D and a subset of dots from the first dithering mask 700A. The subset of dots selected from the first dithering mask 700A may be stacked together to the dot pattern of the fourth dithering mask 70DC to make up the difference portion of the target grayscale value to the grayscale limit (i.e., g−0.25). Because the dot pattern of the fourth dithering mask 700D and the first dithering mask 700A have no overlapping dots, the subset of dots selected from the first dithering mask 700A may be stacked to the dot pattern of the fourth dithering mask 700D without violating the spatio stacking constraints. The principle of selecting dots from other dithering mask for representing target grayscale value in the range of [0.25, 0.5] may be described by the following equations:
G
1|0.25<g<0.5
=G
1|g=0.25
+G
2|g≤0.25 (6)
G
2|0.25<g<0.5
=G
2|g=0.25
+G
3|g≤0.25 (7)
G
3|0.25<g<0.5
=G
3|g=0.25
+G
4|g≤0.25 (8)
G
4|0.25<g<0.5
=G
4|g=0.25
+G
1|g≤0.25 (9)
where, G1, G2, G3, and G4 are the dot set of the first, second, third and fourth subframes, respectively, and g is grayscale level.
In particular embodiments, for representing a target grayscale values in the range of [0.5, 0.75] in the n-th subframe image, the system may need to use all dots in the dot pattern of the n-th dithering mask and select dots from the dithering masks of another two subframes. For example, for representing a target grayscale values in the range of [0.5, 0.75] in the first subframe image, the system may determine a dot pattern that includes all the dots from the first dithering mask 700A (which corresponds to 0.25), all the dots of the second dithering mask 700B (which corresponds to 0.25), and a subset of dots from the third dithering mask 700C. The dots selected from the second dithering mask 700B and the third dithering mask 700C may be stacked together to the dot pattern of the first dithering mask 700A. Because the dot patterns of the first, second and third dithering masks 700A-C have no overlapping dots, the dots selected from the second dithering mask 700B and the third dithering mask 700C may be stacked to the dot pattern of the first dithering mask 700A without violating the spatio stacking constraints.
As another example, for representing a target grayscale value in the range of [0.5, 0.75] in the second subframe image, the system may determine a dot pattern that includes all the dots from the second dithering mask 700B (which corresponds to 0.25), all the dots of the third dithering mask 700C (which corresponds to 0.25), and a subset of dots from the fourth dithering mask 700D. The dots selected from the third dithering mask 700C and the fourth dithering mask 700D may be stacked to the dot pattern of the first dithering mask 700A. Because the dot patterns of the second, third and fourth dithering masks 700B-D have no overlapping dots, the dots selected from the third dithering mask 700B and the fourth dithering mask 700D may be stacked to the dot pattern of the first dithering mask 700B without violating the spatio stacking constraints.
As another example, for representing a target grayscale values in the range of [0.5, 0.75] in the third subframe image, the system may determine a dot pattern that includes all the dots from the third dithering mask 700C (which corresponds to 0.25), all the dots of the fourth dithering mask 700D (which corresponds to 0.25), and a subset of dots from the first dithering mask 700A. The dots selected from the fourth dithering mask 700D and the first dithering mask 700A may be stacked to the dot pattern of the third dithering mask 700C. Because the dot patterns of the third, fourth and first dithering masks 700C, 700D and 700A have no overlapping dots, the dots selected from the fourth dithering mask 700D and the first dithering mask 700A may be stacked to the dot pattern of the third dithering mask 700C without violating the spatio stacking constraints.
As another example, for representing a target grayscale value in the range of [0.5, 0.75] in the fourth subframe image, the system may determine a dot pattern that includes all the dots from the fourth dithering mask 700D (which corresponds to 0.25), all the dots of the first dithering mask 700A (which corresponds to 0.25), and a subset of dots from the second dithering mask 700B. The dots selected from the first dithering mask 700A and the second dithering mask 700B may be stacked to the dot pattern of the fourth dithering mask 700D. Because the dot patterns of the first, second and fourth dithering masks 700A, 700B and 700D have no overlapping dots, the dots selected from the first dithering mask 700A and the second dithering mask 700B may be stacked to the dot pattern of the fourth dithering mask 700D without violating the spatio stacking constraints. The principle of selecting dots from other dithering mask for representing target grayscale value in the range of [0.5, 0.75] may be described by the following equations:
G
1|0.5<g<0.75
=G
1|g=0.25
+G
2|g=0.25
+G
3|g≤0.25 (10)
G
2|0.5<g<0.75
=G
2|g=0.25
+G
3|g=0.25
+G
4|g≤0.25 (11)
G
3|0.5<g<0.75
=G
3|g=0.25
+G
4|g=0.25
+G
1|g≤0.25 (12)
G
4|0.5<g<0.75
=G
4|g=0.25
+G
1|g=0.25
+G
2|g≤0.25 (13)
where, G1, G2, G3, and G4 are the dot set of the first, second, third and fourth subframes, respectively, g is grayscale level.
In particular embodiments, for representing a grayscale value in the range of [0.75, 1] in the n-th subframe image, the system may need to use all dots in the dot pattern of the n-th dithering mask and select dots from the dithering masks of another three subframes. For example, for representing a target grayscale value in the range of [0.75, 1] in the first subframe, the system may determine a dot pattern that includes all the dots from the first dithering mask 700A (which corresponds to 0.25), all the dots of the second dithering mask 700B (which corresponds to 0.25), all the dots of the third dithering mask 700C, and a subset of dots from the fourth dithering mask 700D. The dots selected from the second, third and fourth dithering masks 700B-D may be stacked to the dot pattern of the first dithering mask 700A. Because the dot patterns of the four dithering masks 700A-D have no overlapping dots, the dots selected from the second, third and fourth dithering masks 700B-D may be stacked to the dot pattern of the first dithering mask 700A without violating the spatio stacking constraints.
As another example, for representing a target grayscale value in the range of [0.75, 1] on the second subframe, the system may determine a dot pattern that includes all the dots from the second dithering mask 700B (which corresponds to 0.25), all the dots of the third dithering mask 700C (which corresponds to 0.25), all the dots of the fourth dithering mask 700D, and a subset of dots from the first dithering mask 700A. The dots selected from the third, fourth and first dithering masks (700C-D and 700A) may be stacked to the dot pattern of the second dithering mask 700B. Because the dot patterns of the four dithering masks 700A-D have no overlapping dots, the dots selected from the third, fourth and first dithering masks (700C-D and 700A) may be stacked to the dot pattern of the second dithering mask 700B without violating the spatio stacking constraints.
As another example, for representing a target grayscale value in the range of [0.75, 1] on the third subframe image, the system may determine a dot pattern that includes all the dots from the third dithering mask 700C (which corresponds to 0.25), all the dots of the fourth dithering mask 700D (which corresponds to 0.25), all the dots of the first dithering mask 700A, and a subset of dots from the second dithering mask 700B. The dots selected from the fourth, first and second dithering masks (700D and 700A-B) may be stacked to the dot pattern of the third dithering mask 700C. Because the dot patterns of the four dithering masks 700A-D have no overlapping dots, the dots selected from the fourth, first and second dithering masks 700D and 700A-B may be stacked to the dot pattern of the third dithering mask 700C without violating the spatio stacking constraints.
As another example, for representing a target grayscale value in the range of [0.75, 1] in the fourth subframe image, the system may determine a dot pattern that includes all the dots from the fourth dithering mask 700D (which corresponds to 0.25), all the dots of the first dithering mask 700A (which corresponds to 0.25), all the dots of the second dithering mask 700B, and a subset of dots from the third dithering mask 700C. The dots selected from the first, second and third dithering masks 700A-C may be stacked to the dot pattern of the fourth dithering mask 700D. Because the dot patterns of the four dithering masks 700A-D have no overlapping dots, the dots selected from the first, second and third dithering masks 700A-C may be stacked to the dot pattern of the fourth dithering mask 700D without violating the spatio stacking constraints. The principle of selecting dots from other dithering mask for representing target grayscale value in the range of [0.75, 1] may be described by the following equations:
G
1|0.75<g<1
=G
1|g=0.25
+G
2|g=0.25
+G
3|g=0.25
+G
4|g≤0.25 (14)
G
2|0.75<g<1
=G
2|g=0.25
+G
3|g=0.25
+G
4|g=0.25
+G
1|g≤0.25 (15)
G
3|0.75<g<1
=G
3|g=0.25
+G
4|g=0.25
+G
1|g=0.25
+G
2|g≤0.25 (16)
G
4|0.5<g<0.75
=G
4|g=0.25
+G
1|g=0.25
+G
2|g=0.25
+G
3|g≤0.25 (17)
where, G1, G2, G3, and G4 are the dot set of the first, second, third and fourth subframes, respectively, g is grayscale level. It is notable that the subframe number N=4 is used as an example and the number of subframes is not limited to N=4 and can by any suitable integer number. The systems and methods described in this disclosure may be applicable to any Nnumber of subframes.
In particular embodiments, the system may use an offline process to generate N coupled dithering masks and store the generated N dithering masks in a storage media. During the process of generating subframe images, the system may access the stored dithering masks from the storage media and use the dithering masks to generate N number of subframe images. In particular embodiments, the coupled Nnumber of dithering masks may have a cyclical relationship which allows all the dithering masks to be generated from a single seed mask based on the cyclical relationship. In particular embodiments, the system may only store the seed mask instead of storing all N number of dithering masks and generate all other dithering masks from the seed mask based on the cyclical relationship when they are needed. As a result, the system may reduce the memory usage for storing the dithering mask by a factor of N.
In particular embodiments, the seed mask stored by the system may include the dot pattern having a dot density corresponding the maximum grayscale level (e.g., 1 for normalized grayscale level range [0, 1]). The seed mask may have the same resolution and size with each of the N dithering masks to be generated based on the seed mask. For example, the seed mask and each of the N dithering masks may have the pixel resolution of 100 pixels×100 pixels, 150 pixels×150 pixels, 180 pixels×180 pixels, etc. The dot pattern of the seed mask may include all the dots corresponding to all the pixels of a target region (e.g., a tile region of the same size) of the target image. The threshold values stored in the seed mask may over all the grayscale levels in the quantization range. In particular embodiments, the system may pre-generate the seed mask during an offline process and store the seed mask in a storage media for later use. The pre-generated seed mask may be fixed after being generated and the same seed mask may be used for generating the dithering masks for all target images of a digital content (e.g., all mainframe images of a AR/VR content).
In particular embodiments, for an arbitrary Nnumber of subframes, the system may store a single seed mask (rather than N dithering masks) to reduce the memory usage related to the subframe generating process. For the n-th dithering mask of a total of N dithering mask, the dithering mask may be generated by cyclically permuting the seed mask as described in the following equations:
t
n=mod(t1−kn, 1) (18)
k
n=mod((n−1)g, 1) (19)
where tn is the threshold value of the n-th dithering mask; t1 is the threshold value of the seed mask; kn is the offsetting coefficient; g is the target grayscale value; and mod is the remainder operator. For a target grayscale level g of the n-th subframe image, the system may determine an offsetting coefficient kn based on a remainder of (n−1)·g divided by 1. Then, the system may determine the threshold values of the n-th subframe mask based on the remainder of (t1−kn) divided by 1. The seed mask may include a matrix of threshold values corresponding to all associated dots. The system may repeatedly apply Equation 18 on each of the threshold values to determine the corresponding threshold values of the n-th dithering mask.
As described in earlier sections of this disclosure, for representing grayscale values below the grayscale limit of gmax/N, the system may generate N dithering masks for each of the N subframes and allow the generated N dithering masks to satisfy spatio and temporal stacking constraints. For spatio stacking properties, each dithering mask of the N dithering masks may include a dot pattern having a dot density corresponding to the grayscale limit gmax/N. The dot pattern of dithering mask may include a stack of dot patterns corresponding to all the lower grayscale levels from 0 to gmax/N and the dot pattern of any grayscale level may include all dots of the dot patterns of lower scale levels. For temporal stacking properties, the N dithering masks may be generated in such as way that they have no overlapping dots with each other. For example, for N=4, the system may generate 4 dithering masks whose dot patterns have no overlapping dots with each other. In particular embodiments, the system may divide the grayscale range into N grayscale segments with each segment covering gmax/N grayscale level units (with each grayscale level unit corresponding to an incremental grayscale step) and select the seed mask dots covered in each segment as the dots to be included in a dither mask corresponding to that segment. For example, for N=4 and a seed mask covering the threshold values of [0, 1], the system may determine the first, second, third and fourth dithering masks to include the seed mask dots covered by the threshold segments of 0<tM1<=0.25, 0.25<tM2<=0.5, 0.5<tM2<=0.75 and 0.75<tM2<=1, respectively, where tM1, tM2, tM3 and tM1 are threshold values of the first, second, third and fourth dithering masks. Because of the spatio stacking property, the dot pattern of each grayscale level unit or grayscale level step may not have overlapping dots with any other grayscale unit and the dots selected for each dithering mask in this manner may naturally satisfy the temporal stacking property by having no overlapping dots between any two dithering masks.
In particular embodiments, the system may use the cyclical relationship as described in Equations 18 and 19 to determine the threshold values of N dithering masks based on the threshold values of the seed mask. As an example and not by way of limitation, the system may generate 4 subframe images for each mainframe image. For N=4, the grayscale limit may be determined to be 0.25. For representing the grayscale value 0.25, the system may generate 4 dithering masks from a seed mask using cyclical relationship as described in Equations 18 and 19. As described in earlier sections, the seed mask may have a dot pattern corresponding to the maximum grayscale level of 1 for the normalized grayscale range of [0, 1]. For the dithering masks of the first, second, third and fourth subframes, the offset coefficients of k1, k2, k3, and k4 may be determined to be 0, 0.25, 0.5 and 0.75, respectively, by applying Equation 19 as follows:
k
1=mod((1−1)·0.25, 1)=0 (20)
k
2=mod((2−1)·0.25, 1)=0.25 (21)
k
3=mod((3−1)·0.25, 1)=0.5 (22)
k
3=mod((4−1)·0.25, 1)=0.75 (23)
Consequently, the threshold values of the first, second, third and fourth dithering masks may be determined by applying Equation 18 as follows:
t
M1=mod(tSM1−0, 1) (24)
t
M2=mod(tSM1−0.25, 1) (25)
t
M3=mod (tSM1−0.5, 1) (26)
t
M2=mod(tSM1−0.75, 1) (27)
where tM1, tM2, tM3 and tM1 are threshold values of the first, second, third and fourth dithering masks; tSM is the corresponding threshold values of the seed mask. The system may repeatedly apply these equations to the threshold value of each dot in the seed mask to determine the threshold values of the corresponding dithering masks. As a result, the dithering pattern for the first, second, third and fourth dithering masks may be determined to be 0<tM1<=0.25, 0<tM2<=0.25, 0<tM2<=0.25 and 0<tM2<=0.25, respectively, where tM1, tM2, tM3 and tM1 are threshold values of the first, second, third and fourth dithering masks. In other words, the four dithering masks may have their threshold values shifted by the respective offsetting coefficients to the target threshold range of [0, 0.25]. As a result, each dithering mask may have a dot pattern corresponding the grayscale value of 0.25 and each dithering mask may cover a threshold range of [0, 0.25]. Therefore, for representing any grayscale value less than or equal to the grayscale limit of 0.25, the four subframe images may be generated by applying the four dithering masks to the target mainframe image and the generated subframe images may each satisfy the spatio stacking property and together satisfy the temporal sacking property as described in earlier sections of this disclosure.
In particular embodiments, for representing grayscale values above the grayscale limit of gmax/N, the system may generate N dithering masks from the seed mask using cyclical relationship as described in Equations 18 and 19. The generated N dithering masks may have overlapping dots but the dot patterns of the dithering masks may be selected in a way satisfying the spatio and temporal stacking property and allowing uniform luminance or energy distribution among the N dithering masks. As described in the earlier sections of this disclosure, the dot patterns with overlapping dots may be determined by selecting or borrowing dots from the dithering masks for other subframes. In particular embodiments, the system may determine the dot patterns of the dithering masks with overlapping dots based on the cyclical relationship as described in Equations 18 and 19.
In particular embodiments, the system may determine N grayscale segments within the grayscale range of [0, 1] and have each grayscale segment cover gmax/N grayscale level units (with each grayscale level unit corresponding to an incremental grayscale step). The system may select the seed mask dots covered in each segment as the dots to be included in a dither mask corresponding to that segment. The determination of the grayscale segments and corresponding dot patterns may be performed based on the cyclical relationship by repeatedly applying Equations 18 and 19. As an example and not by way of limitation, for N=4, to represent a target grayscale of 0.6 within a grayscale range of [0, 1], the first, second, third and fourth grayscale segment may be determined to be 0<tSM<=0.6, 0.2<tSM<=0.8, 0.4<tSM<=1, and 0.2<tSM<=0.8, where tSM is the threshold values of the seed mask. The system may select the dots in the seed mask (which covers the grayscale rang of [0, 1]) covered by the four grayscale segments as to be included in the respective dot patterns of the four dithering masks. For example, the first dithering mask may have a dot pattern including dots of the seed mask covered in the grayscale segment of 0<tSM<=0.6. The second dithering mask may have a dot pattern including dots of the seed mask covered in the grayscale segment of 0.21<tSM<=0.8. The third dithering mask may have a dot pattern including dots of the seed mask covered in the grayscale segment of 0.4<tSM<=1. The fourth dithering mask may have a dot pattern including dots of the seed mask covered in the grayscale segment of 0.2<tSM<=0.8.
It is notable that the selection of the dot patterns for the dithering masks is for example purpose and the section of dot patterns is not limited by these segments or ranges. It is also notable that the determination of the dot patterns of the dithering masks may not depend on the order of the grayscale segments or ranges. Any threshold segments or ranges in any order that allow the dot patterns of the dithering masks to satisfy the spatio-temporal stacking property and have uniform luminance distribution among the subframes may be used for determining the dot patterns of the dithering masks. For example, the dot pattern of the first, second, third and fourth dithering mask may include dots of the seed mask covered by the threshold segments of 0<tSM<=0.6, 0.2<tSM<=0.8, 0.2<tSM<0.8, and 0.4<tSM<=1. It is notable that some threshold segments may be determined by a wrap-around operation. For example, since the range of [0.6, 1.2] is out of the grayscale range of [0, 1], the system may use a wrap-around operation to determine the grayscale segments of [0.6, 1] and [0, 0.2] which also cover a grayscale range of the same width.
For N=4, to represent a target grayscale of 0.6 within a grayscale range of [0, 1], the offset coefficients k1, k2, k3, and k4 of the first, second, third and fourth dithering masks may be determined to be 0, 0.6, 0.2 and 0.8, respectively, by applying Equation 19 as follows:
k
1=mod((1−1)·0.6, 1)=0 (20)
k
2=mod((2−1)·0.6, 1)=0.6 (21)
k
3=mod((3−1)·0.6, 1)=0.2 (22)
k
3=mod((4−1)·0.6, 1)=0.8 (23)
Consequently, the threshold values of the first, second, third and fourth dithering masks may be determined by applying Equation 18 as follows:
t
M1=mod (tSM1−0, 1) (24)
t
M2=mod (tSM1−0.6, 1) (25)
t
M3=mod (tSM1−0.2, 1) (26)
t
M2=mod (tSM1−0.8, 1) (27)
where tM1, tM2, tM3 and tM1 are threshold values of the first, second, third and fourth dithering masks; tSM is the corresponding threshold values of the seed mask. The system may repeatedly apply these equations to the threshold value of each dot in the seed mask to determine the threshold values of the corresponding dithering masks. As a result, the dithering pattern for the first, second, third and fourth dithering masks may be determined to be 0<tM1<=0.6, 0<tM2<=0.6, 0<tM2<=0.6 and 0<tM2<=0.6, respectively, where tM1, tM2, tM3 and tM1 are threshold values of the first, second, third and fourth dithering masks. The system may use these four dithering masks to generate four subframe images and each of the subframe images may have a dot pattern which has a dot density corresponding to the target grayscale value of 0.6.
In particular embodiments, the system may use the dithering masks to determine the values to be dithered for arbitrary grayscale values. For a display with evenly-spaced M-bit grayscale levels and N temporal subframes, the system may dither an arbitrary target grayscale value g using the processes described as follows. In particular embodiments, the system may define the least significant bit (LSB) as LSB=1/(2M−1). For example, the LSB values for displays of 8 bits, 6 bits, and 4 bits may be determined to be 1/255, 1/63, and 1/7. The whole part of the grayscale value w may be determined to by w=LSB·floor(g/LSB). The remainder r, which has the range of [0, 1], may be determined by r=(g−w)/LSB. For the n-th subframe, the offset coefficient kn may be determined to be kn=mod((n−1)·r, 1). Consequently, the threshold values of the n-th subframe may be determined by tn=mod(t1−kn, 1). The system may display the whole part of the grayscale value w in the n-th subframe and dither the remainder r temporally to other subframes. For the subframe that receives the dithered remainder, the total displayed grayscale value may be determined as by dn=w+(r>tn)·LSB.
In particular embodiments, the system may use the dithering masks to determine values to be deterred for arbitrary grayscale values. For a display with unevenly-spaced J-bit grayscale levels and N temporal subframes, the system may dither an arbitrary target grayscale value g using the processes described as follows. In particular embodiments, the system may define the least significant bit (LSB) as the difference between the adjacent grayscale levels (i.e., wj+1−wj). The system may determine the index j corresponding to the closest grayscale level wj that is smaller than g which is between the wj and wj+1. The system may determine the remainder r, which is in the range of [0, 1], by r=(g−wj)/(wj−wj+1). For the n-th subframe, the offset coefficient kn may be determined by kn=mod((n−1)·r, 1). Consequently, the threshold values of the n-th subframe may be determined by tn=mod(t1−kn, 1). The system may display the whole part of the grayscale value w in the n-th subframe and dither the remainder r temporally to other subframes. For the subframe that receives the dithered remainder, the total displayed grayscale value may be determined as by dn=wj+(r>tn).
In particular embodiments, to represent any grayscale value g by N subframe images, the system may generate N dithering masks from the seed mask based on the cyclical relationship. The generated N dithering masks may have dot patterns satisfying both spatio stacking property and temporal stacking property. The threshold values of N dithering masks may be determined by shifting the grayscale ranges of the seed mask with the offset coefficients determined using Equation 18. After shifted, the four dithering masks may each have a dot pattern covering a threshold range of 0<tM<=g. The system may use the N dithering masks to generate N subframe images each of which may have a dot density corresponding to the target grayscale value g. The N subframe images may have uniform luminance distribution or energy distribution with each other. As a result, to perform spatio-temporal dithering for any arbitrary number of subframes, the system may only need to store a seed mask and generate all the dithering masks from the seed mask using the cyclical relationship when the dithering masks are needed for generating the subframe images.
It is notable that the systems, methods, and processes for determining the dot patterns based on the remainder operations are for example purpose and the generating of dot patterns of the dithering masks is not limited thereof. As long as the four dithering masks covers N grayscale ranges and the dot patterns are determined in a manner satisfying the spatio and temporal stacking property, the generated dithering masks may be qualified for being used to generate the subframe images. This disclosure covers all suitable systems, methods and processes for generating dot patterns of the dithering masks satisfying the spatio and temporal stacking properties as described in earlier sections of this disclosure.
At step 930, the system may generate a number of images based on the target image and the masks. Each of the image may have a second number of bits per color which is smaller than the first number of bits per color. For example, the system may generate N number of subframe images based on the target image and N dithering masks. Each of the subframe images may have color values with a second number of bits per color (e.g., grayscale values in a second bit length). The second number of bits per color may correspond to the color depth of the display which may be smaller than the first number of bits per color. In other words, the N number of subframe images may have less color depth comparing to the target image. The dithering masks satisfying the temporal stacking constraint may allow the subframe images to have the uniform luminance distribution among the images (e.g., each image has luminosity within a threshold range). In particular embodiments, the N dithering mask may be made available at the same time for a process of generating the subframe images. The system may determine one or more quantitation errors based on one or more color values of the target image and one or more threshold values associated with one of the dithering masks. The system may dither the quantization errors temporally to one or more other subframe images without using an error buffer. A subframe image may be generated by repeatedly applying the corresponding mask to the target image.
At step 940, the system may display the subframe images sequentially on the display for representing the target image. As a result, the subframe images used for representing the target image may have even luminance distribution among the subframe images (e.g., having luminosity within a threshold range). In particular embodiments, the system may generate a seed mask which may include threshold values covering a quantization grayscale range. The system may store the seed mask in a storage media and access the seed mask from the storage media for generating the N dithering masks from the seed mask based on a cyclical relationship, and therefore reduce the storage space usage for generating the subframe images. In particular embodiments, the system may determine a grayscale limit gL based on a maximum grayscale level gMAX and the N number of subframe images for representing the target image (e.g., gL=gMAX/N). When the target grayscale value for a tile region of the subframe images is smaller than the grayscale limit, the corresponding tile regions of the subframe images may be generated based on the non-overlapping dithering masks which include non-overlapping set of dots from each other. Consequently, the corresponding tile regions of the subframe images may include non-overlapping sets of pixels from each other. When a target grayscale value associated with the target image is greater than the grayscale limit, the corresponding tile regions of the subframe images may be generated based on N number of overlapping dithering masks which include overlapping dots incrementally selected from at least another dithering mask. The N number of overlapping masks may be generated by incrementally selecting dots from at least another mask of the plurality of masks. Consequently, the corresponding tile regions of the subframe images may include overlapping sets of pixels which may be determined by incrementally selecting dots from at least another mask of the masks.
Particular embodiments may repeat one or more steps of the method of
This disclosure contemplates any suitable number of computer systems 1000. This disclosure contemplates computer system 1000 taking any suitable physical form. As example and not by way of limitation, computer system 1000 may be an embedded computer system, a system-on-chip (SOC), a single-board computer system (SBC) (such as, for example, a computer-on-module (COM) or system-on-module (SOM)), a desktop computer system, a laptop or notebook computer system, an interactive kiosk, a mainframe, a mesh of computer systems, a mobile telephone, a personal digital assistant (PDA), a server, a tablet computer system, an augmented/virtual reality device, or a combination of two or more of these. Where appropriate, computer system 1000 may include one or more computer systems 1000; be unitary or distributed; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks. Where appropriate, one or more computer systems 1000 may perform without substantial spatial or temporal limitation one or more steps of one or more methods described or illustrated herein. As an example and not by way of limitation, one or more computer systems 1000 may perform in real time or in batch mode one or more steps of one or more methods described or illustrated herein. One or more computer systems 1000 may perform at different times or at different locations one or more steps of one or more methods described or illustrated herein, where appropriate.
In particular embodiments, computer system 1000 includes a processor 1002, memory 1004, storage 1006, an input/output (I/O) interface 1008, a communication interface 1010, and a bus 1012. Although this disclosure describes and illustrates a particular computer system having a particular number of particular components in a particular arrangement, this disclosure contemplates any suitable computer system having any suitable number of any suitable components in any suitable arrangement.
In particular embodiments, processor 1002 includes hardware for executing instructions, such as those making up a computer program. As an example and not by way of limitation, to execute instructions, processor 1002 may retrieve (or fetch) the instructions from an internal register, an internal cache, memory 1004, or storage 1006; decode and execute them; and then write one or more results to an internal register, an internal cache, memory 1004, or storage 1006. In particular embodiments, processor 1002 may include one or more internal caches for data, instructions, or addresses. This disclosure contemplates processor 1002 including any suitable number of any suitable internal caches, where appropriate. As an example and not by way of limitation, processor 1002 may include one or more instruction caches, one or more data caches, and one or more translation lookaside buffers (TLBs). Instructions in the instruction caches may be copies of instructions in memory 1004 or storage 1006, and the instruction caches may speed up retrieval of those instructions by processor 1002. Data in the data caches may be copies of data in memory 1004 or storage 1006 for instructions executing at processor 1002 to operate on; the results of previous instructions executed at processor 1002 for access by subsequent instructions executing at processor 1002 or for writing to memory 1004 or storage 1006; or other suitable data. The data caches may speed up read or write operations by processor 1002. The TLBs may speed up virtual-address translation for processor 1002. In particular embodiments, processor 1002 may include one or more internal registers for data, instructions, or addresses. This disclosure contemplates processor 1002 including any suitable number of any suitable internal registers, where appropriate. Where appropriate, processor 1002 may include one or more arithmetic logic units (ALUs); be a multi-core processor; or include one or more processors 1002. Although this disclosure describes and illustrates a particular processor, this disclosure contemplates any suitable processor.
In particular embodiments, memory 1004 includes main memory for storing instructions for processor 1002 to execute or data for processor 1002 to operate on. As an example and not by way of limitation, computer system 1000 may load instructions from storage 1006 or another source (such as, for example, another computer system 1000) to memory 1004. Processor 1002 may then load the instructions from memory 1004 to an internal register or internal cache. To execute the instructions, processor 1002 may retrieve the instructions from the internal register or internal cache and decode them. During or after execution of the instructions, processor 1002 may write one or more results (which may be intermediate or final results) to the internal register or internal cache. Processor 1002 may then write one or more of those results to memory 1004. In particular embodiments, processor 1002 executes only instructions in one or more internal registers or internal caches or in memory 1004 (as opposed to storage 1006 or elsewhere) and operates only on data in one or more internal registers or internal caches or in memory 1004 (as opposed to storage 1006 or elsewhere). One or more memory buses (which may each include an address bus and a data bus) may couple processor 1002 to memory 1004. Bus 1012 may include one or more memory buses, as described below. In particular embodiments, one or more memory management units (MMUs) reside between processor 1002 and memory 1004 and facilitate accesses to memory 1004 requested by processor 1002. In particular embodiments, memory 1004 includes random access memory (RAM). This RAM may be volatile memory, where appropriate. Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM. This disclosure contemplates any suitable RAM. Memory 1004 may include one or more memories 1004, where appropriate. Although this disclosure describes and illustrates particular memory, this disclosure contemplates any suitable memory.
In particular embodiments, storage 1006 includes mass storage for data or instructions. As an example and not by way of limitation, storage 1006 may include a hard disk drive (HDD), a floppy disk drive, flash memory, an optical disc, a magneto-optical disc, magnetic tape, or a Universal Serial Bus (USB) drive or a combination of two or more of these. Storage 1006 may include removable or non-removable (or fixed) media, where appropriate. Storage 1006 may be internal or external to computer system 1000, where appropriate. In particular embodiments, storage 1006 is non-volatile, solid-state memory. In particular embodiments, storage 1006 includes read-only memory (ROM). Where appropriate, this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these. This disclosure contemplates mass storage 1006 taking any suitable physical form. Storage 1006 may include one or more storage control units facilitating communication between processor 1002 and storage 1006, where appropriate. Where appropriate, storage 1006 may include one or more storages 1006. Although this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
In particular embodiments, I/O interface 1008 includes hardware, software, or both, providing one or more interfaces for communication between computer system 1000 and one or more I/O devices. Computer system 1000 may include one or more of these I/O devices, where appropriate. One or more of these I/O devices may enable communication between a person and computer system 1000. As an example and not by way of limitation, an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these. An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 1008 for them. Where appropriate, I/O interface 1008 may include one or more device or software drivers enabling processor 1002 to drive one or more of these I/O devices. I/O interface 1008 may include one or more I/O interfaces 1008, where appropriate. Although this disclosure describes and illustrates a particular I/O interface, this disclosure contemplates any suitable I/O interface.
In particular embodiments, communication interface 1010 includes hardware, software, or both providing one or more interfaces for communication (such as, for example, packet-based communication) between computer system 1000 and one or more other computer systems 1000 or one or more networks. As an example and not by way of limitation, communication interface 1010 may include a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network or a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network. This disclosure contemplates any suitable network and any suitable communication interface 1010 for it. As an example and not by way of limitation, computer system 1000 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these. One or more portions of one or more of these networks may be wired or wireless. As an example, computer system 1000 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these. Computer system 1000 may include any suitable communication interface 1010 for any of these networks, where appropriate. Communication interface 1010 may include one or more communication interfaces 1010, where appropriate. Although this disclosure describes and illustrates a particular communication interface, this disclosure contemplates any suitable communication interface.
In particular embodiments, bus 1012 includes hardware, software, or both coupling components of computer system 1000 to each other. As an example and not by way of limitation, bus 1012 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these. Bus 1012 may include one or more buses 1012, where appropriate. Although this disclosure describes and illustrates a particular bus, this disclosure contemplates any suitable bus or interconnect.
Herein, a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid-state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate. A computer-readable non-transitory storage medium may be volatile, non-volatile, or a combination of volatile and non-volatile, where appropriate.
Herein, “or” is inclusive and not exclusive, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A or B” means “A, B, or both,” unless expressly indicated otherwise or indicated otherwise by context. Moreover, “and” is both joint and several, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A and B” means “A and B, jointly or severally,” unless expressly indicated otherwise or indicated otherwise by context.
The scope of this disclosure encompasses all changes, substitutions, variations, alterations, and modifications to the example embodiments described or illustrated herein that a person having ordinary skill in the art would comprehend. The scope of this disclosure is not limited to the example embodiments described or illustrated herein. Moreover, although this disclosure describes and illustrates respective embodiments herein as including particular components, elements, feature, functions, operations, or steps, any of these embodiments may include any combination or permutation of any of the components, elements, features, functions, operations, or steps described or illustrated anywhere herein that a person having ordinary skill in the art would comprehend. Furthermore, reference in the appended claims to an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative. Additionally, although this disclosure describes or illustrates particular embodiments as providing particular advantages, particular embodiments may provide none, some, or all of these advantages.