The present disclosure relates to computer graphics systems, and more particularly, to a Micro light emitting diode (MicroLED) display system and the color management of the display device.
One area of computing devices that has grown in recent years is the area of virtual reality (VR) and augmented reality (AR) devices, which use a graphics processing unit (GPU) to render graphics from a computing device to a display device. Such technology may be incorporated into a head-mounted display (HMD) device in the form of eyeglasses, goggles, a helmet, a visor, or other eyewear. As used herein, a HMD device may include a device that generates and/or displays virtual reality images (e.g., from at least one virtual environment input), and/or mixed reality (MR) images or augmented reality (AR) images (e.g., from at least one virtual environment input and one real environment input). In such devices, a scene produced on a display device can be oriented or modified based on user input (e.g., movement of a gamepad button or stick to cause movement of the orientation of the scene, introduction of items into the scene, etc.).
One challenge with incorporating display devices into HMD or mobile devices is the size constraints that limit some of the optical or display components that can be integrated into the HMD devices while miniaturizing the overall size of the HMD devices to improve user mobility. Current HMDs usually use illuminated micro displays such as reflective liquid crystal on silicon (hereafter “LCoS”) or digital light processing (DLP) projectors as they provide a high standard of display performance. These displays offer advantages such as high resolution, a wide color gamut, high brightness and a high contrast ratio. However, such digital projection systems that rely on LCoS or DLP technology require large form factors to create a uniform illumination of panels. Thus, there is a need in the art for improvements in presenting images on a display with miniaturized components without compromising the display quality or user experience.
MicroLED arrays offer a small form factor solution for the HMD image sources since they do not need separate illumination optics. Features of the present disclosure implement a MicroLED display system that incorporates a plurality of monochrome projectors (e.g., three MicroLED projectors) to generate three monochrome images (e.g., red, blue, and green images) that are separately input into a single waveguide of the HMD and combined to form an image that is displayed to the user. By utilizing a single waveguide that includes a plurality of spatially separated input regions (e.g., a region for inputting blue light, a region for inputting red light, and a region for inputting green light), the MicroLED display system of the present disclosure may reduce the form factor of the HMD device because of the reduced number and/or size of optical components, such as a reduced number of plates that may be required to combine the three monochrome images.
In one example, a display device is disclosed. The display device may include an optical waveguide having a plurality of input regions for receiving at least a portion of light corresponding to an image, wherein the plurality of input regions include at least a first input region, a second input region, and a third input region. The display device may also include a plurality of monolithic micro light emitting diode (MicroLED) projectors that each generate a different monochrome color version of the image. In some examples, the plurality of MicroLED projectors may include at least a first projector generating a first version of the image in a first color, a second projector generating a second version of the image in a second color, and a third projector generating a third version of the image in a third color. In some aspects, the first input region of the optical waveguide may be physically aligned with the first projector to receive light corresponding to the first version of the image, the second input region of the optical waveguide may be physically aligned with the second projector to receive light corresponding to the second version of the image, and the third input region of the optical waveguide may be physically aligned with the third projector to receive light corresponding to the third version of the image. In some examples, the optical waveguide may further include an output region configured to guide light from each of the plurality of MicroLED projectors toward a target to make a final image visible to user, wherein the final image is a fully colored image that combines the different monochrome color versions of the image.
In another example, a method for displaying an image on a display device is disclosed. The method may include generating a plurality of monochrome color versions of an image using a plurality of monolithic MicroLED projectors, wherein the plurality of MicroLED projectors includes at least a first projector generating a first version of the image in a first color, a second projector generating a second version of the image in a second color, and a third projector generating a third version of the image in a third color. The method may further include receiving, at an optical waveguide, light inputs from each of the plurality of MicroLED projectors into a plurality of different input regions of the optical waveguide such that a first input region of the optical waveguide is physically aligned with a first color projector to receive light corresponding to the first version of the image, the second input region of the optical waveguide is physically aligned with the second color projector to receive light corresponding to the second version of the image, and the third input region of the optical waveguide is physically aligned with the third color projector to receive light corresponding to the third version of the image. The method may further include outputting, through the optical waveguide, light from each of the plurality of MicroLED projectors toward a target to make a final image visible to user, wherein the final image is a fully colored image that combines the plurality of monochrome color versions of an image.
In another example, another method for calibrating an image on a display device is disclosed. The method may include detecting, at a camera, light rays output from an optical waveguide, wherein the light rays correspond to a plurality of monochrome color versions of an image generated by a plurality of monolithic MicroLED projectors. The method may further include determining a first position of a first version of the image from the plurality of monochrome color versions of the image, wherein the first version of the image is generated in first monochrome color by a first projector. The method may further include determining a second position of a second version of the image from the plurality of monochrome color versions of the image, wherein the second version of the image is generated in second monochrome color by a second projector. The method may further include measuring an image displacement between the first position of the first version of the image and the second position of the second version of the image to calculate an offset value between the first position and the second position, wherein the image displacement is one or both of lateral displacement or angular rotational displacement. The method may further include configuring the display device to adjust at least one image output parameter of the plurality of MicroLED projectors such that the output from the optical waveguide of the plurality of monochrome color versions of the image have an overlapping alignment that meets an alignment threshold.
In another example, an apparatus for calibrating an image on a display device is disclosed. The apparatus may include a memory to store data and instructions. The apparatus may further include a processor in communication with the memory to execute the instructions to detect, at a camera, light rays output from an optical waveguide, wherein the light rays correspond to a plurality of monochrome color versions of an image generated by a plurality of monolithic MicroLED projectors. The apparatus may further include instructions to determine a first position of a first version of the image from the plurality of monochrome color versions of the image, wherein the first version of the image is generated in first monochrome color by a first projector. The apparatus may further include instructions to determine a second position of a second version of the image from the plurality of monochrome color versions of the image, wherein the second version of the image is generated in second monochrome color by a second projector. The apparatus may further include instructions to measure an image displacement between the first position of the first version of the image and the second position of the second version of the image to calculate an offset value between the first position and the second position, wherein the image displacement is one or both of lateral displacement or angular rotational displacement. The apparatus may further include instructions to configure the display device to adjust at least one image output parameter of the plurality of MicroLED projectors such that the output from the optical waveguide of the plurality of monochrome color versions of the image have an overlapping alignment that meets an alignment threshold.
The above presents a simplified summary of one or more implementations of the present disclosure in order to provide a basic understanding of such implementations. This summary is not an extensive overview of all contemplated implementations, and is intended to neither identify key or critical elements of all implementations nor delineate the scope of any or all implementations. Its sole purpose is to present some concepts of one or more implementations of the present disclosure in a simplified form as a prelude to the more detailed description that is presented later.
To the accomplishment of the foregoing and related ends, the one or more aspects comprise the features hereinafter fully described and particularly pointed out in the claims. The following description and the annexed drawings set forth in detail certain illustrative features of the one or more aspects. These features are indicative, however, of but a few of the various ways in which the principles of various aspects may be employed, and this description is intended to include all such aspects and their equivalents.
The disclosed aspects of the present disclosure will hereinafter be described in conjunction with the appended drawings, provided to illustrate and not to limit the disclosed aspects, wherein like designations denote like elements, where a dashed line may indicate an optional component, and in which:
Various aspects are now described with reference to the drawings. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of one or more aspects. It may be evident, however, that such aspect(s) may be practiced without these specific details. Additionally, the term “component” as used herein may be one of the parts that make up a system, may be hardware, firmware, and/or software stored on a computer-readable medium, and may be divided into other components.
The present disclosure provides devices and methods for presentation of images such as virtual reality (VR) or mixed reality (MR)/augmented reality (AR) images on a display that is incorporated into mobile display devices, such as displays implemented for head mounted display (HMD). It should be appreciated by those of ordinary skill in the art that while the present disclosure references HMD, the display techniques implemented herein may be adaptable for any mobile device, including but not limited to, mobile phones, tablets, or laptops.
As discussed above, one challenge with incorporating display devices into mobile devices is the size constraints that limit the components that can be integrated into the display systems while miniaturizing the overall size of the HMD devices or mobile display to improve user mobility. Current HMDs generally use illuminated micro displays such as reflective LCoS or DLP projectors as they provide a high standard of display performance. While these displays offer advantages such as high resolution, a wide color gamut, high brightness and a high contrast ratio, such digital projection systems that rely on LCoS or DLP technology also require large form factors to create a uniform illumination of panels.
In contrast, MicroLED arrays offer a small form factor solution for the HMD image sources since they do not need separate illumination optics. However, the materials required to create semiconductors with either blue-green or red LEDs are generally incompatible in the manufacturing process of LEDs for specific colors (e.g., Indium, Gallium and Nitrogen (InGaN) to produce green, blue and white colors and/or Aluminum Gallium Arsenide (AlGaAs) used to generate red and amber portions of the visible spectrum). Specifically, from the manufacturing perspective, it may be difficult to get all three colors into the same array without either the labor intensive process of picking and placing each LED individually or developing an expensive growth/lithography process.
One solution to the above-identified problem with MicroLED micro displays may be to separate each of the red, green, and blue arrays and combine the different colors with optics. However, optical combiners require large prism cubes to combine three different images together and relay optics to prevent the divergence of the pixels when combining the colors. Thus, such optical systems increase the optics footprint and prevent small form factor HMD development with micro LEDs. Furthermore, the divergence angle of the microLEDs may be so large (E.g., +/−40 degrees) that it may impact the optical power.
Features of the present disclosure implement a MicroLED display system that incorporates a plurality of monochrome projectors (e.g., three MicroLED projectors) to generate three monochrome images (e.g., red, blue, and green images) that are separately input into a single waveguide of the HMD and combined to form an image that is displayed to the user. By utilizing a single waveguide that includes a plurality of spatially separated input regions (e.g., a region for inputting blue light, a region for inputting red light, a region for inputting green light), the MicroLED display system of the present disclosure may reduce the size and form factor of the HMD device because of the reduced number of plates that may be required to combine the three monochrome images and/or because of the reduced number of size of the optical components. However, it should be appreciated by those of ordinary skill in the art that the features of the present disclosure are not limited to plurality of projectors generating monochrome images. Instead, in some examples, a single projector without external illumination may also generate and input the full color image into the waveguide. Such system would provide benefit of simplistic hardware and reduced form factor in comparison to the current digital projection systems that rely on LCoS or DLP technology that require large form factors to create a uniform illumination of panels.
Specifically, in accordance with aspects of the present disclosure, waveguides with gratings may provide a mechanism to combine several images together with the smallest possible form factor. With a separate microLED display for each primary color, the exit pupil of each image/color may be coupled into a single optical waveguide that brings the pupils on top of each other using pupil replication. Further, since the waveguide may be designed to act as a pupil expander of an augmented reality (AR) display, the same optical waveguide may be used to relay the image into the user's eye, thereby saving on additional hardware by using the same integrated waveguide to combine the three colors together and to act as an AR-display. This may include configuring the waveguide to steer the incoming field-of-view (FoV) at least into two or more directions in order to enable pupil expansion with a large FoV.
Additional advantages of the present disclosure may be the hardware simplicity that is required for HMD. For example, by incorporating a plurality of monochrome projectors (e.g., three MicroLED projectors) to generate three monochrome images (e.g., red, blue, and green images) that are separately input into a single waveguide of the HMD, the optics of each MicroLED projectors may be simpler because no mirrors or doublet lenses may be needed to avoid or correct chromatic aberration.
In addition, in some implementations, the present disclosure may further reduce the form factor of the HMD by implementing a color management scheme that may include modifying the color resolution of each MicroLED projector. Specifically, the physical size of the MicroLED projector may depend on the resolution/pixel count of the imaging system. Recognizing that the human eye has the highest sensitivity for green light, the overall form factor of the HMD may be further reduced by lowering the color resolution of blue and/or red colors, while increasing the color resolution of green light for full resolution. To this end, due to lower color resolution requirements of one or more colors (blue and/or red color images), features of the present disclosure may further reduce the size and complexity of hardware of the monolithic MicroLED projectors. Thus, the lower color resolution of the blue and/or red colors may be compensated by a higher resolution of green color to provide a user with a full resolution image while benefiting from reduced form factor of the overall display system. Although the above example discusses lowering the color resolution of blue and/or red color images, it should be appreciated that the modification of the color resolution is not limited to only blue and/or red colors, but instead may be adapted for any number of colors. Additionally, in some examples, the techniques of the present disclosure may be applied to lowering the color resolution of only one color.
Finally, features of the present disclosure also provide techniques of calibrating the image to be displayed via the optical guide by configuring one or more MicroLED projectors. For example, any mechanical mounting issues with the MicroLED projectors may change the angle of the projected image, which in turn may cause pixels from that MicroLED projector to be in misplaced. Thus, any displacement between the three monochrome images may result in a blurry image being displayed on the display device. One technique to solve this problem may include manipulating the incoming image in order to pre-correct the location of one or more pixels in order to account for any disparity. Thus, in some aspects of the manufacturing process, a calibration procedure may be implemented that may utilize a camera to be placed in front of the optical waveguide (e.g., where a user's eye would otherwise be) to identify how each color field is separated from one another, and to provide configuration parameters to calibrate the image from the three MicroLED projectors such that the three images overlap, generating a cohesive final image to be displayed. To this end, features of the present disclosure may include techniques for pre-rotating and offsetting at least one of the three monochromic images to correct for any misalignment in the mechanical assembly of the display device, including the MicroLED projectors.
The following description provides examples, and is not limiting of the scope, applicability, or examples set forth in the claims. Changes may be made in the function and arrangement of elements discussed without departing from the scope of the disclosure. Various examples may omit, substitute, or add various procedures or components as appropriate. For instance, the methods described may be performed in an order different from that described, and various steps may be added, omitted, or combined. Also, features described with respect to some examples may be combined in other examples.
Turning first to
A display device 100 may be configured to provide virtual reality images (e.g., from at least one virtual environment input), and/or mixed reality (MR) images or augmented reality (AR) images (e.g., from at least one virtual environment input and one real environment input). The display device 100, when implemented as a HMD, may comprises a headpiece arranged to be worn on the user's head using a frame 105 (in the manner of conventional spectacles), helmet or other fit system. The purpose of the fit system is to support the display and provide stability to the display and other head borne systems such as tracking systems and cameras.
The display device 100 may include one or more optical components 115 (e.g., one or more lenses), including one or more optical waveguides 130 (see
As illustrated in
The optical components 115 may focus a user's vision on one or more portions of one or more display panels 125, as shown in
In some examples, the optical components 115 may include a left and right optical components (e.g., left optical component 115-L and right optical component 115-R). The optical components 115 may use plate-shaped (usually planar) waveguides 130 for transmitting angular image information to users' eyes as virtual images from image sources (e.g., light engine and/or MicroLED projectors 120) located out of the user's line of sight 135. The image information may be input near one end of the waveguides 130 and is output near another end of the waveguides 130.
In some examples, the image information may propagate along the optical waveguides 130 as a plurality of angularly related beams that are internally reflected along the waveguide. The optical waveguide 130 can be either a hollow pipe with reflective inner surfaces or an integrator rod with total or partial internal reflection. Additionally or alternatively, the optical waveguide 130 may be a single optical waveguide with different input regions for inputting different color lights from the three MicroLED projectors 120 or multiple optical waveguides or plates stacked on top of each other such that each optical waveguide may input a single color light (e.g., first optical waveguide for accepting red color light from a first MicroLED projector, a second optical waveguide for accepting a blue color light from a second MicroLED projector, and a third optical waveguide for accepting a green color light from the third MicroLED projector). In either instance, the optical waveguide 130 may include an inside surface (facing the user's eye) and an outside surface (facing the ambient environment), with both the inside and outside surfaces being exposed to air or another lower refractive index medium. As such the optical waveguide 130 may be at least partially transparent so that the user can also view the ambient environment through the waveguide.
Diffractive optics may be used for injecting the image information into the waveguides through a first range of incidence angles that are internally reflected by the waveguides as well as for ejecting the image information through a corresponding range of lower incidence angles for relaying or otherwise forming an exit pupil behind the waveguides 130 in a position that can be aligned with the users' eyes 135. Both the waveguides 130 and the diffractive optics at the output end of the waveguides may be at least partially transparent so that the user can also view the ambient environment through the waveguides 130, such as when the image information is not being conveyed by the waveguides or when the image information does not fill the entire field of view.
As discussed above, in some examples, the optical waveguide 130 may be a single waveguide that may include plurality of input regions for receiving at least a portion of light corresponding to an image from the one or more monolithic MicroLED projectors 120. In such instances, the ability of a single optical waveguide 130 to accept input from three different color lights produced by three MicroLED projectors 120 (e.g., 120-a, 120-b, and 120-c) may reduce the number of physical plates that may be used, and thereby reduce the physical size of the display device. The plurality of input regions may include at least a first input region 140-a, a second input region 140-b, and a third input region 140-c. In some instances, the first input region 140-a of the optical waveguide 130 may be physically aligned with the first projector 120-a to receive light corresponding to the first version of the image, the second input region 140-b of the optical waveguide 130 may be physically aligned with the second projector 120-b to receive light corresponding to the second version of the image, and the third input region 140-c of the optical waveguide 130 may be physically aligned with the third projector 120-c to receive light corresponding to the third version of the image. Further, the optical waveguide 130 may include an output region 145 configured to guide light from each of the plurality of MicroLED projectors 120 toward a target to make a final image visible to user's eye 135, wherein the final image is a fully colored image that combines the different monochrome color versions of the image.
In some aspects, the first input region 140-a, the second input region 140-b, and the third input region 140-c may be configured to independently receive the first version of the image, the second version of the image, and the third version of the image, respectively, and the optical waveguide 130 may include a plurality of diffractive optical elements configured to combine the first version of the image, the second version of the image, and the third version of the image to form the final image. In other words, features of the present disclosure provide techniques for each of the plurality of monochrome images (e.g., red, blue, and green) to be combined within the single optical waveguide 130 without requiring separate hardware for the optical combiner, e.g., large prisms to combine three different images together and relay optics to prevent the divergence of the pixels when combining the colors.
In some aspects, each input region 140 of the optical waveguide 130 may include a different grating structure. For example, the first input region 140-a of the optical waveguide 130 may include a first grating structure configured to cause a first amount of phase change upon reflection of the light corresponding to the first version of the image, while the second input region 140-b of the optical waveguide 130 may include a second grating structure configured to cause a second amount of phase change upon reflection of the light corresponding to the second version of the image. Even further, in such instance, the third input region 140-c of the optical waveguide 130 may include a third grating structure configured to cause a third amount of phase change upon reflection of the light corresponding to the version of the image, wherein the first amount, the second amount, and the third amount are different amounts.
By configuring each input region 140 with a different grating structure to separately reflect the light by varying degree of phase change for each version of the image, the aspects of the present disclosure may allow for the three monochrome images to overlap and align within the waveguide to produce a final image that is a fully colored image. Specifically, in some instances, the first grating structure, the second grating structure, and the third grating structure may be arranged such that the light associated with each of the first version of the image, the second version of the image, and the third version of the image overlaps within the optical waveguide to produce the final image. Thus, in such instances, as illustrated in
Alternatively, in other examples—as illustrated in
Continuing with discussion with reference to
Though not shown in
Referring next to
As illustrated, the image to be projected may be represented in angular space where every angle reflects light at the input location of first pixel 202 and second pixel 204. The imaging unit 200 (or MicroLED projector 120) may include a display panel 205 that may display the image to be projected within the optical waveguide 130. Each dashed line in
In this respect, the optical waveguide may reflect light associated with each pixel (e.g., first light rays 215 associated with the first pixel 205, and second light rays 220 associated with the second pixel 210) such that the image may be displayed to the user based on the where the eye would be at the output region of the optical waveguide.
Turning next to
In some aspects, the output port 310 of each MicroLED projector 120 may be is optically coupled (but not necessarily physically coupled) to each of the plurality of input regions 140 of the optical waveguide 130. For example, the first input region 140-a of the optical waveguide 130 may be physically aligned and optically coupled (but not physically coupled) with the first MicroLED projector 120-a to receive light corresponding to the first version of the image output from the first output port 310-a of the first MicroLED projector 120-a. Similarly, the second input region 140-b of the optical waveguide 130 may be physically aligned and optically coupled with the second MicroLED projector 120-b to receive light corresponding to the second version of the image output from the second output port 310-b of the second MicroLED projector 120-b. Finally, the third input region 140-c of the optical waveguide 130 may be physically aligned and optically coupled with the third MicroLED projector 120-c to receive light corresponding to the third version of the image output from the third output port 310-c of the third MicroLED projector 120-c.
The optical waveguide 130 may further include an output region 145 configured to guide light from each of the plurality of MicroLED projectors 120 toward a target (e.g., a user's eye) to make a final image visible to user, wherein the final image is a fully colored image that combines the different monochrome color versions of the image. In some instances, each of the plurality of input regions 140 may independently receive each of the three versions of the monochrome images from the plurality of MicroLED projectors 120 such that the optical waveguide 140 may combine the first version of the image, the second version of the image, and the third version of the image to form the final image by ensuring that the light from each of the different versions of the image overlap.
The optical waveguide 130 may achieve the overlapping image by combination of different grating structures for different input regions, and by nature of the replication process that allow the images to stack up on top of each other upon entering the optical waveguide 130. As discussed above, in some instances, minor calibration corrections may be required to be performed in order to ensure that the three different versions of the image overlap to form a high resolution image to be output from the output region 145.
For the calibration process, features of the present disclosure may utilize a camera in position of an eye to detect light rays output from an optical waveguide 130. The light rays may correspond to a plurality of monochrome color versions of an image generated by a plurality of monolithic MicroLED projectors. Upon detecting the light rays output from the output region 145, the image calibration component 735 (see
For example, if the image calibration component 735 detects a pixel displacement of 20 microns between the first version of the image (e.g., the version produced by green MicroLED projector) and a second version of the image (e.g., the version produced by the blue MicroLED projector), the camera may be configured to detect such displacement and configure one of the MicroLED projectors 120 (e.g., the green MicroLED projector) to offset the green images by 5 pixels. Thus, instead of starting the green image at pixel position 0, the calibration process may offset the green image to pixel 5 position such that the green image and the blue image align or overlap when output from the optical waveguide 130.
Additionally, in some examples, each input region 140 of the optical waveguide 130 may have a different grating structure to cause different variation of phase change upon reflection of light that is input into the optical waveguide 130. For example, in some instances, the first input region 140-a of the optical waveguide 130 may include a first grating structure configured to cause a first amount of phase change upon reflection of the light corresponding to the first version of the image, the second input region 140-b of the optical waveguide 130 may include a second grating structure configured to cause a second amount of phase change upon reflection of the light corresponding to the second version of the image, and the third input region 140-c of the optical waveguide 130 may include a third grating structure configured to cause a third amount of phase change upon reflection of the light corresponding to the version of the image. In some aspects, of the first amount, the second amount, and the third amount may be different amounts that result in all three versions of the image to overlap within the optical guide 130.
In other instances, such as in
In such instances, there may be a number of reasons that the projectors 120 may be positioned on the opposite sides of the waveguide 130. First, in general, the lateral dimensions of the projector 120 may be larger than the output stream. Thus, in order to handle the spatial constraints of mounting the projectors within the display device, it may be necessary to physically separate the projectors 120 to different portions of the optical waveguide. However, it should be appreciated that such separation may not only be limited to positioning the projectors 120 on opposite sides, but may also be on the same side but different ends of the optical waveguide 130. Second, the projectors may be spatially separated in order to allow multiple colors (e.g., blue and red) to work with the same waveguide parameters (e.g., to be projected onto the same grating regions). Thus, in such instance, the first projector 120-a and the second projector 120-b may be physically aligned with the first input region 140-a and second input region 140-b that both share the same grating structure. Meanwhile, the third projector 120-c may be physically aligned with the third input region 140-c that is spatially separated from the first and second input regions (140-a and 140-b). Third, by placing two projectors on one side and the remaining projector on the other side of the waveguide, the opposing input regions may overlap, thereby reducing the area taken up by the input regions and/or the space taken up by the projectors. Such architecture allows the display device 100 to achieve smaller form factor and simpler hardware configuration.
Referring next to
At block 505, the method 500 may include generating a plurality of monochrome color versions of an image using a plurality of monolithic MicroLED projectors. In some examples, the plurality of MicroLED projectors may include at least a first projector generating a first version of the image in a first color, a second projector generating a second version of the image in a second color, and a third projector generating a third version of the image in a third color. In some examples, the first projector may generate the first version of the image in only red color tone, the second projector may generate the second version of the image in only blue color tone, and the third projector may generate the third version of the image in only green color tone. Aspects of block 505 may be performed by a plurality of MicroLED projectors 305-a, 305-b, and 305-c in conjunction with rendering component 730 described with reference to
At block 510, the method 500 may include receiving, at an optical waveguide, light inputs from each of the plurality of MicroLED projectors into a plurality of different input regions of the optical waveguide such that a first input region of the optical waveguide is physically aligned with a first color projector to receive light corresponding to the first version of the image, the second input region of the optical waveguide is physically aligned with the second color projector to receive light corresponding to the second version of the image, and the third input region of the optical waveguide is physically aligned with the third color projector to receive light corresponding to the third version of the image. In some examples, the optical waveguide may have a plurality of internally reflective surfaces to enable light rays of a plurality of different colors generated by the plurality of MicroLED projectors to propagate through its substrate by total internal reflection.
Further, in some examples, the first input region, the second input region, and the third input region may be configured to independently receive the first version of the image, the second version of the image, and the third version of the image, respectively. Further, in such situation, the optical waveguide may include a plurality of diffractive optical elements configured to combine the first version of the image, the second version of the image, and the third version of the image to form the final image.
For example, the first input region of the optical waveguide may include a first grating structure configured to cause a first amount of phase change upon reflection of the light corresponding to the first version of the image. Further, in some examples, the second input region of the optical waveguide may include a second grating structure configured to cause a second amount of phase change upon reflection of the light corresponding to the second version of the image, and the third input region of the optical waveguide may include a third grating structure configured to cause a third amount of phase change upon reflection of the light corresponding to the version of the image, wherein the first amount, the second amount, and the third amount are different amounts. In such instances, the first grating structure, the second grating structure, and the third grating structure may be arranged such that the light associated with each of the first version of the image, the second version of the image, and the third version of the image overlaps within the optical waveguide to produce the final image
In other instances, the first input region and the second input region of the optical waveguide may have a first grating structure which causes light to change phase upon reflection by a first amount, and the third input region of the optical waveguide may have a second grating structure which causes light to change phase upon reflection by a second amount. In other words, in some examples, two colors (e.g., light associated with the first color and the second color image, where the first color is red and second color is blue) may share the same waveguide parameters (e.g., grating structure). In such situations, features of the present disclosure may allow use of a single grating regions for two projectors (e.g., projectors generating the red color monochrome image and the blue color monochrome image) and a separate waveguide parameter for the third projector (e.g., projector generating the green color monochrome image).
By sharing the waveguide parameters for at least two projectors, features of the present disclosure may further reduce the size of the HMD. In order to ensure such implementation, aspects of the present disclosure may configure the first projector and the second projector to be physically positioned at a first side of the optical waveguide such that the light from the first projector and the second projector is projected towards the first input grating structure, and the third projector may be physically positioned at a second side of the optical waveguide such that the light from the third projector is projected towards the second grating structure of the optical waveguide, wherein the second side is opposite the first side. In some examples, the third input region may overlap with at least one of the first input region or the second input region. Aspects of block 510 may be performed by the display such as an optical waveguide 52 described with references to
At block 515, the method 500 may include outputting, through the optical waveguide, light from each of the plurality of MicroLED projectors toward a target to make a final image visible to user, wherein the final image is a fully colored image that combines the plurality of monochrome color versions of an image. Aspects of the 515 may also be performed by the display such an optical waveguide 52 that may include an output region for allowing light to exit to user's eye displaying the rendered image.
Turning next to
At block 605, the method 600 may include detecting, at a camera, light rays output from an optical waveguide, wherein the light rays correspond to a plurality of monochrome color versions of an image generated by a plurality of monolithic MicroLED projectors. In some examples, the plurality of MicroLED projectors may include at least the first projector for generating images in the first monochrome color, the second projector for generating images in the second monochrome color, and a third projector for generating images in a third monochrome color. Aspects of block 605 may be performed by the user interface component 720 described with reference to
At block 610, the method 600 may include determining a first position of a first version of the image from the plurality of monochrome color versions of the image, wherein the first version of the image is generated in first monochrome color by a first projector. In some examples, the optical waveguide receives the first version of the image at a first input region of the optical waveguide. The first input region of the optical waveguide may be physically aligned with the first projector. Aspects of block 610 may be performed by the image calibration component 735 described with reference to
At block 615, the method 600 may include determining a second position of a second version of the image from the plurality of monochrome color versions of the image, wherein the second version of the image is generated in second monochrome color by a second projector. In some examples, the optical waveguide may receive the second version of the image at a second input region of the optical waveguide. The second input region of the optical waveguide may be physically aligned with the second projector. In some instances, the first input region and the second input region are configured to independently receive the first version of the image and the second version of the image, respectively, and wherein the optical waveguide may further include a plurality of diffractive optical elements configured to combine the first version of the image and the second version of the image to form a final image. In other words, features of the present disclosure allow the three monochrome images to be merged into a single image within the single waveguide without requiring prisms and other hardware to combine the three separate images prior to inputting the light into the optical waveguide. Aspects of block 610 may be performed by the image calibration component 735 described with reference to
At block 620, the method 600 may include measuring an image displacement between the first position of the first version of the image and the second position of the second version of the image to calculate an offset value between the first position and the second position, wherein the image displacement is one or both of lateral displacement or angular rotational displacement. Aspects of block 620 may be performed by the image displacement component 740 described with reference to
At block 625, the method 600 may include configuring the display device to adjust at least one image output parameter of the plurality of MicroLED projectors such that the output from the optical waveguide of the plurality of monochrome color versions of the image have an overlapping alignment that meets an alignment threshold. In some examples, configuring the display device may adjust at least one image output parameter includes correcting pixel spacing between the first position of the first version of the image and the second position of the second version of the image by shifting a start position of at least one of the first version of the image or the second version of the image.
In some examples, configuring the display device may further include adjusting a color resolution of the plurality of MicroLED projectors, including lowering the color resolution of the first monochrome color generated by the first projector and/or the second monochrome color generated by the second projector, while increasing the color resolution of the third monochrome color generated by the third projector. For example, in instances the first monochrome color may be red color tone of the image, the second monochrome color may be blue color tone of the image, and the third monochrome color is green color tone of the image, the display device may be able to lower the resolution for blue and/or red, and increase the resolution for the green color. Because the human eye has the highest sensitivity for green light, the display device may be able to display a high resolution image by compensating the lower blue and/or red resolution with an image processing with the higher green resolution. Due to lower blue and/or red resolutions, the form factor of the blue and/or red imaging system may be significantly reduced (e.g., reduction of factor of four in some instances). Thus, the smaller form factor for two of the three projectors may allow the HMD to have a smaller size and form factor. Aspects of block 625 may be performed by the projector configuration component 745 described with reference to
Referring now to
The image display device 400 may include a processor 405 for carrying out one or more processing functions (e.g., methods 500 and 600) described herein. The processor 705 may include a single or multiple set of processors or multi-core processors. Moreover, the processor 705 can be implemented as an integrated processing system and/or a distributed processing system.
The image display device 700 may further include memory 710, such as for storing local versions of applications being executed by the processor 705. In some aspects, the memory 710 may be implemented as a single memory or partitioned memory. In some examples, the operations of the memory 710 may be managed by the processor 705. Memory 710 can include a type of memory usable by a computer, such as random access memory (RAM), read only memory (ROM), tapes, magnetic discs, optical discs, volatile memory, non-volatile memory, and any combination thereof. Additionally, the processor 705, and memory 710 may include and execute operating system (not shown).
Further, the image display device 700 may include a communications component 715 that provides for establishing and maintaining communications with one or more parties utilizing hardware, software, and services as described herein. Communications component 715 may carry communications between components on image display device 705. The communications component 715 may also facilitate communications with external devices to the image display device 700, such as to electronic devices coupled locally to the image display device 700 and/or located across a communications network and/or devices serially or locally connected to image display device 700. For example, communications component 715 may include one or more buses operable for interfacing with external devices. In some examples, communications component 715 establish real-time video communication events, via the network, with another user(s) of the communication system operating their own devices running their own version of the communication client software in order to facilitate augmented reality.
The image display device 700 may also include a user interface component 720 operable to receive inputs from a user of display device 700 and further operable to generate outputs for presentation to the user. User interface component 720 may include one or more input devices, including but not limited to a navigation key, a function key, a microphone, a voice recognition component, joystick or any other mechanism capable of receiving an input from a user, or any combination thereof. Further, user interface component 720 may include one or more output devices, including but not limited to a speaker, headphones, or any other mechanism capable of presenting an output to a user, or any combination thereof.
The image display device 700 may include a rendering component 730 that controls the light engine(s) to generate an image visible to the wearer of the HMD, i.e. to generate slightly different 2D or 3D images that are projected onto the waveguide so as to create the impression of 3D structure. In some examples, the rendering component 730 may be configured to generate a plurality of monochrome color versions of an image using a plurality of monolithic MicroLED projectors, wherein the plurality of MicroLED projectors includes at least a first projector generating a first version of the image in a first color, a second projector generating a second version of the image in a second color, and a third projector generating a third version of the image in a third color. The rendering component 730 may also input light, at an optical waveguide, inputs from each of the plurality of MicroLED projectors into a plurality of different input regions of the optical waveguide such that a first input region of the optical waveguide is physically aligned with a first color projector to receive light corresponding to the first version of the image, the second input region of the optical waveguide is physically aligned with the second color projector to receive light corresponding to the second version of the image, and the third input region of the optical waveguide is physically aligned with the third color projector to receive light corresponding to the third version of the image.
The image display device 700 may further include a display 52 that may be an example of the optics 115 or waveguide 210 described with reference to
The image display device 700 may further an image calibration component 735. In some examples, the image calibration component 735 may be configured to detect, at a camera, light rays output from an optical waveguide, wherein the light rays correspond to a plurality of monochrome color versions of an image generated by a plurality of monolithic MicroLED projectors. The image calibration component 735 may further be configured to determine a first position of a first version of the image from the plurality of monochrome color versions of the image, wherein the first version of the image is generated in first monochrome color by a first projector. The image calibration component 735 may further be configured to determine a second position of a second version of the image from the plurality of monochrome color versions of the image, wherein the second version of the image is generated in second monochrome color by a second projector. The image calibration component 735, and more particularly, the image displacement component 740 may be configured to measure an image displacement between the first position of the first version of the image and the second position of the second version of the image to calculate an offset value between the first position and the second position, wherein the image displacement is one or both of lateral displacement or angular rotational displacement.
The image calibration component 735 may also include a projector configuration component 745 to configure the display device to adjust at least one image output parameter of the plurality of MicroLED projectors such that the output from the optical waveguide of the plurality of monochrome color versions of the image have an overlapping alignment that meets an alignment threshold. In some examples, configuring the display device may adjust at least one image output parameter including correcting pixel spacing between the first position of the first version of the image and the second position of the second version of the image by shifting a start position of at least one of the first version of the image or the second version of the image.
As used in this application, the terms “component,” “system” and the like are intended to include a computer-related entity, such as but not limited to hardware, firmware, a combination of hardware and software, software, or software in execution. For example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration, both an application running on a computing device and the computing device can be a component. One or more components can reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers. In addition, these components can execute from various computer readable media having various data structures stored thereon. The components may communicate by way of local and/or remote processes such as in accordance with a signal having one or more data packets, such as data from one component interacting with another component in a local system, distributed system, and/or across a network such as the Internet with other systems by way of the signal.
Furthermore, various aspects are described herein in connection with a device, which can be a wired device or a wireless device. A wireless device may be a cellular telephone, a satellite phone, a cordless telephone, a Session Initiation Protocol (SIP) phone, a wireless local loop (WLL) station, a personal digital assistant (PDA), a handheld device having wireless connection capability, a computing device, or other processing devices connected to a wireless modem. In contract, a wired device may include a server operable in a data centers (e.g., cloud computing).
It is understood that the specific order or hierarchy of blocks in the processes/flow charts disclosed is an illustration of exemplary approaches. Based upon design preferences, it is understood that the specific order or hierarchy of blocks in the processes/flow charts may be rearranged. Further, some blocks may be combined or omitted. The accompanying method claims present elements of the various blocks in a sample order, and are not meant to be limited to the specific order or hierarchy presented.
The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects. Thus, the claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any aspect described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects. Unless specifically stated otherwise, the term “some” refers to one or more. Combinations such as “at least one of A, B, or C,” “at least one of A, B, and C,” and “A, B, C, or any combination thereof” include any combination of A, B, and/or C, and may include multiples of A, multiples of B, or multiples of C. Specifically, combinations such as “at least one of A, B, or C,” “at least one of A, B, and C,” and “A, B, C, or any combination thereof” may be A only, B only, C only, A and B, A and C, B and C, or A and B and C, where any such combinations may contain one or more member or members of A, B, or C. All structural and functional equivalents to the elements of the various aspects described throughout this disclosure that are known or later come to be known to those of ordinary skill in the art are intended to be encompassed by the claims. Moreover, nothing disclosed herein is intended to be dedicated to the public regardless of whether such disclosure is explicitly recited in the claims. No claim element is to be construed as a means plus function unless the element is expressly recited using the phrase “means for.”
It should be appreciated to those of ordinary skill that various aspects or features are presented in terms of systems that may include a number of devices, components, modules, and the like. It is to be understood and appreciated that the various systems may include additional devices, components, modules, etc. and/or may not include all of the devices, components, modules etc. discussed in connection with the figures.
The various illustrative logics, logical blocks, and actions of methods described in connection with the embodiments disclosed herein may be implemented or performed with a specially-programmed one of a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general-purpose processor may be a microprocessor, but, in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration. Additionally, at least one processor may comprise one or more components operable to perform one or more of the steps and/or actions described above.
Further, the steps and/or actions of a method or algorithm described in connection with the aspects disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, a hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. An exemplary storage medium may be coupled to the processor, such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. Further, in some aspects, the processor and the storage medium may reside in an ASIC. Additionally, the ASIC may reside in a user terminal. In the alternative, the processor and the storage medium may reside as discrete components in a user terminal. Additionally, in some aspects, the steps and/or actions of a method or algorithm may reside as one or any combination or set of codes and/or instructions on a machine readable medium and/or computer readable medium, which may be incorporated into a computer program product.
In one or more aspects, the functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored or transmitted as one or more instructions or code on a computer-readable medium. Computer-readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another. A storage medium may be any available media that can be accessed by a computer. By way of example, and not limitation, such computer-readable media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. Also, any connection may be termed a computer-readable medium. For example, if software is transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave may be included in the definition of medium. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc where disks usually reproduce data magnetically, while discs usually reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
While aspects of the present disclosure have been described in connection with examples thereof, it will be understood by those skilled in the art that variations and modifications of the aspects described above may be made without departing from the scope hereof. Other aspects will be apparent to those skilled in the art from a consideration of the specification or from a practice in accordance with aspects disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
9274338 | Robbins et al. | Mar 2016 | B2 |
9627437 | Ulmer | Apr 2017 | B1 |
9715067 | Brown et al. | Jul 2017 | B1 |
20030160706 | Endress | Aug 2003 | A1 |
20100134520 | Coe-Sullivan | Jun 2010 | A1 |
20110242661 | Simmonds | Oct 2011 | A1 |
20140232759 | Simmonds et al. | Aug 2014 | A1 |
20160252724 | Nikkhoo | Sep 2016 | A1 |
20170184776 | El-Ghoroury | Jun 2017 | A1 |
20180131926 | Shanks et al. | May 2018 | A1 |
20180182275 | Ahmed et al. | Jun 2018 | A1 |
20180284460 | Cheng et al. | Oct 2018 | A1 |
20190227319 | Trail et al. | Jul 2019 | A1 |
Entry |
---|
“The Future of MicroLED Displays using Next-Generation Technologies”, In White Paper of Plessey, 5 Pages, Mar. 2018. |
Kress, et al., “Diffractive and Holographic Optics as Optical Combiners in Head Mounted Displays”, In Proceedings of the ACM Conference on Pervasive and Ubiquitous Computing Adjunct Publication, pp. 1479-1482, Sep. 8, 2013. |
“Non Final Office Action Issued in U.S. Appl. No. 16/281,805”, dated Mar. 18, 2020, 13 Pages. |
Strickland, Keith, “The future of MicroLED Displays using Next-generation Technologies”, In SID Symposium Digest of Technical Papers, vol. 49, Issue 1, May 2018, pp. 696-697. |