The present disclosure relates to visual displays, and in particular to visual display devices using pupil-replicating lightguides.
Visual displays provide information to viewer(s) including still images, video, data, etc. Visual displays have applications in diverse fields including entertainment, education, engineering, science, professional training, advertising, to name just a few examples. Some visual displays, such as TV sets, display images to several users, and some visual display systems, such s near-eye displays or NEDs, are intended for individual users.
An artificial reality system generally includes an NED, for example a headset or a pair of glasses, configured to present content to a user. The NED may display virtual objects or combine images of real objects with virtual objects, as in virtual reality (VR), augmented reality (AR), or mixed reality (MR) applications. For example, in an AR system, a user may view both images of virtual objects (e.g., computer-generated images (CGIs)) and the surrounding environment by seeing through a “combiner” component. The combiner of a wearable display is typically transparent to external light but includes some light routing optics to direct the display light into the user's field of view.
Because a display of HMD or NED is usually worn on the head of a user, a large, bulky, unbalanced, and/or heavy display device with a heavy battery would be cumbersome and uncomfortable for the user to wear. Head-mounted display devices require compact and efficient optical train that conveys an image generated by a microdisplay or a beam scanner to eyes of a user.
Exemplary embodiments will now be described in conjunction with the drawings, in which:
While the present teachings are described in conjunction with various embodiments and examples, it is not intended that the present teachings be limited to such embodiments. On the contrary, the present teachings encompass various alternatives and equivalents, as will be appreciated by those of skill in the art. All statements herein reciting principles, aspects, and embodiments of this disclosure, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
As used herein, the terms “first”, “second”, and so forth are not intended to imply sequential ordering, but rather are intended to distinguish one element from another, unless explicitly stated. Similarly, sequential ordering of method steps does not imply a sequential order of their execution, unless explicitly stated. In
Near-eye displays and augmented reality displays may use pupil-replicating lightguides to expand image light carrying a projected image over an eyebox of the display, i.e., over an area where a user's eye may be located during normal operation of the display. A pupil-replicating lightguide may include a plano-parallel slab of a transparent material propagating the image light in a zigzag pattern by total internal reflection (TIR) from the waveguide's top and bottom surfaces.
One drawback of pupil-replicating lightguides is that by spreading the image light over an extended area, a considerable portion of the image light coupled into the pupil-replicating lightguide never reaches the eye pupil, illuminating instead the user's eyes and face. This causes a considerable reduction of the overall light utilization efficiency. A poor light utilization efficiency may be compensated by using a brighter light source; however, this requires a larger and heavier battery, which increases the size and weight of the display making it uncomfortable to wear for extended periods of time. Furthermore, when the lightguide is a part of an augmented reality system, the image light from an image source internal to the augmented reality system has to compete in brightness with outside light visible by the display user. The augmented image brightness may need to be increased by several orders of magnitude for the generated imagery to be visible in broad daylight. It is therefore highly desirable to improve the light utilization efficiency by a pupil-replicating lightguide.
In accordance with this disclosure, a pupil-replicating lightguide of a display device may include an out-coupling grating having an out-coupling efficiency tunable in a spatially-selective, time-variant manner, enabling the distribution of the image light portions at the eyebox to match the eye pupil position and eye orientation. Thus, the image light distribution may be optimized to account for the eye position and orientation, as well as for a currently displayed portion of the field of view (FOV). The goal is to not send the light energy to eyebox locations where the eye pupil is not present, and/or to not send the image light energy into FOV portions currently not displayed or otherwise non-visible by the viewer.
In accordance with the present disclosure, there is provided a display device comprising an image projector for providing image light carrying an image in angular domain, and a pupil-replicating lightguide coupled to the image projector and comprising an out-coupling grating for out-coupling spaced apart portions of the image light towards an eyebox. An out-coupling efficiency of the out-coupling grating is tunable in a spatially-selective manner for providing a configurable distribution of the image light portions at the eyebox.
The display device may include a controller operably coupled to the image projector and the out-coupling grating and configured to: cause the image projector to provide a first field of view (FOV) portion; increase the out-coupling efficiency of a first portion of the out-coupling grating to increase a first portion of the image light corresponding to the first FOV portion; cause the image projector to provide a second, different FOV portion; and increase the out-coupling efficiency of a second portion of the out-coupling grating to increase a second portion of the image light corresponding to the second FOV portion. The first FOV portion may include a first line of the image in angular domain, and the second FOV portion may include a second, different (not necessarily subsequent) line of the image in angular domain. The first FOV portion may include a first section of the image in angular domain, and the second FOV portion may include a second, different (not necessarily subsequent) section of the image in angular domain. Each section may include several lines.
The display device may further include an eye tracking system for determining a position of a pupil of a user's eye at the eyebox. In such embodiments, the controller may be operably coupled to the image projector, the out-coupling grating, and the eye tracking system, and configured to: cause the eye tracking system to determine the position of the pupil of the user's eye; cause the image projector to provide a first field of view (FOV) portion; and increase the out-coupling efficiency of a first portion of the out-coupling grating to increase a first portion of the image light at the position of the pupil determined by the eye tracking system, the first image light portion corresponding to the first FOV portion.
In some embodiments, the out-coupling grating may include at least one of a polarization volume hologram (PVH) grating, a tunable Pancharatnam-Berry phase (PBP) liquid crystal (LC) grating, a fluidic surface-relief grating, etc. The image projector may include a scanning image projector, for example.
In accordance with the present disclosure, there is provided a display device comprising a scanning image projector for scanning a light beam to provide a line of an image in angular domain, and a pupil-replicating lightguide coupled to the image projector and comprising an out-coupling grating for out-coupling spaced apart portions of the light beam towards an eyebox. An out-coupling efficiency of the out-coupling grating is tunable in a spatially-selective manner. The display device may further include a controller operably coupled to the scanning image projector and the out-coupling grating and configured to: cause the scanning image projector to scan the light beam from a first pixel to a second pixel of the line of the image; and tune the out-coupling efficiency of the out-coupling grating to provide a high-efficiency portion of the out-coupling grating running in coordination with scanning of the light beam by the scanning image projector, to keep a first portion of the light beam at a first location at the eyebox during the scanning, wherein the first portion is out-coupled by the high-efficiency portion of the out-coupling grating. In some embodiments, an x-coordinate x(t) of the high-efficiency portion is expressed as x(t)=D*tan(θ(t)), wherein D is a distance between the out-coupling grating and the first location, θ(t) is an instantaneous scanning angle of the light beam, and t is time.
In some embodiments, the display device further includes an eye tracking system operably coupled to the controller for determining a position of a pupil of a user's eye at the eyebox, wherein the first location is at the pupil position determined by the eye tracking system. The controller may be further configured to adjust the first location upon determining, using the eye tracking system, that the pupil position has shifted. The controller may be further configured to adjust the first location upon determining, using the eye tracking system, that a gaze direction of the user's eye has shifted.
In accordance with the present disclosure, there is further provided a method for providing a line of an image in angular domain to a first location at an eyebox. The method includes angularly scanning a light beam while modulating its brightness to provide the line of the image in angular domain; providing the scanned light beam to a pupil-replicating lightguide comprising an out-coupling grating having an out-coupling efficiency tunable in a spatially-selective manner; and tuning the out-coupling efficiency of the out-coupling grating to provide a high-efficiency portion of the out-coupling grating running in coordination with the scanning of the light beam, to keep a first portion of the light beam at a first location at the eyebox during the scanning, wherein the first portion is out-coupled by the high-efficiency portion of the out-coupling grating. In some embodiments, an x-coordinate x(t) of the high-efficiency portion is expressed as x(t)=D*tan(θ(t)), wherein D is a distance between the out-coupling grating and the first location, θ(t) is an instantaneous scanning angle of the light beam, and t is time.
In some embodiments, the method further includes using an eye tracking system to determine a position of a pupil of a user's eye at the eyebox, wherein the first location is at the pupil position determined by the eye tracking system. The first location may be adjusted upon determining, using the eye tracking system, that the pupil position has shifted.
Referring now to
A pupil-replicating lightguide 120 is coupled to the image projector 110. The pupil-replicating lightguide 120 may be implemented in a slab of transparent material 121 for conveying the image light 112 by a series of alternating zigzag reflections from top 131 and bottom 132 surfaces of the slab 121, e.g. total internal reflections (TIRs). An in-coupler 122, e.g. an in-coupling grating, may be provided for in-coupling the image light 112 into the slab 121.
The pupil-replicating lightguide 120 further includes an out-coupling grating 124 for out-coupling spaced apart portions of the image light towards an eyebox 126. An out-coupling efficiency of the out-coupling grating 124 is tunable in a spatially-selective manner for providing a configurable distribution of the image light portions at the eyebox 126. Herein and throughout the rest of the specification, the terms “switchable”, “tunable”, and “variable” may be used interchangeably. These terms mean that the grating strength, blazing angle, etc., may be controlled by applying an external control signal. In the illustrated example, the grating efficiency spatial distribution includes a high-efficiency region 125 which moves or “runs” from left to right, as indicated by an arrow 128.
The display device 100 may further include a controller 130 operably coupled to the image projector 110 and the pupil-replicating lightguide 120, specifically to its out-coupling grating 124 of the pupil-replicating lightguide 120. The controller 130 may be configured to tune the spatial distribution of the out-coupling efficiency of the out-coupling grating 124 in coordination with operating the image projector 110, to optimize the out-coupling of the image light portions carrying currently displayed FOV portion to a user's eye 127. For example, the controller 130 may cause the image projector 110 to provide a first FOV portion 141, and increase the out-coupling efficiency of a first portion 151 of the out-coupling grating 124 to increase a first portion 161 of the image light 112, the first image light portion 161 corresponding to the first FOV portion 141. The controller may cause the high-efficiency region 125 to shift or run along the out-coupling grating 124 in sync or in coordination with the FOV scanning by the image projector 110. By the time the image projector 110 provides a second, different FOV portion 142, the controller 130 increases the out-coupling efficiency of a second portion 152 of the out-coupling grating 124 to increase a second portion 162 of the image light 112 corresponding to the second FOV portion 142. This enables the out-coupling grating 124 to always be optimized for out-coupling the FOV portion that is currently being provided by the image projector 110, thereby improving the overall light utilization efficiency.
A bottom portion of
The scanning may be optimized not only for section-by-section and line-by-line scanning, but also for pixel-by-pixel scanning at a high enough speed of tunability of the out-coupling grating 124.
A bottom portion of
Turning to
The display device 300 may further include a controller 330 operably coupled to the image projector 110 and the out-coupling grating 124 of the pupil-replicating lightguide 120. The controller 330 may be configured to tune the spatial distribution of the out-coupling efficiency of the out-coupling grating 124 in coordination with operating the image projector 110, to increase the out-coupling of the spaced apart image light portions carrying the currently displayed FOV portion to the eyebox 126, in the following manner. The scanning image projector 110 scans the collimated light beam 109 generated by the light source 102. The controller 330 tunes the out-coupling efficiency of the out-coupling grating 124 to run a high-efficiency portion 351 of the out-coupling grating 124 along the slab 121 and in coordination with the scanning of the collimated light beam 109 by the scanning image projector 110. The controller 330 is configured to keep a first portion 361 of the light beam at a location 381 of the eye 127 in the eyebox 126 during the scanning. As illustrated in
The controller 330 may scan the collimated light beam 109 in accordance with a pre-defined scanning trajectory. By way of a non-limiting example, considering a on-dimensional scanning for simplicity, the scanning angle θ of the collimated light beam 109 may be described by a scanning function 8(t), i.e. 8(t) is an instantaneous scanning angle of the collimated light beam 109 at the time moment t. It follows from the geometry of
x(t)=D*tan(θ(t)), (1)
where D is a distance between the out-coupling grating 124 and the eye location 381. As the angle θ of the collimated light beam 109 is scanned, the high-efficiency portion 351 of the out-coupling grating 124 “travels” to a second location 351′, while the corresponding out-coupled image light portion 361′ remains directed at the eye location 381. The eye location 381 at any given moment of time may be determined by using the eye tracking system 340. When the eye location 381 changes, the controller 330 may adjust the high-efficiency portion 351 movement function x(t) accordingly, to keep the out-coupled image light at the updated eye location at all times.
The adjustment of the movement function x(t) depending on the eye location is illustrated in
The adjustment of the movement function x(t) depending on the eye gaze direction is illustrated in
Referring to
The method 500 may further include using an eye tracking system to determine (508) a position of a pupil of a user's eye at the eyebox. The determined position is taken to be a position at which the out-coupled light beam portion is to be kept during the tuning (506). The determined position may be adjusted (510) upon determining, by the eye tracking system, that the pupil position and/or the gaze direction of the user's eye has shifted.
Non-limiting examples of spatially switchable/tunable gratings usable in lightguides and displays of this disclosure will now be presented. Referring first to
A second substrate 602 is spaced apart from the first substrate 601. The second substrate 602 supports a second conductive layer 612. A cell is formed by the first 611 and second 612 conductive layers. The cell is filled with an LC fluid, forming an LC layer 608. The LC layer 608 includes nematic LC molecules 610, which may be oriented by an electric field across the LC layer 608. The electric field may be provided by applying a voltage V to the first 611 and second 612 conductive layers or conductive electrodes. At least one of the first 611 and second 612 conductive layers may be pixelated to provide the spatially selective application of the voltage for spatial selectivity of the tuning.
The surface-relief grating structure 604 may be formed from a polymer with an isotropic refractive index np of about 1.5, for example. The LC fluid has an anisotropic refractive index. For light polarization parallel to a director of the LC fluid, i.e. to the direction of orientation of the nematic LC molecules 610, the LC fluid has an extraordinary refractive index ne, which may be higher than an ordinary refractive index no of the LC fluid for light polarization perpendicular to the director. For example, the extraordinary refractive index ne may be about 1.7, and the ordinary refractive index no may be about 1.5, i.e. matched to the refractive index np of the surface-relief grating structure 604.
When the voltage Vis not applied (left side of
Referring now to
ϕ(x)=πx/T=πx sin θ/λo (1)
where λo is the wavelength of impinging light, T is a pitch of the PBP LC switchable grating 700, and θ is a diffraction angle given by
θ=sin−1(λo/T) (2)
The azimuthal angle ϕ varies continuously across the surface of an LC layer 704 parallel to XY plane as illustrated in
In
Turning to
Boundary LC molecules 907b at the top surface 905 of the LC layer 904 may be oriented at an angle to the top surface 905. The boundary LC molecules 907b may have a spatially varying azimuthal angle, e.g. linearly varying along X-axis parallel to the top surface 905, as shown in
The boundary LC molecules 907b define relative phases of the helical structures 908 having the helical period p. The helical structures 908 form a volume grating comprising helical fringes 914 tilted at an angle ϕ, as shown in
The helical nature of the fringes 914 of the volume grating makes the PVH grating 900 preferably responsive to light of polarization having one particular handedness, e.g. left- or right-circular polarization, while being substantially non-responsive to light of the opposite handedness of polarization. Thus, the helical fringes 914 make the PVH grating 900 polarization-selective, causing the PVH grating 900 to diffract light of only one handedness of circular polarization. This is illustrated in
Referring now to
At least one of the first 1021 and second 1022 electrode structures may be pixelated/segmented/patterned for imposing a spatially variant electric field onto the 1001 and second 1002 fluids. For example, in 10A and 10B, the first electrode 1021 is patterned, and the second electrodes 1022 is not patterned, i.e. the second electrodes 1022 is a backplane electrode. In the embodiment shown, both the first 1021 and second 1022 electrodes are substantially transparent. For example, the first 1021 and second 1022 electrodes may be indium tin oxide (ITO) electrodes. The individual portions of a patterned electrode may be individually addressable. In some embodiments, the patterned electrode 1021 may be replaced with a continuous, non-patterned electrode coupled to a patterned dielectric layer for creating a spatially non-uniform electric field across the first 1001 and second 1002 fluids.
The thickness of the first 1021 and second 1022 electrodes may be e.g. between 10 nm and 50 nm. The materials of the first 1021 and second 1022 electrodes besides ITO may be e.g. indium zinc oxide (IZO), zinc oxide (ZO), indium oxide (TO), tin oxide (TO), indium gallium zinc oxide (IGZO), etc. The first 1001 and second 1002 fluids may have a refractive index difference of at least 0.1, and may be as high as 0.2 and higher. One of the first 1001 or second 1002 fluids may include polyphenylether, 1,3-bis(phenylthio)benzene, etc. The first 1011 and/or second 1012 substrates may include e.g. fused silica, quartz, sapphire, etc. The first 1011 and/or second 1012 substrates may be straight or curved, and may include vias and other electrical interconnects. The applied voltage may be varied in amplitude and/or duty cycle when applied at a frequency of between 100 Hz and 100 kHz. The applied voltage can change polarity and/or be bipolar. Individual first 1001 and/r second 1002 fluid layers may have a thickness of between 0.5-5 micrometers, more preferably between 0.5-2 micrometer.
To separate the first 1001 and second 1002 fluids, surfactants containing one hydrophilic end functional group and one hydrophobic end functional group may be used. The examples of a hydrophilic end functional group are hydroxyl, carboxyl, carbonyl, amino, phosphate, sulfhydryl. The hydrophilic functional groups may also be anionic groups such as sulfate, sulfonate, carboxylates, phosphates, for example. Non-limiting examples of a hydrophobic end functional group are aliphatic groups, aromatic groups, fluorinated groups. For example, when polyphenyl thioether and fluorinated fluid may be selected as a fluid pair, a surfactant containing aromatic end group and fluronirated end group may be used. When phenyl silicone oil and water are selected as the fluid pair, a surfactant containing aromatic end group and hydroxyl (or amino, or ionic) end group may be used. These are only non-limiting examples.
Referring now to
The purpose of the eye-tracking cameras 1104 is to determine position and/or orientation of both eyes of the user. The eyebox illuminators 1110 illuminate the eyes at the corresponding eyeboxes 1112, allowing the eye-tracking cameras 1104 to obtain the images of the eyes, as well as to provide reference reflections i.e. glints. The glints may function as reference points in the captured eye image, facilitating the eye gazing direction determination by determining position of the eye pupil images relative to the glints images. To avoid distracting the user with the light of the eyebox illuminators 1110, the latter may be made to emit light invisible to the user. For example, infrared light may be used to illuminate the eyeboxes 1112.
Turning to
In some embodiments, the front body 1202 includes locators 1208 and an inertial measurement unit (IMU) 1210 for tracking acceleration of the HMD 1200, and position sensors 1212 for tracking position of the HMD 1200. The IMU 1210 is an electronic device that generates data indicating a position of the HMD 1200 based on measurement signals received from one or more of position sensors 1212, which generate one or more measurement signals in response to motion of the HMD 1200. Examples of position sensors 1212 include: one or more accelerometers, one or more gyroscopes, one or more magnetometers, another suitable type of sensor that detects motion, a type of sensor used for error correction of the IMU 1210, or some combination thereof. The position sensors 1212 may be located external to the IMU 1210, internal to the IMU 1210, or some combination thereof.
The locators 1208 are traced by an external imaging device of a virtual reality system, such that the virtual reality system can track the location and orientation of the entire HMD 1200. Information generated by the IMU 1210 and the position sensors 1212 may be compared with the position and orientation obtained by tracking the locators 1208, for improved tracking accuracy of position and orientation of the HMD 1200. Accurate position and orientation is important for presenting appropriate virtual scenery to the user as the latter moves and turns in 3D space.
The HMD 1200 may further include a depth camera assembly (DCA) 1211, which captures data describing depth information of a local area surrounding some or all of the HMD 1200. The depth information may be compared with the information from the IMU 1210, for better accuracy of determination of position and orientation of the HMD 1200 in 3D space.
The HMD 1200 may further include an eye tracking system 1214 for determining orientation and position of user's eyes in real time. The obtained position and orientation of the eyes also allows the HMD 1200 to determine the gaze direction of the user and to adjust the image generated by the display system 1280 accordingly. The determined gaze direction and vergence angle may be used to adjust the display system 1280 to reduce the vergence-accommodation conflict. The direction and vergence may also be used for displays' exit pupil steering as disclosed herein. Furthermore, the determined vergence and gaze angles may be used for interaction with the user, highlighting objects, bringing objects to the foreground, creating additional objects or pointers, etc. An audio system may also be provided including e.g. a set of small speakers built into the front body 1202.
Embodiments of the present disclosure may include, or be implemented in conjunction with, an artificial reality system. An artificial reality system adjusts sensory information about outside world obtained through the senses such as visual information, audio, touch (somatosensation) information, acceleration, balance, etc., in some manner before presentation to a user. By way of non-limiting examples, artificial reality may include virtual reality (VR), augmented reality (AR), mixed reality (MR), hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include entirely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, somatic or haptic feedback, or some combination thereof. Any of this content may be presented in a single channel or in multiple channels, such as in a stereo video that produces a three-dimensional effect to the viewer. Furthermore, in some embodiments, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, for example, create content in artificial reality and/or are otherwise used in (e.g., perform activities in) artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a wearable display such as an HMD connected to a host computer system, a standalone HMD, a near-eye display having a form factor of eyeglasses, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
The present disclosure is not to be limited in scope by the specific embodiments described herein. Indeed, other various embodiments and modifications, in addition to those described herein, will be apparent to those of ordinary skill in the art from the foregoing description and accompanying drawings. Thus, such other embodiments and modifications are intended to fall within the scope of the present disclosure. Further, although the present disclosure has been described herein in the context of a particular implementation in a particular environment for a particular purpose, those of ordinary skill in the art will recognize that its usefulness is not limited thereto and that the present disclosure may be beneficially implemented in any number of environments for any number of purposes. Accordingly, the claims set forth below should be construed in view of the full breadth and spirit of the present disclosure as described herein.
This application claims priority from U.S. Provisional Patent Application No. 63/286,349 entitled “Active Gratings in Pupil-Replicated Displays and Illuminators”, U.S. Provisional Patent Application No. 63/286,230 entitled “Active Fluidic Optical Element”, both filed on Dec. 6, 2021 and incorporated herein by reference in their entireties; from U.S. Provisional Patent Application No. 63/341,416 entitled “Active Eyebox Solutions and Applications” filed on May 12, 2022, and from U.S. Provisional Patent Application No. 63/392,403 filed Jul. 26, 2022 entitled “DISPLAY DEVICE WITH RUNNING OUT-COUPLING GRATING, which are incorporated herein by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
63286349 | Dec 2021 | US | |
63286230 | Dec 2021 | US | |
63341416 | May 2022 | US | |
63392403 | Jul 2022 | US |