This disclosure relates generally to optics, and in particular to lenses.
Systems with different focal points are useful in both imaging and display applications. Generally, different focal points are accomplished with adjusting refractive lenses with respect to one another.
Non-limiting and non-exhaustive embodiments of the invention are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified.
Embodiments of systems and methods of sensing structured light with time of flight (TOF) with duty cycle modulation are described herein. In the following description, numerous specific details are set forth to provide a thorough understanding of the embodiments. One skilled in the relevant art will recognize, however, that the techniques described herein can be practiced without one or more of the specific details, or with other methods, components, materials, etc. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring certain aspects.
Reference throughout this specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, the appearances of the phrases “in one embodiment” or “in an embodiment” in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
Throughout this specification, several terms of art are used. These terms are to take on their ordinary meaning in the art from which they come, unless specifically defined herein or the context of their use would clearly suggest otherwise.
In some implementations of the disclosure, the term “near-eye” may be defined as including an element that is configured to be placed within 50 mm of an eye of a user while a near-eye device is being utilized. Therefore, a “near-eye optical element” or a “near-eye system” would include one or more elements configured to be placed within 50 mm of the eye of the user.
In aspects of this disclosure, visible light may be defined as having a wavelength range of approximately 380 nm-700 nm. Non-visible light may be defined as light having wavelengths that are outside the visible light range, such as ultraviolet light and infrared light. Infrared light having a wavelength range of approximately 700 nm-1 mm includes near-infrared light. In aspects of this disclosure, near-infrared light may be defined as having a wavelength range of approximately 700 nm-1.6 μm.
In aspects of this disclosure, the term “transparent” may be defined as having greater than 90% transmission of light. In some aspects, the term “transparent” may be defined as a material having greater than 90% transmission of visible light.
In the context of head mounted displays, it may be advantageous to have an optical system with more than one focal length. This disclosure includes various implementations of a multi-focal optical system. The disclosed multi-focal optical systems combine spatial polarization of display light to illuminate a Pancharatnam-Berry Phase (PBP) lens. When the display light has a first polarization orientation, the PBP lens focuses the display light at a first focal length. And, when the display light has a second polarization (different from the first polarization orientation), the PBP lens focuses the display light at a second focal length that is different than the first focal length. In some implementations, the multi-focal system can provide four different focal lengths for display light. These and other embodiments are described in more detail in connections with
The different polarization orientations of the display light propagate to a super achromatic quarter waveplate (SAQWP) 130. The quarter waveplate (QWP) is configured to shift the polarization axis of incident light by π/4. Therefore, incident linearly polarized light may be converted to circularly polarized light by the QWP. The QWP may be made of birefringent materials such as quartz, organic material sheets, or liquid crystal, for example.
The display light propagating toward the PBP lens has right-hand circular polarization (RHCP) if the display light was emitted by the first display pixels (and thus encountered the first pixels of the pixelated polarization array), in the illustrated implementation. The display light propagating toward the PBP lens has left-hand circular polarization (LHCP) if the display light was emitted by the second display pixels (and thus encountered the second pixels of the pixelated polarization array), in the illustrated implementation. Thus, the system can be driven by processing logic (not particularly illustrated) to emit RHCP display light or LHCP by driving the first display pixels or the second display pixels, respectively.
After encountering the SAQWP 130, the display light encounters the PBP lens 140 (also known as a geometric phase lens). The PBP lens 140 is configured to focus the first polarized light having the first polarization orientation to a first focal length and focus the second polarized light having the second polarization orientation to a second focal length.
The left portion of
The multi-focal system of
In operation, processing logic may be configured to (1) drive the multi-focal system to a first focal length (e.g. focal length 421) by driving the first display pixels to generate the first display light (solid lines light in
Focal length 423 is greater than focal length 421. The focal length of the second display light (dashed lines light) in
A second implementation of the disclosure includes Time-of-Flight (TOF) with Duty Cycle Modulation. In the context of eye-tracking, fringe patterns have been used to illuminate an eyebox region with a fringe pattern and then an image is captured of the fringe pattern to achieve three-dimensional (3D) sensing of an eye occupying the eyebox region. Image processing may be performed on the image of the fringe pattern to generate eye data that is used to determine a depth and/or position of the eye in the eyebox region, for example. A fringe pattern may be projected onto an eye by a projection system of a head mounted device (e.g. smartglasses or Augmented Reality, Mixed Reality, or Virtual Reality headset) and a camera of the head mounted device may capture the image of the fringe pattern for depth sensing analysis.
In fringe based structured light sensing, one of the key questions is how many different fringe patterns need to be illuminated onto the target.
Indirect Time-of-Flight (iTOF) has also been used in eye-tracking contexts for measuring depth. However, iTOF may struggle with depth precision due to modulation speed. For example, it is difficult to achieve ˜100 micron resolution unless the modulation frequency is above 2 GHz.
In implementations of the disclosure, pulse width modulation (PWM) based iTOF illumination is utilized for fringe pattern projection, as shown in
In
Camera 1150 captures an eye-tracking image of fringe pattern 1130 illuminating the eye and/or face of a user. The spatial frequency of fringe pattern(s) 1130 illuminating eye and/or face is extracted from the eye-tracking image using image processing techniques.
The phase of returning light that is reflected by the eye and/or face is also measured using iTOF. A portion of the laser light that illuminates the eye or face will be reflected/scattered back to MEMS scanner 1105 and become incident on beam splitter 1103. Beam Splitter 1103 directs at least a portion of the returning light to photodiode 1109 as iTOF light 1107. Photodiode 1109 generates a TOF signal (e.g. an electrical current) that may be mixed with a local oscillator signal to generate a phase signal. This phase signal indicates the TOF of the laser light and the TOF of the laser light can be used to determine the depth of the eye/face from the iTOF system.
Therefore, system 1100 generates (1) spatial frequency data from the eye-tracking image of camera 1150; and (2) TOF information from photodiode 1109 by measuring the phase of returning light.
In some implementations, the TOF information provides a rough depth of the eye/face for a given scan point of MEMS scanner 1105. For example, the TOF information may resolve depth at a particular scan point to about 1 cm. The rough depth from the TOF information can then be further refined using the spatial frequency data from the eye-tracking image captured by camera 1150. For example, the spatial frequency data combined with the TOF information may resolve the depth at the particular scan point to approximately 100 microns. By getting both the TOF information and the frequency data from the same fringe illumination projection, eye depth data accuracy can be increased while running at lower frequency (e.g. 200 kHz instead of 2 GHz) and displaying/projecting fewer different fringe illumination patterns to the eyebox region. Processing speed and power consumption may also decrease as a result since less image processing is required to analyze fewer fringe illumination patterns.
In
In a third implementation of this disclosure includes Optical Vortex Arrays for Adaptive Dynamic Illumination. Structured illumination based three-dimensional (3D) sensing relies on illuminating the objects with known patterns of light. Signal processing algorithms are then used to compute the 3D profile of the object. In one group of implementations, stereo cameras can be used to correlate two images captured from different angles. If no distinct intensity-based features are present (for example white sclera in the eyeball) this becomes problematic. Alternatively, a stereo capture method can be enhanced by projecting structured light patterns that create the features that are needed for depth reconstruction. In another group of implementations, the object is illuminated with a known pattern and the image for the object is captured with a single camera. Multiple patterns can be used here and the reconstruction can be done using classical approaches such as phase shifting profilometry or with ML machine learning-based approaches. The key problem with these methods is that different objects can benefit from the use of different illumination patterns and the switching between these patterns typically require complex projection systems.
As a part of this disclosure, the Applicant introduces utilizing correlated vortex arrays to illuminate the object for profilometry purposes. Correlated vortex arrays can be generated using a static or dynamic phase mask in conjunction with an illumination source. The pattern has a high depth of field that is desirable for measuring objects such as an eye. Also, the patterns can be dynamically changed to adjust for different fitment of augmented reality (AR) and/or virtual reality (VR) device. Additionally, the area which is illuminated can be dynamically controlled for improving power efficiency. Dynamic pattern control can be achieved through the use of a single static phase mask and by modulating the coherence properties of the light source. Alternatively, a dynamic phase mask in conjunction with adjustable light source could be used. Finally, polarization modulation at the light source level can be employed for switching between two or more pre-set illumination patterns.
A fourth implementation of this disclosure includes non-line-of-sight retinal imaging. Retinal imaging is of great value for eye tracking contexts, for example. Traditional eye trackers usually use the eye surface information for eye tracking, such as information form cornea, iris, sclera, and pupil. However, the retina is the most direct tissue to receive light signals. With retina imaging, display(s) in VR and AR devices can achieve further miniaturization, lower power consumption, and higher quality. However, traditional retinal imaging usually assumes lines of sight, where the camera and lasers directly image the retina. However, due to eye rotation, the retina may be outside a camera's direct line of sight, leading to the system only working for a very constrained gaze range.
In this fourth implementation of the disclosure, non-line-of-sight technologies for retinal imaging is described. Time-of-flight (TOF) based non-line-of-sight (NLOS) imaging systems have recently demonstrated impressive results. TOF principles takes a light pulse to travel along a direct path. NLOS uses multi-bounce light paths to measure the visual appearance of a 3D shape.
In the illumination optical path, the laser emits laser light that propagates through the beam splitter to illuminate the galvo mirror(s). The galvo mirrors are driven to direct the laser light to different points on the display lens that may be included in a head mounted device. At least a portion of that light reflects off the display lens and propagates through the pupil to illuminate the retina.
In the return optical path, at least a portion of the illumination light reflects/scatters off the retina as returning light. The returning light propagates back through the pupil and encounters the display lens. The display lens reflects at least a portion of the returning light back to the galvo mirror(s) which directs the light to the beam splitter. The beam splitter directs at least a portion of the light to the SPAD camera.
Thus, the display lens may be used as a scatter wall. By acquiring these timing measurements for different laser positions on the wall, the 3D geometry and appearance of the retina can be reconstructed. Additionally, as the eye rotates, the galvo mirror(s) can be driven to direct the laser light to different positions on the display lens to reflect through the pupil and onto the retina to be reflected/scattered back to the SPAD camera on the returning optical path. Thus, this implementation allows for imaging of the eye even for different eye rotations.
Embodiments of the invention may include or be implemented in conjunction with an artificial reality system. Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, haptic feedback, or some combination thereof, and any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Additionally, in some embodiments, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, e.g., create content in an artificial reality and/or are otherwise used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head-mounted display (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
The term “processing logic” in this disclosure may include one or more processors, microprocessors, multi-core processors, Application-specific integrated circuits (ASIC), and/or Field Programmable Gate Arrays (FPGAs) to execute operations disclosed herein. In some embodiments, memories (not illustrated) are integrated into the processing logic to store instructions to execute operations and/or store data. Processing logic may also include analog or digital circuitry to perform the operations in accordance with embodiments of the disclosure.
A “memory” or “memories” described in this disclosure may include one or more volatile or non-volatile memory architectures. The “memory” or “memories” may be removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules, or other data. Example memory technologies may include RAM, ROM, EEPROM, flash memory, CD-ROM, digital versatile disks (DVD), high-definition multimedia/data storage disks, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information for access by a computing device.
Networks may include any network or network system such as, but not limited to, the following: a peer-to-peer network; a Local Area Network (LAN); a Wide Area Network (WAN); a public network, such as the Internet; a private network; a cellular network; a wireless network; a wired network; a wireless and wired combination network; and a satellite network.
Communication channels may include or be routed through one or more wired or wireless communication utilizing IEEE 802.11 protocols, short-range wireless protocols, SPI (Serial Peripheral Interface), I2C (Inter-Integrated Circuit), USB (Universal Serial Port), CAN (Controller Area Network), cellular data protocols (e.g. 3G, 4G, LTE, 5G), optical communication networks, Internet Service Providers (ISPs), a peer-to-peer network, a Local Area Network (LAN), a Wide Area Network (WAN), a public network (e.g. “the Internet”), a private network, a satellite network, or otherwise.
A computing device may include a desktop computer, a laptop computer, a tablet, a phablet, a smartphone, a feature phone, a server computer, or otherwise. A server computer may be located remotely in a data center or be stored locally.
The processes explained above are described in terms of computer software and hardware. The techniques described may constitute machine-executable instructions embodied within a tangible or non-transitory machine (e.g., computer) readable storage medium, that when executed by a machine will cause the machine to perform the operations described. Additionally, the processes may be embodied within hardware, such as an application specific integrated circuit (“ASIC”) or otherwise.
A tangible non-transitory machine-readable storage medium includes any mechanism that provides (i.e., stores) information in a form accessible by a machine (e.g., a computer, network device, personal digital assistant, manufacturing tool, any device with a set of one or more processors, etc.). For example, a machine-readable storage medium includes recordable/non-recordable media (e.g., read only memory (ROM), random access memory (RAM), magnetic disk storage media, optical storage media, flash memory devices, etc.).
The above description of illustrated embodiments of the invention, including what is described in the Abstract, is not intended to be exhaustive or to limit the invention to the precise forms disclosed. While specific embodiments of, and examples for, the invention are described herein for illustrative purposes, various modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize.
These modifications can be made to the invention in light of the above detailed description. The terms used in the following claims should not be construed to limit the invention to the specific embodiments disclosed in the specification. Rather, the scope of the invention is to be determined entirely by the following claims, which are to be construed in accordance with established doctrines of claim interpretation.
This application claims priority to U.S. Provisional Application No. 63/430,774 filed Dec. 7, 2022 and U.S. Provisional Application No. 63/429,779 filed Dec. 2, 2022, which are hereby incorporated by reference.
Number | Date | Country | |
---|---|---|---|
63430774 | Dec 2022 | US | |
63429779 | Dec 2022 | US |