The present disclosure relates to optical components and modules, and in particular to microlens arrays and other components usable in wavefront sensors and display systems using same.
Micro-optics have many applications in areas such as imaging, remote sensing, display systems, optical communications, optical data processing, and so on. Micro-optics enable significant size and weight reduction of optical systems. Micro-optics may be produced inexpensively in large numbers using such processes as stack fabrication and dicing, injection molding, etc.
Micro-optics, such as arrays of microlenses for example, may be used in visual displays and arrayed photodetectors for increasing light efficiency, controlling field of view, and improving spatial directivity. Head mounted displays (HMD), helmet mounted displays, and near-eye displays (NED) are being used increasingly for displaying virtual reality (VR) content, augmented reality (AR) content, mixed reality (MR) content, and the like. Such displays are finding applications in diverse fields including entertainment, education, training and biomedical science, to name just a few examples. The displayed VR/AR/MR content can be three-dimensional (3D) to enhance the experience and to match virtual objects to real objects observed by the user. External environment of a near-eye display may be tracked in real time, and the displayed imagery may be dynamically adjusted depending on the environment, as well as user's head orientation and gaze direction. To sense the environment, various systems may be deployed, e.g. special outward-facing camera systems.
Compact and efficient outside environment monitoring systems may greatly benefit a near-eye display by enabling the user to be immersed into the real-world environment. However, many modern outside monitoring and tracking systems are bulky and heavy. Because a display of HMD or NED is usually worn on the head of a user, a large, bulky, unbalanced, and/or heavy display device would be cumbersome and may be uncomfortable for the user to wear.
Exemplary embodiments will now be described in conjunction with the drawings, in which:
While the present teachings are described in conjunction with various embodiments and examples, it is not intended that the present teachings be limited to such embodiments. On the contrary, the present teachings encompass various alternatives and equivalents, as will be appreciated by those of skill in the art. All statements herein reciting principles, aspects, and embodiments of this disclosure, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
As used herein, the terms “first”, “second”, and so forth are not intended to imply sequential ordering, but rather are intended to distinguish one element from another, unless explicitly stated. Similarly, sequential ordering of method steps does not imply a sequential order of their execution, unless explicitly stated.
One difference of a depth camera as compared to a regular camera is that an image obtained by a depth camera includes not only brightness and/or color information of an object being imaged, but also depth information, i.e. a three-dimensional shape of the object, or of a portion of the object visible to the camera and, in some cases, distance to the object being imaged. A depth camera may obtain information about distance and shape of visible objects by detecting not only optical power density and spectral distribution of an incoming light field, but also a wavefront shape of the light field.
Light field wavefront shape can be measured by using a wavefront sensor. A wavefront sensor may be constructed by placing a microlens array in front of a photodetector array, and processing photodetector array data to measure location of focal spots produced by individual microlenses relative to pixels of the photodetector array. Widespread use of microlens-based wavefront sensors has been hindered by high manufacturing costs, in particular high manufacturing costs of suitable microlens arrays. It is therefore highly desirable to produce high-quality, small-size microlenses inexpensively and with high yield.
In accordance with the present disclosure, an array of microlenses may be manufactured by nanoimprinting a fringe pattern on a suitable substrate capable of keeping the shape of the nanoimprint, e.g. using an imprint resist or elastomer that can be thermally or UV cured after nanoimprinting, followed by an optional reactive ion etching of the nanoimprinted resist layer. Such process allows one to obtain arrays of very small, precisely manufactured microlenses. When the nanoimprinted pattern includes flat binary patterns, very thin lenses may be obtained, much lower than equivalent refractive microlenses.
In accordance with the present disclosure, there is provided a microlens array component comprising a substrate and an array of microlenses formed on the substrate by nanoimprint lithography. Each microlens of the array comprises a plurality of concentric ridges extending from the substrate and separated by concentric grooves. A ratio F of a width of the concentric ridges to a pitch p of the concentric ridges is a function of a radial distance r from a microlens center to the concentric ridges.
In some embodiments, the microlens array component includes an imprint resist layer supported by the substrate, wherein the array of microlenses is formed in the imprint resist layer. The concentric grooves may include air or some filling material. The concentric ridges may be circular, elliptical, square, etc., and may have rectangular, trapezoidal, oval, etc. cross-section. The concentric ridges may have substantially the same height. The substrate of the microlens array component may be flat or curved.
In some embodiments, an effective refractive index n of each microlens of the array of microlenses is the following function of the radial distance r: n(r)=nRF(r)+nG(1−F(r)), where nR is a refractive index of the concentric ridges, and nG is refractive index of the concentric grooves. Each microlens may have a phase profile comprising a plurality of concentric phase profile segments having an amplitude of 2π and adding up to a parabolic phase profile. In some embodiments, each microlens has a phase profile
where f is a focal length of the microlens, λ is wavelength of impinging light, and ϕ(0) is a phase at the microlens center. In some embodiments, a height of the concentric ridges is less than 1700 nm; the pitch p of the concentric ridges is less than 600 nm; and/or each microlens of the array of microlenses is no greater than 0.1 mm.
In accordance with the present disclosure, there is provided a mold for manufacturing a microlens array component. The mold includes an array of inverted microlenses. Each inverted microlens of the array of inverted microlenses comprises concentric mold ridges extending from the mold and separated by concentric mold grooves. A ratio F′ of a width of the concentric mold grooves to a pitch p′ of the concentric mold grooves is a function of a radial distance r′ from the inverted microlens center to the concentric mold grooves. The concentric mold ridges may have a substantially same height.
In accordance with the present disclosure, there is further provided a method of manufacturing a microlens array component. The method includes forming an imprint resist layer on a substrate, obtaining a mold comprising an array of inverted microlenses, and imprinting the imprint resist layer with the mold so as to form an array of microlenses in the imprint resist layer. Each inverted microlens of the array of inverted microlenses comprises concentric mold ridges extending from the mold and separated by concentric mold grooves, wherein a ratio F′ of a width of the concentric mold grooves to a pitch p′ of the concentric mold grooves is a function of a radial distance r′ from an inverted microlens center to the concentric mold grooves. Each microlens of the array of microlenses comprises a plurality of concentric imprint ridges extending from the substrate and separated by concentric imprint grooves, wherein a ratio F of a width of the concentric imprint ridges to a pitch p of the concentric imprint ridges is a function of a radial distance r from the microlens center to the concentric imprint ridges, and F′(r′)=F(r) at r′=r.
In some embodiments, an effective refractive index n of each microlens of the array of microlenses is the following function of the radial distance r: n(r)=nRF(r)+nG(1−F(r)), where nR is a refractive index of the concentric ridges, and nG is refractive index of the concentric grooves. Each microlens may have a phase profile comprising a plurality of concentric phase profile segments having an amplitude of 2p and adding up to a parabolic profile. For example, each microlens may have a phase profile
where f is a focal length of the microlens, l is wavelength of impinging light, and f(0) is a phase at the microlens center. In some embodiments, the plurality of concentric imprint ridges comprises circular imprint ridges. The method may further include reactive ion etching the imprint resist layer after imprinting with the mold.
Referring now to
The array of microlenses 104 may be formed by nanoimprinting, e.g. by depositing an imprint resist layer on the substrate, imprinting the imprint resist layer with a suitable mold having nano-scale ring pattern, and curing the imprint resist. Various methods of forming arrays of microlenses will be considered in more detail further below. The concentric grooves 108 may be filled with air or with a planarizing layer, not shown.
The microlenses 104 may be of any suitable shape, e.g. circular as illustrated, elliptical, rectangular, square, etc. The shape of the microlenses 104 does not need to be tied to the shape of the concentric grooves 106, e.g. the concentric grooves 106 may be circular, while the shape of the microlenses 104 may be square, for example. The microlenses 104 may be disposed on the substrate 102 in a rectangular pattern as shown, in honeycomb pattern, rhombic pattern, etc. The concentric ridges 106 may all have substantially same height h (
In some embodiments, the duty cycle F may determine the effective local refractive index n(r) as follows:
n(r)=nRF(r)+nG(1−F(r)),
where nR is the refractive index of the concentric ridges 106 and nG is the refractive index of the concentric grooves 108. If the concentric grooves 108 contain air, then nG=1.0.
Dependence of the effective refractive index n on profile height h and duty cycle F of the nanoimprinted pattern of the microlens 104 is illustrated in
The “folded” phase profile is illustrated in
where f is the focal length, λ is wavelength of light, and ϕ(0) is the phase delay at the microlens center.
The phase function ϕ(r) may be broken into profile segments 302A, 302B, 302C, 302D, and 302E. The segments 302B, 302C, 302D, and 302E may be shifted down by an integer number of 2π, to form a folded phase profile 300′ comprising a plurality of concentric phase profile segments 302B′, 302C′, 302D′, and 302E′ having an amplitude of 2π and adding up to the parabolic phase profile 304. The folded phase profile 300′ may be represented by a function
The folded phase profile 300′ enables a considerable overall thickness reduction of the microlenses 104, because its amplitude does not exceed 2π.
A general process of nanoimprinting is illustrated in
To obtain the desired microlens shape imprinted into the substrate 400, each inverted microlens of the array of inverted microlenses of the mold 440 may include concentric mold ridges 446 (
F′(r)=F(r) at r′=r (3)
In the embodiment shown, the concentric mold ridges 442 have a substantially same height h′.
Nanoimprinting process enables printing of features with characteristic size of less than 1 micrometer, typically tens to hundreds nanometers. This enables the production of very compact microlenses. Referring back to
Referring now to
A mold is obtained (504), e.g. micromachined in a firm substrate using e-beam nanolithography or another suitable method. The mold geometry may be selected to be inverse to that of an optical component to be manufactured, e.g. as has been explained above with reference to
The imprint resist layer is imprinted (506) with the mold by applying pressure and/or heating above the glass transition temperature of the imprint resist material. While the pressure is applied, the imprint resist layer is cured (508) to preserve the imprinted shape. Heating and/or UV illumination may be used to cure the imprint resist layer. Adhesion between the mold and the imprint resist may be controlled to enable the imprinted pattern to be eventually released (510) from the mold. The microlens or array of microlenses may be formed in the imprint resist layer.
In some embodiments, the pattern imprinted into the polymer layer may be transferred to the underlying substrate. The pattern transfer may be performed e.g. by reactive ion etching. Briefly, the released imprinted pattern is bombarded with ions reactive with the substrate. Exposed areas of the substrate will be etched away, while areas of the substrate protected with the resist will not be etched. Alternatively, the resist layer may also be etched by the reactive ions, at the same or a different rate, depending on chemical composition. When all of the imprint resist layer is etched away to the level of substrate, the pattern nanoimprinted into the resist layer is effectively transferred into the substrate because the exposed areas of the substrate had more time to be etched than areas protected by the imprint resist layer. Thus, the end product includes the desired pattern, e.g. a microlens array pattern, be imprinted in the substrate itself. The remaining imprint resist layer, if any, may then be stripped away.
Referring to
The operation of the wavefront sensor 600 is illustrated in
Referring to
Referring to
Referring to
A fast photodetector 908 may be provided to receive light pulses 904′ reflected from the object 805. The photodetector 908 may include, for example, a fast photodiode capable of detecting the reflected light pulses 904′ with temporal resolution sufficient for optical rangefinding purposes. A controller 910 may be operably coupled to the wavefront sensor 600, the light source 902, and the photodetector 904.
The controller 910 may be configured to operate the light source 902 to emit a probing light pulse 904 towards the object 805. The controller 910 may receive an electric pulse 912 from photodetector, the electric pulse 912 corresponding to a light pulse 904′ reflected from the object 805. The controller 910 may determine a distance to the object 805 from a time delay between emitting the probing light pulse 904 and receiving the electric pulse generated by the photodetector 908 upon receiving the reflected light pulse 904′. The controller 910 may also be configured to receive the image frame 802 from the wavefront sensor 600. The image frame 802 includes images of the light spots focused by corresponding microlenses 604 of the array of microlenses 610 upon illumination with the reflected light pulse 904′, or upon illumination with another light source. Then, the controller 910 may obtain a local wavefront tilt at each microlens 610 from a position of the corresponding light spot in the image frame 802.
The controller 910 may then reconstruct the total wavefront and optical power density distribution of the light beam reflected from the object 805 and impinging onto the wavefront sensor 600. Information related to a distance to the object 805 and/or shape of the object 805 may be obtained from the reconstructed data. For example, the controller 910 may obtain a wavefront radius of the reflected light pulse from the obtained local wavefront tilts at each microlens 604. The distance to the object 805 may be determined from the wavefront radius. In some embodiments, the controller 910 may be configured to obtain a 3D profile of the object from wavefront radiae of reflected light pulses 904′ corresponding to the succession of probing light pulses 904. To that end, the controller 910 may operate the light source 902 to emit a succession of the probing light pulses 904, and may operate the optical scanner 906 to scan the succession of probing light pulses 904 over the object 805. In some embodiments, the light source 902 may be used to merely illuminate the object 805 for detection by the wavefront sensor 600. The light source 902, therefore, does not need to be a pulsed light source; it may provide continuous-wave illuminating light, e.g. near-infrared light, to illuminate the object 805.
Turning to
The display device 1000 may further include a controller 1008 operably coupled to the image sources 1002 for providing image frames to be displayed to the left and right eyes of the user placed at the eyeboxes 1005. An eye tracker 1010 may be operably coupled to the controller 1008 for providing a real-time information about user eye's position and/or orientation. The controller 1008 may be configured to determine the user's current gaze direction from that information, and adjust the image frames to be displayed to the user, for a more realistic immersion of the user into virtual or augmented environment.
The display device 1000 may further include an imaging optical rangefinder 1014, e.g. the imaging optical rangefinder 900 of
Embodiments of the present disclosure may include, or be implemented in conjunction with, an artificial reality system. An artificial reality system adjusts sensory information about outside world obtained through the senses such as visual information, audio, touch (somatosensation) information, acceleration, balance, etc., in some manner before presentation to a user. By way of non-limiting examples, artificial reality may include virtual reality (VR), augmented reality (AR), mixed reality (MR), hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include entirely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, somatic or haptic feedback, or some combination thereof. Any of this content may be presented in a single channel or in multiple channels, such as in a stereo video that produces a three-dimensional effect to the viewer. Furthermore, in some embodiments, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, for example, create content in artificial reality and/or are otherwise used in (e.g., perform activities in) artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a wearable display such as an HMD connected to a host computer system, a standalone HMD, a near-eye display having a form factor of eyeglasses, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
Referring to
In some embodiments, the front body 1102 includes locators 1108 and an inertial measurement unit (IMU) 1110 for tracking acceleration of the HMD 1100, and position sensors 1112 for tracking position of the HMD 1100. The IMU 1110 is an electronic device that generates data indicating a position of the HMD 1100 based on measurement signals received from one or more of position sensors 1112, which generate one or more measurement signals in response to motion of the HMD 1100. Examples of position sensors 1112 include: one or more accelerometers, one or more gyroscopes, one or more magnetometers, another suitable type of sensor that detects motion, a type of sensor used for error correction of the IMU 1110, or some combination thereof. The position sensors 1112 may be located external to the IMU 1110, internal to the IMU 1110, or some combination thereof.
The locators 1108 are traced by an external imaging device of a virtual reality system, such that the virtual reality system can track the location and orientation of the entire HMD 1100. Information generated by the IMU 1110 and the position sensors 1112 may be compared with the position and orientation obtained by tracking the locators 1108, for improved tracking accuracy of position and orientation of the HMD 1100. Accurate position and orientation is important for presenting appropriate virtual scenery to the user as the latter moves and turns in 3D space.
The HMD 1100 may further include a depth camera assembly (DCA) 1111, which captures data describing depth information of a local area surrounding some or all of the HMD 1100. To that end, the DCA 1111 may include a laser radar (LIDAR), or a similar device. The depth information may be compared with the information from the IMU 1110, for better accuracy of determination of position and orientation of the HMD 1100 in 3D space.
The HMD 1100 may further include an eye tracking system 1114 for determining orientation and position of user's eyes in real time. The obtained position and orientation of the eyes also allows the HMD 1100 to determine the gaze direction of the user and to adjust the image generated by the display system 1180 accordingly. In one embodiment, the vergence, that is, the convergence angle of the user's eyes gaze, is determined. The determined gaze direction and vergence angle may also be used for real-time compensation of visual artifacts dependent on the angle of view and eye position. Furthermore, the determined vergence and gaze angles may be used for interaction with the user, highlighting objects, bringing objects to the foreground, creating additional objects or pointers, etc. An audio system may also be provided including e.g. a set of small speakers built into the front body 1102.
Referring to
As described above with reference to
The I/O interface 1115 is a device that allows a user to send action requests and receive responses from the console 1190. An action request is a request to perform a particular action. For example, an action request may be an instruction to start or end capture of image or video data or an instruction to perform a particular action within an application. The I/O interface 1115 may include one or more input devices, such as a keyboard, a mouse, a game controller, or any other suitable device for receiving action requests and communicating the action requests to the console 1190. An action request received by the I/O interface 1115 is communicated to the console 1190, which performs an action corresponding to the action request. In some embodiments, the I/O interface 1115 includes an IMU that captures calibration data indicating an estimated position of the I/O interface 1115 relative to an initial position of the I/O interface 1115. In some embodiments, the I/O interface 1115 may provide haptic feedback to the user in accordance with instructions received from the console 1190. For example, haptic feedback can be provided when an action request is received, or the console 1190 communicates instructions to the I/O interface 1115 causing the I/O interface 1115 to generate haptic feedback when the console 1190 performs an action.
The console 1190 may provide content to the HMD 1100 for processing in accordance with information received from one or more of: the IMU 1110, the DCA 1111, the eye tracking system 1114, and the I/O interface 1115. In the example shown in
The application store 1155 may store one or more applications for execution by the console 1190. An application is a group of instructions that, when executed by a processor, generates content for presentation to the user. Content generated by an application may be in response to inputs received from the user via movement of the HMD 1100 or the I/O interface 1115. Examples of applications include: gaming applications, presentation and conferencing applications, video playback applications, or other suitable applications.
The tracking module 1160 may calibrate the AR/VR system 1150 using one or more calibration parameters and may adjust one or more calibration parameters to reduce error in determination of the position of the HMD 1100 or the I/O interface 1115. Calibration performed by the tracking module 1160 also accounts for information received from the IMU 1110 in the HMD 1100 and/or an IMU included in the I/O interface 1115, if any. Additionally, if tracking of the HMD 1100 is lost, the tracking module 1160 may re-calibrate some or all of the AR/VR system 1150.
The tracking module 1160 may track movements of the HMD 1100 or of the I/O interface 1115, the IMU 1110, or some combination thereof. For example, the tracking module 1160 may determine a position of a reference point of the HMD 1100 in a mapping of a local area based on information from the HMD 1100. The tracking module 1160 may also determine positions of the reference point of the HMD 1100 or a reference point of the I/O interface 1115 using data indicating a position of the HMD 1100 from the IMU 1110 or using data indicating a position of the I/O interface 1115 from an IMU included in the I/O interface 1115, respectively. Furthermore, in some embodiments, the tracking module 1160 may use portions of data indicating a position or the HMD 1100 from the IMU 1110 as well as representations of the local area from the DCA 1111 to predict a future location of the HMD 1100. The tracking module 1160 provides the estimated or predicted future position of the HMD 1100 or the I/O interface 1115 to the processing module 1165.
The processing module 1165 may generate a 3D mapping of the area surrounding some or all of the HMD 1100 (“local area”) based on information received from the HMD 1100. In some embodiments, the processing module 1165 determines depth information for the 3D mapping of the local area based on information received from the DCA 1111 that is relevant for techniques used in computing depth. In various embodiments, the processing module 1165 may use the depth information to update a model of the local area and generate content based in part on the updated model.
The processing module 1165 executes applications within the AR/VR system 1150 and receives position information, acceleration information, velocity information, predicted future positions, or some combination thereof, of the HMD 1100 from the tracking module 1160. Based on the received information, the processing module 1165 determines content to provide to the HMD 1100 for presentation to the user. For example, if the received information indicates that the user has looked to the left, the processing module 1165 generates content for the HMD 1100 that mirrors the user's movement in a virtual environment or in an environment augmenting the local area with additional content. Additionally, the processing module 1165 performs an action within an application executing on the console 1190 in response to an action request received from the I/O interface 1115 and provides feedback to the user that the action was performed. The provided feedback may be visual or audible feedback via the HMD 1100 or haptic feedback via the I/O interface 1115.
In some embodiments, based on the eye tracking information (e.g., orientation of the user's eyes) received from the eye tracking system 1114, the processing module 1165 determines resolution of the content provided to the HMD 1100 for presentation to the user on the electronic display 1125. The processing module 1165 may provide the content to the HMD 1100 having a maximum pixel resolution on the electronic display 1125 in a foveal region of the user's gaze. The processing module 1165 may provide a lower pixel resolution in other regions of the electronic display 1125, thus lessening power consumption of the AR/VR system 1150 and saving computing resources of the console 1190 without compromising a visual experience of the user. In some embodiments, the processing module 1165 can further use the eye tracking information to adjust where objects are displayed on the electronic display 1125 to prevent vergence-accommodation conflict and/or to offset optical distortions and aberrations.
The hardware used to implement the various illustrative logics, logical blocks, modules, and circuits described in connection with the aspects disclosed herein may be implemented or performed with a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general-purpose processor may be a microprocessor, but, in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration. Alternatively, some steps or methods may be performed by circuitry that is specific to a given function.
The present disclosure is not to be limited in scope by the specific embodiments described herein. Indeed, other various embodiments and modifications, in addition to those described herein, will be apparent to those of ordinary skill in the art from the foregoing description and accompanying drawings. Thus, such other embodiments and modifications are intended to fall within the scope of the present disclosure. Further, although the present disclosure has been described herein in the context of a particular implementation in a particular environment for a particular purpose, those of ordinary skill in the art will recognize that its usefulness is not limited thereto and that the present disclosure may be beneficially implemented in any number of environments for any number of purposes. Accordingly, the claims set forth below should be construed in view of the full breadth and spirit of the present disclosure as described herein.