The present invention relates generally to a head-mounted projection display, and more particularly, but not exclusively to a polarized head-mounted projection display including a light engine and a compact, high-performance projection lens for use with reflective microdisplays.
The head mounted projection display (HMPD), as an alternative to the conventional eyepiece-based head mounted display (HMD), has attracted much interest in recent years, because it offers the ability to design a wide field of view (FOV), low distortion and ergonomically compact optical see-through head mounted display (OST-HMD). Like most OST-HMDs, however, one of the limiting factors for the HMPD technology is its low image brightness and contrast, which limits the feasibility to apply such information to outdoor or well-lit indoor environments such as operating rooms. Due to the multiple beamsplitting through a beamsplitter and low retroreflectivity of typical retroreflective materials, the overall efficiency of a HMPD is around 4%. For instance, with a miniature backlit active matrix liquid crystal display (AMLCD) as the image source, the luminance of the observed image is estimated to be 4 cd/m2, while the average luminance of a well-lit indoor environment is over 100 cd/m2. As a result, the low-brightness image of HMPDs will appear washed out in such well-lit environments. In fact, most optical see-through HMDs, including HMPD, are typically operated under a dimmed lighting condition.
To address this problem, a polarized head-mounted projection display (p-HMPD) was proposed (H. Hua and C. Gao, “A polarized head-mounted projective displays,” Proceedings of 2005 IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 32-35, October 2005) and a prototype based on a pair of transmissive AMLCDs was designed recently (H. Hua, C. Gao “Design of a bright polarized head-mounted projection display” Applied Optics, Vol. 46, Issue 14, pp. 2600-2610, May 2007). A pair of 1.3″ color AMLCDs was used as the image sources which have a resolution of (640*3)*480 pixels. 1.4″ Alphalight™ RGB LED panels (Teledyne Inc., Los Angeles, Calif.) were used as the backlighting sources. By carefully manipulating the polarization states of the light propagating through the system, a p-HMPD can potentially be three times brighter than a traditional non-polarized HMPD design using the same microdisplay technologies. A schematic design of a monocular p-HMPD configuration is illustrated in
The image on the LCD display is projected through the projection lens, forming a real intermediate image. The light from the LCD is manipulated to be S-polarized so that its polarization direction is matched with the high-reflection axis of the polarized beamsplitter (PBS). After the projected light is reflected by the PBS, it is retroreflected back to the same PBS by a retroreflective screen. The depolarization effect by the retroreflective screen is less than 10% within ±20 degrees and is less than 20% up to ±30 degrees. As a result, the retroreflected light remains dominantly the same polarization as its incidence light. In order to achieve high transmission through the PBS after the light is retroreflected back, a quarter-wave retarder is placed between the PBS and the retroreflective screen. By passing through the quarter wave retarder twice, the incident S-polarized light is converted to P-polarization and transmits through the PBS with high efficiency. Thus the projected image from the microdisplay can be then observed at the exit pupil of the system where the eye is placed.
However, since a transmissive LCD microdisplay has a low transmission efficiency of around 5%, the overall performance of the first p-HMPD prototype is still unsatisfactory in a well-lit environment. Furthermore, owing to its inherent low pixel fill factor, a transmissive AMLCD microdisplay typically has a relatively low resolution. Accordingly, it would be an advance in the field of head-mounted projection displays to provide a head-mounted projection display which has higher luminance while maintaining high contrast.
In one of its aspects, the present invention provides a compact, telecentric projection lens for use in a head-mounted projection display system. The projection lens may include a plurality of lens elements configured to have an overall length that is no more than about two times the effective focal length of the projection lens. In addition, the plurality of lens elements may be configured so that the projection lens is telecentric in image space. As one measure of the compactness the projection lens may have an overall length that is no more than about 85% larger than the effective focal length of the projection lens. Further, the projection lens may have a back focal length that is about 40% larger than the effective focal length. To further facilitate the use of the projection lens in a head-mounted display, the projection lens may also be lightweight and have a ratio of the weight of the projection lens to the square of the F-number of the projection lens of less than about 2 to 1. For example, in one configuration the present invention provides a projection lens with a F-number of 2.1 and a weight of only 8.2 g.
In another of its aspects the present invention provides a telecentric optical illumination system for use with a reflective microdisplay. The illumination system may include a source of optical radiation having a first polarization state and a polarized beamsplitter disposed at a location to receive optical radiation from the optical radiation source. The beamsplitter may further be oriented relative to the source of optical radiation to reflect the received optical radiation. In addition, the illumination system may include an optical retarder disposed at a location relative to the beamsplitter to receive optical radiation reflected by the beamsplitter and oriented to convert the polarization state of the received optical radiation to be circularly polarized. A reflector having optical power may also be disposed at a location to receive the circularly polarized optical radiation from the optical retarder and to reflect the received optical radiation back through the retarder and beamsplitter. The reflector may be a concave spherical reflector. Still further, in order to provide an optical illumination system that is telecentric in image space, the illumination system may include a pupil disposed at the source of optical radiation.
In yet another of its aspects, the present invention provides a head-mounted projection display system which may include the aforementioned illumination system, a microdisplay disposed one reflector focal length away from the reflector for receiving optical radiation from the reflector, and a telecentric projection lens. The projection lens may assume the configuration of the aforementioned projection lens.
The foregoing summary and the following detailed description of the preferred embodiments of the present invention will be best understood when read in conjunction with the appended drawings, in which:
In order to realize the above-mentioned benefits afforded by a reflective microdisplay 120 in the context of a head-mounted projection display, suitable optical designs are required for both a light engine 100, to deliver uniform, telecentric illumination to the microdisplay 120, and for a telecentric projection lens 300 to collect the light reflected by the microdisplay 120 for delivery to the user,
Turning now to the overall system, design began by selection of an exemplary microdisplay 120 from which the performance requirements of both the light engine 100 and projection lens 300 may be derived. Major properties of several candidate microdisplay technologies were evaluated, including AMLCD, organic light emitting displays (OLEDs), liquid crystal on silicon (LCOS), and ferroelectric Liquid-crystal-on-Silicon (FLCOS) were considered. For use in the examples provided below, the SXGA-R2D FLCOS microdisplay kit (Forth Dimensional Displays Limited, Birmingham, UK) was selected as an exemplary microdisplay for use in the present invention due to its high resolution, optical efficiency, and compactness. However, though a FLCOS microdisplay is used for illustrative purposes in the designs presented below, it is understood that the present invention is not limited to such a microdisplay, but may be used in conjunction with suitable other displays, such as, reflective microdisplay types like LCOS, MEMS, Texas Instruments DLP, and so forth, for example.
The usage of a FLCOS microdisplay 120 makes the prototype design of a HMPD quite different from previous designs of HMPD optics. One of the key differences is the requirement for the custom-designed light engine 100 to illuminate the microdisplay 120. The FLCOS microdisplay 120 is considered as the combination of a mirror and an electrically switchable quarter-wave retarder formed by the liquid crystal layers. The FLCOS microdisplay 120 works most efficiently when the illumination rays are normally incident upon the display surface. To ensure the high contrast of the output image, it is recommended to limit the incident angle to the range of ±16 degrees, which imposes a critical requirement on the design of both the light engine 100 and projection lens 300. The key requirements for the light engine 100 include: (1) that the illumination system be image-space telecentric to ensure that for every pixel on the display surface the incident chief ray is normal to the display surface; and (2) that the cone angle of the ray bundle is smaller than 16 degrees.
Likewise, owing to the reflective nature of the microdisplay 120, the output ray bundles from every pixel of the microdisplay 120 are telecentric with a cone angle smaller than 16 degrees. To efficiently collect rays from the microdisplay 120 and form a projected image with uniform illumination, the projection lens 300 must be image-space telecentric. To the contrary, a projection system using backlit AMLCDs, which have a relatively large viewing angle, and thus relaxed requirement for the angle of incidence ray bundles, can relax the telecentric constraint to gain compactness.
Design of a Compact Light Engine
Light source selection is the first step in the light engine design. Based on the design requirement of a HMPD, there are several constraints on the source selection and the light engine design. First, safety is a primary concern in any head-mounted device. Therefore sources with low power consumption and low heat dissipation are highly desired. Secondly, compactness and low weight are always critical for HMD systems. Finally, in order to generate an image with high brightness and uniformity across the whole FOV, the illumination on the microdisplay 120 should be uniform and bright. A 0.5″ Alphalight™ color LED panel 102 (Teledyne Inc., Los Angeles, Calif.) is selected for the p-HMPD prototype design. The LED panel 102 is compatibly driven by the color sequential technique of the FLCOS displays. Table 1 summarizes the major specifications of the microdisplay 120 and LED panel 102. The luminance distribution of the LED panel 102,
The basic idea of the design of the light engine 100 is to place the LED panel 102 at the front focal point of a positive lens (or reflector) and the microdisplay 120 at the back focal point. In order to make a compact system, a concave spherical reflector 106, a PBS 104 and a quarter wave retarder 108 are used to fold the length of the system in half. The microdisplay 120 is disposed at the conjugate position of the LED panel 102, and both the microdisplay 120 and LED panel 102 are at the focal point of the reflector 106. With a polarizer 114 in front of the LED panel 102, S-polarized light from the LED panel 102 is transmitted through the polarizer 114 and is reflected by the PBS 104. A quarter-wave retarder 108 is placed between the reflector 106 and PBS 104 and its fast axis is at 45 degrees with the polarization direction of S-polarized light. By passing through the retarder 108 twice, the reflected light by the reflector 106 becomes P-polarized light and will transmit the PBS 104 to illuminate the microdisplay 120 with high efficiency.
In this design, the LED panel 102 itself can be taken as the pupil of the system to form an image-space telecentric system, where the ray bundle received by the microdisplay 120 is symmetric with the display normal. With both LED panel 102 and microdisplay 120 at the focal point of the reflector 106, the light distribution on the microdisplay 120 is the Fourier transform of that of the LED panel 102. Thus the spatial distribution on the microdisplay 120 can be derived as
where Edisplay(x,y) is the illuminance at (x,y) on the display assuming the center of the display is at the origin, LLED(θ) is the luminance of LED panel 102 as a function of angle, SLED is the area of the LED panel 102 and f is the focal length of the reflector 106. Across the microdisplay 120, the ratio of the luminance at the center of the microdisplay 120 to that at the edge is
while D is the diagonal size of the microdisplay 120. To get better uniformity on the microdisplay 120, a reflector 106 with larger focal length is preferred. But meanwhile, larger focal length will result in a less compact structure and a smaller solid angle with lower luminance efficiency. Considering all these factors, a reflector 106 with 35 mm focal length and 35 mm diameter was selected. As a result, the ratio of the maximum luminance to the minimum luminance on the microdisplay 120 is 1:0.82, and the cone angle of the ray bundle on the microdisplay 120 is within 8.6 degrees. The light within the cone angle of 18 degrees emitted by the LED panel 102 can be collected by the reflector 106 to illuminate the microdisplay 120 while the light emitted at a larger angle is wasted.
In order to further improve the light efficiency and uniformity of the light engine 100, a mirror-based tapered light pipe 110 was designed to recycle the light with emission angles larger than 18 degrees.
Numerical Simulation
To determine the parameters of the tapered light pipe 110 and to examine the light efficiency and uniformity of the light engine 100, the light engine 100 was modeled using LightTools® (Optical Research Associates, Pasadena, Calif.). In the simulation, the total power of the source was set to be 1 lumen. A light receiver was placed on the microdisplay 120 to estimate the efficiency of the light engine 100 and to evaluate the light distribution on the microdisplay 120. Through the simulation, it is shown that the light engine 100 has higher uniformity and light efficiency when the tilt angle, a, of the mirror is 18 degrees. By balancing the performance and space constraint of the light engine 100, the mirror length, t, is selected to be 8 mm.
As indicated by the simulation results, with a mirror based light pipe 110, the light efficiency has increased from 8.93% to 12.3%, and the non-uniformity, quantified by the average standard deviation of the illuminance distribution across the display area, has reduced from 5.61% to 2.15%. Thus, the system with the light pipe 110 has higher efficiency and better uniformity than without.
Design of a Compact, High-Performance Projection Lens
Based on the design of the light engine 100, a lightweight and compact image-telecentric projection lens 300 is designed. In this section, the design process of the projection lens 300 is described and the performance of the projection lens 300 is analyzed.
Projection Lens Specification
Although projection optics do not scale as much with the increase of FOV as eyepiece-type optics, which makes designing wide FOV, optical see-through HMPD systems relatively easier than conventional HMDs, there are a few factors in HMPD systems that impose limits on the FOV. First of all, the use of a planar PBS 116 or a regular beamsplitter in front of the eye, which is oriented at 45 degrees with the optical axis, sets up the FOV upper limit of 90 degrees. Furthermore, a wide FOV requires a large size of PBS 116 and retarder 118 and consequently challenges the compactness and lightweight of the display system. The limit of allowable PBS and retarder dimensions is set by the interpupilary distance (IPD), which is in the range of 55 to 75 mm for over 95% of the population. Thirdly, previous investigation on retroreflective materials shows that the retroreflectance of currently available materials drops off significantly for light incident at angles beyond ±35°. A FOV beyond 70 degrees will inevitably cause vignetting-like effect and compromise image uniformity. Finally, the angular resolution of the microdisplay 120 degrades with the increase of the FOV. Taking into account these factors, a target was set to design the projection system with a FOV of 55 degrees, which corresponds to an effective focal length (EFL) of 21.6 mm for the selected FLCOS microdisplay 120.
In addition to being image-space telecentric, the projection lens 300 must have a large back focal length (BFL) to ensure enough space for the PBS 104 which is placed between the microdisplay 120 and projection lens 300. Based on the light engine design, the BFL is chosen to be at least 30.5 mm. Thus this projection lens 300 is also a reverse telephoto lens. (It is understood that, in the designs below, the microdisplay 120 is located at the image plane, hence the reason that the BFL represents the space allotted for the PBS 104 placed therebetween.)
Since a user's eye is positioned at the conjugate position to the entrance pupil of the projection lens 300, the entrance pupil diameter is very critical for comfortable observation. Typically it is suggested the pupil diameter of the projection system for HMPDs should be around 10-12 mm. This range of pupil size allows an eye swivel of about ±21° up to 26.5° within the eye sockets without causing vignetting or loss of image with a typical 3-mm eye pupil in the lighting conditions provided by HMPDs. Furthermore, it allows a ±5 mm to 6 mm IPD tolerance for different users without the need to mechanically adjust the IPD of the binocular optics. Considering the short focal length of the optics, a target entrance pupil with a diameter of at least 10 mm was selected, which leads to a projection system with an F/# of 2.16.
A lens 10 from Nanba U.S. Pat. No. 6,236,521 was used as the starting point of the design, with
The projection lens 210 is rotationally symmetric, and thus the optimization is only necessary over half of the full FOV for the radial direction. Three representative wavelengths (i.e., 486 nm, 589 nm and 656 nm) were set with the weights of 1, 2 and 1, respectively. Five fields, corresponding to 0, 7°, 14°, 21° and 27.5°, respectively, were used in the optimization process to represent the half FOV. The weights of the five fields were adjusted in the optimization process to balance the MTF performances across the entire FOV. During the optimization, all surface curvatures, surface thicknesses, and coefficients of aspheric surfaces were set to be variables. Several constraints were set to satisfy the specifications of the overall system and each individual lens element, including the EFL, BFL, OAL, distortion requirements, and the center thickness of individual elements, for example. The telecentric requirement was satisfied by setting the exit pupil distance to be at least 210 mm from the image plane. This distance corresponds to a deviation of the chief ray by 3° from a perfectly telecentric system, which yields a good balance between the overall optical performance and the system compactness considering the difficulty in designing a perfectly telecentric lens with a short OAL.
One of the major problems of the lens 210 in
This lens 220 is composed of five glass elements 221-225 (with aspheres on two surfaces 227, 229) and weighs about 38.7 grams, which needs to be significantly reduced to obtain a lightweight p-HMPD system. Plastic materials were selected to replace the glass elements 221-225. Considering that the density of glass is about three times of most plastics, it is expected that the weight of the lens 220 would drop to around 10 grams. The drawback of using the plastic materials is also obvious. While there are many choices of optical glass, from low to high refractive index and low to high Abbe number, for plastics only very limited number of materials are available for diamond turning fabrication.
The initial target was to replace the glass elements 223-225 on the right side of the stop with the plastics, since they contributed most of the weight due to their large aperture. Polystyrene with low Abbe number and Cyclic Olefin Polymer (COC) with relatively high Abbe number were selected to replace the glass materials of the doublet 223, 224. After this replacement, the optimization was rerun to achieve the desired specifications of the lens 220. Since the last element 225 has the highest optical power among all the elements in the lens 220 and high Abbe number, COC was selected for the last element 225. Unfortunately, after the plastic replacement the system had much worse performance compared with the system before replacing the last glass element 225. Chromatic aberration dominates the resulting system. To effectively correct the residual chromatic aberration, a diffractive optical element (DOE) was added to the system.
A DOE can be viewed as a material with large dispersion but opposite in sign to conventional materials, i.e., the Abbe number of a DOE is approximately −3.5 for the visible spectrum. This DOE with negative Abbe number would help to correct the chromatic aberration. The substrate shape of a diffractive surface DOE can be spherical, planar being a special case, or aspheric. The commonly used orders of diffraction are 0, −1, or 1. The +1 order of diffraction was adopted. The diffractive surface can be used on any of the lens surfaces as long as it can help to correct the chromatic aberration. However, after a few trials, it was found to be most effective to place the diffractive surface on the left surface 227 of the last element 225. The DOE quadratic coefficient was set to be a variable up to 12 orders in the optimization process. Finally, to further reduce the lens weight, the first two glass elements 221, 222 on the left of the stop were replaced with Acrylic and Polystyrene, respectively.
Considering the fact that a doublet requires higher cost in the fabrication process, the doublet 223, 224 was split into two single elements. The split of the doublet 223, 224 also offered extra freedom in the optimization, and helped improve the overall performance. Finally, through several rounds of optimization, a telecentric lens 300 with an OAL of 34 mm and a total weight of 8.2 grams was obtained. The ratio of the weight of the lens to the square of the F-number was 1.86 to 1.
The lens 300 includes five lens elements 310, 320, 330, 340, 350 ordered left-to-right from object space to image space. The first two lens elements 310, 320 form a first lens group 315 having very little optical power (e.g., the focal length of the lens group 315 is two orders of magnitude larger than that of the lens 300, see Tables 5, 9) disposed to the left of the stop. The first lens element 310 is a negative meniscus lens having two surfaces that are concave towards image space, with the surface 311 closest to image space being aspheric. The second lens element 320 is a positive lens disposed adjacent the first lens element 310. The third through fifth lens elements 330, 340, 350 are disposed on the right side of the stop, with the third element 330 closest to the stop being a negative lens and the fourth and fifth elements 340, 350 closest to the image plane being positive lenses. The first surface 331 of the third lens element 330 closest to the stop is aspheric as is the first surface 351 of the fifth lens element 350. In addition, the first surface 351 of the fifth lens element 350 includes the aforementioned DOE.
Performance Analysis of the Projection Lens
The diffraction efficiency of a DOE drops as its physical features become finer near the edge.
The optical performance of the optimized lens 300 is assessed on the image plane at the five representative field angles for three different wavelengths. The spot diagrams are shown in
p-HMPD Prototype
With the design of the light engine 100 and projection lens 300 complete, a prototype of the new p-HMPD was built,
Considering both the ergonomic and aesthetic factors, the optics 100, 120, 300, were mounted vertically so that the width of the helmet was around the average width of the adult head. In the vertical direction, the optics were mounted according to the shape of the head, and the associated electronics were mounted on the top of the helmet. A main drawback of the vertical mount is that a ghost image of the ground formed by the PBS 104 is overlaid with the projected image, which leads to reduced image contrast. This problem, however, can be solved by blocking the optical path from the ground.
To make the system more compact and lighter, the mount of the light engine 100 with the microdisplays 120 was fabricated separately and then integrated with the shell as a whole. The lens 300 position relative to the microdisplay 120 is adjustable to provide a projected image with adjustable magnification.
The helmet shells were fabricated using rapid prototyping techniques, in which physical models are fabricated layer by layer directly from a 3D CAD model. The helmet shells were assembled and attached to an off-the-shelf headband that offers head-size adjustment. The front and side views of the prototype are shown in
These and other advantages of the present invention will be apparent to those skilled in the art from the foregoing specification. Accordingly, it will be recognized by those skilled in the art that changes or modifications may be made to the above-described embodiments without departing from the broad inventive concepts of the invention. It should therefore be understood that this invention is not limited to the particular embodiments described herein, but is intended to include all changes and modifications that are within the scope and spirit of the invention as set forth in the claims.
This application is a continuation application of U.S. application Ser. No. 16/685,204 filed on, Nov. 15, 2019, which is a continuation application of U.S. application Ser. No. 16/170,860 filed on, Oct. 25, 2018, which is a continuation application of U.S. application Ser. No. 15/092,831 filed on, Apr. 7, 2016, which is a continuation application of U.S. application Ser. No. 13/955,076 filed on, Jul. 31, 2013, which is divisional application of U.S. application Ser. No. 12/863,771 filed on Oct. 29, 2010, which is a 371 application of International Application No. PCT/US2009/31606 filed Jan. 21, 2009, which claims the benefit of priority of U.S. Provisional Application No. 61/011,789, filed on Jan. 22, 2008, the entire contents of which applications are incorporated herein by reference.
This invention was made with government support under Grant No. IIS0534777 awarded by NSF. The government has certain rights in the invention.
Number | Name | Date | Kind |
---|---|---|---|
3632184 | King | Jan 1972 | A |
3992084 | Nakamura | Nov 1976 | A |
4468101 | Ellis | Aug 1984 | A |
4669810 | Wood | Jun 1987 | A |
4753522 | Nishina | Jun 1988 | A |
4863251 | Herloski | Sep 1989 | A |
5109469 | Duggan | Apr 1992 | A |
5172272 | Aoki | Dec 1992 | A |
5172275 | Dejager | Dec 1992 | A |
5416315 | Filipovich | May 1995 | A |
5436763 | Chen | Jul 1995 | A |
5526183 | Chen | Jun 1996 | A |
5572229 | Fisher | Nov 1996 | A |
5621572 | Fergason | Apr 1997 | A |
5625495 | Moskovich | Apr 1997 | A |
5699194 | Takahashi | Dec 1997 | A |
5701202 | Takahashi | Dec 1997 | A |
5706136 | Okuyama | Jan 1998 | A |
5818632 | Stephenson | Oct 1998 | A |
5880711 | Tamada | Mar 1999 | A |
5880888 | Schoenmakers | Mar 1999 | A |
5917656 | Hayakawa | Jun 1999 | A |
5959780 | Togino | Sep 1999 | A |
6008781 | Furness | Dec 1999 | A |
6023373 | Inoguchi | Feb 2000 | A |
6028606 | Kolb | Feb 2000 | A |
6034823 | Togino | Mar 2000 | A |
6198577 | Kedar | Mar 2001 | B1 |
6201646 | Togino | Mar 2001 | B1 |
6236521 | Nanba | May 2001 | B1 |
6239915 | Takagi | May 2001 | B1 |
6243199 | Hansen | Jun 2001 | B1 |
6271972 | Kedar | Aug 2001 | B1 |
6384983 | Yamazaki | May 2002 | B1 |
6396639 | Takayoshi | May 2002 | B1 |
6404561 | Isono | Jun 2002 | B1 |
6404562 | Ota | Jun 2002 | B1 |
6433376 | Kim | Aug 2002 | B2 |
6433760 | Vaissie | Aug 2002 | B1 |
6493146 | Inoguchi | Dec 2002 | B2 |
6510006 | Togino | Jan 2003 | B1 |
6563648 | Gleckman | May 2003 | B2 |
6646811 | Inoguchi | Nov 2003 | B2 |
6653989 | Nakanishi | Nov 2003 | B2 |
6671099 | Nagata | Dec 2003 | B2 |
6731434 | Hua | May 2004 | B1 |
6829113 | Togino | Dec 2004 | B2 |
6963454 | Martins | Nov 2005 | B1 |
6999239 | Martins | Feb 2006 | B1 |
7152977 | Ruda | Dec 2006 | B2 |
7177083 | Holler | Feb 2007 | B2 |
7230583 | Tidwell | Jun 2007 | B2 |
7249853 | Weller-Brophy | Jul 2007 | B2 |
7405881 | Shimizu | Jul 2008 | B2 |
7414791 | Urakawa | Aug 2008 | B2 |
7522344 | Curatu | Apr 2009 | B1 |
8467133 | Miller | Jun 2013 | B2 |
8503087 | Amirparviz | Aug 2013 | B1 |
8511827 | Hua | Aug 2013 | B2 |
9201193 | Smith | Dec 2015 | B1 |
9239453 | Cheng | Jan 2016 | B2 |
9310591 | Hua | Apr 2016 | B2 |
9581818 | Suzuki | Feb 2017 | B2 |
9720232 | Hua | Aug 2017 | B2 |
9874760 | Hua | Jan 2018 | B2 |
10495859 | Hua | Dec 2019 | B2 |
10593507 | Hua | Mar 2020 | B2 |
10739578 | Hua | Aug 2020 | B2 |
10805598 | Hua | Oct 2020 | B2 |
10809533 | Cheng | Oct 2020 | B2 |
11150449 | Hua | Oct 2021 | B2 |
11181746 | Hua | Nov 2021 | B2 |
20010009478 | Yamazaki | Jul 2001 | A1 |
20010048561 | Heacock | Dec 2001 | A1 |
20020015116 | Park | Feb 2002 | A1 |
20020041446 | Nagaoka | Apr 2002 | A1 |
20020060850 | Takeyama | May 2002 | A1 |
20020063913 | Nakamura | May 2002 | A1 |
20020067467 | Dorval | Jun 2002 | A1 |
20020114077 | Javidi | Aug 2002 | A1 |
20020181115 | Massof | Dec 2002 | A1 |
20030076591 | Ohmori | Apr 2003 | A1 |
20030090753 | Takeyama | May 2003 | A1 |
20040136097 | Park | Jul 2004 | A1 |
20040164927 | Suyama | Aug 2004 | A1 |
20040196213 | Tidwell | Oct 2004 | A1 |
20040218243 | Yamazaki | Nov 2004 | A1 |
20040233551 | Takahashi | Nov 2004 | A1 |
20050036119 | Ruda | Feb 2005 | A1 |
20050179868 | Seo | Aug 2005 | A1 |
20050248849 | Urey | Nov 2005 | A1 |
20060028400 | Lapstun | Feb 2006 | A1 |
20060119951 | McGuire | Jun 2006 | A1 |
20070109505 | Kubara | May 2007 | A1 |
20070246641 | Baun | Oct 2007 | A1 |
20080036853 | Shestak | Feb 2008 | A1 |
20080094720 | Yamazaki | Apr 2008 | A1 |
20080291531 | Heimer | Nov 2008 | A1 |
20090115842 | Saito | May 2009 | A1 |
20090168010 | Vinogradov | Jul 2009 | A1 |
20090256943 | Kondo | Oct 2009 | A1 |
20100091027 | Oyama | Apr 2010 | A1 |
20100109977 | Yamazaki | May 2010 | A1 |
20100208372 | Heimer | Aug 2010 | A1 |
20100271698 | Kessler | Oct 2010 | A1 |
20100289970 | Watanabe | Nov 2010 | A1 |
20110037951 | Hua | Feb 2011 | A1 |
20110043644 | Munger | Feb 2011 | A1 |
20110075257 | Hua | Mar 2011 | A1 |
20110090389 | Saito | Apr 2011 | A1 |
20110109880 | Nummela | May 2011 | A1 |
20110221656 | Haddick | Sep 2011 | A1 |
20120013988 | Hutchin | Jan 2012 | A1 |
20120019557 | Aronsson | Jan 2012 | A1 |
20120050891 | Seidl | Mar 2012 | A1 |
20120057129 | Durnell | Mar 2012 | A1 |
20120081800 | Cheng | Apr 2012 | A1 |
20120113092 | Bar-Zeev | May 2012 | A1 |
20120160302 | Citron | Jun 2012 | A1 |
20120162549 | Gao | Jun 2012 | A1 |
20120242697 | Border | Sep 2012 | A1 |
20120262802 | Huang | Oct 2012 | A1 |
20130100524 | Magarill | Apr 2013 | A1 |
20130112705 | McGill | May 2013 | A1 |
20130182317 | Takahashi | Jul 2013 | A1 |
20130187836 | Cheng | Jul 2013 | A1 |
20130222896 | Komatsu | Aug 2013 | A1 |
20130258461 | Sato | Oct 2013 | A1 |
20130285885 | Nowatzyk | Oct 2013 | A1 |
20130286053 | Fleck | Oct 2013 | A1 |
20130300634 | White | Nov 2013 | A1 |
20130329304 | Hua | Dec 2013 | A1 |
20140009845 | Cheng | Jan 2014 | A1 |
20140035959 | Paul | Feb 2014 | A1 |
20140049833 | Totani | Feb 2014 | A1 |
20140071539 | Gao | Mar 2014 | A1 |
20140300869 | Hirsch | Oct 2014 | A1 |
20140347361 | Alpaslan | Nov 2014 | A1 |
20140361957 | Hua | Dec 2014 | A1 |
20150168802 | Bohn | Jun 2015 | A1 |
20150177445 | Takagi | Jun 2015 | A1 |
20150201176 | Graziosi | Jul 2015 | A1 |
20150208061 | Yang | Jul 2015 | A1 |
20150212321 | Zhao | Jul 2015 | A1 |
20150277129 | Hua | Oct 2015 | A1 |
20150346495 | Welch | Dec 2015 | A1 |
20150363978 | Maimone | Dec 2015 | A1 |
20160011419 | Gao | Jan 2016 | A1 |
20160085075 | Cheng | Mar 2016 | A1 |
20160239985 | Haddick et al. | Aug 2016 | A1 |
20160320620 | Maimone | Nov 2016 | A1 |
20170078652 | Hua | Mar 2017 | A1 |
20170102545 | Hua | Apr 2017 | A1 |
20170202633 | Liu | Jul 2017 | A1 |
20180045949 | Hua | Feb 2018 | A1 |
20190346686 | Hua | Nov 2019 | A1 |
20200103654 | Cheng | Apr 2020 | A1 |
20200110250 | Hua | Apr 2020 | A1 |
20200192118 | Hua | Jun 2020 | A1 |
20200278550 | Hua | Sep 2020 | A1 |
20200386993 | Hua | Dec 2020 | A1 |
20200393676 | Hua | Dec 2020 | A1 |
20200393677 | Hua | Dec 2020 | A1 |
20210006773 | Hua | Jan 2021 | A1 |
20210014473 | Hua | Jan 2021 | A1 |
20210088795 | Cheng | Mar 2021 | A1 |
20210103148 | Hua | Apr 2021 | A1 |
20210255464 | Hua | Aug 2021 | A1 |
Number | Date | Country |
---|---|---|
2019204862 | Jul 2019 | AU |
1252133 | May 2000 | CN |
101359089 | Feb 2009 | CN |
101424788 | May 2009 | CN |
110770633 | Feb 2020 | CN |
110879468 | Mar 2020 | CN |
110914741 | Mar 2020 | CN |
0408344 | Jan 1991 | EP |
1102105 | May 2001 | EP |
2928034 | Aug 2009 | FR |
02200074 | Aug 1990 | JP |
H03101709 | Apr 1991 | JP |
08160345 | Jun 1996 | JP |
H09218375 | Aug 1997 | JP |
H09297282 | Nov 1997 | JP |
H1013861 | Jan 1998 | JP |
H10307263 | Nov 1998 | JP |
H11326820 | Nov 1999 | JP |
2000249974 | Sep 2000 | JP |
2001013446 | Jan 2001 | JP |
2001066543 | Mar 2001 | JP |
2001145127 | May 2001 | JP |
2001238229 | Aug 2001 | JP |
2002148559 | May 2002 | JP |
2003241100 | Aug 2003 | JP |
2006091333 | Apr 2006 | JP |
2006276884 | Oct 2006 | JP |
2007101930 | Apr 2007 | JP |
2010072188 | Apr 2010 | JP |
2014505381 | Feb 2014 | JP |
9923647 | May 1999 | WO |
2004079431 | Sep 2004 | WO |
2007002694 | Jan 2007 | WO |
2007085682 | Aug 2007 | WO |
2007002694 | Dec 2007 | WO |
2007140273 | Dec 2007 | WO |
2008089417 | Jul 2008 | WO |
2011134169 | Nov 2011 | WO |
2012064546 | May 2012 | WO |
2012118573 | Sep 2012 | WO |
2013112705 | Aug 2013 | WO |
2014062912 | Apr 2014 | WO |
2015134738 | Sep 2015 | WO |
2015134740 | Sep 2015 | WO |
2015184409 | Dec 2015 | WO |
2016033317 | Mar 2016 | WO |
2018052590 | Mar 2018 | WO |
2021242667 | Dec 2021 | WO |
Entry |
---|
Armitage, David, Ian Underwood, and Shin-Tson Wu. Introduction to Microdisplays. Chichester, England: Wiley, 2006. |
International Search Report dated Jun. 18, 2010 in corresponding international application PCT/US2010/031799. |
Wartenberg, Philipp, “EyeCatcher, the Bi-directional OLED Microdisplay,” Proc. of SID 2011. |
R.J. Jacob, “The use of eye movements in human-computer interaction techniques: what you look at is what you get”, ACM Transactions on Information Systems, 9(2), 152-69, 1991. |
Loschky, L.C. and Wolverton, G.S., “How late can you update gaze-contingent multiresolutional displays without detection” ACM Trans. Mult. Comp. Comm, and App. 3, Nov. 2007. |
Reingold, E.M., L.C. Loschky, G.W. McConkie and D.M. Stampe, “Gaze-contingent multiresolutional displays: An integrative review,” Hum. Factors 45, 307-328 (2003). |
Written Opinion dated Jun. 18, 2010 in corresponding international application PCT/US2010/031799. |
Geisler, W.S., J.S. Perry and J. Najemnik, “Visual search: The role of peripheral information measured using gaze-contingent displays,” J. Vision 6, 858-873 (2006). |
Duchowski, A.T., and A. Coltekin, “Foveated gaze-contingent displays for peripheral LOD management, 3D visualization, and stereo imaging,” ACM Trans. on Mult. Comp., Comm., and App. 3, 1-21, (2007). |
Akeley et al., “A Stereo Display Prototype with Multiple Focal Distances,” ACM Trans. Graphics 23:804-813 (2004). |
Cruz-Neira et al., ‘Surround-Screen Projection-Based Virtual Reality: the Design and Implementation of the CAVE,’ Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques pp. 135-142, ACM SIGGRAPH, ACM Press (1993). |
Davis et al., “Accommodation to Large Disparity Stereograms,” Journal of AAPOS 6:377-384 (2002). |
Downing et al., “A Three-Color, Solid-State, Three-Dimensional Display,” Science 273:1185-1189 (1996). |
Edgar et al., “Visual Accommodation Problems with Head-Up and Helmet-Mounted Displays,” Displays 15:68-75 (1994). |
Favalora et al., “100 Million-Voxel Volumetric Display,” Proc. SPIE 4712:300-312 (2002). |
Graham-Rowe, “Liquid Lenses Make a Splash,” Nature-Photonics pp. 2-4 (2006). |
International Search Report dated Feb. 10, 2011 from PCT/CN2010/072376. |
Heanue et al., “Volume Holographic Storage and Retrieval of Digital Data,” Science 265:749-752 (1994). |
Hua, “Merging the Worlds of Atoms and Bits: Augmented Virtual Environments,” Optics and Photonics News 17:26-33 (2006). |
Hua et al., ‘Design of a Bright Polarized Head-Mounted Projection Display’ Applied Optics 46:2600-2610 (2007). |
Inoue et al., “Accommodative Responses to Stereoscopic Three-Dimensional Display,” Applied Optics, 36:4509-4515 (1997). |
Kuiper et al., “Variable-Focus Liquid Lens for Miniature Cameras,” Applied Physics Letters 85:1128-1130 (2004). |
Kuribayashi, et al., “A Method for Reproducing Apparent Continuous Depth in a Stereoscopic Display Using “Depth-Fused 3D” Technology” Journal of the Society for Information Display 14:493-498 (2006). |
Liu et al., “An Optical See-Through head Mounted Display with Addressable Focal Planes,” IEEE Computer Society, pp. 33-42 (2008). |
Liu et al., “Time-Multiplexed Dual-Focal Plane Head-Mounted Display with a Liquid Lens,” Optics Letters 34:1642-1644 (2009). |
Liu et al., ‘A Novel Prototype for an Optical See-Through Head-Mounted Display with Addressable Focus Cues,’ IEEE Transactions on Visualization and Computer Graphics 16:381-393 (2010). |
Liu et al., “A Systematic Method for Designing Depth-Fused Multi-Focal Plane Three-Dimensional Displays,” Optics Express 18:11562-11573 (2010). |
Written Opinion dated Feb. 10, 2011 from PCT/CN2010/072376. |
McQuaide et al., “A Retinal Scanning Display System That Produces Multiple Focal Planes with a Deformable Membrane Mirror,” Displays 24:65-72 (2003). |
Mon-Williams et al., “Binocular Vision in a Virtual World: Visual Deficits Following the Wearing of a Head-Mounted Display,” Ophthalmic Physiol. Opt. 13:387-391 (1993). |
“OLED-XL Microdisplays,” eMagin 5 pages (2010). |
Optical Research Associates, http://www.opticalres.com, 2 pages (obtained Jan. 26, 2011). |
Rolland et al., “Multifocal Planes Head-Mounted Displays,” Applied Optics 39:3209-3215 (2000). |
Schowengerdt et al., “True 3-D Scanned Voxel Displays Using Single or Multiple Light Sources,” J. Soc. Info. Display 14:135-143 (2006). |
Sheedy et al., “Performance and Comfort on Near-Eye Computer Displays,” Optometry and Vision Science 79:306-312 (2002). |
Shibata et al., “Stereoscopic 3-D Display with Optical Correction for the Reduction of the Discrepancy Between Accommodation and Convergence,” Journal of the Society for Information Display 13:665-671 (2005). |
Shiwa et al., “Proposal for a 3-D Display with Accommodative Compensation: 3DDAC,” Journal of the Society for Information Display 4:255-261 (1996). |
Sullivan, “A Solid-State Multi-Planar Volumetric Display,” SID Symposium Digest of Technical Papers 34:354-356 (2003). |
Examination Report dated Apr. 29, 2011 from corresponding GB Application No. GB1012165.5. |
Varioptic, “Video Auto Focus and Optical Image Stabilization,” http://vvww.varioptic.com/en/home.html, 2 pages (2008). |
Wann et al., Natural Problems for Stereoscopic Depth Perception in Virtual Environments, Vision Res. 35:2731-2736 (1995). |
Watt et al., “Focus Cues Affect Perceived Depth,” J Vision 5:834-862 (2005). |
Love et al. (High Speed switchable lens enables the development of a volumetric stereoscopic display. Aug. 2009, Optics Express. vol. 17, No. 18, pp. 15716-15725.). |
Hidenori Kuriyabashi, Munekazu Date, Shiro Suyama, Toyohiko HatadaJ. of the SID 14/5, 2006 pp. 493-498. |
Xinda Hu and Hong Hua, “High-resolution optical see-through multi-focal-plane head-mounted display using freeform optics,” Optics Express,22(11): 13896-13903, Jun. 2014. |
Xin Shen, Yu-Jen Wang, Hung-Shan Chen, Xiao Xiao, Yi-Hsin Lin, and Bahram Javidi, “Extended depth-of-focus 3D micro integral imaging display using a bifocal liquid crystal lens,” Optics Letters, vol. 40, issue 4, pp. 538-541 (Feb. 9, 2015). |
S. Hong, J. Jang, and B. Javidi,“Three-dimensional volumetric object reconstruction using computational integral imaging,” Journal of Optics Express, on-line Journal of the Optical Society of America, vol. 12, No. 3, pp. 483-491, Feb. 9, 2004. |
R. Schulein, M. DaneshPanah, and B. Javidi, “3D imaging with axially distributed sensing,” Journal of Optics Letters, vol. 34, Issue 13, pp. 2012-2014, Jul. 1, 2009. |
S. Kishk and B. Javidi, “Improved Resolution 3D Object Sensing and Recognition using time multiplexed Computational Integral Imaging,” Optics Express, on-line Journal of the Optical Society of America, vol. 11, No. 26, pp. 3528-3541, Dec. 29, 2003. |
R. Schulein, C. Do, and B. Javidi, “Distortion-tolerant 3D recognition of underwater objects using neural networks,” Journal of Optical Society of America A, vol. 27, No. 3, pp. 461-468, Mar. 2010. |
C. Manh Do, R. Martinez-Cuenca, and B. Javidi, “Three-dimensional object-distortion-tolerant recognition for integral imaging using independent component analysis,” Journal of Optical Society of America A 26, issue 2, pp. 245-251 (Feb. 1, 2009). |
S. Hong and B. Javidi, “Distortion-tolerant 3D recognition of occluded objects using computational integral imaging,” Journal of Optics Express, vol. 14, Issue 25, pp. 12085-12095, Dec. 11, 2006. |
Christopher M. Bishop, Neural Networks for Pattern Recognition, Oxford University Press, Inc. New York, NY 1995. |
‘Fresnel Lenses’ downloaded from http://www.fresneltech.com on Jun. 8, 2011. Copyright Fresnel Technologies, Inc., 2003. |
M. Daneshpanah, B. Javidi, and E. Watson, “Three dimensional integral imaging with randomly distributed sensors,” Journal of Optics Express, vol. 16, Issue 9, pp. 6368-6377, Apr. 21, 2008. |
Yano, S., Emoto, M., Mitsuhashi, T., and Thwaites, H., “A study of visual fatigue and visual comfort for 3D HDTV/HDTV images,” Displays, 23(4), pp. 191-201, 2002. |
D.M. Hoffman, A.R. Girshick, K. Akeley, and M.S. Banks, “Vergence-Accommodation Conflicts Hinder Visual Performance and Cause Visual Fatigue,” J. Vision, 8(3), 1-30, (2008). |
C. B. Burckhardt, “Optimum parameters and resolution limitation of integral photography,” J. Opt. Soc. Am. 58, 71-76 (1968). |
T. Okoshi, “Optimum design and depth resolution of lens-sheet and projection-type three-dimensional displays,” Appl. Opt. 10, 2284-2291 (1971). |
F. Okano, H. Hoshino, J. Arai y I. Yuyama, “Real-time pickup method for a three-dimensional image based on integral photography,” Appl. Opt. 36, 1598-1603 (1997). |
Zhang, R., “8.3: Design of a Compact Light Engine for FLCOS Microdisplays in a p-HMPD system”, Society for Information Display 2008 International Symposium, Seminar and Exhibition (SID2008), Los Angeles, CA, May 2008. |
J. Arai, et al., “Depth-control method for integral imaging,” Feb. 1, 2008 / vol. 33, No. 3 / Optics Letters. |
H. Hua, “Sunglass-like displays become a reality with freeform optical technology,” SPIE Newsroom, 2012. |
H. Hua, X. Hu, and C. Gao, “A high-resolution optical see-through head-mounted display with eyetracking capability,” Optics Express, Nov. 2013. |
D. Cheng, Y.Wang, H. Hua, and M. M. Talha, Design of an optical see-through headmounted display with a low f-number and large field of view using a free-form prism, App. Opt. 48 (14), pp. 2655-2668, 2009. |
A. Jones, I. McDowall, Yamada H., M. Bolas, P. Debevec, Rendering for an Interactive 360° Light Field Display ACM Transactions on Graphics (TOG)—Proceedings of ACM SIGGRAPH 2007, 26(3), 2007. |
Tibor Balogh, “The HoloVizio System,” Proceedings of SPIE, VOI 6055, 2006. |
Y. Takaki, Y. Urano, S. Kashiwada, H. Ando, and K. Nakamura, “Super multi-view winshield display for long-distance image information presentation,” Opt. Express, 19, 704-16, 2011. |
Winterbottom, M., et al., ‘Helmet-Mounted Displays for use in Air Force Training and Simulation’, Human Effectiveness Directorate, Nov. 2005, pp. 1-54. |
Schowengerdt, B. T., and Seibel, E. J., “True 3-D scanned voxel displays using single or multiple light sources,” Journal of SID, 14(2), pp. 135-143, 2006. |
X. Hu and H. Hua, “Design and assessment of a depth-fused multi-focal-plane display prototype,” Journal of Display Technology, Dec. 2013. |
Suyama, S., Ohtsuka, S., Takada, H., Uehira, K., and Sakai, S., “Apparent 3D image perceived from luminance-modulated two 2D images displayed at different depths,” Vision Research, 44: 785-793, 2004. |
J. Hong, S. Min, and B. Lee, “Integral floating display systems for augmented reality,” Applixed Optics, 51(18):4201-9, 2012. |
A. Malmone, and H. Fuchs, “Computational augmented reality eyeglasses,” Proc. of ISMAR 2012. |
H. Mukawa, K. Akutsu, I. Matsumura, S. Nakano, T. Yoshida, M. Kuwahara, and K. Aiki, A full-color eyewear display using planar waveguides with reflection volume holograms, J. Soc. Inf. Display 19 (3), pp. 185-193, 2009. |
Cakmakci, O., et al., ‘Head-Worn Displays: A Review’. Journal of Display Technology, vol. 2, No. 3, Sep. 2006, pp. 199-216. |
M. Marti-nez-Corral, H. Navarro, R. Marti-nez-Cuenca, G. Saavedra, and B. Javidi, “Full parallax 3-D TV with programmable display parameters,” Opt. Phot. News 22, 50-50 (2011). |
J. S. Jang and B. Javidi, “Large depth-of-focus time-multiplexed three-dimensional integral imaging by use of lenslets with non-uniform focal lengths and aperture sizes,” Opt. Lett. vol. 28, pp. 1924-1926 (2003). |
Chih-Wei Chen, Myungjin Cho, Yi-Pai Huang, and Bahram Javidi, “Improved viewing zones for projection type integral imaging 3D display using adaptive liquid crystal prism array,” IEEE Journal of Display Technology, 2014. |
Xiao Xiao, Bahram Javidi, Manuel Martinez-Corral, and Adrian Stern , “Advances in Three-Dimensional Integral Imaging: Sensing, Display, and Applications,” Applied Optics, 52(4):. 546-560,2013. |
J. S. Jang, F. Jin, and B. Javidi, “Three-dimensional integral imaging with large depth of focus by use of real and virtual image fields,” Opt. Lett. 28:1421-23, 2003. |
S. Bagheri and B. Javidi, “Extension of Depth of Field Using Amplitude and Phase Modulation of the Pupil Function,” Journal of Optics Letters, vol. 33, No. 7, pp. 757-759, Apr. 1, 2008. |
M. Lucente, “Interactive three-dimensional holographic displays: seeing the future in depth,” Computer Graphics, 31(2), pp. 63-67, 1997. |
P. A. Blanche, et al, “Holographic three-dimensional telepresence using large-area photorefractive polymer”, Nature, 468, 80-83, Nov. 2010. |
G. Wetzstein et al., “Tensor Displays: Compressive light field synthesis using multilayer displays with directional backlighting,” ACM Transactions on Graphics, 31(4), 2012. |
Blundell, B. G., and Schwarz, A. J., “The classification of volumetric display systems: characteristics and predictability of the image space,” IEEE Transaction on Visualization and Computer Graphics, 8(1), pp. 66-75, 2002. |
Jisoo Hong, et al., “Three-dimensional display technologies of recent interest: Principles, Status, and Issues,” Applied Optics (Dec. 1, 2011) 50(34):106. |
J. Y. Son, W.H. Son, S.K. Kim, K.H. Lee, B. Javidi, “Three-Dimensional Imaging for Creating Real-World-Like Environments,” Proceedings of IEEE Journal, vol. 101, issue 1, pp. 190-205, Jan. 2013. |
R. R.Martinez-Cuenca et al, H. Navarro, G. Saavedra, B. Javidi, and M. Martinez-Corral, “Enhanced viewing-angle integral imaging by multiple-axis telecentric relay system,” Optics Express, vol. 15, Issue 24, pp. 16255-16260, Nov. 21, 2007. |
A. Castro, Y. Frauel, and B. Javidi, “Integral imaging with large depth of field using an asymmetric phase mask,” Journal of Optics Express, vol. 15, Issue 16, pp. 10266-10273 (Aug. 2007). |
International Search Report dated Jan. 29, 2014 in corresponding international application PCT/US2013/065422. |
GB Examination Report corresponding to GB 1012165.5 dated Jun. 28, 2011. |
European Search Report dated Aug. 14, 2015 in corresponding EP application 13740989.2. |
International Search Report and Written Opinion dated Nov. 24, 2015 in corresponding PCT application PCT/US2015/047163. |
International Search Report dated Mar. 9, 2009 with regard to International Patent Application No. PCT/US2009/031606. |
Xin et al., “Design of Secondary Optics for IRED in active night vision systems,” Jan. 10, 2013, vol. 21, No. 1, Optics Express, pp. 1113-1120. |
S. Nikzad, Q. Yu, A. L. Smith, T. J. Jones, T. A. Tombrello, S. T. Elliott, “Direct detection and imaging of low-energy electrons with delta-doped charge-coupled devices,” Applied Physics Letters, vol. 73, p. 3417, 1998. |
Written Opinion of the International Searching Authority dated Mar. 9, 2009 with regard to International Patent Application No. PCT/US2009/031606. |
European Search Report dated Apr. 28, 2016 from EP application 13847218.8. |
Xinda Hu et al: “48.1: Distinguished Student Paper: A Depth-Fused Multi-Focal-Plane Display Prototype Enabling Focus Cues in StereoscopicDisplays”, SID International Symposium. Digest of Technical Papers, vol. 42, No. I, Jun. 1, 2011 (Jun. 1, 2011), pp. 691-694, XP055266326. |
Hu and Hua, “Design and tolerance of a freeform optical system for an optical see-through multi-focal plane display,” Applied Optics, 2015. |
A. Yabe, “Representation of freeform surface suitable for optimization,” Applied Optics, 2012. |
Krueerke, Daniel, “Speed May Give Ferroelectric LCOS Edge in Projection Race,” Display Devices Fall '05. Copyright 2005 Dempa Publications, Inc. pp. 29-31. |
J. P. Rolland and Hong Hua. “Head-mounted display systems,” in Encyclopedia of Optical Engineering. R. Barry Johnson and Ronald O. Driggers, Eds, (2005). |
Zhang, R., et al., “Design of a Polarized Head-Mounted Projection Display using FLCOS Microdisplays”, Proc. of SPIE vol. 6489, 64890B-1. (2007). |
Azuma, R., et al., ‘Recent advances in augmented reality’, IEEE Computer Graphics App;. 21, 34-47 (2001). |
Bajura, M., et al., “Merging virtual objects with the real world: seeing ultrasound imagery within the patient” in Proceedings of ACM SIGGRAPH (ACM, Chicago, 1992), pp. 203-210. |
Caudell, T., et al., “Augmented reality: an application of heads-up display technology to manual manufacturing processes” in Proceedings of Hawaii International Conferences on Systems Sciences (Hawaii, 1992), pp. 659-669. |
Rolland, J.P., et al., ‘Optical versus video see-through head mounted displays in medical visualization’, Presence' Teleoperators and Virtual Environments 9, 287-309 (2000). |
H. Hua., A. Girardot, C. Gao. J. P. Rolland. ‘Engineering of head-mounted projective displays’. Applied Optics. 39 (22), pp. 3814-3824. (2000). |
R. N. Berry, L. A. Riggs, and C. P. Duncan, “The relation of vernier and depth discriminations to field brightness,” J. Exp. Psychol. 40, 349-354 (1950). |
Biocca, et al., “Virtual eyes can rearrange your body: adapting to visual displacement in see-through, head-mounted displays”, Presence: Teleoperators and Virtual Environments 7, 262-277 (1998). |
R. Kijima and T. Ojika, “Transition between virtual environment and workstation environment with projective headmounted display,” in Proceedings of IEEE VR 1997 (IEEE, 1997), pp. 130-137. |
H. Hua, C. Gao, and J. P. Rolland, ‘Study of the Imaging properties of retroreflective materials used in head-mounted projective displays (HMPDs),’ Proc. SPIE4711,194-201 (2002). |
H. Hua and C. Gao, “A polarized head-mounted projective display,” in Proceedings of IEEE and ACM International Symposium on Mixed and Augmented Reality 2005 (IEEE, 2005), pp. 32-35. |
Bunkenburg, J. ‘Innovative Diffractive Eyepiece for Helmet-Mounted Display.’ SPIE vol. 3430. pp. 41-49 Jul. 1998. |
N. Kawakami, M. Inami, D. Sekiguchi, Y. Yangagida, T. Maeda, and S. Tachi, ‘Object-oriented displays: a new type of display systemsfrom immersive display to object-oriented displays,’ in Proceedings of IEEE SMC 1999, IEEE International Conference on Systems, Man, and Cybernetics, vol. 5, pp. 1066-1069. |
J. E. Melzer's: ‘Overcoming the field-of- view/resolution invariant in head-mounted displays’ PROC. SPIE vol. 3362, 1998, p. 284. |
M. Inami, N. Kawakami, D. Sekiguchi, Y. Yanagida, T. Maeda, and S. Tachi, “Visuo-haptic display using head-mounted projector,” in Proceedings of IEEE Virtual Reality 2000, pp. 233-240. |
H. Hua, Y. Ha, and J. P. Rolland, ‘Design of an ultra-light and compact projection lens,’ Appl. Opt. 42, 1-12 (2003), pp. 97-107. |
H. Hua, C. Gao, F. Biocca, and J. P. Rolland, “An ultra-light and compact design and implementation of head-mounted projective displays,” in Proceedings of IEEE VR 2001, pp. 175-182. |
Y. Ha, H. Hua, R. Martins, and J. P. Rolland, “Design of a wearable wide-angle projection color display,” in Proceedings of International Optical Design Conference 2002 (IODC, 2002), pp. 67-73. |
J. P. Rolland, F. Biocca, F. Hamza-Lup, Y. Ha, and R. Martins, “Development of head-mounted projection displays for distributed, collaborative, augmented reality applications,” Presence: Teleoperators and Virtual Environments 14, 528-549 (2005). |
J. L. Pezzaniti and R. A. Chipman, “Angular dependence of polarizing beam-splitter cubes,” Appl. Opt. 33, 1916-1929 (1994). |
R. Azuma, A Survey of Augmented Reality in Presence; Teleoperators and Virtual Environments 6. 4, 355-385, (1997). |
R. Martins, V. Shaoulov, Y. Ha, and J. P. Rolland, “Projection based head-mounted displays for wearable computers,” Proc. SPIE 5442, 104-110 (2004). |
C. Curatu, H. Hua, and J. P. Rolland, “Projection-based headmounted display with eye-tracking capabilities,” Proc. SPIE 5875, 587050J (2005). |
R. Kijima, K. Haza, Y. Tada, and T. Ojika, “Distributed display approach using PHMD with infrared camera,” in Proceedings of IEEE 2002 Virtual Reality Annual International Symposium (IEEE, 2002), pp. 1-8. |
L. G. Brown's: ‘Applications of the Sensics panoramic HMD’ SID Symposium Digest vol. 39, 2008, p. 77. |
M. Inami, N. Kawakami, and S. Tachi, ‘Optical camouflage using retro-reflective projection technology,’ in Proceedings of ISMAR 2003 {ISMAR, 2003). |
L. Davis, J. P. Rolland, F. Hamza-Lup, Y. Ha, J. Norfleet, and C. Imielinska, ‘Enabling a continuum of virtual environment experiences,’ IEEE Comput. Graphics Appl. 23, pp. 10-12 Mar./Apr. 2003. |
H. Hua, L. Brown, and C. Gao, “Anew collaborative infrastructure: SCAPE,” in Proceedings of IEEE VR 2003 (IEEE, 2003), pp. 171-179. |
H. Hua, L. Brown, and C. Gao, “SCAPE: supporting stereoscopic collaboration in augmented and projective environments,” IEEE Comput. Graphics Appl. 24, 66-75 (2004). |
H. Hua, L. Brown, and C. Gao, “System and interface framework for SCAPE as a collaborative infrastructure,” Presence: Teleoperators and Virtual Environments 13, 234-250 (2004). |
L. Brown and H. Hua, “Magic lenses for augmented virtual environments,” IEEE Comput. Graphics Appl. 26, 64-73 (2006). |
M. Gutin: ‘Automated design and fabrication of ocular optics’ PROC. SPIE 2008, p. 7060. |
M. Robinson. J. Chen, and G. Sharp, Polarization Engineering for LCD Projection. John Wiley & Sons, Ltd. England, 2005. |
Zhang, R., et al., “Design of a Polarized Head-Mounted Projection Display Using Ferroelectric Liquid-Crystal-on-Silicon Microdisplays”, Applied Optics, vol. 47, No. 15, May 20, 2008, pp. 2888-2896. |
D. Cheng, Y. Wang, H. Hua, and M. M. Talha, “Design of an optical see-through head-mounted display with a low f-number and large field of view using a freeform prism,” Appl. Opt., 48(14):2655-2668, 2009. |
D. Cheng, Y. Wang, H. Hua, J. Sasian, “Design of a wide-angle, lightweight head-mounted display using free-form optics tiling,” Opt. Lett., 36(11):2098-100, 2011. |
P. Gabbur, H. Hua, and K. Barnard, ‘A fast connected components labeling algorithm for real-time pupil detection,’ Mach. Vision Appl., 21(5):779-787, 2010. |
H. Hua, P. Krishnaswamy, and J.P. Rolland, ‘Video-based eyetracking methods and algorithms in head-mounted displays,’ Opt. Express, 14(10):4328-50, May 2006. |
H. Hua, C. Pansing, and J.P. Rolland, “Modeling of an eye-imaging system for optimizing illumination schemes in an eye-tracked head-mounted display,” Appl. Opt., 46(31):7757-75, Oct. 2007. |
Yamazaki et al, “Thin wide-field-of-view HMD with free-form-surface prism and applications”, Proc. SPIE 3639, Stereoscopic Displays and Virtual Reality Systems VI, 453 (May 24, 1999). |
T. Ando, K. Yamasaki, M. Okamoto, and E. Shimizu, “Head Mounted Display using holographic optical element,” Proc. SPIE, vol. 3293, 183 (1998). |
O. Cakmakci, B. Moore, H. Foroosh, and J. P. Rolland, “Optimal local shape description for rotationally non-symmetric optical surface design and analysis,” Opt. Express 16, 1583-1589 (2008). |
Curatu, C., J.P. Rolland, and Hong Hua, “Dual purpose lens for an eye-tracked projection head-mounted display,” Proceedings of International Optical Design Conference, Vancouver, Canada, Jun. 2006. |
J. G. Droessler, D. J. Rotier, “Tilted cat helmet-mounted display,” Opt. Eng., vol. 29, 849 (1990). |
A. T. Duchowski, “Incorporating the viewer's Point-Of-Regard (POR) in gaze-contingent virtual environments”, SPIE-Int. Soc. Opt. Eng. Proceedings of Spie—the International Society for Optical Engineering, vol. 3295, 1998, pp. 332-343. |
Duchowski, A., “Eyetracking Methodology: theory and practice,” Publisher: Springer, 2003. |
H. Hoshi, N. Taniguchi, H. Morishima, T. Akiyama, S. Yamazaki and A. Okuyama, “Off-axial HMD optical system consisting of aspherical surfaces without rotational symmetry,” SPIE vol. 2653, 234 (1996). |
Hua, H., C. Pansing, and J. P. Rolland, “Modeling of an eye-imaging system for optimizing illumination schemes in an eye-tracked head-mounted display,” Applied Optics, 46(32): 1-14, Nov. 2007. |
K. Iwamoto, S. Katsumata, K. Tanie, “An eye movement tracking type head mounted display for virtual reality system: -evaluation experiments of a prototype system”, Proceedings of 1994 IEEE International Conference on Systems, Man, and Cybernetics. Humans, Information and Technology (Cat. No. 94CH3571-5). IEEE. Part vol. 1, 1994, pp. 13-18 vol. 1. New York, NY, USA. |
K. Iwamoto, K. Tanie, T. T. Maeda, “A head-mounted eye movement tracking display and its image display method”, Systems & Computers in Japan, vol. 28, No. 7, Jun. 30, 1997, pp. 89-99. Publisher: Scripta Technica, USA. |
Hua et al., “Compact eyetracked optical see-through head-mounted display”, Proc. SPIE 8288, Stereoscopic Displays and Applications XXIII, 82881F (Feb. 9, 2012). |
M. D. Missig and G. M. Morris, “Diffractive optics applied to eyepiece design,” Appl. Opt. 34, 2452-2461 (1995). |
J. P. Rolland, “Wide-angle, off-axis, see-through head-mounted display,” Opt. Eng., vol. 39, 1760 (2000). |
M. L. Thomas, W. P. Siegmund, S. E. Antos, and R. M. Robinson, “Fiber optic development for use on the fiberoptic helmet-mounted display”, Helmet-Mounted Displays, J. T. Carollo, ed., Proc. SPIE 116, 90-101, 1989. |
Laurence R. Young, David Sheena, “Survey of eye movement recording methods”, Behavior Research Methods & Instrumentation, 7(5), 397-429, 1975. |
Rolland, J. P., A. Yoshida, L. D. Davis and J. H. Reif, “High-resolution inset head-mounted display,” Appl. Opt. 37, 4183-93 (1998). |
Hua, H. “Integration of eye tracking capability into optical see-through head-mounted displays,” Proceedings of SPIE (Electronic Imaging 2001), pp. 496-503, Jan. 2001. |
Hoshi, et al, “Off-axial HMD optical system consisting of aspherical surfaces without rotational symmetry,” Proc. SPIE 2653, Stereoscopic Displays and Virtual Reality Systems III, 234 (Apr. 10, 1996). |
S. Feiner, 2002, “Augmented reality: A new way of seeing,” Scientific American, No. 54, 2002. |
K. Ukai and P.A. Howardth, “Visual fatigue caused by viewing stereoscopic motion images: background, theories, and observations,” Displays, 29(2), pp. 106-116, 2008. |
X. Hu and H. Hua, “Design and tolerance of a free-form optical system for an optical see-hrough multi-focal-plane display,” Applied Optics, 54(33): 9990-9, 2015. |
H. Hua and B. Javidi, “A 3D integral imaging optical see-through head-mounted display”, Optics Express, 22(11): 13484-13491, 2014. |
W. Song, Y. Wang. D. Cheng, Y. Liu, “Light field head-mounted display with correct focus cue using micro structure array,” Chinese Optics Letters, 12(6): 060010, 2014. |
B. T. Schowengerdt, M. Murari, E. J. Seibel, “Volumetric display using scanned fiber array,” SID Symposium Digest of Technical Papers, 2010. |
Hu, X., Development of the Depth-Fused Multi-Focal Plane Display Technology, Ph.D. Dissertation, College of Optical Sciences, University of Arizona, 2014. |
S. Ravikumar, K. Akeley, and M. S. Banks, “Creating effective focus cues in multi-plane 3D displays,” Opt. Express 19, 20940-20952, 2011. |
T. Peterka, R. Kooima, D. Sandin, A. Johnson, J. Leigh, T. DeFanti, “Advances in the Dynallax solid-state dynamic parallax barrier autostereoscopi visualization display system,” IEEE Trans. Visua. Comp. Graphics, 14(3): 487-499, 2008. |
Dewen Cheng et al.; “Large field-of-view and high resolution free-form head-mounted display”; SPIE-OSA/ vol. 7652 Jun. 2018. |
G. Lippmann, “Epreuves reversibles donnant la sensation du relief,” Journal of Physics (Paris) 7, 821-825 (1908). |
Huang et al., “An integral-imaging-based head-mounted light field display using a tunable lens ,;1nd aperture array.” Journal of the Society for Information Display Mar. 1, 2017; p. 199-201. |
Full Certified Translation of Reference JP 02200074. |
Full Certified Translation of Reference JP008160345. |
Cheol-Joong Kim et al, “Depth plane adaptive integral imaging using a varifocal liquid lens array”, Applied Optics, OSA, vol. 54, No. 10, Apr. 1, 2015 (Apr. 1, 2015) , pp. 2565-2571. |
Xin Shen et al: “Large depth of focus dynamic micro integral imaging for optical see-through augmented reality display using a focus-tunable lens”, Applied Optics, vol. 57, No. 7, Mar. 1, 2018 (Mar. 1, 2018), p. B184. |
Martinez-Cuenca R et al: “Progress in 3-D Multiperspective Display by Integral Imaging”, Proceedings of the IEEE, IEEE. New York, us, vol. 97, No. 6, Jun. 1, 2009 (Jun. 1, 2009), pp. 1067-1077. |
Kim Cheoljoong et al: “Depth-enhanced integral imaging display system with time-multiplexed depth planes using a varifocal liquid lens array”, Proceedings of SPIE, IEEE, US, vol. 9385, Mar. 11, 2015 (Mar. 11, 2015), pp. 93850D-93850D. |
Huan Deng et al: “The Realization of Computer Generated Integral Imaging Based on Two Step Pickup Method”, Photonics and Optoelectronic (SOPO), 2010 Symposium ON, IEEE, Piscataway, NJ, USA, Jun. 19, 2010 (Jun. 19, 2010), pp. 1-3. |
H. Hua, “Enabling focus cues in head-mounted displays,” Proceedings of the IEEE 105(5), 805-824 (2017). |
G. E. Favalora, “Volumetric 3D displays and application infrastructure,” Computer, 38(8), 37-44 (2005). |
H. Yu, K. Lee, J. Park, and Y. Park, “Ultrahigh-definition dynamic 3D holographic display by active control of volume speckle fields,” Nature Photonics 11(3), 186 (2017). |
G. Li, D. Lee, Y. Jeong, J. Cho, and B. Lee, “Holographic display for see-through augmented reality using mirror-lens holographic optical element,” Opt. Letters 41(11), 2486-2489 (2016). |
S. B. Kim and J. H. Park, “Optical see-through Maxwellian near-to-eye display with an enlarged eyebox,” Opt. Letters 43(4), 767-770 (2018). |
D. Lanman and D. Luebke, “Near-eye light field displays,” ACM Trans. Graph. 32(6), 1-10 (2013). |
H. Huang and H. Hua, “High-performance integral-imaging-based light field augmented reality display using freeform optics,” Opt. Express 26(13), 17578-17590 (2018). |
B. Liu, X. Sang, X. Yu, X. Gao, L. Liu, C. Gao, P. Wang, Y. Le, and J. Du, “Time-multiplexed light field display with 120-degree wide viewing angle”. Opt. Express 27(24), pp. 35728-35739 (2019). |
H. Huang and H. Hua, “Generalized methods and strategies for modeling and optimizing the optics of 3D head-mounted light field displays,” Opt. Express 27(18), 25154-25171 (2019). |
H. Huang and H. Hua, “Systematic characterization and optimization of 3D light field displays,” Opt. Express 25(16), 18508-18525 (2017). |
J. H. Park, S. W. Min, S. Jung, and B. Lee. “Analysis of viewing parameters for two display methods based on integral photography.” Applied Optics 40, No. 29 5217-5232 (2001). |
X. Wang, Y. Qin, H. Hua, Y. H. Lee, and S. T. Wu. “Digitally switchable multi-focal lens using freeform optics.” Opt. Express 16;26(8):11007-17(2018). |
X. Wang, and H. Hua. “Digitally Switchable Micro Lens Array for Integral Imaging.” SID Symposium Digest of Technical Papers. vol. 51. No. 1. (2020). |
M. Xu and H. Hua, “Finite-depth and varifocal head-mounted display based on geometrical lightguide,” Opt. Express 28(8), 12121-12137 (2020). |
Jason Geng: “Three-dimensional display technologies”, Advances in Optics and Photonics, vol. 5, No. 4, Nov. 22, 2013 (Nov. 22, 2013), pp. 456-535. |
Number | Date | Country | |
---|---|---|---|
20220003975 A1 | Jan 2022 | US |
Number | Date | Country | |
---|---|---|---|
61011789 | Jan 2008 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12863771 | US | |
Child | 13955076 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16685204 | Nov 2019 | US |
Child | 17478033 | US | |
Parent | 16170860 | Oct 2018 | US |
Child | 16685204 | US | |
Parent | 15092831 | Apr 2016 | US |
Child | 16170860 | US | |
Parent | 13955076 | Jul 2013 | US |
Child | 15092831 | US |