The present invention is related to a near-eye displaying method capable of multiple depths of field imaging; more particularly, a near-eye displaying method capable of making light emitted by any two pixels to intersect and create foci at different locations so that the output image has multiple depths of field.
As the demand for real-time information rises, the importance of on-demand data transmission is also increased. Near-eye displays (NED) are often portable, and can be easily incorporated into other devices and can transmit images, colors, texts and/or sound data at any time; therefore, it is a primary choice for portable information device or on-demand data transmission purposes. Near-eye displays are often implemented for military or governmental uses in the past. Currently, the near-eye display industry seeks expansion in the consumer sector. Meanwhile, the entertainment industry also sees the market potential in near-eye displays; for example, home entertainment system and gaming software developers have been putting effort into research and development of near-eye displays.
Currently, a typical near-eye display includes head-mounted display (HMD), which can project image directly into users' eyes. This type of display can emulate bigger displays to overcome the shortcomings of the displays in mobile devices. The head-mounted display can also be applied to virtual reality or augmented reality uses.
Near-eye displays can be further categorized into two types: immersive display and see-through display. In virtual reality (VR) environment, an immersive display can be implemented to enable composite images to completely cover the visual field of a user. In augmented reality (AR) environment, a see-through display is implemented; and therefore, texts, side notes or images can be overlapped with real images. In the field of augmented reality display technology, a transparent panel (implemented via optical or electro-optical means) is often used in a see-through display. This enables the user of the near-eye display to see both virtual images and real images in the same time.
However, since human eyes cannot focus on objects placed at a very close distance (for example, when a user is wearing glasses and using a magnify lens as a reading aid, a distance within the range of the magnify lens and the glasses is considered “close distance”); therefore, the near-eye display needs to be calibrated and adjusted to avoid image being out of focus so as to provide a comfortable using experience for the users. The traditional near-eye displays rely on complex and heavy optical assembly to adjust the focus of the image; however, since near-eye display is usually worn on the user's head, heavier near-eye displays oftentimes cannot be accepted by the users.
To overcome the above mentioned shortcomings, if one can enable at least two light beams emitted by at least two separate pixels to intersect and focus to produce a clear image, heavy optical assembly would no longer be necessary; furthermore, the manufacturing cost arisen from the optical assembly would be eliminated.
The present invention provides a near-eye displaying method capable of multiple depths of field imaging, which reduces manufacturing cost and increases displaying efficiency. The present invention allows lights emitted by any two pixels to intersect and create foci at different locations so that the output image has multiple depths of field.
To achieve the aforementioned purpose, the present invention discloses a near-eye displaying method capable of multiple depths of field imaging, characterized in that the method comprises the steps of:
In some embodiments of the present invention, the self-emissive display comprises an active light source including an organic light-emitting diode, a micro light emitting diode, a quantum dot light emitter or a laser.
In some embodiments of the present invention, the self-emissive display is a transparent display or a non-transparent display.
In some embodiments of the present invention, the collimator is a microlens, a flat metalens or a liquid crystal spatial light modulator.
In some embodiments of the present invention, the flat metalens has the function of a diopter lens for collimating a direction of the light.
In some embodiments of the present invention, the liquid crystal spatial light modulator comprises a plurality of liquid crystal cells, an alignment of a liquid crystal within the liquid crystal cells can be changed by altering a driving voltage applied to the liquid crystal cells so that a direction of an incident light from every pixel is collimated.
In some embodiments of the present invention, the collimated light direction altering unit is a microlens, a flat metalens, or a liquid crystal spatial light modulator.
In some embodiments of the present invention, the microlens enables at least two collimated lights to intersect and focus.
In some embodiments of the present invention, the flat metalens comprises a plurality of areas having bumps for enabling at least two collimated lights to intersect and focus.
In some embodiments of the present invention, two different areas having bumps are utilized to enable at least two collimated lights to intersect and focus at different locations to create an image having multiple depths of field.
In some embodiments of the present invention, an area having bumps is utilized to enable at least two collimated lights to intersect and focus at different locations to create an image having multiple depths of field.
In some embodiments of the present invention, the liquid crystal spatial light modulator comprises a plurality of liquid crystal cells, an alignment of a liquid crystal in the liquid crystal cells can be changed by altering a driving voltage applied to the liquid crystal cells so as to manipulate a direction of the collimated light and enable at least two collimated lights to intersect and focus.
In some embodiments of the present invention, a driving voltage of at least two liquid crystal cells can be changed to enable at least two collimated lights to intersect and focus at different locations to create an image having multiple depths of field.
In some embodiments of the present invention, a driving voltage of at least one liquid crystal cell can be changed to enable at least two collimated lights to intersect and focus at different locations to create an image having multiple depths of field.
In some embodiments of the present invention, the pixel is a single pixel or a collection of pixels comprising a plurality of pixels.
The near-eye displaying method capable of multiple depths of field imaging according to the present invention has the following advantages over the prior art:
Embodiments of the present invention will now be described, by way of example only, with reference to the accompanying drawings.
According to the aforementioned method, a self-emissive display 1 utilized technology that enables self-emission; and the self-emissive display 1 may be a transparent display or a non-transparent display. The self-emissive display 1 may comprise active light sources such as organic light-emitting diodes (OLED), micro light emitting diodes (micro LED), quantum dot light emitters, or lasers . . . etc.
The collimator may be a microlens, a liquid crystal spatial light modulator (LCSLM) or a flat metalens. The following further explains the different types of collimators:
The collimated light direction altering unit is a microlens, a liquid crystal spatial light modulator (LCSLM) or a flat metalens. The following further explains the different types of collimated light direction altering units:
(1) Microlens:
(2) Liquid Crystal Spatial Light Modulator (LCSLM):
(3) Flat Metalens:
When generating image having multiple depths of field, different combinations of collimators and collimated light direction altering units can be used. The combinations are described as the following:
As shown in
In another aspect of the present invention, a single element can be utilized to collimate and change the direction of the light. The description is as the following:
The near-eye displaying method capable of multiple depths of field imaging according to the present invention has the following advantages over the prior art:
Although particular embodiments of the present invention have been described in detail for purposes of illustration, various modifications and enhancements may be made without departing from the spirit and scope of the present invention. Accordingly, the present invention is not to be limited except as by the appended claims.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CN2018/077715 | 3/1/2018 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/165620 | 9/6/2019 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6961993 | Oohata | Nov 2005 | B2 |
8866184 | Yamagata et al. | Oct 2014 | B2 |
9860522 | Lapstun | Jan 2018 | B2 |
9906781 | Fujimaki | Feb 2018 | B2 |
9964767 | Son et al. | May 2018 | B2 |
10429650 | Tsai et al. | Oct 2019 | B2 |
10663735 | Tsai et al. | May 2020 | B2 |
20020064032 | Oohata | May 2002 | A1 |
20050067944 | Masuda et al. | Mar 2005 | A1 |
20130285885 | Nowatzyk et al. | Oct 2013 | A1 |
20140056003 | Frattalone et al. | Feb 2014 | A1 |
20140183515 | Yamagata et al. | Jul 2014 | A1 |
20160219272 | Fujimaki | Jul 2016 | A1 |
20170082263 | Byrnes et al. | Mar 2017 | A1 |
20170111635 | Miyazaki et al. | Apr 2017 | A1 |
20170219828 | Tsai et al. | Aug 2017 | A1 |
20170255011 | Son et al. | Sep 2017 | A1 |
20170269358 | Luebke et al. | Sep 2017 | A9 |
20180031842 | Cai et al. | Feb 2018 | A1 |
20180067456 | Zhang et al. | Mar 2018 | A1 |
20180247922 | Robin et al. | Aug 2018 | A1 |
20190018244 | Tan | Jan 2019 | A1 |
20190361243 | Tsai et al. | Nov 2019 | A1 |
Number | Date | Country |
---|---|---|
102629667 | Aug 2012 | CN |
103792826 | May 2014 | CN |
103885582 | Jun 2014 | CN |
104246578 | Dec 2014 | CN |
102629667 | Mar 2015 | CN |
105008983 | Oct 2015 | CN |
105449125 | Mar 2016 | CN |
105739094 | Jul 2016 | CN |
105911791 | Aug 2016 | CN |
106057843 | Oct 2016 | CN |
106292240 | Jan 2017 | CN |
106526864 | Mar 2017 | CN |
106873161 | Jun 2017 | CN |
107015361 | Aug 2017 | CN |
107490862 | Dec 2017 | CN |
107561697 | Jan 2018 | CN |
107561702 | Jan 2018 | CN |
107664840 | Feb 2018 | CN |
108431666 | Aug 2018 | CN |
105449125 | Nov 2018 | CN |
106932916 | Jul 2021 | CN |
3423885 | Jan 2019 | EP |
H05142511 | Jun 1993 | JP |
H11234705 | Aug 1999 | JP |
2002090704 | Mar 2002 | JP |
2002118124 | Apr 2002 | JP |
2007147718 | Jun 2007 | JP |
2011145607 | Jul 2011 | JP |
2015521298 | Jul 2015 | JP |
2016020929 | Feb 2016 | JP |
2007027157 | Feb 2017 | JP |
2017529557 | Oct 2017 | JP |
2017538959 | Dec 2017 | JP |
509969 | Nov 2002 | TW |
201416717 | May 2014 | TW |
201510573 | Mar 2015 | TW |
I607243 | Dec 2017 | TW |
201805688 | Feb 2018 | TW |
2015170080 | Nov 2015 | WO |
2016163231 | Oct 2016 | WO |
2017089676 | Jun 2017 | WO |
2017151199 | Sep 2017 | WO |
2018001321 | Jan 2018 | WO |
2018013307 | Jan 2018 | WO |
2018126677 | Jul 2018 | WO |
2019166018 | Sep 2019 | WO |
Entry |
---|
Taiwan Office Action issued in TW107106033 dated Mar. 21, 2019. |
Taiwan Office Action issued in TW107106033 dated Sep. 17, 2019. |
International Preliminary Report on Patentability dated Jun. 10, 2020 in International Patent Application No. PCT/CN2018/077715, filed on Mar. 1, 2018 with English translation of International Preliminary Report. |
International Search Report and Written Opinion dated Dec. 6, 2018 in International Patent Application No. PCT/CN2018/077715, filed on Mar. 1, 2018 with English translation of International Search Report, 2 pages. |
Office Action of Japan counterpart application dated Jan. 11, 2022. |
European Search Report issued in EP 19760179.2 dated Mar. 23, 2021. |
European Search Report issued in EP 18907507.0 dated Sep. 24, 2021. |
Supplementary European Search Report, dated Jun. 13, 2022 in corresponding application EP 19760179.2. |
Supplementary European Search Report, dated Nov. 7, 2022 in corresponding application EP 18907507.0. |
Taiwanese Office Action, dated Sep. 7, 2022, in a counterpart Taiwanese patent application, No. TW 108106943. |
Chinese Office Action, dated Feb. 16, 2023 in a counterpart Chinese patent application, No. CN 201880090627.7. |
International Search Report and Written Opinion dated May 29 2019 in International Patent Application No. PCT/CN2019/076752, filed on Mar. 1, 2019 with English translation of International Search Report, 2 pages. |
Office action issued in a corresponding U.S. Appl. No. 16/976,526 dated Aug. 28, 2023. |
Number | Date | Country | |
---|---|---|---|
20210003900 A1 | Jan 2021 | US |