This disclosure relates generally to ophthalmic devices and techniques, and in particular but not exclusively, relates to ophthalmic devices and techniques for controlling myopia.
Myopia, or nearsightedness, is a refractive effect that causes people to be able to focus on objects near to them, while objects far away are blurry. Typical treatment for myopia is to wear a negatively powered spectacle or contact lens. Myopia is caused by an eyeball length that is longer than normal. This extended shape causes multiple problems in addition to defocus. The process which determines the length of a human's eye has both genetic and environmental factors. Experiments in multiple species, including humans, have shown that a normal eye attempts a process known as emmetropization to control its growth in a closed-loop fashion such that it grows to the appropriate length where visual stimulus from the environment is focused on the retina.
Myopia currently affects approximately 30% of the human population and this number is anticipated to significantly expand through 2050. Myopia is particularly acute in East Asian countries where genetic factors and cultural norms may work together to drive particularly high myopia rates in children. In fact, overwhelming majorities of college-aged, urban populations in East Asia suffer from near-sightedness. The increasing prevalence of myopia is believed to be associated with increased near work while the eye is growing in adolescence. A major contributor to near work is not just books, but screen time associated with personal computing devices.
The optical geometry of an emmetropic eye 100 is remarkably maintained to within microns of optimal alignment despite the eye representing a complex multi-component optical system that increases by roughly 50% in length from birth to adulthood. This alignment is achieved using feedback growth signals that encourage or discourage growth based upon small amounts of hyperopic or myopic defocus experienced during growth years. As mentioned above, these feedback growth signals/cues are believed to be responsible for controlling the emmetropization process. However, exposure to prolonged, daily periods of near-field vision tasks (e.g., regular screen time on a portable computing device) and other aspects of modern life may alter this feedback loop, thereby preventing the appropriate feedback growth signals. If myopia is allowed to progress too far, it is correlated with more serious conditions later in life such as retinal detachment, glaucoma, macular degeneration, cataracts, as well as other deleterious conditions.
Conventional approaches to treating or controlling the onset of myopia fail to show consistently high degrees of effectivity and are often accompanied by undesirable side effects. Such undesirable side effects include blurred vision in a portion of the visual field (e.g., as caused by bifocal/multifocal lenses), an inability to achieve near-field focus (e.g., as caused by atropine eye drops), or temporary vision impairment along with stable, though not permanent, correction associated with mechanical tissue reshaping (e.g., orthokeratology).
An effective approach to treating or controlling the onset of myopia (or hyperopia) in a safe, effective, and cost-efficient manner is desirable not only to treat a current ophthalmic condition, but also could save large numbers of the population from suffering significant visual impairment later in life.
Non-limiting and non-exhaustive embodiments of the invention are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified. Not all instances of an element are necessarily labeled so as not to clutter the drawings where appropriate. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the principles being described.
Embodiments of a system, apparatus, and method for using simulated longitudinal chromatic aberration (LCA) to drive or encourage emmetropization are described herein. In the following description numerous specific details are set forth to provide a thorough understanding of the embodiments. One skilled in the relevant art will recognize, however, that the techniques described herein can be practiced without one or more of the specific details, or with other methods, components, materials, etc. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring certain aspects.
Reference throughout this specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, the appearances of the phrases “in one embodiment” or “in an embodiment” in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
Chromatic aberration is the failure of a lens (such as the human crystalline lens) to focus all wavelengths of light to a single point. There are two types of chromatic aberration: longitudinal (also referred to as axial) and transverse (also referred to as lateral). Longitudinal chromatic aberration (LCA) occurs when different wavelengths or colors of multi-color light are brought to focus at different offset distances from the lens. The crystalline lens and cornea of the human eye induce LCA. However, humans use LCA to beneficial ends. In fact, LCA is a significant driver of accommodation, helping our eyes bring objects into focus. It is also known that the chromatic information provided by LCA is involved in the process of emmetropization. For example, experiments show that animals deprived of chromatic bandwidth, grown in either short or long wavelengths exclusively, do not achieve emmetropization.
Accordingly, embodiments described herein use a simulated LCA to trigger desirable feedback growth cues to encourage or drive emmetropization of the eye. The simulated LCA is achieved by a form of chromatic blurring (i.e., wavelength dependent blurring) that selectively blurs one or more color channels or wavelengths differently than others. LCA as defined herein includes wavelength dependent blurring that may be applied across the visual spectrum, applied to a single color channel (e.g., just blue), or applied discontinuously across multiple wavelength bands (e.g., blurring blue and red color channels but not a green color channel). In some embodiments that treat myopia, shorter wavelength colors (e.g., blues) are more blurred relative to longer wavelength colors (e.g., reds). In some instances, there may be benefits to chromatic blurring red and blue, while leaving green substantially unblurred (or at least less blurred). The amount or degree of blurring between the red and blue channels may be equivalent, blue may be blurred to a greater extent than red, or in select scenarios red may be blurred to a greater extent than blue. In various therapeutic embodiments, the amount of chromatic blurring induced may range between 0 and 5 diopters of equivalent optical defocus, though approximately 3 diopters is anticipated to be suitable in certain therapeutic embodiments. The chromatic blurring may be achieved using spatial blurring (e.g., spreading chromatic pixel data), applying spatial frequency filtering, or potentially using phase shifting techniques. In various embodiments, the chromatic blurring is implemented computationally to produce the simulated LCA. The blurring of shorter wavelength color channels relative to the longer wavelength color channels in a color image simulates the effect of an optical myopic defocus, which is known to help slow or prevent the onset of myopia, and potentially reverse myopia. The simulated LCA described herein achieves the desirable effects of myopic defocus without having to defocus the image light use lensing power (e.g., refractive, diffractive, or specular lensing) as is required by conventional myopic treatment. In various embodiments, the simulated LCA may be generated entirely in software facilitating the use of virtually any electronic display (e.g., televisions, portable computing devices, etc.) to provide emmetropization therapy. In other words, the techniques described herein provide a safe and inexpensive way of transforming the very devices (e.g., personal computing devices such as smart phones, tables, and laptops) that are major contributing factors to the increasing prevalence of myopia into devices capable of beneficially providing emmetropization therapy, or at least, offsetting the negative effects of their use.
An eye that has unduly elongated along the axis running from the cornea to retina 210 through lens 205 becomes myopic (e.g., see elongation of axial length 305 along axis 310 in
Alternatively, just the eye's peripheral vision may be myopically defocused while leaving the eye's central vision unmaligned. It is believed that the beneficial feedback growth cues are still adequately stimulated with just peripheral visual stimulation. Since the majority of human acuity resided in the central vision, peripheral visual stimulation that only defocuses the peripheral vision may be more comfortable and less intrusive for the end user and thus tolerated for longer durations.
Display 405 may be implemented with a variety of different color display technologies. Display 405 may be a liquid crystal display (LCD), an organic light emitting diode (OLED) display, or otherwise. In particular, personal computing devices such as smart phones, tablet computers, laptops, desktop computers etc. are well suited to implement the techniques described herein. It is noteworthy that display 405 does not require expensive lenses for optically defocusing color image 420 emitted from display 405. Rather, the myopic defocus is simulated in software and/or dedicated hardware logic.
In the illustrated embodiment, system 400 includes a camera 410 for tracking a gazing direction 425 of eye 401. Camera 410 may be an external camera that mounts to display 405 or an integrated camera. Although
Controller 415 is coupled to display 405 and camera 410 to choreograph their operation. Controller 415 may be implemented as a general-purpose processor that executes software instructions stored in a memory, as hardware logic (e.g., application specific integrated circuit, field programmable gate array, etc.), or a combination of both. Controller 415 may be a separate module that couples to display 405 and camera 410, or integrated circuitry/logic that is disposed within a single computing device. For example,
During operation, controller 415 receives image data 430 corresponding to a color image, selectively blurs at least one color channel of image data 430 in at least a portion of the color image to provide simulated LCA in that portion of the color image, and then displays color image 420 with the simulated LCA to provide an emmetropization therapy. Image data 430 may represent a variety of data types, such as color pictures or a video data stream for a video.
The illustrated embodiment of control architecture 500 includes an image portion selection module 502, an image decomposer module 505, a blurring module 510, and an image reconstitution module 515. The function of each of these modules is described below. However, it should be appreciated that image decomposer module 505, blurring module 510, and an image reconstitution module 515 form a rendering pipeline 501 that may be implemented using a variation of the ChromaBlur software described in “ChromaBlur: Rendering Chromatic Eye Aberration Improves Accommodation and Realism” by Cholewiak et al. Rendering pipeline 501 may be implemented at the software level or at a lower level of hardware integration (field programmable gate array, application specific integrated circuit, etc.). For example, rendering pipeline 501 may be integrated into the display hardware itself and applied across all media types displayed on the screen. In this case, the blurring scheme described may manipulate the image at a hardware level encoding such as PAL, NTSC, SECAM, Display Serial Interface, Digital Visual Interface, or otherwise.
Image portion selection module 502 operates to select the portion of color image 420 to which chromatic blurring is to be applied. This portion may represent the entire color image 420. Alternatively, this portion may be a peripheral portion 625 that surrounds a central or fixation region 620 (see
Image decomposer module 505 decomposes image data 430 into multiple color channels. In one embodiment, image decomposer module 505 decomposes image data 430 into three color channels (e.g., red channel, green channel, and a blue channel). Of course, other color models may be implemented. Blurring module 510 selectively blurs the image data in one or more color channels in the select image portion. The blurring is a chromatic blurring that only blurs the selected color component(s) of image pixels falling within the region of color image 420 designated by image portion selection module 502. The chromatic blurring may be applied to a single color channel (e.g., just blue channel) or multiple color channels (e.g., blue and green channels). In one embodiment, the blurring is applied as a gradient blur where the shorter wavelength color channels receive greater blurring than the longer wavelength color channels. For example, a blue channel may receive the greatest chromatic blur, the green channel may receive less or no chromatic blur while the red channel does not receive any chromatic blur. In another embodiment, the blurring is applied only to blue and red, leaving green in sharp focus. The chromatic blur may be implemented by spatially spreading image pixels of color image 420 in only the select color channels. For example, a white image pixel (or any color image pixel) may be decomposed into component colors and manipulated to have a sharp red center with a blurred fringe (i.e., spatial spreading) of blue and green components. The same technique holds true for black and white images (e.g., a page of text). Fonts may even be calculated with a predetermined amount of LCA, which may be applied at least across a portion of a field of text within image 420.
After the selected image pixels have been chromatically blurred, the image data of the color channels are recombined to reconstitute color image 420 with the simulated LCA in the selected image portions. Image reconstitution module 515 implements this recombining process to generate the therapeutic image 520 (i.e., color image 420 with simulated LCA).
In a process block 705, controller 415 receives image data 430 corresponding to color image 420. Image data 430 may represent a variety of different types of image data such as digital images (e.g., jpeg, png, tiff, gif, etc) or a video data stream (e.g., mpeg, MP4, H.264, H.265, AAC, AVI, etc.). In a process block 710, rendering pipeline 501 decomposes image data 430 into multiple color channels corresponding to different color wavelengths (e.g., red, green, and blue channels).
If an entire portion of color image 420 is to be chromatically blurred (decision block 715), then process 700 continues to a process block 720 where blurring module 510 chromatically blurs (e.g., spatial color blurring) image data corresponding to one or more of the color channels across the entire color image 420. Correspondingly, if only a peripheral portion of color image 420 is to be chromatically blurred (decision block 715), then process 700 continues to a process block 725. In process block 725, a peripheral portion 625 of color image 420 is chromatically blurred in one or more color channels while a central portion 620 is left unblurred.
The use of peripheral visual stimulation for emmetropization therapy may be applied to a stationary peripheral portion of color image 420 disposed around a stationary central portion of color image 420. In this case eye tracking is not used (decision block 730) and process 700 continues to process block 750. However, if eye tracking is used (decision block 730), then peripheral region 625 is a dynamic peripheral region that surrounds a fixation region 620 representing the user's central vision. As the user's scans their gaze about color image 420, fixation region 620 changes. As such, controller 415 uses camera 410 to track the eye's gazing direction 425 (process block 735) and identify the location of fixation region 620 within color image 420 in real-time based upon the determined gazing direction 425 (process block 740). As gazing direction 425 changes, the location of fixation region 620 and dynamic peripheral region 625 is revised/adjusted to account for the changing location of fixation region 620 (process block 745). These revisions may be made to ensure the user's central vision with higher acuity receives a sharp image while peripheral region 625 having one or more blurred color channels (i.e., simulated LCA) is incident upon the user's peripheral vision. In other embodiments, the user's full field vision may be blurred without any central zone or eye tracking.
In a process block 750, image reconstitution module 515 reconstitutes the image data of the various color channels to generate the color image 420 with simulated LCA in the selected portions or regions of color image 420. As mentioned above, this color image 420 with simulated LCA may also referred to as therapeutic image 520. Finally, in a process block 755 therapeutic image 520 is output from display 405 (or 605) to provide emmetropization therapy to the eye.
The principles discussed above to drive emmetropization of an eye may be performed in software using simulated LCA as discussed above, or performed optically using optical lensing to induce myopic defocus across a user's full FOV (see
The processes explained above are described in terms of computer software and hardware. The techniques described may constitute machine-executable instructions embodied within a tangible or non-transitory machine (e.g., computer) readable storage medium, that when executed by a machine will cause the machine to perform the operations described. Additionally, the processes may be embodied within hardware, such as an application specific integrated circuit (“ASIC”) or otherwise.
A tangible machine-readable storage medium includes any mechanism that provides (i.e., stores) information in a non-transitory form accessible by a machine (e.g., a computer, network device, personal digital assistant, manufacturing tool, any device with a set of one or more processors, etc.). For example, a machine-readable storage medium includes recordable/non-recordable media (e.g., read only memory (ROM), random access memory (RAM), magnetic disk storage media, optical storage media, flash memory devices, etc.).
The above description of illustrated embodiments of the invention, including what is described in the Abstract, is not intended to be exhaustive or to limit the invention to the precise forms disclosed. While specific embodiments of, and examples for, the invention are described herein for illustrative purposes, various modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize.
These modifications can be made to the invention in light of the above detailed description. The terms used in the following claims should not be construed to limit the invention to the specific embodiments disclosed in the specification. Rather, the scope of the invention is to be determined entirely by the following claims, which are to be construed in accordance with established doctrines of claim interpretation.
This application claims the benefit of U.S. Provisional Patent Application No. 63/066,959, filed Aug. 18, 2020, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
63066959 | Aug 2020 | US |