The present invention relates to method and system for registering an object, such as an iris of an eye, so that the object has a known orientation during a medical procedure.
The industry of ophthalmic surgical devices has seen rapid growth over the past couple decades. The use of Excimer lasers in procedures such as LASIK and PRK has become standard practice, and currently cataract surgery is undergoing a similar revolution with femtosecond lasers. In any ophthalmic surgery involving astigmatism correction, it is necessary to account for cyclotorsion, which is a significant rotation of the eye within the socket when a person transitions from standing or sitting up to lying down, as well as any small variations in head tilt or other patient-system alignment parameters. Generally speaking, diagnostic imaging for treatment planning is performed with the patient in an upright position while surgery is performed with the patient lying down, which opens the door for cyclotorsion to cause significant alignment error if not properly accounted for. Thus, in order to reliably use any information from a diagnostic imaging device (such as astigmatism axis) for incision planning, the ocular rotation difference between the diagnostic device and the surgical device must be determined so that the coordinate systems of the devices can be properly aligned.
Historically, cyclotorsion is accounted for by making ink marks along either the “vertical” or “horizontal” axis of the eye when the patient is standing up and using those ink marks as the reference axis when performing the surgical procedure. However, in the context of LASIK procedures, the VISX (Abbott Medical Optics) was the first to switch over to an automatic registration method using the iris patterns of the patient, which requires no ink marks and no manual intervention by the surgeon whatsoever. Automatic iris registration involves a surgical laser system receiving a picture of the patient's eye as seen by the diagnostic device when the treatment was planned, taking its own picture of the patient's eye, and registering the alignment between these two pictures using the iris patterns. To date, these algorithms have only been used in situations where the patient's eye has not undergone drug induced pupil dilation. Existing algorithms are all landmark based, meaning that specific points of interest in the iris are identified in both pre-treatment and treatment images and the registration is performed by matching these points between the two images.
However, in order to be usable in cataract surgery applications, it is necessary to have the patient's eye to undergo drug induced pupil dilation. Astigmatism correcting procedures as a part of cataract surgery typically involve partial thickness arcuate incisions, full thickness clear corneal incisions, tonic intra-ocular lenses, or a combination of the three. There is the possibility for all of these methods to greatly benefit from accurately accounting for cyclotorsion using automatic iris registration.
One aspect of the present invention regards a method that includes illuminating an eye with light at a first time and a second time and generating a first image of the eye based on the light that illuminates the eye at the first time. The method further includes generating a second image of the eye based on the light that illuminates the eye at the second time. The method further includes positioning a laser source relative to the eye, wherein the laser source generates a therapeutic laser beam to be directed to the eye, wherein the first time is just prior to the therapeutic laser beam being directed to the eye and the second time is prior to the first time. The method further includes correcting orientation of the laser source relative to the eye based on a correlation function that is defined for the first and second images of the eye.
A second aspect of the present invention regards a laser therapeutic system that include a laser source that emits a therapeutic laser beam toward an eye and a laser control system that is in communication with the laser source, wherein the laser control system controls one or more parameters of the therapeutic laser beam. The system includes an analyzer that has a light source that illuminates the eye and one or more detectors that receive reflected light off from the eye and generate images of the eye. The analyzer further includes a memory containing computer executable instructions and a processor in communication with the laser control system, the one or more detectors and the memory, wherein the processor receives the images of the eye and executes the computer executable instructions so as to generate a signal based on a correlation function that is defined for the images of the eye and the signal is able to correct orientation of the therapeutic laser beam relative to the eye. Furthermore, the signal is received by the laser control system, which uses the signal to change orientation of the therapeutic laser beam relative to the eye.
One or more of the above mentioned aspects of the present invention provide the advantage of allowing for registration of an iris when drug induced pupil dilation is used for medical procedures, such as cataract surgery applications and astigmatism correcting procedures as a part of cataract surgery typically involve partial thickness arcuate incisions, full thickness clear corneal incisions, toric intra-ocular lenses, or a combination of the three.
One or more of the above mentioned aspects of the invention provide the advantage of accurately accounting for cyclotorsion.
As schematically shown in
In communication with the laser source 102 and laser control system 104 is an analyzer 110. The analyzer 110 includes a light source 112 that illuminates the eye 108. One or more detectors or cameras 114 receive light reflected off the eye 108 and generate images of the eye 108. One image of the eye 108 is a pre-treatment image in that it is taken prior to the patient's eye 108 being subjected to the therapeutic laser beam 106. A second image of the eye 108 is a treatment image and is taken substantially at the time the eye 108 is treated by the therapeutic laser beam 106. The pretreatment and treatment images are stored in a recording medium, such as a memory 116, and are processed in a processor 118, which is in communication with the controller 104, memory 116 and light source 112. An example of an analyzer 110 that can be used is the Topcon CA-200F Corneal Analyzer manufactured by Topcon based in Japan.
The processor 118 executes instructions stored in the memory 116 so that an algorithm is performed in a very different approach from that used by existing algorithms. The algorithm proposed here is a global correlation algorithm, in which the registration is based on a correlation function that is defined for the pre-treatment and treatment images without singling out particular points in the iris. In operation, the eye 108 is imaged by the analyzer 110 prior to drug-induced dilation. Next, the eye 108 undergoes a laser procedure, such as cataract surgery, using the laser source 102 and laser control system 104. The basic steps/processes for the process or algorithm 200 are schematically shown in
In operation, the algorithm(s) related to processes 202-212 listed above are stored in the memory 116 as computer executable instructions, wherein the processor 118 executes the instructions so as to process the pre-treatment and treatment images so as to generate a signal that is able to correct the orientation of the therapeutic laser beam. Such signal is sent to the controller 104 which controls the optics 102 and laser source 103 so as to generate a correctly oriented laser beam 106.
Boundary Detection—Process 202
The easiest boundary to find is the pupil-iris boundary, as this boundary is extremely strong and the pupil itself is uniformly dark. An elliptical fit to the boundary is first found by approximating the center with a histogram method, performing a radial edge filter from this center on edges extracted with the standard canny algorithm, extracting up to 4 circles with a RANSAC algorithm, and combining matching circles together into an elliptical fit. An additional algorithm is used to fine-tune the result even further, which is basically a simplified implementation of Active Contours or Snakes. This algorithm takes as input the image and a previously found elliptical fit to the pupil boundary, and “explores” the image in the neighborhood of the boundary at several values of theta, finding the location that maximizes the radial component of the gradient of intensity values in the image for each theta. This builds a list of points that describe the boundary point by point in polar coordinates (with the origin remaining the center of the previously found ellipse). A simple Gaussian smoothing is then performed on this list of points to enforce continuity. The smoothed list of points is then taken to be pupil boundary.
To find the iris-sclera boundary in the diagnostic image of
To find the iris-sclera boundary in the treatment image, the ellipse describing the limbus in the diagnostic image is transferred to the treatment image by scaling the two radii of the ellipse according to the differing resolutions of the two cameras, assuming no cyclotorsion in placing the axis of the ellipse, and assuming that in the treatment image the limbus will be approximately concentric with the dilated pupil. This constitutes a good initial approximation, which is then improved upon by first using the same snakes algorithm that is used for the pupil boundary and then fitting an ellipse to the resulting set of points.
Often, images taken at a diagnostic device, such as analyzer 110, have some degree of eyelid or eyelash interference concealing a portion of the iris. To mask out these regions from consideration in the registration algorithm, eyelid/iris boundaries must be segmented in an image obtained from analyzer 110, such as shown in
If the threshold for minimum average intensity is met in either one of these three tests, the pixel remains white; otherwise the pixel is made black. Then, a circular mask is applied to mask out areas that are too close to the top and bottom borders of the image to be considered, and the classical erode algorithm is applied to thin out the eyelid/eyelash interference region as well as get rid of any lingering undesired edges, resulting in the image of
Finally, a bottom-up filter is applied to the upper eyelid region resulting in the image of
Filtering and Unwrapping the Iris—Process 204
The iris during dilation is approximated by a rubber sheet model, such that the iris in the non-dilated eye is assumed to basically be a stretched out version of the iris in the dilated eye. In this approximation, a pseudopolar mapping is carried out to unwrap the iris into a rectangular image in which the dimensions represent angle and distance from the inner (pupil) boundary. If the boundary detection is perfect, then the top row of this image will perfectly represent the pupil boundary and the bottom row will perfectly represent the sclera boundary. The size of the averaging area used to fill each pixel in the unwrapped image increases linearly as a function of distance from the pupil center. Obviously, there is technically information loss associated with this approach, with the amount of information loss increasing with distance from the pupil center. However, this loss does not have any noticeable impact, and in fact running registration algorithms on these images rather than the original images results in both a cleaner implementation and faster run time.
After unwrapping, the images are filtered with a Difference-Of-Gaussians (DOG) technique. This technique simply involves subtracting a severely blurred version of the image from a slightly blurred version of the image, which is in effect a band pass filter in the frequency domain. The result is increased signal strength of the iris fibers. An example of an unwrapped, DOG filtered iris is shown in
Feature Extraction—Process 206
A feature vector is built for each unwrapped iris image, with the content of the feature vector being derived from gradient information from the structure tensor and Gabor filters. Thus, the components of the image feature vector are themselves “local” feature vectors with one Gabor filter component and one structure tensor component, and each of these two components are vectors themselves. Gabor filters are used by J. Daugman in his iris recognition algorithms (see http://www.cl.cam.ac.uk/˜jgd1000/csvt.pdf and U.S. and U.S. Pat. No. 5,291,560, the entire contents of which are incorporated herein by reference). The information extracted from the Gabor filter is a point in the complex plane which is computed by convolving a 2D Gabor wavelet with an area of the iris, according to the formula below:
Where α, β, and ω are wavelet size and frequency parameters, (r0,θ0) is the point about which the area of the iris being considered is centered, and I is the intensity value of the unwrapped iris image at a given point. In discrete form, this equation is applied as follows:
Where θ−, θ+, r−, and r+ denote the boundaries of the shell-like region over which the computation is done. For unwrapped images, φ becomes x, ρ becomes y, and the region is rectangular rather than shell-like. This allows for a simple and computationally fast implementation, which is to set r0 and θ0 to zero and fill a 2D array with values according to the above equations with the image intensity values removed, for each the real part and imaginary part, and then convolve these 2D arrays with the images. This yields, at every pixel of each image, a 2D vector with components for the real and imaginary part of the result of centering a gabor filter on that pixel.
Similarly, the structure tensor is used to extract gradient information in the local neighborhood of each pixel. The entries in the 2×2 matrix representing the structure tensor are filled by averaging the derivative-based quantity over the entire neighborhood. Then, the eigenvalues and eigenvectors are extracted from the resulting matrix. The eigenvectors and eigenvalues give the dominant gradient direction and a measure of the strength of the gradient, respectively.
Measuring Correlation Strength—Process 208
Consider the filtered, unwrapped image taken at the time of surgery I1 and the filtered, unwrapped image taken prior to surgery I2. Define an inner product for the structure part of the feature vectors of the two images given a particular hypothesized angle of cyclotorsion δ and a radial shear function ξδ(x) (to allow room for errors in boundary detection and the rubber sheet model approximation) as follows:
Similarly, we define an inner product for the Gabor wavelet part of the feature vectors as follows:
With
When doing this computation, w needs to be large enough to prevent ξδ(x) from being completely chaotic but not so large as to ruin the whole point of allowing a varying radial offset. For example, 10° has been observed to work well. Once the function ξδ(x) is computed for each δ, the inner products as defined in the equations at the beginning of this section can readily be computed.
A strong correlation corresponds to large values of both inner products. The domain of both inner products is [−1, +1]—thus, the net correlation is based on the average of the two inner products. Over a range of +−18°, a reasonable biological limit for cyclotorsion, the net correlation is computed from the average of the two inner products.
An example of correlation measurements as a function of proposed cyclotorsion angle is shown in
Extracting and Applying the Angle of Cyclotorsion—Processes 210 and 212
The angle of cyclotorsion is the angle that produces the maximum correlation strength between the extracted features, which corresponds to the global maximum of the curve C in
τ=100*(1−101-r)
Where r is the ratio between the global maximum and the next largest local maximum present in the correlation function after Gaussian smoothing. For example, in the smoothed correlation function for the plot in
In the left-handed natural coordinate system of the images, the cyclotorsion angle computed tells what value of angle in the topographer image was lined up with the zero angle in the treatment image. In the right-handed coordinate system (where counter clockwise corresponds to positive values of theta), this is equivalent to how the topographer image would be rotated to line up with the treatment image. This is the number needed, because treatment was planned in the frame of reference of the topographer image. Thus, for a cyclotorsion angle of 9.5 degrees, the compass defining the angular coordinate system on the laser system should be rotated by +9.5 degrees. This angle of rotation is calculated by processor 118 and conveyed to the controller 104 which rotates the laser beam 106.
As shown in
Some significant statistical validation is possible as well. One example method for doing this is to extend the allowable range of cyclotorsion angles all the way out to +180° to verify that the global maxima remains at the same location and examine how strong of a maxima it is relative to the correlation strengths of all other cyclotorsion angles. A couple of examples are shown in
Accounting for cyclotorsion is of paramount importance in any refractive surgery focused on correcting and/or compensating for astigmatism. An algorithm is proposed that should be able to solve this problem in cataract surgery either in the context of astigmatism-correcting incisions or assistance in tonic IOL positioning. One exciting aspect of applying iris registration to cataract surgery is the prospect of releasing the first iris registration algorithm that can accommodate drug induced pupil dilation.
From the foregoing description, one skilled in the art can readily ascertain the essential characteristics of this invention, and without departing from the spirit and scope thereof, can make various changes and/or modifications of the invention to adapt it to various usages and conditions.
This application is a continuation of U.S. patent application Ser. No. 14/444,366 filed Jul. 28, 2014, which application claims pursuant to 35 U.S.C. § 119(e) the benefit of U.S. provisional application Ser. No. 61/891,149, filed Oct. 15, 2013, the entire disclosure of each of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4538608 | L'Esperance | Sep 1985 | A |
4764930 | Bille | Aug 1988 | A |
4901718 | Bille | Feb 1990 | A |
4907586 | Bille | Mar 1990 | A |
5098426 | Sklar | Mar 1992 | A |
5246435 | Bille | Sep 1993 | A |
5355181 | Ashizaki | Oct 1994 | A |
5439462 | Bille | Aug 1995 | A |
5480396 | Simon | Jan 1996 | A |
6004314 | Wei | Dec 1999 | A |
6099522 | Knopp | Aug 2000 | A |
6197018 | O'Donnell | Mar 2001 | B1 |
6312422 | Dubnack | Jun 2001 | B1 |
6322556 | Gwon | Nov 2001 | B1 |
6325792 | Swinger | Dec 2001 | B1 |
7655002 | Myers | Feb 2010 | B2 |
8262646 | Frey | Sep 2012 | B2 |
8382745 | Naranjo-Tackman | Feb 2013 | B2 |
8394084 | Optimedica | Mar 2013 | B2 |
8403921 | Optimedica | Mar 2013 | B2 |
8425497 | Optimedica | Apr 2013 | B2 |
8465478 | Frey | Jun 2013 | B2 |
8480659 | Frey | Jul 2013 | B2 |
8500723 | Frey | Aug 2013 | B2 |
8617146 | Frey | Dec 2013 | B2 |
8758332 | Frey | Jun 2014 | B2 |
8801186 | Frey | Aug 2014 | B2 |
9180051 | Frey | Nov 2015 | B2 |
9375349 | Frey | Jun 2016 | B2 |
9545338 | Frey | Jan 2017 | B2 |
9968485 | Potter | May 2018 | B2 |
10213340 | Gray | Feb 2019 | B2 |
10709610 | Morley | Jul 2020 | B2 |
11090190 | Morley | Aug 2021 | B2 |
20020103478 | Gwon | Aug 2002 | A1 |
20030223037 | Chernyak | Dec 2003 | A1 |
20050137586 | Grray | Jun 2005 | A1 |
20070173794 | Frey | Jul 2007 | A1 |
20070185475 | Frey | Aug 2007 | A1 |
20080287928 | Arnoldussen | Nov 2008 | A1 |
20100004641 | Frey | Jan 2010 | A1 |
20110190739 | Frey | Aug 2011 | A1 |
20120016350 | Myers | Jan 2012 | A1 |
20160095752 | Srinivasan | Apr 2016 | A1 |
20160302971 | Frey | Oct 2016 | A1 |
20170290703 | Teuma | Oct 2017 | A1 |
20180085256 | Gray | Mar 2018 | A1 |
Number | Date | Country |
---|---|---|
WO 03102498 | Dec 2003 | WO |
Entry |
---|
Oct. 30, 2015, WIPO, PCT Opinion PCT/US2015/042381. |
Sep. 25, 2005, Chernyak, Method For Wavefront Registration, IE Transactions On Biomedical Engineering, vol. 52, No. 12, Dec. 2005. |
Number | Date | Country | |
---|---|---|---|
20220142816 A1 | May 2022 | US |
Number | Date | Country | |
---|---|---|---|
61891149 | Oct 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14444366 | Jul 2014 | US |
Child | 17373375 | US |