Imaging-controlled laser surgical system

Information

  • Patent Grant
  • 9622913
  • Patent Number
    9,622,913
  • Date Filed
    Wednesday, May 18, 2011
    13 years ago
  • Date Issued
    Tuesday, April 18, 2017
    7 years ago
Abstract
An imaging-based laser system can include a laser-beam system, configured to generate and scan a beam of laser pulses with an adjustable laser-power parameter to points of a scan-pattern in an eye, and an imaging-based laser-controller, configured to image a layer in the eye, to control the scanning of the beam of laser pulses to the points of the scan-pattern, and to control a laser-power parameter of the laser pulses according to the distance of the points of the scan-pattern from the imaged layer.
Description
TECHNICAL FIELD

This patent document describes a system and method for controlling a laser in an ophthalmic procedure. In more detail, this patent document describes an imaging-controlled laser system for controlling the power of a pulsed ophthalmic laser during capsulotomy and cataract procedures, among others.


BACKGROUND

Laser systems have become essential for ophthalmic surgery. They have been employed in corneal procedures for some time now with high precision and therefore considerable success. In very recent times applications for other ophthalmic procedures have been contemplated, including cataract procedures.


Lasers can be used for forming high precision cuts. These cuts are created by focusing or directing a rapid sequence of laser pulses to a scan-pattern or point-pattern. The points of the scan-pattern often form a line or layer and the laser pulses are directed to these points by a scanning system that includes deflection devices, mirrors and lenses whose alignment can be changed very quickly. In typical laser systems the pulses can have a duration or pulse length in the nanosecond, picosecond, or even femtosecond range. The pulse repetition rate can be in the kHz to hundreds of kHz range.


The power or energy of the laser pulses can be chosen to exceed a so-called photodisruption threshold. Laser pulses with a power above this threshold can disrupt the ophthalmic tissue at the target points, inducing the formation of bubbles. Lines or layers of these bubbles can weaken the mechanical connection between the tissue-portions on the opposite sides of the bubbles. Often the weakening is substantial, effectively cutting the tissue. Therefore, a subsequent manual procedure can completely separate the tissue portions with ease.


One ophthalmic procedure which could benefit from using such a high precision laser cutting system is cataract surgery. A typical cataract surgery involves a capsulotomy step and a lysis or lens fragmentation step. During lysis, energy is applied to a lens nucleus to liquefy it. During lens fragmentation, or phaco-fragmentation, the nucleus of the lens can be cut into several pieces by scanning the laser along cutting surfaces to enable the subsequent piece-by-piece removal of the nucleus. The capsulotomy involves forming a circular cut on the anterior portion of the capsular bag of the lens to allow the surgeon to access and remove the cut-up pieces of the nucleus.


To optimize surgical laser systems for these complex ophthalmic procedures is a great challenge. However, the optimization promises great returns in terms of the precision and efficacy of the surgical procedures.


SUMMARY

One of the challenges of laser cataract surgery is that the procedures of capsulotomy and lens fragmentation can interfere with each other. In advanced laser systems the precision of the surgery can be enhanced by imaging the ophthalmic target tissue prior to the surgery and guide the laser pulses based on the image. If the lens fragmentation is performed first, then, as a surgical by-product, the capsule is expanded considerably and unevenly by the substantial amount of bubbles formed inside the capsule. Therefore, after the lens fragmentation, the capsule and lens has to be imaged for a second time to guide the subsequent circular cut of the capsulotomy. However, imaging the severely photodisrupted and distorted lens can be challenging. Also, the repeated imaging procedure consumes precious surgical time, increasing the discomfort of the patient, potentially undermining the precision of the entire procedure.


On the other hand, if the capsulotomy is performed first, it creates a substantial amount of bubbles in the anterior region of the lens and in the anterior aqueous chamber of the eye. The amount of bubbles is especially high if the lens is in a tilted position before the procedure, as explained below. These bubbles can increase the scattering of the laser pulses of the subsequent lens fragmentation considerably as the subsequent pulses are directed to the inside of the lens and thus propagate through the bubble-rich anterior region. The increased scattering can again potentially undermine the precision of the cataract procedure.


Thus, both sequences of the lens fragmentation and capsulotomy have drawbacks, as the first step can reduce the precision and control of the subsequent step. Therefore, laser systems that reduce, resolve, or eliminate one or more of these drawbacks can offer advantages.


Embodiments of the present invention can provide advantageous functionalities in view of these challenges. In particular, an embodiment of an imaging-based laser system can include a laser-beam system, configured to generate and scan a beam of laser pulses with an adjustable laser-power parameter to points of a scan-pattern in an eye, and an imaging-based laser-controller, configured to image a layer in the eye, to control the scanning of the beam of laser pulses to the points of the scan-pattern, and to control a laser-power parameter of the laser pulses according to the distance of the points of the scan-pattern from the imaged layer.


An implementation of an imaging-based laser system can include a laser that generates and directs a beam of laser pulses into an eye, an imaging system that images a capsule layer of the eye, and a laser control system that controls the laser to direct the beam to spots within a tracking band of the imaged capsule layer with a laser-power parameter above a photo-disruption threshold, and to spots outside the tracking band of the imaged capsule layer with a laser-power parameter below a photo-disruption threshold, wherein the image-based laser system is configured to perform a capsulotomy before a lysis or lens- or phaco-fragmentation during a cataract procedure.


An implementation of an image-guided ophthalmic laser system can include a laser engine, configured to generate laser pulses, a beam modifier, configured to modify a laser-power parameter of the laser pulses, a laser scanner, configured to direct the laser pulses to scanning-points in an eye, an imaging system, configured to image a region in the eye, and a pattern generator, coupled to the imaging system, the beam modifier and the laser scanner, configured to generate coordinates of the scanning-points for the laser scanner, and to associate a laser-power parameter with the scanning-points depending on a distance of the scanning-points from a target-pattern.


In some implementations, a method of performing an imaging-controlled ophthalmic procedure can include imaging a layer in an eye, generating coordinates of points of a scan-pattern, determining a distance of the points of the scan-pattern from the imaged layer, and associating laser-power parameters with the points based on the determined distance.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates an embodiment of a surgical laser system with an imaging-controlled laser system



FIGS. 2A-D illustrate embodiments of the laser-beam system.



FIGS. 3A-E illustrate embodiments of the imaging-based laser controller.



FIGS. 4A-B illustrate the scan-patterns for non-tilted and tilted lenses.



FIGS. 5A-B illustrate traditional scan-patterns for non-tilted and tilted lenses as a function of a scanning variable.



FIGS. 6A-H illustrate a scan-pattern along a circular scan with a distance-dependent laser-power parameter.



FIG. 7 illustrates a determination of the z-depth of the imaged layer by using a model curve.



FIG. 8A-B illustrate methods of cataract surgery with the lens fragmentation and capsulotomy in different sequences.



FIG. 9 illustrates a method of cataract surgery with an imaging-controlled laser system in detail.



FIG. 10 illustrates a multi-extrema tracking-band laser scan-pattern after lens-fragmentation expanded the lens capsule in a non-uniform manner.



FIGS. 11A-D illustrate scan-patterns for tilted chop cuts.



FIGS. 12A-B illustrate scan-patterns for tilted volume cuts.





DETAILED DESCRIPTION

Implementations and embodiments described in this patent document offer improvements for the above described challenges.



FIG. 1 illustrates an imaging-based laser system 100, including a laser-beam system 110 to generate and scan a beam of laser pulses with an adjustable laser-power parameter to points of a scan-pattern in an eye 1, and an imaging-based laser-controller 120 to image a layer in the eye, to control the scanning of the beam of laser pulses to the points of the scan-pattern, and to control a laser-power parameter of the laser pulses according to the distance of the points of the scan-pattern from the imaged layer. The laser-controller 120 can perform these functions by sending a power control signal and a scanning control signal to the laser-beam system 110, for example.


The laser beam of the laser-beam system 110 can be guided into the main optical pathway at a beam-splitter 132-1 that can redirect the beam to an objective 134. The beam can propagate through the objective 134 and through a patient interface 136 to enter into the surgical eye 1.


The surgery can be assisted by imaging the eye 1 with various techniques. A visible imaging light can be used to create a video image that is processed by a video microscope 138. In addition, the imaging-based laser-controller 120 can shine an imaging beam on the eye and form an image based on the returned image beam. This imaging beam can be coupled into and out of the main optical path by a beam-splitter 132-2.



FIGS. 2A-D illustrate various embodiments of the laser-beam system 110.



FIG. 2A illustrates that embodiments of the laser-beam system 110 can include a laser engine 112 to generate the beam of laser pulses, a beam attenuator 114 to modify the laser-power parameter of the laser pulses, and a beam scanner 116 to direct the beam of laser pulses to the points of the scan-pattern in the eye. The laser engine 112 can generate laser pulses with a duration of nanoseconds, picoseconds or even femtoseconds, i.e. in the 10−9-10−15 sec range. These pulses can be generated at a repetition rate in a wide range of frequencies: from 0.1 kHz to 1,000 kHz, or in a range of 1 kHz to 500 kHz, or in some implementations in the 10 kHz to 100 kHz range. The power control signal of the laser-controller 120 can be coupled into the beam attenuator 114 and the scanning control signal of the laser-controller 120 can be coupled into the beam scanner 116.


The beam attenuator 114 can include a Pockels cell, a polarizer-assembly, a mechanical shutter, an electro-mechanical shutter, or an energy wheel. Each of these implementations can modify a laser-power parameter of the laser pulses. The laser-power parameter can be a pulse energy, a pulse power, a pulse length or a pulse repetition rate of the laser pulses, among others. The beam attenuator 114 can modify one or more of these laser-power parameters. In a simple implementation, the beam attenuator 114 can shutter or block selected laser pulses. In another, a polarizer assembly can reduce the power of selected laser pulses by adjusting the relative angle of subsequent polarizing filters.


In the embodiment of FIG. 2A, the beam attenuator 114 can be located between the laser engine 112 and the beam scanner 116 in the path of the laser beam.



FIG. 2B illustrates and embodiment in which the beam attenuator 114 is at least partially integrated into the laser engine 112. In some cases, the beam attenuator 114 can be part of the laser engine 112. For example, a Pockels cell within the laser engine 112 can be the beam attenuator 114.



FIG. 2C illustrates and embodiment in which the beam attenuator 114 is located after the beam scanner 116 in the path of the laser beam.


Finally, FIG. 2D illustrates an embodiment in which the beam attenuator 114 and the beam scanner 116 are at least partially integrated.



FIGS. 3A-E illustrate various embodiments of the imaging-based laser-controller 120.



FIG. 3A illustrates that the laser-controller 120 can include an imaging system 122 to image the imaged layer in the eye and a pattern generator 124 to generate coordinates of the points of the scan-pattern, to associate laser-power parameters with the points depending on the distance of the points from the imaged layer, and to signal the generated coordinates of the points and the corresponding laser-power parameters to the laser-beam system 110. In some implementations, the imaging system 122 can image any ophthalmic target in the anterior or posterior segment of the eye, targets from the cornea to the retina.


The pattern generator 124 can signal the generated coordinates of the points of the scan-pattern to the beam scanner 116 with a scanning control signal. Further, the pattern generator 124 can signal the laser-power parameters corresponding to the points of the scan-pattern to the beam attenuator 114 with a power control signal. The laser-power parameter can be a pulse energy, a pulse power, a pulse length or a pulse repetition rate of the laser pulses.


The imaging system 122 can include an ophthalmic coherence tomography (OCT) system, a Scheimpflug imaging system, a scanning imaging system, a single shot imaging system, an ultrasound imaging system, and a video imaging system. Here, the scanning imaging systems can create the image by scanning an imaging beam, whereas single shot imaging systems can acquire imaging information about an imaged area or volume in a single shot. The OCT system can be a time-domain OCT, a frequency-domain OCT, or a spectrometer-based OCT system, among others.



FIG. 3B illustrates that in some implementations the laser-controller 120 can include an image-analyzer 126. The image analyzer 126 can receive the image of the imaged layer from the imaging system 122, perform an analysis of the imaged layer as described below and forward the result of the analysis to the pattern generator 124.



FIG. 3C illustrates that in some implementations the image analyzer 126 can be at least partially integrated with the imaging system 122. FIG. 3D illustrates that in some implementations the image analyzer 126 can be at least partially integrated with the pattern generator 124.



FIG. 3E illustrates that in some embodiments, the laser system 100 can include an operator-interface 128 that can be coupled to one or more of the imaging system 122, the pattern generator 124 and the image analyzer 126.



FIGS. 4A-B set the stage to illustrate the operation of the laser system 100. The imaging system 122 can image the imaged layer in an image region that can be based on a loop, an arc, a line, or a two-dimensional pattern transverse to a z-axis of the imaging system, and extends to a depth range Dimage along the z-axis of the imaging system. The imaging system 122 can support a determination of a z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image-scan.



FIG. 4A illustrates that the imaging system 122 can perform an imaging relevant for a capsulotomy step of a cataract procedure. The schematic cross section illustrates the anterior segment of the eye 1. The outermost layer is a cornea 210. A crystalline lens 220 is located behind the cornea 210, separated from it by an aqueous anterior chamber 230. The crystalline lens 220 is encapsulated in a thin capsule or capsular bag 222. The lens 220 is held in place by ciliary muscles 240. These muscles 240 also adjust the shape of the crystalline lens 220 as needed for bringing objects into focus.


As it has been described above, in order to facilitate the removal of a fragmented nucleus of the lens 220, the cataract surgery typically involves creating a circular capsulotomy cut 250 on the capsular bag 222. As a first step, the imaging system 122 can create an image 252 of the anterior segment of the eye by scanning along a scanning circle 254 and imaging the eye in a depth-range Dimage, defining an image-cylinder 260-i.



FIG. 5A illustrates that the image 252 typically includes an image 256 of the imaged anterior capsule layer of the lens 220 “unfolded” along a scanning variable, such as an angle along the circumference of the scanning circle 254. If a z-axis of the lens 220 is aligned with a z-axis of the laser system 100, the image 256 of the imaged layer is a flat line, indicating an essentially constant z-depth.


In other implementations, the image 252 can include the image of other ophthalmic targets, including corneal layers, portions of the sclera and even retinal layers. The zero depth level can be defined in a large number of ways, using a lens of the objective 134, a reference mirror of the imaging system 122, a level of the patient interface 136, or a level of an ophthalmic structure, such as the cornea 210.


By analyzing the image 252, a surgeon can recognize the image 256 of the imaged layer. Based on the z-depth of the imaged layer, the surgeon can decide where to direct the cutting laser beam to form the capsulotomy cut 250. The cutting laser beam is typically scanned along the same scanning circle 254 to form a cut-cylinder 260-c with a depth-range Dcut, typically smaller than Dimage. This way the placement of the cut-cylinder 260-c benefits maximally from the information contained in the image 252, and in particular in the image 256 of the imaged layer. The capsulotomy cut 250 is formed where the cut-cylinder 260-c intersects the lens capsule 222. In practice, the cut cylinder 260-c is often formed as a stack of bubble-circles, where the individual circles are created by directing the laser pulses along a circular scan-pattern at a fixed z-depth to cause photodisruption, followed by the formation of a similar circle at a slightly lesser z-depth.


In some typical cases, the image depth-range Dimage can be 5-10 millimeters, whereas the cut depth-range Dcut can be in the range of 50-200 microns, in some cases 75-150 microns, sometimes approximately 100 microns.


It is noted that the bubbles of the cut-cylinder 260-c can scatter and deflect laser pulses applied in subsequent surgical steps. For example, in a cataract surgery the capsulotomy can be followed by the lens fragmentation or lysis. The bubbles of the cut-cylinder 260-c can negatively impact the precision and efficiency of this subsequent lens-fragmentation by scattering the lens-fragmenting laser pulses.


Fortunately, when a z-axis of the lens 220 is parallel to a z-axis of the laser system 100, the depth range Dcut of the cut cylinder 260-c can be as little as 100 microns, creating only a limited number of bubbles. Thus, in the case of a well-aligned lens 220, the bubbles of the cut-cylinder 260-c introduce only a limited amount of scatter for the subsequent lens fragmentation laser pulses.



FIG. 4B illustrates, however, that in the typical surgical case the crystalline lens 220 can be tilted. This situation can occur for a variety of reasons. For example, the weight of the objective 134 can push the lens 220 sideways upon docking to the eye 1. Or, applying suction at the patient interface 136 to immobilize the eye 1 can lead to a tilting of the lens 220 as well.



FIG. 5B illustrates the image 252 of such a tilted lens 220 unfolded along the angular scanning variable of the scanning circle 254. In contrast to the non-tilted case of FIG. 5A, the image 256 of the tilted imaged layer can exhibit substantial sinusoidal oscillations. The amplitude of these oscillations can be as much as 300-500 microns. To make sure that the capsular bag 222 is cut everywhere along this sinusoid, the cut-cylinder 260-c can be formed with a much enlarged depth-range Dcut, exceeding the amplitude of the sinusoid. In the above example, Dcut can be 400-600 microns to be sure that the capsular bag 222 was cut along the entire sinusoid. Clearly, this approach may create 4-6 times more photodisrupted bubbles during capsulotomy than the procedure for a non-tilted lens. Capsulotomy bubbles in such an increased number can scatter the laser pulses of the subsequent lens fragmentation to a substantial degree, threatening its precision and efficacy.



FIGS. 6A-H illustrate that some implementations of the laser system 100 can substantially reduce the number of photodisrupted bubbles by generating bubbles only in a narrow proximity of the imaged layer.


As described above, this outcome can be achieved, for example, by the imaging-based laser-controller 120 imaging the capsular bag 222, controlling the scanning of the beam of laser pulses to the points of the scan-pattern, and controlling a laser-power parameter of the laser pulses according to the distance of the points of the scan-pattern from the imaged layer.



FIGS. 6A-B illustrate that as the laser pulses are directed to points of the scan-pattern, the laser controller 120 can modify or adjust a laser-power parameter of the pulses. In particular, when a laser pulse is directed to a point of the scan pattern that is within a Dcut distance from the image 256 of the imaged layer along the z axis, the laser-controller 120 can adjust its laser-power parameter to a high value, e.g. above a photodisruption threshold. Whereas, when a laser pulse is directed to a point of the scan pattern that is farther than Dcut from the image 256 of the imaged layer, the laser-controller 120 can adjust its laser-power parameter value to a low value, such as below a photodisruption threshold.


The just-described method creates bubbles only in a Dcut proximity of the imaged layer and therefore substantially reduces the number of bubbles to a value close to the number of bubbles for a well-aligned lens. For this reason, the scattering of the subsequent lens-fragmenting laser pulses by these capsulotomy bubbles is substantially reduced. Using the earlier value of Dcut being 400-600 microns for a tilted lens and 100 microns for a non-tilted lens, the present method may reduce the scattering of the lens-fragmenting bubbles by a factor of 4-6: a considerable gain in precision and control.



FIG. 6A illustrates the implementation when the scanning of the capsulotomy laser pulses of the scan-pattern is performed along the z-axis for fixed points of the circular scan. FIG. 6B illustrates the implementation when the scanning is performed along the circular scan with a fixed z-depth. This implementation can be used to create the above mentioned stacked circles. In either implementation, the points with high laser-power are placed within a tracking band 257 with a z-extent of Dcut.



FIGS. 6C-E illustrate the implementation when the laser pulses are scanned at fixed z-depths along the circular scan. A tracking band 257 can be defined as the set of points of the scan-pattern that are within the preselected distance Dcut from the image 256 of the imaged layer.



FIGS. 6D-E illustrate the laser power parameter of the pulses along the circular scan at two selected z-depths of 3600 microns and 3650 microns in an unfolded representation. The laser-controller 120 can control the laser power of the pulses that are directed to points inside the tracking band 257 to be above a photo-disruption threshold, and the laser power of the pulses that are directed to points outside the tracking band 257 to be below the photo-disruption threshold. In this embodiment, photodisrupted bubbles are only generated at points within the tracking band 257, achieving the above functionality of the laser system 100.



FIG. 6F expresses the same operation in a folded representation. Here the value of the laser power parameter is shown as a function of the angular scanning variable (typically the angle), projected on the scanning circle 254 itself. Again, for those points of the scan-pattern that lie within the tracking band 257, the laser power is high—indicated by a thick line—whereas for those points that lie outside the tracking band 257, the laser power is low.



FIGS. 6G-H illustrate a related implementation, where the laser-power controller 120 controls the laser power parameter as a function of the distance of the points from the imaged layer, wherein the laser-power is a decreasing function of the distance. FIG. 6G illustrates the implementation where this function is essentially a two-valued step-function. FIG. 6H illustrates the implementation where this function is a continuous function, its value decaying with the increasing distance from the imaged layer. In some implementations, it may be easier to control the laser power in the continuous manner of FIG. 6H.


The above-outlined implementations depend on the knowledge of the distance between the points of the scan-pattern and the imaged layer. Three stages are involved in determining this distance. First, the identity of the imaged layer is identified in the image 252 to determine the image 256 of the imaged layer. Then, the z-depth coordinate of the imaged layer is determined. Finally, the distance of the imaged layer and the points of the scan-pattern can be determined, for example, by taking the difference of the z-depth coordinates of the points of the scan-pattern and the imaged layer at the corresponding angular scanning coordinates, such as at the same angle.


Concerning the first step, the raw image 252 does not isolate or identify the imaged layer explicitly. Thus, establishing the identity of the imaged layer may necessitate an analysis of the image 252. As discussed earlier, this analysis of the image can be performed by the imaging system 122, the pattern generator 124, or the image analyzer 126, possibly assisted by an input from a system operator through the operator interface 128.



FIG. 7 illustrates that the imaging system 122 can support the identification of the imaged layer and the determination of its z-depth coordinates in different ways. In some implementations the laser system 100 can include the operator interface 128 and the imaging system 122 can support the identification of the imaged layer using an input from an operator through the operator interface 128.


For example, on a graphical user interface, or GUI, the operator interface 128 can prompt the operator to fit a model curve 258 to the spots in the image 252 representing the imaged layer. Since in the case of a tilted ellipsoid-shaped lens the image 256 of the imaged layer is typically a sinusoidal curve, the operator interface 128 can display a generic sinusoidal curve 258 on the GUI and prompt the operator to fit this model curve 258 to the layer-spots in the image 252. Once the operator fitted the model curve 258 to the layer-spots in the image 252, the model curve 258 can serve as the image 256 of the imaged layer.


The operator can achieve this task through various approaches: by shifting the model curve 258 by an Xshift in the X direction (i.e. adjusting the angle along the circular scan) and by shifting the model curve 258 by a Yshift in the Y direction (i.e. adjusting the z-depth coordinate). In other implementations the operator can be prompted to adjust the scale of the model curve 258 to the scale of the sinusoidally located layer-spots in the image 252, i.e. to rescale the z-depth of the model curve 258 to fit the z-depth of the layer-spots. Many other fitting techniques can be implemented to achieve analogous functionalities.


The operator interface 128 can receive the input from the operator in many different ways, including through a keyboard, a touch-screen, a computer-communication channel, an external memory, a flash-drive, an internet connection, a speech-recognition apparatus or a wireless connection.


In other implementations, the determination of the identity and the z-depth of the imaged layer can be performed by the laser system 100 without the input of a surgeon or operator. In particular, the imaging system 122 can be configured to determine the identity and then the z-depth coordinate of the imaged layer by a processor or micro-computer performing a feature-recognition analysis of the image 252. For example, the imaging system 122 can determine the identity and coordinates of the imaged layer by locating local maxima of the gradient of the spot intensity. In other implementations, an edge-recognition algorithm can be used. In these implementations, the imaging system 122 can identify the manifold of the maximum-gradient points as the image 256 of the imaged layer without resorting to fitting a model curve 258. In some implementations, of course, the imaging system 122 can make use of a model curve 258 to identify the image 256 of the imaged layer.


In the above implementations, once the identity of the imaged layer has been determined in the image 252, the z-depth coordinates of the imaged layer can be determined in a straightforward manner, for example, by counting the pixels in the image 252, or using a reference or a look-up table.


For the image analysis, the imaging system 122 can utilize a result of a pre-surgery measurement, statistical data, video image data, ophthalmic coherence tomography image data, or a model-based computation during the determination of the z-depth.


Once the z-depth of the imaged layer has been determined, the imaging system 122 can forward the z-depth and the corresponding scanning coordinates of the imaged layer to the pattern generator 124 to carry out the last stage, the determination of the distance between the imaged layer and the points of the scan-pattern, generated by the pattern generator 124. This stage can be carried out, for example, by subtracting the z-depth coordinates of the points of the scan-pattern from the z-depth coordinates of the imaged layer that correspond to the same scanning variable, such as the same scanning angle.


Finally, having determined the distance of the points of the scan-pattern from the imaged layer, the pattern generator 124 can associate a laser-power parameter above a photodisruption threshold with those points that are closer to the imaged layer than a predetermined distance, and associate a laser-power parameter below a photodisruption threshold with those points that are farther from the imaged layer than the predetermined distance, as described in relation to FIGS. 6A-H.


In some implementations, the imaging system 122 only captures the image 252 but does not identify the imaged layer or determine its z-depth coordinates. In these embodiments, the imaging system 122 can simply forward the unprocessed image 252 to the pattern generator 124 without analyzing it. The pattern generator 124 can receive the image 252, identify the imaged layer and determine the z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image scan.


As above, in some implementations, the pattern generator 124 can determine the z-depth of the imaged layer by performing a feature-recognition analysis of the received image 252. In other implementations, the pattern generator 124 can receive an operator input through the operator interface 128 during the process of determining the z-depth of the imaged layer, as described before.


In these implementations, once the z-depth coordinates of the imaged layer have been determined, the pattern generator 124 can define a tracking band 257 as a manifold of the points of the scan-pattern that are within a predefined distance from the coordinates of the imaged layer. Then the pattern generator 124 can associate a laser-power parameter above a photodisruption threshold with points of the scan-pattern inside the tracking band 257, and a laser-power parameter below a photodisruption threshold with points of the scan-pattern outside the tracking band 257.


Yet other implementations of the laser controller 120 may include an image analyzer 126 that can determine the z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image-scan. As was illustrated in FIGS. 3B-D, the image analyzer 126 can be self-standing or at least partially integrated with the imaging system 122 or the pattern generator 124.


The image analyzer 126 can identify the imaged layer and determine the z-depth coordinate of the imaged layer by performing a feature-recognition analysis of the image 252. In other implementations, the image analyzer 126 can determine the z-depth coordinate by making use of an operator input through an operator-interface 128.


The operation of the laser system 100 can be demonstrated on the example of the capsulotomy procedure, where the imaged layer is the lens capsule 222 between the lens 220 and the aqueous anterior chamber 230. In this case, the scan-pattern corresponds to the cut-cylinder 260-c intersecting the lens capsule 222 at the capsulotomy cut 250. The pattern generator 124 can associate a photodisruptive laser-power parameter with points inside a tracking band 257 related to the intersection 250 of the cut-cylinder 260-c and the lens capsule 222, and a non-photodisruptive laser-power parameter with points outside the tracking band 257.



FIG. 8A illustrates a first cataract procedure 300 performed without the benefits of the laser system 100. The cataract procedure 300 can be practiced when the capsulotomy generates an excessive number of bubbles as in FIGS. 4B-5B. To prevent excessive scattering by these capsulotomy bubbles, the lens fragmentation is performed prior to the capsulotomy. In detail, the cataract procedure 300 can include a first imaging 310 of the capsule 222, performed by an OCT procedure, followed by a lens fragmentation 320. During the lens fragmentation 320 the capsule 222 expands because of the large number of bubbles generated in the crystalline lens 220. The fragments of the lens 220 are removed through an opening, cut into the capsule 222 by a capsulotomy 340. However, since the capsule 222 has expanded during the lens fragmentation 320, the results of the first imaging 310 are not reliable anymore. Therefore, the capsulotomy 340 has to be preceded by a second imaging 330. The second imaging 330 can take up precious surgical time and increase the discomfort of the patient. Both of these factors can endanger the efficacy of the cataract procedure 300.



FIG. 8B illustrates a cataract procedure 350 with an embodiment of the laser system 100. Since the laser system 100 is capable of creating only a limited number of bubbles during the capsulotomy, the capsulotomy can be performed before the lens fragmentation. This change of sequence can reduce the surgical time to a considerable degree and thus increase the precision of the cataract procedure substantially.


In some detail, the cataract procedure 350 can include an imaging 360 of the capsule 222, e.g. by an OCT imaging system, followed by a capsulotomy 370, and completed by a lens fragmentation 380. Since the capsulotomy 370 does not deform the lens 220, there is no need for a second imaging, in contrast to the procedure 300.



FIG. 9 illustrates an imaging-controlled cataract method 400 in more detail. The method 400 can include an imaging 410 of an imaged ophthalmic layer in an imaged region of an eye, followed by an identifying 420 of the coordinates of the imaged layer from the image. These tasks can be performed, for example, by the imaging system 122 of the imaging-based laser-controller 120. The identifying 420 can include performing a feature-recognition analysis. In other cases, it can include receiving an operator-input through an operator interface 128. These tasks can be performed by the imaging system 122, the pattern generator 124 or the image analyzer 126.


Next, the method 400 can include a generating 430 of coordinates of points of a scan-pattern, and a determining 440 of a distance of the points of the scan-pattern from the imaged layer. These steps can be performed for example, by the pattern generator 124.


The method 400 can further include an associating 450 of laser-power parameters with the generated points based on their determined distance. The tasks 420 to 450 can include receiving possible inputs 422-452 from an operator of the laser system 100 through the operator interface 128.


The method can also include a signaling 460 of the generated coordinates of the points of the scan-pattern to the beam scanner 116 and a signaling 470 of the corresponding laser-power parameters to the beam attenuator 114.



FIG. 10 illustrates the case of surgical relevance when the lens capsule 222 has an uneven shape. This situation can arise in different circumstances. For example, the docking of the patient interface 136 can cause considerable deformation of the anterior segment of the eye 1. Or an ophthalmic trauma or a prior lens fragmentation procedure can result in an uneven lens shape. In any of these circumstances, the laser system 100 can be capable of analyzing an image 256 of the imaged layer that exhibits more than two local extrema. Visibly, a simple sinusoidal model curve 258 is insufficient to identify the imaged layer and to determine its z-depth coordinate in this case. Therefore, embodiments of the imaging system 122, the pattern generator 124 or the image analyzer 126 can be capable of recognizing the imaged layer and determine its z-depth coordinate even in this more challenging case, for example, by using sophisticated feature-recognition software. Having determined and characterized the image 256 of the imaged layer can allow the pattern generator 124 to define the tracking band 257 to associate laser-power parameters with the spots of the scan-pattern accordingly.



FIGS. 11A-D illustrate that the imaging system 122 of the laser system 100 can image a region in the eye, the pattern generator 124 can generate coordinates of points of a scan-pattern for the beam scanner 116, and associate a laser-power parameter with the points of the scan-pattern depending on a distance of the points from a target-pattern.


An example for such a target pattern can be a chop pattern 500, including the chop-planes 500-X and 500-Y. Such chop patterns 500 can be used for lens fragmentation. FIG. 11A illustrates the case when the z-axis of the lens 220 is aligned with the z-axis of the laser system 100. In this case the chop-planes 500-X and 500-Y are also parallel to the z-axis of the laser system 100.



FIG. 11B illustrates that if the lens 220 is tilted relative to the z-axis of the laser system 100, as illustrated e.g. in FIG. 4B, then the chop planes 500-Xt and 500-Yt can be tilted as well. Since the scan-pattern often includes a first manifold of points at a first fixed z-depth, followed by a second manifold at a slightly lesser z-depth, the scan-pattern of tilted chop-planes with laser systems that cannot adjust the power of the laser pulses would create cuts into the capsular bag 222, leading to massive surgical complications.


In contrast, embodiments of the laser system 100 can associate laser-parameters depending on the distance of the points of the scan-pattern from the chop planes 500-Xt and 500-Yt.



FIGS. 11C-D illustrate the points of the scan-pattern with low and high laser power, generated by the pattern generator 124 to form the tilted 500-Xt and 500-Yt chop planes. Visibly, creating cuts by adjusting the power of the laser pulses depending on their proximity to the target-pattern can avoid cutting into the capsular bag—a major surgical advantage.



FIG. 11D illustrates clearly that, as it was the case of the tracking band 257, a photodisruptive laser-power parameter can be associated with scan-points that are closer to the target-pattern 500-Xt and 500-Yt than a predetermined distance Dcut, and a non-photodisruptive laser-power parameter with the scan-points that are farther from the target-pattern than the predetermined distance Dcut.


In other implementations, the cutting surface can be a circular surface-segment, a spiral surface-segment, a corneal access cut and a limbal relaxing cut.



FIGS. 12A-B illustrate that in some cases the target pattern 260-2 can be a target volume with an axis tilted relative to an optical axis of the laser system 100. Here, the scan pattern includes cylindrical patterns 260-1, and the laser-power parameter of the points of this scan-pattern is adjusted to form a tilted volume cut 260-2. Such a utility can be useful for correcting a refractive property of the lens 220, for example.


In some implementations, the pattern generator 124 can be configured to associate the laser-power parameters with the points of the scan-pattern depending additionally on a distance of the points from an ophthalmic layer, imaged by the imaging system 122.


While this specification contains many specifics, these should not be construed as limitations on the scope of the invention or of what can be claimed, but rather as descriptions of features specific to particular embodiments. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features can be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination can be directed to a subcombination or variation of a subcombination.

Claims
  • 1. An imaging-based laser system, comprising: a laser-beam system, including a laser engine, configured to generate a beam of laser pulses,a beam attenuator, configured to modify a laser-power parameter of the laser pulses, wherein the laser-power parameter is one of a pulse energy, a pulse power, a pulse length and a pulse repetition rate, anda beam scanner, configured to scan the beam to points of a cylindrical scan-pattern in an eye; andan imaging-based laser-controller, configured to: image a layer in the eye that is tilted relative to an optical axis of the laser system,determine z-depths of a sequence of points in the cylindrical scan-pattern that correspond to the imaged layer in the eye,generate a tracking band within the cylindrical scan pattern defining a cut to be made in the eye, wherein a lower boundary of the tracking band has a non-uniform z-depth that varies according to the determined z-depths of the sequence of points corresponding to the imaged layer,cause the beam scanner to scan the beam of laser pulses to the points of the cylindrical scan-pattern, andcause the beam attenuator to control the laser-power parameter of the laser pulses such that a laser power parameter of laser pulses in the tracking band is above a photo-disruption threshold, and a laser power parameter of laser pulses outside the tracking band is below the photo-disruption threshold.
  • 2. The laser system of claim 1, the beam attenuator comprising at least one of: a Pockels cell, a polarizer-assembly, a mechanical shutter, an electro-mechanical shutter, and an energy wheel.
  • 3. The laser system of claim 1, wherein: the beam attenuator is disposed between the laser engine and the beam scanner in a path of the beam.
  • 4. The laser system of claim 1, wherein: the beam attenuator is disposed after the beam scanner in a path of the beam.
  • 5. The laser system of claim 1, wherein: the beam attenuator is part of the laser engine.
  • 6. The laser system of claim 1, wherein: the beam attenuator and the beam scanner are at least partially integrated.
  • 7. The laser system of claim 1, the laser-controller comprising: an imaging system, configured to image the imaged layer in the eye; anda pattern generator, configured to: generate coordinates of each point within the cylindrical scan-pattern,associate a particular laser-power parameter with each point in the cylindrical scan pattern based on the tracking band, andsignal the generated coordinates of each point to the beam scanner, andsignal the particular laser-power parameter of each point to the beam attenuator.
  • 8. The laser system of claim 7, the imaging system comprising: at least one of an ophthalmic coherence tomography system, a Scheimpflug imaging system, a scanning imaging system, a single shot imaging system, an ultrasound imaging system, and a video imaging system.
  • 9. The laser system of claim 7, wherein: the imaging system is configured to image the imaged layer in an image region, wherein the image region is based on one of a loop, an arc, a line, and a two-dimensional pattern transverse to an axis of the imaging system, andextends to an image depth along the axis of the imaging system.
  • 10. The laser system of claim 7, wherein: the imaging system is configured to support a determination of a z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image-scan.
  • 11. The laser system of claim 10, wherein: the laser system comprises an operator interface; andthe imaging system is configured to support the determination of the z depth coordinate of the imaged layer using an input from an operator through the operator interface.
  • 12. The laser system of claim 11, wherein: the operator interface is configured to assist the operator to fit a model curve to the image of the imaged layer.
  • 13. The laser system of claim 11, wherein: the operator interface is capable of receiving the operator-input from at least one of a keyboard, a touch-screen, a computer-communication channel, an external memory, a flash-drive, an internet connection, a speech-recognition apparatus and a wireless connection.
  • 14. The laser system of claim 10, wherein: the imaging system is configured to determine the z-depth coordinate of the imaged layer by performing a feature-recognition analysis of the image of the imaged layer.
  • 15. The laser system of claim 14, wherein: the imaging system is configured to utilize at least one of a result of a pre-surgery measurement, statistical data, video image data, ophthalmic coherence tomography image data,and a model-based computation during the determination of the z-depth.
  • 16. The laser system of claim 10, wherein: the imaging system is configured to forward the z-depth and scanning coordinates of the imaged layer to the pattern generator; andthe pattern generator is configured to determine the distance of the points of the scan-pattern from the imaged layer based on the forwarded coordinates of the imaged layer and the generated coordinates of the points,to associate a first laser-power parameter above a photodisruption threshold with a first set of points closer to the imaged layer than a predetermined distance, andto associate a second laser-power parameter below a photodisruption threshold with a second set of points farther from the imaged layer than the predetermined distance.
  • 17. The laser system of claim 10, wherein: the imaging system is configured to forward the z-depth and scanning coordinates of the imaged layer to the pattern generator; andthe pattern generator is configured to determine the distance of the points of the scan-pattern from the imaged layer based on the forwarded coordinates of the imaged layer and the generated coordinates of the points, andto associate with the coordinates of the points a laser-power parameter that is a decreasing function of the distance of the points from the imaged layer.
  • 18. The laser system of claim 7, wherein: the imaging system is configured to forward the image of the imaged layer to the pattern generator; andthe pattern generator is configured to receive the image from the imaging system, andto determine a z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image scan.
  • 19. The laser system of claim 18, wherein: the pattern generator is configured to determine the z-depth of the imaged layer in part by performing a feature-recognition analysis of the received image of the imaged layer.
  • 20. The laser system of claim 18, wherein: the pattern generator is configured to receive an operator input through an operator interface during the process of determining the z-depth of the imaged layer.
  • 21. The laser system of claim 20, wherein: the operator interface is capable of receiving the operator-input from at least one of a keyboard, a touch-screen, a computer-communication channel, an external memory, a flash-drive, an internet connection, a speech-recognition apparatus and a wireless connection.
  • 22. The laser system of claim 18, wherein: the pattern generator is configured to: generate the tracking band as a manifold of points within a predefined distance from the coordinates of the imaged layer;associate a laser-power parameter above a photodisruption threshold with points of the scan-pattern inside the tracking band, andto associate a laser-power parameter below a photodisruption threshold with points of the scan-pattern outside the tracking band.
  • 23. The laser system of claim 7, the laser controller comprising: an image analyzer, configured to determine a z-depth coordinate of the imaged layer corresponding to a scanning coordinate along an image-scan.
  • 24. The laser system of claim 23, wherein: the image analyzer is configured to determine the z-depth coordinate of the imaged layer by performing a feature-recognition analysis of the image of the imaged layer.
  • 25. The laser system of claim 23, wherein: the image analyzer is configured to determine the z-depth coordinate of the imaged layer by receiving an operator input through an operator-interface.
  • 26. The laser system of claim 23, wherein: the image analyzer is at least partially integrated with one of the imaging system and the pattern generator.
  • 27. The laser system of claim 1, wherein: the imaged layer is a lens capsule between a lens of an eye and an aqueous anterior chamber of the eye;the scan-pattern corresponds to a cylindrical capsulotomy cut intersecting the lens capsule; andthe imaging-based laser controller is configured to: associate a photodisruptive laser-power parameter with points inside a tracking band related to the intersection of the cylindrical capsulotomy cut and the lens capsule, andassociate a non-photodisruptive laser-power parameter with points outside the tracking band.
  • 28. The laser system of claim 27, wherein: the laser system is configured to perform a capsulotomy before a lens fragmentation during a cataract procedure.
  • 29. The laser system of claim 27, wherein: the imaging-based laser controller is configured to to analyze an image of the capsule boundary layer with more than two local extrema by at least one of a pattern generator and an image analyzer.
US Referenced Citations (231)
Number Name Date Kind
4164222 Prokhorov et al. Aug 1979 A
4198143 Karasawa Apr 1980 A
4235529 Kawase et al. Nov 1980 A
4465348 Lang Aug 1984 A
4520816 Schachar et al. Jun 1985 A
4533222 Ishikawa Aug 1985 A
4538608 L'Esperance, Jr. Sep 1985 A
4554917 Tagnon Nov 1985 A
4638801 Daly et al. Jan 1987 A
4764005 Webb et al. Aug 1988 A
4881808 Bille et al. Nov 1989 A
4901718 Bille et al. Feb 1990 A
4907586 Bille et al. Mar 1990 A
5048946 Sklar et al. Sep 1991 A
5049147 Danon Sep 1991 A
5054907 Sklar et al. Oct 1991 A
5098426 Sklar et al. Mar 1992 A
5112328 Taboada et al. May 1992 A
5139022 Lempert Aug 1992 A
5246435 Bille et al. Sep 1993 A
5255025 Volk Oct 1993 A
5286964 Fountain Feb 1994 A
5321501 Swanson et al. Jun 1994 A
5336215 Hsueh et al. Aug 1994 A
5391165 Fountain et al. Feb 1995 A
5439462 Bille et al. Aug 1995 A
5493109 Wei et al. Feb 1996 A
5549632 Lai Aug 1996 A
5656186 Mouron et al. Aug 1997 A
5738676 Hammer et al. Apr 1998 A
5779696 Berry et al. Jul 1998 A
5795295 Hellmuth et al. Aug 1998 A
5936706 Takagi Aug 1999 A
5954648 Van Der Brug Sep 1999 A
5954711 Ozaki et al. Sep 1999 A
5994690 Kulkarni et al. Nov 1999 A
6004314 Wei et al. Dec 1999 A
6095648 Birngruber et al. Aug 2000 A
6099522 Knopp et al. Aug 2000 A
6137585 Hitzenberger et al. Oct 2000 A
6254595 Juhasz et al. Jul 2001 B1
6288784 Hitzenberger et al. Sep 2001 B1
6314311 Williams et al. Nov 2001 B1
6317616 Glossop Nov 2001 B1
6337925 Cohen et al. Jan 2002 B1
6377349 Fercher Apr 2002 B1
6379005 Williams et al. Apr 2002 B1
6451009 Dasilva et al. Sep 2002 B1
6454761 Freedman Sep 2002 B1
6497701 Shimmick et al. Dec 2002 B2
6529758 Shahidi Mar 2003 B2
6579282 Bille et al. Jun 2003 B2
6623476 Juhasz et al. Sep 2003 B2
6687010 Horii et al. Feb 2004 B1
6730074 Bille et al. May 2004 B2
6741359 Wei et al. May 2004 B2
6751033 Goldstein et al. Jun 2004 B2
6755819 Waelti Jun 2004 B1
6763259 Hauger et al. Jul 2004 B1
6769769 Podoleanu et al. Aug 2004 B2
6775007 Izatt et al. Aug 2004 B2
6863667 Webb et al. Mar 2005 B2
6887232 Bille May 2005 B2
6899707 Scholler et al. May 2005 B2
6932807 Tomita et al. Aug 2005 B1
6991629 Juhasz et al. Jan 2006 B1
7006232 Rollins et al. Feb 2006 B2
7018376 Webb et al. Mar 2006 B2
7027233 Goldstein et al. Apr 2006 B2
7061622 Rollins et al. Jun 2006 B2
7072045 Chen et al. Jul 2006 B2
7072047 Westphal et al. Jul 2006 B2
7079254 Kane et al. Jul 2006 B2
7102756 Izatt et al. Sep 2006 B2
7113818 Podoleanu et al. Sep 2006 B2
7126693 Everett et al. Oct 2006 B2
7130054 Ostrovsky et al. Oct 2006 B2
7133137 Shimmick Nov 2006 B2
7139077 Podoleanu et al. Nov 2006 B2
7145661 Hitzenberger Dec 2006 B2
7148970 de Boer Dec 2006 B2
7184148 Alphonse Feb 2007 B2
7207983 Hahn et al. Apr 2007 B2
7248371 Chan et al. Jul 2007 B2
7268885 Chan et al. Sep 2007 B2
7280221 Wei Oct 2007 B2
7307733 Chan et al. Dec 2007 B2
7310150 Guillermo et al. Dec 2007 B2
7312876 Chan et al. Dec 2007 B2
7319566 Prince et al. Jan 2008 B2
7329002 Nakanishi Feb 2008 B2
7330270 O'Hara et al. Feb 2008 B2
7330273 Podoleanu et al. Feb 2008 B2
7335223 Obrebski Feb 2008 B2
7336366 Choma et al. Feb 2008 B2
7342659 Horn et al. Mar 2008 B2
7347548 Huang et al. Mar 2008 B2
7352444 Seams et al. Apr 2008 B1
7355716 de Boer et al. Apr 2008 B2
7364296 Miller et al. Apr 2008 B2
7365856 Everett et al. Apr 2008 B2
7365859 Yun et al. Apr 2008 B2
7370966 Fukuma et al. May 2008 B2
7371230 Webb et al. May 2008 B2
7372578 Akiba et al. May 2008 B2
7388672 Zhou et al. Jun 2008 B2
7390089 Loesel et al. Jun 2008 B2
7400410 Baker et al. Jul 2008 B2
7402159 Loesel et al. Jul 2008 B2
7426037 Ostrovsky et al. Sep 2008 B2
7433046 Everett et al. Oct 2008 B2
7452077 Meyer et al. Nov 2008 B2
7452080 Wiltberger et al. Nov 2008 B2
7461658 Jones et al. Dec 2008 B2
7466423 Podoleanu et al. Dec 2008 B2
7470025 Iwanaga Dec 2008 B2
7477764 Haisch Jan 2009 B2
7480058 Zhao et al. Jan 2009 B2
7480059 Zhou et al. Jan 2009 B2
7488070 Hauger et al. Feb 2009 B2
7488930 Ajgaonkar et al. Feb 2009 B2
7492466 Chan et al. Feb 2009 B2
7503916 Shimmick Mar 2009 B2
7508525 Zhou et al. Mar 2009 B2
7535577 Podoleanu et al. May 2009 B2
7537591 Feige et al. May 2009 B2
7557928 Ueno Jul 2009 B2
7575322 Somani Aug 2009 B2
7593559 Toth et al. Sep 2009 B2
7602500 Izatt et al. Oct 2009 B2
7604351 Fukuma et al. Oct 2009 B2
7614744 Abe Nov 2009 B2
7630083 de Boer et al. Dec 2009 B2
7631970 Wei Dec 2009 B2
7633627 Choma et al. Dec 2009 B2
7643152 de Boer et al. Jan 2010 B2
7797119 De Boer et al. Sep 2010 B2
7813644 Chen et al. Oct 2010 B2
7898712 Adams et al. Mar 2011 B2
8262646 Frey et al. Sep 2012 B2
8382745 Naranjo-Tackman et al. Feb 2013 B2
8394084 Palankar et al. Mar 2013 B2
20010022648 Lai Sep 2001 A1
20020013574 Elbrecht et al. Jan 2002 A1
20020082466 Han Jun 2002 A1
20020097374 Payne et al. Jul 2002 A1
20020133145 Gerlach et al. Sep 2002 A1
20020198516 Knopp Dec 2002 A1
20030090674 Zeylikovich et al. May 2003 A1
20030206272 Cornsweet et al. Nov 2003 A1
20040039378 Lin Feb 2004 A1
20040059321 Knopp et al. Mar 2004 A1
20040151466 Crossman-Bosworth et al. Aug 2004 A1
20040243233 Phillips Dec 2004 A1
20050010109 Faul Jan 2005 A1
20050015120 Seibel et al. Jan 2005 A1
20050021011 LaHaye Jan 2005 A1
20050173817 Fauver et al. Aug 2005 A1
20050192562 Loesel et al. Sep 2005 A1
20050201633 Moon et al. Sep 2005 A1
20050203492 Nguyen et al. Sep 2005 A1
20050215986 Chernyak et al. Sep 2005 A1
20050284774 Mordaunt Dec 2005 A1
20050286019 Wiltberger et al. Dec 2005 A1
20050288745 Andersen et al. Dec 2005 A1
20060020172 Luerssen et al. Jan 2006 A1
20060077346 Matsumoto Apr 2006 A1
20060100613 McArdle et al. May 2006 A1
20060179992 Kermani Aug 2006 A1
20060187462 Srinivasan et al. Aug 2006 A1
20060195076 Blumenkranz et al. Aug 2006 A1
20060206102 Shimmick Sep 2006 A1
20070013867 Ichikawa Jan 2007 A1
20070121069 Andersen et al. May 2007 A1
20070126985 Wiltberger et al. Jun 2007 A1
20070129709 Andersen et al. Jun 2007 A1
20070129775 Mordaunt et al. Jun 2007 A1
20070147730 Wiltberger et al. Jun 2007 A1
20070173791 Raksi Jul 2007 A1
20070173794 Frey et al. Jul 2007 A1
20070173795 Frey et al. Jul 2007 A1
20070185475 Frey et al. Aug 2007 A1
20070189664 Andersen et al. Aug 2007 A1
20070216909 Everett et al. Sep 2007 A1
20070219541 Kurtz Sep 2007 A1
20070230520 Mordaunt et al. Oct 2007 A1
20070282313 Huang et al. Dec 2007 A1
20070291277 Everett et al. Dec 2007 A1
20070299429 Amano Dec 2007 A1
20080033406 Andersen et al. Feb 2008 A1
20080049188 Wiltberger et al. Feb 2008 A1
20080055543 Meyer et al. Mar 2008 A1
20080056610 Kanda Mar 2008 A1
20080071254 Lummis et al. Mar 2008 A1
20080088795 Goldstein et al. Apr 2008 A1
20080100612 Dastmalchi et al. May 2008 A1
20080281303 Culbertson et al. Nov 2008 A1
20080281413 Culbertson et al. Nov 2008 A1
20080319427 Palanker Dec 2008 A1
20090012507 Culbertson et al. Jan 2009 A1
20090088734 Mordaunt Apr 2009 A1
20090125005 Chernyak et al. May 2009 A1
20090131921 Kurtz et al. May 2009 A1
20090149742 Kato et al. Jun 2009 A1
20090157062 Hauger et al. Jun 2009 A1
20090161827 Gertner et al. Jun 2009 A1
20090168017 O'Hara et al. Jul 2009 A1
20090177189 Raksi Jul 2009 A1
20090268161 Hart et al. Oct 2009 A1
20100004641 Frey et al. Jan 2010 A1
20100004643 Frey et al. Jan 2010 A1
20100007848 Murata Jan 2010 A1
20100022994 Frey et al. Jan 2010 A1
20100022995 Frey et al. Jan 2010 A1
20100022996 Frey et al. Jan 2010 A1
20100042079 Frey et al. Feb 2010 A1
20100110377 Maloca et al. May 2010 A1
20100137850 Culbertson et al. Jun 2010 A1
20100292678 Frey et al. Nov 2010 A1
20100324543 Kurtz et al. Dec 2010 A1
20110022036 Frey et al. Jan 2011 A1
20110106066 Bor May 2011 A1
20110118609 Goldshleger et al. May 2011 A1
20110166557 Naranjo-Tackman et al. Jul 2011 A1
20110184392 Culbertson et al. Jul 2011 A1
20110184395 Schuele et al. Jul 2011 A1
20110196350 Friedman et al. Aug 2011 A1
20110202046 Angeley et al. Aug 2011 A1
20110319873 Raksi et al. Dec 2011 A1
20120089134 Horvath et al. Apr 2012 A1
20120274903 Sayeram et al. Nov 2012 A1
Foreign Referenced Citations (9)
Number Date Country
1444946 Aug 2004 EP
9808048 Feb 1998 WO
03062802 Jul 2003 WO
2006074469 Jul 2006 WO
2007084694 Jul 2007 WO
2007106326 Sep 2007 WO
2007130411 Nov 2007 WO
2010075571 Jul 2010 WO
2011011788 Jan 2011 WO
Non-Patent Literature Citations (58)
Entry
Arimoto et al., “Imaging Properties of Axicon in a Scanning Optical System,” Nov. 1, 1992, Applied Optics, 31 (31):6652-6657.
Birngruber et al., “In-Vivo Imaging of the Development of Linear and Non-Linear Retinal Laser Effects Using Optical Coherence Tomography in Correlation with Histopathological Findings,” 1995, Proc. SPIE 2391:21-27.
Chinn, S.R., et al., “Optical coherence tomography using a frequency-tunable optical source”, Optics Letters, 22 (5):340-342, Mar. 1997.
European Search Report, European Patent Application No. 10191057.8, mailed Mar. 16, 2011, to be published by the USPTO.
Fercher et al., “Eye-Length Measurement by Interferometry With Partially Coherent Light,” Mar. 1988, Optics Letters, 13(3):186-188.
Fercher et al., “Measurement of Intraocular Distances by Backscattering Spectral Interferometry,” May 15, 1995, Optics Comm. 117:43-48.
Hee, M., et al., “Femotosecond transillumination optical coherence tomography”, Optics Letters, 18(12):950-952, Jun. 1993.
Huber, R., et al., “Three-dimensional and C-mode OCT imaging with a compact, frequency swept laser source at 1300 nm”, Optics Express, 13(26):10523-10538, Dec. 2005.
Izatt et al., “Micron-Resolution Biomedical Imaging With Optical Coherence Tomography,” Oct. 1993, Optics & Photonics News, pp. 14-19.
Kamensky, V., et al., “In situ monitoring of the middle IR laser ablation of a cataract-suffered human lens by optical coherent tomography”, Proc. SPIE, 2930:222-229, 1996.
Kamensky, V., et al., “Monitoring and animation of laser ablation process in cataracted eye lens using coherence tomography”, Proc. SPIE, 2981:94-102, 1997.
Massow, O., et al., “Optical coherence tomography controlled femtosecond laser microsurgery system”, Proceedings of the SPIE—Optical Coherence Tomography and Coherence Techniques III, vol. 6627, pp. 662717(1)-662717(6), Aug. 2007.
Ohmi, M., et al., “In-situ Observation of Tissue Laser Ablation Using Optical Coherence Tomography”, Optical and Quantum Electronics, 37(13-15):1175-1183, Dec. 2005.
Ostaszewski et al., “Risley prism Beam Pointer”, Proc. of SPIE, vol. 6304, 630406-1 thru 630406-10.
PCT International Search Report for International Application No. PCT/US2011/023710 mailed Aug. 24, 2011.
PCT International Search Report for International Application No. PCT/US2011/025332 mailed Sep. 16, 2011.
PCT International Search Report for International Application No. PCT/US2010/056701 mailed Jan. 12, 2011.
PCT International Search Report for International Application No. PCT/US2008/075511 mailed Mar. 12, 2009.
Sarunic, M., et al., “Instantaneous complex conjugate resolved spectral domain and swept-source OCT using 3×3 fiber couplers”, Optics Express, 13(3):957-967, Feb. 2005.
Sarunic, M., et al., “Real-time quadrature projection complex conjugate resolved Fourier domain optical coherence tomography”, Optics Letters, 31(16):2426-2428, Aug. 2006.
Sarunic, M., et al., “Imaging the Ocular Anterior Segment With Real-Time, Full-Range Fourier-Domain Optical Coherence Tomography”, Archives of Ophthalmology, 126(4):537-542, Apr. 2008.
Stern et al., “Femtosecond Optical Ranging of Corneal Incision Depth,” Jan. 1989, Investigative Ophthalmology & Visual Science, 30(1):99-104.
Swanson et al., “In vivo retinal imaging by optical coherence tomography”, Optics Letters, 18(21), 1864-1866, Nov. 1993.
Tao, Y., et al., “High-speed complex conjugate resolved retinal spectral domain optical coherence tomography using sinusoidal phase modulation”, Optics letters, 32(20):2918-2920, Oct. 2007.
Wojtkowski et al., “In Vivo Human Retinal Imaging by Fourier Domain Optical Coherence Tomography,” Jul. 2002, Journal of Biomedical Optics 7(3):457-463, 7 pages.
Yun, S.H., et al., “Wavelength-swept fiber laser with frequency shifted feedback and resonantly swept intra-cavity acoustooptic tunable filter”, IEEE Journal of Selected Topics in Quantum Electronics, 3(4):1087-1096, Aug. 1997.
PCT International Search Report corresponding to PCT Application Serial No. PCT/US2011/051466 dated Apr. 10, 2012.
Partial International Search Report corresponding to PCT Application Serial No. PCT/US2012/035927 dated Aug. 3, 2012.
PCT International Search Report dated Jun. 29, 2012 for International Application No. PCT/US2011/051360, filed Sep. 13, 2011.
PCT International Search Report and Written Opinion dated Feb. 9, 2012 for International Application Serial No. PCT/US2011/040223.
Aslyo-Vogel et al., “Darstellung von LTK-Läsionen durch optische Kurzkohärenztomographie (OCT) und Polarisationsmikroskopie nach Sirius-Rot-Fäbung”, Ophthalmologe, pp. 487-491, 7-97.
Bagayev et al., “Optical coherence tomography for in situ monitoring of laser corneal ablation”, Journal of Biomedical Optics, 7(4), pp. 633-642 (Oct. 2002).
Blaha et al., “The slit lamp and the laser in ophthalmology—a new laser slit lamp”, SPIE Optical Instrumentation for Biomedical Laser Applications, vol. 658, pp. 38-42, 1986.
Boppart, S., et al., “Intraoperative Assessment of Microsurgery with Three-dimensional Optical Coherence Tomography”, Radiology, 208(1):81-86, Jul. 1998.
Davidson, “Analytic Waveguide Solutions and the Coherence Probe Microscope”, Microelectronic Engineering, 13, pp. 523-526, 1991.
Drexler, W., et al., “Measurement of the thickness of fundus layers by partial coherence tomography”, Optical Engineering, 34(3):701-710, Mar. 1995.
Dyer, R, et al., “Optical Fibre Delivery and Tissue Ablation Studies using a Pulsed Hydrogen Fluoride Laser”, Lasers in Medical Science, 7:331-340, 1992.
Fercher et al., “In Vivo Optical Coherence Tomography”, American Journal of Ophthalmology, 116(1), pp. 113-114, 1993.
Fujimoto, J., et al., :Biomedical Imaging using Optical Coherent Tomography, 1994, 67.
Hammer, D., “Ultrashort pulse laser induced bubble creation thresholds in ocular media”, SPIE, 2391:30-40, 1995.
Hauger, C., et al., “High speed low coherence interferometer for optical coherence tomography”, Proceedings of SPIE, 4619:1-9, 2002.
Hee, M., et al., “Optical Coherence tomography of the Human Retina”, Arch Ophthalmol, 113:325-332; Mar. 1995.
Hitzenberger et al., “Interferometric Measurement of Corneal Thickness With Micrometer Precision”, American Journal of Ophthalmology, 118:468-476, Oct. 1994.
Hitzenberger, C., et al., “Retinal layers located with a precision of 5 μm by partial coherence interferometry”, SPIE, 2393:176-181, 1995.
Itoh et al., “Absolute measurements of 3-D shape using white-light interferometer”, SPIE Interferometry: Techniques and Analysis, 1755:24-28, 1992.
Izatt et al., “Ophthalmic Diagnostics using Optical Coherence Tomography”, SPIE Ophthalmic Technologies, 1877:136-144, 1993.
Izatt, J., et al., “Micrometer-Scale Resolution Imaging of the Anterior Eye In vivo With Optical Coherence Tomography”, Arch Ophthalmol, 112:1584-1589, Dec. 1994.
Jean, B., et al., “Topography assisted photoablation”, SPIE, vol. 3591:202-208, 1999.
Kamensky, V., et al., “In Situ Monitoring of Laser Modification Process in Human Cataractous Lens and Porcine Cornea Using Coherence Tomography”, Journal of biomedical Optics, 4(1), 137-143, Jan. 1999.
Lee et al., “Profilometry with a coherence scanning microscope”, Applied Optics, 29(26), 3784-3788, Sep. 10, 1990.
Lubatschowski, “The German Ministry of Research and education funded this OCT guided fs laser surgery in Sep. 2005”, http://www.laser-zentrum-hannover.de/download/pdf/taetigkeitsbericht2005.pdf.
Massow, O., et al., “Femotosecond laser microsurgery system controlled by OCT”, Laser Zentrum Hannover e.V., The German Ministry of education and research,19 slides, 2007.
Puliafito, Carmen, “Final technical Report: Air Force Grant #F49620-93-I-03337(1)” dated Feb. 12, 1997, 9 pages.
Ren, Q., et al., “Axicon: A New Laser Beam Delivery System for Corneal Surgery”, IEEE Journal of Quantum Electronics, 26(12):2305-2308, Dec. 1990.
Ren, Q., et al., “Cataract Surgery with a Mid-Infrared Endo-laser System”, SPIE Ophthalmic Technologies II, 1644:188-192, 1992.
Thompson, K., et al., “Therapeutic and Diagnostic Application of Lasers in Ophthalmology”, Proceedings of the IEEE, 80(6):838-860, Jun. 1992.
Thrane, L, et al., “Calculation of the maximum obtainable probing depth of optical coherence tomography in tissue”, Proceedings of SPIE, 3915:2-11, 2000.
Wisweh, H., et al., “OCT controlled vocal fold femtosecond laser microsurgery”, Laser Zentrum Hannover e.V., The German Ministry of education and research, Grants: 13N8710 and 13N8712; 23 slides, 2008.
Related Publications (1)
Number Date Country
20120296319 A1 Nov 2012 US