The present invention relates to ophthalmological devices and methods for the treatment of glaucoma, ocular hypertension (OHT), and other diseases.
In a trabeculoplasty procedure, a radiation source irradiates the trabecular meshwork in an eye of a patient with one or more treatment beams, thus lowering the intraocular pressure in the eye.
Geffen, Noa, et al., “Transscleral selective laser trabeculoplasty without a gonioscopy lens,” Journal of glaucoma 26.3 (2017): 201-207 describes a study to investigate results of selective laser trabeculoplasty (SLT) performed directly on the sclera without a gonioscopy lens.
US Patent Application Publication 2015/0366706 to Belkin, whose disclosure is incorporated herein by reference, describes an apparatus including a probe and a processor. The probe is positioned adjacent to an eye of a patient and is configured to irradiate a trabecular meshwork of the eye with one or more optical beams. The processor is configured to select one or more target regions of the trabecular meshwork, and to control the probe to irradiate the selected target regions with the optical beams.
There is provided, in accordance with some embodiments of the present invention, a system including a radiation source and a controller. The controller is configured to display a live sequence of images of an eye of a patient, and, while displaying the sequence of images, cause the radiation source to irradiate the eye with one or more aiming beams, which are visible in the images. The controller is further configured to, subsequently to causing the radiation source to irradiate the eye with the aiming beams, receive a confirmation input from a user, and, in response to receiving the confirmation input, treat the eye by causing the radiation source to irradiate respective target regions of the eye with a plurality of treatment beams.
In some embodiments, the system further includes:
a focusing lens; and
one or more beam-directing elements,
and the controller is configured to cause the radiation source to irradiate the eye with the treatment beams by firing the treatment beams at the beam-directing elements through the focusing lens, such that the beams are focused by the focusing lens prior to being directed, by the beam-directing elements, toward the respective target regions.
In some embodiments, the aiming beams impinge on at least part of each of the target regions.
In some embodiments, the controller is further configured to superimpose, on each of the images, a marker passing through each of the target regions.
In some embodiments, the marker is elliptical.
In some embodiments, at least part of each of the target regions is located within 1 mm of a limbus of the eye.
In some embodiments, the controller is further configured to:
superimpose a marker on each of the images, and
prior to treating the eye, by processing the images, verify respective positions of the aiming beams with respect to the marker,
and the controller is configured to treat the eye in response to verifying the positions of the aiming beams.
In some embodiments, the controller is configured to verify the positions of the aiming beams by verifying that the aiming beams overlap the marker.
In some embodiments, the controller is configured to verify the positions of the aiming beams by verifying that the aiming beams lie outside the marker.
In some embodiments, the controller is configured to treat the eye such that respective edges of the treatment beams impinge on respective portions of the eye over which the marker is superimposed.
In some embodiments, the marker is elliptical.
In some embodiments, the controller is further configured to:
prior to displaying the live images, display a still image of the eye,
identify an elliptical portion of the eye in the still image, based on input from the user, and
in response to identifying the elliptical portion of the eye, superimpose an elliptical marker over the elliptical portion of the eye in each of the images.
In some embodiments, the controller is configured to superimpose the elliptical marker over the elliptical portion of the eye by:
subsequently to identifying the elliptical portion of the eye, identifying an offset from a center of a limbus of the eye to a center of the elliptical portion in the still image, and
for each image of the images:
In some embodiments, the controller is configured to identify the elliptical portion of the eye by:
displaying, over the still image, (i) the elliptical marker, and (ii) a rectangle circumscribing the elliptical marker, and
subsequently to displaying the elliptical marker and the rectangle, in response to the user adjusting the rectangle, adjusting the elliptical marker such that the elliptical marker remains circumscribed by the rectangle, until the elliptical marker is superimposed over the portion of the eye.
In some embodiments, the controller is further configured to identify a limbus of the eye in the still image, and the controller is configured to display the elliptical marker over the limbus.
In some embodiments, the system further includes a camera configured to:
acquire the images, and
acquire a still image of the eye, prior to acquiring the images,
and the controller is further configured to:
In some embodiments, the system further includes one or more beam-directing elements,
the controller is configured to treat the eye by aiming the beam-directing elements at the target regions in sequence and firing the treatment beams at the beam-directing elements, and
the controller is further configured to inhibit the beam-directing elements from being aimed at the static region even while none of the treatment beams is being fired.
In some embodiments, the controller is configured to identify the static region by:
receiving, from the user, a limbus-locating input indicating a location of the limbus in the still image, and
identifying the static region based on the location of the limbus.
In some embodiments,
the images are first images and the aiming beams are first aiming beams,
the system further includes a camera configured to acquire multiple second images of the eye while treating the eye, and
the controller is configured to treat the eye by iteratively:
In some embodiments, the controller is configured to verify the position by verifying that a distance between the second aiming beam and a respective one of the target regions is less than a predefined threshold.
In some embodiments, the controller is configured to fire the respective one of the treatment beams at the respective one of the target regions.
In some embodiments, the system further includes an illumination source, and the controller is further configured to cause the illumination source to intermittently flash visible light at the eye such that the light illuminates the eye at least during respective acquisitions of the second images.
In some embodiments, a peak average intensity of the light over a duration of each of the flashes is between 0.003 and 3 mW/cm2.
In some embodiments, the controller is configured to cause the illumination source to flash the light at a frequency of at least 60 Hz.
In some embodiments, the frequency is at least 100 Hz.
In some embodiments, the system further includes an illumination source, and the controller is further configured to cause the illumination source to illuminate the eye with near-infrared light at least during respective acquisitions of the second images.
In some embodiments, the controller is further configured to cause the illumination source to intermittently flash visible light at the eye while treating the eye.
In some embodiments, the system further includes an optical unit including the radiation source and a plurality of beam emitters,
and the controller is further configured to, prior to causing the radiation source to irradiate the eye with the aiming beams, cause the beam emitters to shine a plurality of range-finding beams on the eye, the range-finding beams being shaped to define different respective portions of a predefined composite pattern such that the predefined composite pattern is formed on the eye only when the optical unit is at a predefined distance from the eye.
In some embodiments, the range-finding beams are shaped to define two perpendicular shapes, and the predefined composite pattern includes a cross.
In some embodiments, the system further includes an optical unit including the radiation source, and the controller is configured to cause the radiation source to irradiate the target regions while the optical unit is directed obliquely upward toward the eye and the eye gazes obliquely downward toward the optical unit.
In some embodiments, the system further includes a wedge, and the optical unit is directed obliquely upward toward the eye by virtue of being mounted on the wedge.
There is further provided, in accordance with some embodiments of the present invention, a system, including a wedge, an optical unit mounted on the wedge such that the optical unit is directed obliquely upward, the optical unit including a radiation source, and a controller. The controller is configured to treat an eye of a patient by causing the radiation source to irradiate respective target regions of the eye with a plurality of treatment beams while the eye gazes obliquely downward toward the optical unit.
There is further provided, in accordance with some embodiments of the present invention, a method including displaying a live sequence of images of an eye of a patient. The method further includes, while displaying the sequence of images, irradiating the eye with one or more aiming beams, which are visible in the images. The method further includes, subsequently to irradiating the eye with the aiming beams, receiving a confirmation input from a user, and in response to receiving the confirmation input, treating the eye by irradiating respective target regions of the eye with a plurality of treatment beams.
The present invention will be more fully understood from the following detailed description of embodiments thereof, taken together with the drawings, in which:
Embodiments of the present invention provide an automated trabeculoplasty device configured to perform a trabeculoplasty procedure on an eye safely and efficiently. The trabeculoplasty device comprises a controller and an optical unit, which comprises a radiation source, a camera, and beam-directing elements. As described in detail below, the controller is configured to control the radiation source and the beam-directing elements in response to feedback from the camera, such that the beam-directing elements direct beams of radiation, which are emitted by the radiation source, toward the appropriate locations on the eye. The emitted beams of radiation include both treatment beams, which irradiate the trabecular meshwork of the eye, and aiming beams, which are used to help aim the treatment beams.
Typically, prior to the procedure, the controller displays a live video of the eye in which two ellipses are superimposed over the eye: an inner ellipse, which marks the limbus of the eye, and an outer ellipse, displaced from the inner ellipse by a small distance, which passes through or near each of the target regions that are to be irradiated by the treatment beams. The controller further simulates the procedure by sweeping an aiming beam over the outer ellipse, typically such that the aiming beam impinges on at least part of each target region. Advantageously, this simulation may help the physician visualize the path along the eye that is to be targeted by the treatment beams, i.e., the path along which the target regions lie. After the physician confirms the targeted path along the eye, the controller causes the radiation source to fire the treatment beams at the target regions.
It is noted that since each beam of radiation generally impinges on the eye with a non-infinitesimal spot size, the present application generally describes each beam as impinging on a “region” of the eye, whose area is a function of the spot size, rather than impinging at a “point” on the eye. Thus, for example, the present application refers to “target regions,” rather than “target points.” Nonetheless, in the context of the present application, including the claims, references to calculating the location of a target region may refer to implicitly calculating the location of the region by calculating the location of a single point within the region, such as the point at the center or edge of the region at which the center or edge (respectively) of the beam is to be aimed. (Even if, subsequently, the center or edge of the beam deviates slightly from the calculated point, the present application, including the claims, may consider the beam to have impinged on the calculated target region.)
Typically, prior to simulating the procedure as described above, the controller acquires a still image of the eye, and identifies the limbus in the still image. The controller then superimposes the aforementioned inner ellipse over the limbus. Subsequently, the controller allows the physician to modify the position and/or shape of the inner ellipse, such that the inner ellipse marks the limbus per the physician's definition thereof. (Since the limbus is generally not well defined, the location of the limbus per the physician may differ slightly from the location automatically identified by the controller.) For example, the controller may circumscribe the inner ellipse by a rectangle, and then allow the physician to adjust the ellipse by dragging the sides or corners of the circumscribing rectangle.
As the present inventors have observed, the trabecular meshwork may be irradiated most effectively when the treatment beams impinge on the eye at or near the limbus, which may be identified by the user as described above or automatically identified by the controller. Hence, in some embodiments of the present invention, the controller causes the radiation source to target the limbus or a portion of the eye near the limbus. For example, at least part of each target region may be located within 1 mm (e.g., within 400 microns) of the limbus. As a specific example of the above, the center of each target region may be located within 1 mm (e.g., within 400 microns) of the limbus, such that the center of each treatment beam impinges on the eye within 1 mm (e.g., within 400 microns) of the limbus.
During both the simulated treatment and the actual treatment, the camera acquires images of the eye at a relatively high frequency (e.g., at a frequency greater than 40 Hz or 50 Hz), and the controller tracks motion of the eye by identifying the center of the limbus in each of the acquired images. In response to identifying the center of the limbus, during the simulated treatment, the controller may move the inner and outer ellipses such that the inner ellipse remains positioned over the limbus as defined by the physician, and the outer ellipse remains at a constant distance from the inner ellipse, even as the eye moves. Similarly, during the procedure, the controller may calculate the center or edge of each target region by adding the appropriate (x, y) offset to the identified limbus center. Advantageously, due to this feedback process, the safety and efficacy of the procedure is greatly improved.
Moreover, as an additional safety measure, the controller may define a region, referred to herein as a “forbidden zone.” in the aforementioned still image. The forbidden zone encompasses the pupil of the eye, along with, typically, a portion of the eye surrounding the pupil. The forbidden zone is static, in that it is defined in terms of the field of view (FOV) of the camera, and is not adjusted even in response to detected motion of the eye. The controller may then prevent any of the treatment beams from striking the forbidden zone. Moreover, the controller may prevent the beam-directing elements from being aimed at the forbidden zone, even while the radiation source is inactive. Thus, the retina of the eye is protected from any potential (though unlikely) stray beams.
In some embodiments, the trabeculoplasty device further comprises a visible light source, and the controller is configured to cause the visible light source to flash visible light at the eye such that the visible light is on at least while each image is acquired. Advantageously, the flash of light reduces the time needed to acquire the image, such that the position of the target region calculated responsively to the image does not move significantly before the aiming beam or treatment beam is fired at the target region. Moreover, the flash may constrict the pupil of the eye, thus further protecting the retina from any potential stray beams.
Typically, the light is flashed at a sufficiently high frequency, and/or each pulse of light has a sufficiently long duration, such that the flashing is unnoticeable to the patient. Nonetheless, the total energy of the flashed light is low enough such that the light does not damage the retina.
Alternatively, to reduce the time required for image acquisition without discomforting the patient, the eye may be illuminated with near-infrared light. In addition, optionally, visible light may be flashed at the eye, such that the visible light is on while the images are acquired and/or between image acquisitions.
Embodiments of the present invention further provide a technique to facilitate positioning the trabeculoplasty device at the correct distance (or “range”) from the eye. Conventionally, this type of positioning is performed by aiming two circular range-finding beams at the eye from the device, and moving the device toward or away from the eye until the two beams overlap. However, as the present inventors have observed, for several reasons, it may be difficult to use this technique for positioning the trabeculoplasty device; for example, the sclera is covered by a conjunctiva that may distort and reflect the range-finding beams, thus making it difficult to discern that the beams overlap. Hence, in embodiments of the present invention, the range-finding beams are given different respective shapes, such that the beams form a particular pattern only when the trabeculoplasty device is positioned at the correct distance from the eye. For example, the range-finding beams may be shaped as perpendicular ellipses, such that the range-finding beams form a cross over the eye only at the correct range.
In some embodiments, to reduce obstruction of the sclera by the upper eyelid, the optical unit of the trabeculoplasty device is mounted on a wedge, such that the camera and radiation source are directed obliquely upward. The patient's gaze is then directed, obliquely downward, toward the optical unit, such that the upper portion of the patient's sclera is exposed.
Although the present description relates mainly to a trabeculoplasty procedure, the techniques described herein may also be applied to automatic photocoagulation procedures, iridotomy procedures, capsulectomy procedures, lens removals, or any other relevant ophthalmological procedures. The target of the radiation may include the trabecular meshwork and/or any other suitable portion of the eye, such as the endothelial stem cells or Schlemm's canal cells of the eye. Embodiments of the present invention may be used to treat glaucoma, ocular hypertension (OHT), and other diseases.
Reference is initially made to
Trabeculoplasty device 21 comprises an optical unit 30. Optical unit 30 comprises a radiation source 48, which is configured to irradiate an eye 25 of a patient 22 with both aiming beams and treatment beams as described herein. Optical unit 30 further comprises one or more beam-directing elements, comprising, for example, one or more galvo mirrors 50 (which may be referred to collectively as a “galvo scanner”) and/or a beam combiner 56. Before the firing of each beam 52 from radiation source 48, or while the beam is being emitted, a controller 44 aims the beam-directing elements at the desired target region on eye 25 such that the beam is directed, by the beam-directing elements, toward the target region. For example, the beam may be deflected by galvo mirrors 50 toward beam combiner 56, which may then deflect the beam through an aperture 58 at the front of the optical unit such that the beam impinges on the target region. Each beam emitted by the radiation source may have an elliptical (e.g., circular) shape, a square shape, or any other suitable shape.
Typically, the radiation source comprises two lasers: one for firing aiming beams as described herein, and another for firing treatment beams as described herein. As a purely illustrative example, the treatment laser may comprise an Ekspla™ NL204-0.5K-SH laser (modified, for example, to include an attenuator, energy meter, and mechanical shutter), while the aiming laser may comprise a Laser Components™ FP-D-635-1DI-C-F laser. Typically, both the aiming beams and the treatment beams comprise visible light.
Alternatively or additionally to a laser, the radiation source may comprise any other suitable emitter configured to emit radiation belonging to any suitable portion of the electromagnetic spectrum, including, for example, microwave radiation, infrared radiation, X-ray radiation, gamma radiation, or ultraviolet radiation.
In some embodiments, each beam 52 passes through a beam expander (not shown), which expands and then re-collimates the beam, prior to reaching the galvo scanner. In such embodiments, optical unit 30 typically comprises an F-theta lens 51, configured to focus each beam subsequently to the direction of the beam by the galvo scanner.
In other embodiments, a focusing lens is disposed between the radiation source and the galvo scanner; for example, the aforementioned beam expander may comprise a focusing lens instead of a collimating lens, or the optical unit may comprise a focusing lens in addition to the beam expander. In such embodiments, each of the beams is focused by the focusing lens prior to being directed by the beam-directing elements, such that F-theta lens 51 may not be needed.
Optical unit 30 further comprises a camera 54. Before and during the procedure, camera 54 acquires multiple images of the patient's eye, typically at a relatively high frequency. Controller 44 processes these images and, in response thereto, controls radiation source 48 and the beam-directing elements, as described below with reference to
Typically, optical unit 30 further comprises an illumination source 60 comprising, for example, one or more light emitting diodes (LEDs), such as a ring of LEDs surrounding aperture 58. In such embodiments, controller 44 may cause illumination source 60 to intermittently flash light at the eye, as further described below with reference to
Optical unit 30 is mounted onto an XYZ stage 32, which is controlled by a control mechanism 36, such as a joystick. Using control mechanism 36, a user of system 20, such as an ophthalmological surgeon or another physician, may position the optical unit at the appropriate position prior to treating the eye of the patient. In some embodiments, XYZ stage 32 comprises locking elements configured to inhibit motion of the stage following the positioning of the stage.
In some embodiments, XYZ stage 32 comprises one or more motors, and control mechanism 36 is connected to interface circuitry 46. As the user manipulates the control mechanism, interface circuitry 46 translates this activity into appropriate electronic signals, and outputs these signals to controller 44. In response to the signals, the controller controls the motors of the XYZ stage. In other embodiments, XYZ stage 32 is controlled manually by manipulating the control mechanism.
Typically, before the radiation source fires any beams at the eye, the user, using control mechanism 36, positions the optical unit at a predefined distance D from the eye. To facilitate this positioning, the optical unit may comprise a plurality of beam emitters 62 (comprising, for example, respective laser diodes), which are configured to shine a plurality of range-finding beams 64 on the eye, e.g., such that the angle between the beams is between 30 and 100 degrees. As further described below with reference to
System 20 further comprises a headrest 24, which is mounted onto a horizontal surface 38, such as a tray or table top. Headrest 24 comprises a forehead rest 26 and a chinrest 28. During the trabeculoplasty procedure, patient 22 presses his forehead against forehead rest 26 while resting his chin on chinrest 28.
In some embodiments, headrest 24 further comprises an immobilization strap 27, configured to secure the patient's head from behind and thus keep the patient's head pressed against the headrest. Immobilization strap 27 may comprise a single segment extending from the headrest at one side of the head and configured to fasten to the headrest at the other side of the head, or two segments that extend from the headrest at opposite sides of the head and are configured to fasten to one another behind the head. Optionally, the immobilization strap may comprise a sensor configured to detect when the immobilization strap is properly fastened. For example, fastening the immobilization strap may cause an electrical circuit to be closed, and the sensor may then detect the flow of electric current through the circuit and generate an output (e.g., by lighting an LED) responsively thereto.
In some embodiments, headrest 24 further comprises one or more sensors, which may be disposed, for example, on the forehead rest or chinrest. Each of these sensors may be configured to generate an output indicating whether the patient's head is resting on the headrest as required. Examples of suitable sensors include capacitive, resistive, and piezoelectric sensors. Alternatively or additionally, the headrest may comprise one or more switches or force-sensitive resistors, such as the Sparkfun™ 9375.
In some embodiments, to contain any radiation reflected by the eye, a physical block is placed around the eye. For example, a hood may be placed over the chinrest and/or over the patient's head. Alternatively or additionally, a hood may be coupled to the face of device 21.
In some embodiments, device 21 further comprises a base unit 34, which is mounted onto surface 38, and XYZ stage 32 is mounted onto base unit 34. In such embodiments, controller 44 and interface circuitry 46 may be disposed within the base unit. In other embodiments, the XYZ stage is mounted directly onto surface 38.
Typically, as shown in
In some embodiments, as shown in
Alternatively or additionally to using wedge 40, the oblique orientation of the optical path may be achieved by tilting the patient's head backward. For example, forehead rest 26 and/or chinrest 28 may comprise an adjustable-length strap, and the patient's head may be tilted backward by adjusting the length of the strap. (For example, the forehead strap may be constricted.) To facilitate this adjustment, the adjustable-length strap may comprise a worm-type drive, a hook-and-loop fastener, snaps, locking pins, knots, and/or any other suitable mechanism.
In other embodiments, the patient's head is tilted slightly forward, e.g., by angling headrest 24 (or at least chinrest 28) toward the optical unit, such that the patient's head rests more securely on the headrest.
System 20 further comprises a monitor 42, configured to display the images of the eye acquired by the camera, as described in detail below with reference to
In some embodiments, monitor 42 is connected directly to controller 44 over a wired or wireless communication interface. In other embodiments, monitor 42 is connected to controller 44 via an external processor, such as a processor belonging to a standard desktop computer.
It is emphasized that the configuration shown in
In some embodiments, at least some of the functionality of controller 44, as described herein, is implemented in hardware, e.g., using one or more Application-Specific Integrated Circuits (ASICs) or Field-Programmable Gate Arrays (FPGAs). Alternatively or additionally, controller 44 may perform at least some of the functionality described herein by executing software and/or firmware code. For example, controller 44 may comprise a central processing unit (CPU) and random access memory (RAM). Program code, including software programs, and/or data may be loaded into the RAM for execution and processing by the CPU. The program code and/or data may be downloaded to the controller in electronic form, over a network, for example. Alternatively or additionally, the program code and/or data may be provided and/or stored on non-transitory tangible media, such as magnetic, optical, or electronic memory. Such program code and/or data, when provided to the controller, produce a machine or special-purpose computer, configured to perform the tasks described herein.
In some embodiments, the controller comprises a system on module (SOM), such as the Varisite™ DART-MX8M.
In some embodiments, controller 44 is disposed externally to device 21. Alternatively or additionally, the controller may cooperatively perform at least some of the functionality described herein with another, external processor.
Reference is now made to
By way of introduction, the procedure illustrated in
In step A, the user positions optical unit 30 (
For example, as shown in
In other embodiments, only a single range-finding beam is emitted, and a computer-generated pattern is superimposed over the images of the eye. When the optical unit is at the correct distance, the range-finding beam and the computer-generated pattern overlap or form composite pattern 66.
In response to observing pattern 66, the user indicates to the controller that the optical unit is at the correct distance from the eye. For example, the user may click an appropriate button on GUI 68. In response to this input, the controller proceeds to step B of the pre-treatment procedure.
In step B, the controller displays a still image 71 of the eye. Subsequently, based on input from the user, the controller identifies an elliptical (e.g., circular or almost circular) portion of the eye, such as the limbus 69 of the eye. For example, the controller may identify the portion of the eye in response to the user superimposing an elliptical marker 78 over the portion of the eye. The position of marker 78 may then be used to compute the respective positions of the treatment-beam target regions, as further described below.
For example, the controller may display, over the still image, both marker 78 and a rectangle 80 circumscribing (or “bounding”) the marker. Subsequently, the user may adjust rectangle 80, e.g., by dragging the sides or corners of the rectangle using a mouse or touch screen. (In some embodiments, the system allows the user to toggle between a rough and fine adjustment of the rectangle.) In response to the user adjusting the rectangle, the controller may adjust marker 78 such that the marker remains circumscribed by the rectangle, until the marker is superimposed over the limbus as defined by the user (or over another portion of the eye). Subsequently, the user may indicate to the controller (e.g., via GUI 68) that the marker is superimposed over the limbus as defined by the user.
In some embodiments, the controller superimposes two horizontal lines tangent to the top and bottom extremities of marker 78, respectively, and two vertical lines tangent to the left and right extremities of marker 78, respectively, without necessarily causing the lines to intersect each other and thus define a rectangle. In such embodiments, the user may adjust marker 78 by dragging the lines.
Typically, prior to allowing the user to adjust marker 78, the controller, using an edge-detection algorithm or any other suitable image-processing technique, identifies the limbus in the still image and then displays marker 78 over the limbus. (It is noted that the controller may approximate the form of the limbus by any suitable shape, such as an elliptical shape aligned with the vertical and horizontal axes or rotated by any suitable angle.) Advantageously, by initializing the placement of marker 78 in this manner, the time required to adjust the marker is reduced. (Since the limbus is generally not a well-defined feature, the location of the limbus as identified by the user typically differs slightly from the location of the limbus as identified initially by the controller; hence, as presently described, the user is allowed to adjust the marker.)
Alternatively or additionally to adjusting the rectangle, the user may directly adjust marker 78 by inputting relevant parameters. For example, for an elliptical (e.g., circular) marker, the user may input the coordinates of the center of the marker and one or two diameters of the marker. Alternatively or additionally, the user may adjust the marker by adjusting an input to the limbus-identification algorithm (such as a threshold for edge detection) that is executed by the controller. As yet another option, the user may manipulate marker 78 directly.
In alternative embodiments, marker 78 is not shown at all. In such embodiments, the user may indicate the position of the limbus by dragging the rectangle or lines that would bound the marker if the marker were shown. As yet another alternative, for greater precision, a non-elliptical marker having another shape that more precisely corresponds to the shape of limbus 69 may be used instead of elliptical marker 78.
Typically, prior to the execution of the pre-treatment procedure illustrated in
For example, the user may specify an elliptical path of target regions adjacent to the limbus, by specifying the number of target regions and the distance from the limbus (or from the center thereof) at which the center or edge of each of the target regions is to be located. Alternatively, the user may specify one or more arced paths, by specifying, in addition to the aforementioned parameters, (i) an angular span of each arc, and (ii) the location of each arc. (For example, the user may specify a 180 degree arc around the bottom or top half of the limbus, or respective 90 degree arcs at the top and bottom.) Given this input, and given the location of the limbus as indicated by the user, the controller calculates the respective positions of the target regions, typically relative to the center of the limbus as identified by the controller. (In some embodiments, the controller calculates the position of the ellipse or arc specified by the user, but does not calculate the specific positions of the target regions on the ellipse or arc until after the performance of step C, described below.)
As a purely illustrative example, the user may specify that the center or edge of each target region is to be at a distance of d1 from the limbus as marked by the user, at a different respective angle θi relative to the center of the limbus. The user may then, during step B, adjust marker 78 such that the center of the marker is at (x0+Δx, y0+Δy), wherein (x0, y0) is the center of the limbus as identified by the controller. In such a case, assuming that marker 78 is a circle with radius R, the controller may compute the offset from the limbus center of the center or edge of each target region as (Δx+(R+d1)cos(θi), Δy+(R+d1)sin(θi)). (It is noted that d1 may be zero, i.e., the center or edge of each target region may coincide with the limbus as marked by the user, such that the respective centers or edges (respectively) of the treatment beams impinge on the limbus as marked by the user.) Subsequently, during the procedure, as further described below with reference to
Typically, in step B, the controller also identifies, based on the still image, a static region 76 in the field of view (FOV) of the camera—also referred to herein as a “forbidden zone”—that includes the pupil 74 of the eye, typically along with a “buffer” that includes a significant portion of the cornea 72 of the eye surrounding pupil 74. Typically, the size of the buffer is set based on the maximum expected movement of the eye.
In some embodiments, region 76 is identified based on the location of the limbus as automatically identified by the controller or as marked by the user. For example, the controller may identify region 76 as the set of all points in the FOV located inside the limbus at more than a predefined distance from the limbus. Alternatively, for example, the controller may identify the point at the center of the limbus or the center of the pupil, and then center region 76 at this center point. In such embodiments, region 76 may have any suitable shape, such as an elliptical or rectangular shape, and may have any suitable size. The significance of region 76 is described below with reference to
Following step B, the controller proceeds to step C, in which the trabeculoplasty procedure is simulated. In response to viewing the simulation, the user may provide a confirmation input to the controller. e.g., by clicking an appropriate button (such as a “START” button) in GUI 68. This input confirms that the controller should proceed with the procedure.
More specifically, in step C, the controller displays a live sequence of images (i.e., a live video) of the eye, and, while displaying the sequence of images, irradiates the eye with one or more aiming beams 84, which are visible in the images. Typically, the aiming beams are red; for example, each aiming beam may have a wavelength of between 620 and 650 nm. In some embodiments, the color of the aiming beams is different from that of the treatment beams; for example, whereas the aiming beams may be red, the treatment beams may be green, having a wavelength of between 515 and 545 nm (e.g., 532 nm), for example.
While irradiating the eye with the aiming beams, the controller controls the beam-directing elements such that, if the treatment beams were to be fired, the treatment beams would impinge on the calculated target regions. Thus, the respective centers of the aiming beams may coincide, sequentially, with the center of each target region. Alternatively, if F-theta lens 51 (
In some embodiments, the controller sweeps a single aiming beam along the eye, such that the aiming beam impinges on at least part of each target region. In other embodiments, the controller fires a plurality of aiming beams, such that each aiming beam impinges on at least part of a different respective one of the target regions.
Typically, while performing the simulation, the controller superimposes marker 78 over the portion of the eye that was identified in step B. To compensate for any movement of the eye, the controller typically identifies the center of the limbus in each of the images, and places marker 78 at the appropriate offset from the limbus. For example, if the final position of the center of marker 78 in the still image (step B) is (x0+Δx, y0+Δy), the controller may place marker 78 at an offset of (Δx, Δy) from the center of the limbus in each of the live images.
Alternatively or additionally to superimposing marker 78, the controller may superimpose, on each of the images, another marker 82 passing through (e.g., through the center of) or near each target region. The position of marker 82 may be adjusted responsively to motion of the eye, by maintaining marker 82 at the proper offset from marker 78. For example, if the center of each target region is to be at a distance of d1 from the limbus as marked by the user, marker 82 may be kept at a distance of d1 from marker 78. In some embodiments, marker 82 is a different color from that of marker 78.
Typically, while performing the simulation, the controller verifies that each of the aiming beams was properly directed by the beam-directing elements. For example, the controller may process a feedback signal from the encoders for galvo mirrors 50. Alternatively or additionally, the controller, by processing the images, may verify the respective positions of the aiming beams with respect to marker 78, marker 82, and/or any other suitable marker superimposed on each of the images. For example, the controller may verify that each aiming beam (e.g., the center of each aiming beam) overlaps marker 82, and/or that the edge of each aiming beam touches marker 78. (In the context of the present application, including the claims, the “edge” of a beam may be defined in terms of the knife-edge measure, the 1/e2 width measure, the full width at half maximum measure, or any other suitable measure.) As another example, the controller may verify that the center or edge of each aiming beam is positioned at the appropriate distance from marker 78.
In response to verifying the positions of the aiming beams, the controller may proceed with the trabeculoplasty procedure, provided that the user provides the aforementioned confirmation input.
In some embodiments, if the user does not confirm the simulation, the treatment is aborted. In other embodiments, the user may (e.g., via GUI 68) adjust the path followed by the aiming beams. This adjustment may be performed by returning to step B and adjusting marker 78, and/or by adjusting the distance from marker 78 at which each target region is to be located. In such embodiments, the simulation may be repeated for each new path defined by the user, until the user confirms the path.
In response to receiving the aforementioned confirmation input from the user, the controller treats the eye by irradiating the target regions with respective treatment beams. The peak power of the treatment beams is much higher than that of the aiming beams; furthermore, typically, the wavelength of the treatment beams is better suited for treating the trabecular meshwork of the eye, relative to the wavelength of the aiming beams.
More specifically, during the treatment, the controller continues to sweep an aiming beam through the target regions, or to fire respective aiming beams at the target regions, while acquiring images of the eye. As further described below with reference to
Typically, the controller causes each of the treatment beams to impinge on the eye outside region 76 (
Typically, while acquiring each of the images during the treatment procedure, the controller causes illumination source 60 (
Typically, the light is flashed at a frequency that is sufficiently high such that the patient does not notice the flashing, but rather, perceives steady illumination. For example, the light may be flashed at a frequency of at least 60 Hz, such as at least 100 Hz. (In such embodiments, the duration of each flash (or “pulse”) is typically less than 3 ms, such as less than 2 ms or 1 ms.) Since the frequency of the flashing may be higher than the frame rate (i.e., the frequency at which the images are acquired), some of the flashes may occur between image acquisitions. For example, the flashing frequency may be an integer multiple of the frequency at which images are acquired, such that the flashing is synchronized with the image acquisition. As a purely illustrative example, with a frame rate of 60 Hz, the flashing frequency may be 120 Hz or 180 Hz.
Alternatively, the light may be flashed at a lower frequency, but the duration of each flash may be increased such that steady illumination is perceived. For example, if the patient perceives flickering with a flashing frequency of 100 Hz and a 20% duty cycle, the duty cycle may be increased to 40% by increasing the pulse width without changing the frequency.
In some embodiments, illumination source 60 is configured to emit near-infrared light. In such embodiments, near-infrared light may be shone continuously during the treatment, or at least while the images are acquired, in order to reduce the required camera exposure time without disturbing the patient. Optionally, illumination source 60 may also flash visible light at the eye during and/or between the image acquisitions, so as to further reduce the required exposure time and/or to constrict the pupil.
Some further details regarding the trabeculoplasty procedure are now provided with reference to
To begin the procedure after approval of the simulated procedure by the user, the controller, at an imaging-and-locating step 88, flashes light at the eye, uses the camera to acquire an image of the eye during the flash, and locates the center of the limbus in the acquired image. Subsequently, at a target-calculating step 90, the controller calculates the position of the next target region, by adding the appropriate (x, y) offset to the location of the center of the limbus. After verifying this position, the target region is irradiated, as further described below. The controller then acquires another image, calculates the position of the next target region, verifies the position, and irradiates the target. In this manner, the controller iteratively irradiates the target regions.
More specifically, for each calculated target region, the controller checks, at a first target-checking step 92, whether the target region lies (even partly) in the forbidden zone, which, it will be recalled, is a static region in the FOV of the camera. (To perform this check, the controller does not necessarily explicitly calculate the boundaries of the target region; for example, the controller may check whether the point at the center of the target region lies more than a predefined distance—equivalent to or slightly greater than the radius of the aiming beam or treatment beam—from the border of the forbidden zone.) If not, the controller performs a second target-checking step 94, at which—provided that the target region was preceded by a previous target region—the controller checks whether the target region is at an acceptable distance from the previous target region. For example, the controller may check whether the distance between the target region and the previous target region is less than a predefined threshold, indicating that the eye is relatively still. If the target region is not at an acceptable distance from the previous target region, or if the target region is in the forbidden zone, the controller returns to imaging-and-locating step 88.
If the calculated target region passes both first target-checking step 92 and second target-checking step 94, the controller aims the beam-directing elements at the target region, at an aiming step 96. Subsequently, the controller, at an aiming-beam-firing step 98, fires an aiming beam at the beam-directing elements, such that the aiming beam is directed toward the target region by the beam-directing elements. Alternatively, a single aiming beam may be continuously emitted, such that there is no need to perform aiming-beam-firing step 98.
Subsequently, the controller performs imaging-and-locating step 88. The controller then checks, at a limbus-center-checking step 100, whether the center of the limbus moved (relative to the most-recently acquired image) by more than a predefined threshold. If yes, the controller returns to target-calculating step 90, and recalculates the location of the target region with respect to the center of the limbus. Otherwise, the controller, at an aiming-beam-identifying step 102, identifies the aiming beam in the image.
Subsequently to identifying the aiming beam, the controller checks, at a first aiming-beam-checking step 106, whether the aiming beam is in the forbidden zone. If the aiming beam is in the forbidden zone—indicating rapid movement of the eye or a failure in the system—the controller terminates the procedure. Otherwise, the controller checks, at a second aiming-beam-checking step 108, whether the distance between the aiming beam and the calculated target region is within a predefined threshold. If not, the controller returns to target-calculating step 90. Otherwise, the controller fires the treatment beam, at a treatment-beam-firing step 110, such that the treatment beam impinges on the target region.
Typically, in addition to identifying and verifying the position of the aiming beam, the controller checks each image for any obstructions that may be obstructing the target region, including, for example, an eyelid, eyelashes, a finger, growths (such as pterygium), blood vessels, or a speculum. In the event that an obstruction is identified, the target region may be shifted to avoid the obstruction. Alternatively, the target region may be skipped entirely, or the treatment procedure may be terminated.
In general, obstructions may be identified using any suitable image-processing techniques, optionally in combination with input from the user. For example, prior to the treatment procedure, the user may select (e.g., with reference to the still image) one or more portions of the eye that constitute potential obstructions. Subsequently, the controller may use template matching, edge detection, or any other suitable techniques—including, for example, identifying changes between successive images—to identify the selected portions of the eye. Such techniques may also be used to identify other static or dynamic obstructions that were not necessarily identified in advance by the user. (It is noted that the definition of “obstruction” may vary between applications; for example, whereas in some applications a particular blood vessel may constitute an obstruction, in other cases it may be desired to irradiate the blood vessel.)
Following treatment-beam-firing step 110, the controller checks, at a final checking step 112, whether all of the target regions have been treated. If yes, the controller terminates the procedure. Otherwise, the controller returns to target-calculating step 90.
Advantageously, the time between the acquisition of each image and the firing of the treatment beam is typically less than 15 ms, e.g., less than 10 ms. In some embodiments, this delay is reduced is even further, by firing the treatment beam between aiming step 96 and aiming-beam-firing step 98 (or, if a single aiming beam is continuously emitted, between aiming step 96 and imaging-and-locating step 88), instead of after second aiming-beam-checking step 108. (In such embodiments, the aiming beam is used to verify post facto that the treatment beam was fired correctly.)
In some embodiments, a separate routine executed by the controller monitors the time from each image acquisition. If this time exceeds a predefined threshold (such as a threshold between 10 and 15 ms), the treatment beam is not fired until after the next image is acquired and the target position is recalculated.
It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof that are not in the prior art, which would occur to persons skilled in the art upon reading the foregoing description.
The present application claims the benefit of (i) U.S. Provisional Appl. No. 62/692,868, entitled “Direct laser selective trabeculoplasty Process (DSLT) and Safeties,” filed Jul. 2, 2018, (ii) U.S. Provisional Appl. No. 62/739,238, entitled “Eye tracking flash illumination,” filed Sep. 30, 2018, and (iii) U.S. Provisional Appl. No. 62/748,461, entitled “Crossed ranging beams,” filed Oct. 21, 2018. The respective disclosure of each of the aforementioned references is incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/IB2019/055564 | 7/1/2019 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/008323 | 1/9/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
2635502 | Richards | Apr 1953 | A |
3594072 | Feather | Jul 1971 | A |
4587257 | DeSantis | May 1986 | A |
4641349 | Flom et al. | Feb 1987 | A |
4718418 | L'Esperance | Jan 1988 | A |
4848894 | Buser et al. | Jul 1989 | A |
4941093 | Marshall | Jul 1990 | A |
4966452 | Shields et al. | Oct 1990 | A |
5123902 | Muller et al. | Jun 1992 | A |
5152760 | Latina | Oct 1992 | A |
5370641 | O'Donnell, Jr. | Dec 1994 | A |
5479222 | Volk et al. | Dec 1995 | A |
5549596 | Latina | Aug 1996 | A |
5598007 | Bunce et al. | Jan 1997 | A |
5786883 | Miller et al. | Jul 1998 | A |
5865830 | Parel et al. | Feb 1999 | A |
5982789 | Marshall et al. | Nov 1999 | A |
6027216 | Guyton et al. | Feb 2000 | A |
6033396 | Huang et al. | Mar 2000 | A |
6059772 | Hsia et al. | May 2000 | A |
6096029 | O'Donnell, Jr. | Aug 2000 | A |
6099521 | Shadduck | Aug 2000 | A |
6099522 | Knopp | Aug 2000 | A |
6146375 | Juhasz et al. | Nov 2000 | A |
6210399 | Parel et al. | Apr 2001 | B1 |
6258082 | Lin | Jul 2001 | B1 |
6263879 | Lin | Jul 2001 | B1 |
6267752 | Svetliza | Jul 2001 | B1 |
6267756 | Feuerstein et al. | Jul 2001 | B1 |
6319274 | Shadduck | Nov 2001 | B1 |
6325792 | Swinger et al. | Dec 2001 | B1 |
6454763 | Motter et al. | Sep 2002 | B1 |
6514241 | Hsia et al. | Feb 2003 | B1 |
6530916 | Shimmick | Mar 2003 | B1 |
6569104 | Ono et al. | May 2003 | B2 |
6673062 | Yee et al. | Jan 2004 | B2 |
6676655 | McDaniel | Jan 2004 | B2 |
6685317 | Su et al. | Feb 2004 | B2 |
6698886 | Pollack et al. | Mar 2004 | B2 |
6736806 | Ruiz et al. | May 2004 | B2 |
6899707 | Scholler et al. | May 2005 | B2 |
6942656 | Pawlowski et al. | Sep 2005 | B2 |
6948815 | Neuberger | Sep 2005 | B2 |
6979328 | Baerveldt et al. | Dec 2005 | B2 |
7027233 | Goldstein et al. | Apr 2006 | B2 |
7252661 | Nguyen et al. | Aug 2007 | B2 |
7282046 | Simon | Oct 2007 | B2 |
7353829 | Wachter et al. | Apr 2008 | B1 |
7371230 | Webb et al. | May 2008 | B2 |
7693259 | Gertner | Apr 2010 | B2 |
7792249 | Gertner et al. | Sep 2010 | B2 |
8004764 | Artsyukhovich et al. | Aug 2011 | B2 |
8048065 | Grecu et al. | Nov 2011 | B2 |
8109635 | Allon et al. | Feb 2012 | B2 |
8403921 | Patankar et al. | Mar 2013 | B2 |
8442185 | Gertner et al. | May 2013 | B2 |
8465478 | Frey et al. | Jun 2013 | B2 |
8475433 | Mrochen et al. | Jul 2013 | B2 |
8545020 | Liesfeld et al. | Oct 2013 | B2 |
8568393 | Palanker | Oct 2013 | B2 |
8630388 | Gertner et al. | Jan 2014 | B2 |
8679100 | Raksi et al. | Mar 2014 | B2 |
8708491 | Frey et al. | Apr 2014 | B2 |
8709029 | Griffis, III et al. | Apr 2014 | B2 |
8771261 | Andersen et al. | Jul 2014 | B2 |
8811657 | Teiwes et al. | Aug 2014 | B2 |
8845625 | Angeley et al. | Sep 2014 | B2 |
8903468 | Peyman | Dec 2014 | B2 |
8920407 | Raksi et al. | Dec 2014 | B2 |
8939965 | Liesfeld et al. | Jan 2015 | B2 |
8968279 | Arnoldussen | Mar 2015 | B2 |
8995618 | Gertner | Mar 2015 | B2 |
9055896 | Amthor et al. | Jun 2015 | B2 |
9192780 | McDaniel | Nov 2015 | B2 |
9220407 | Yam et al. | Dec 2015 | B2 |
9351878 | Muehlhoff et al. | May 2016 | B2 |
9480599 | Degani et al. | Nov 2016 | B2 |
9495743 | Angeley et al. | Nov 2016 | B2 |
9504609 | Kurtz | Nov 2016 | B2 |
9532712 | Liesfeld et al. | Jan 2017 | B2 |
9622911 | Rubinfeld et al. | Apr 2017 | B2 |
9782232 | Papac | Oct 2017 | B1 |
9849032 | Schuele et al. | Dec 2017 | B2 |
9849034 | Artsyukhovich et al. | Dec 2017 | B2 |
9877633 | Zhao et al. | Jan 2018 | B2 |
9889043 | Frey et al. | Feb 2018 | B2 |
9968483 | Takeda et al. | May 2018 | B2 |
10022457 | Peyman | Jul 2018 | B2 |
10064757 | Berlin | Sep 2018 | B2 |
10143590 | Dick et al. | Dec 2018 | B2 |
10244991 | Shademan et al. | Apr 2019 | B2 |
10258507 | Gonzalez et al. | Apr 2019 | B2 |
10278865 | Luttrull et al. | May 2019 | B2 |
10299961 | Luttrull et al. | May 2019 | B2 |
10363169 | Belkin et al. | Jul 2019 | B2 |
10441465 | Hart et al. | Oct 2019 | B2 |
10449091 | Angeley et al. | Oct 2019 | B2 |
10456209 | Peyman | Oct 2019 | B2 |
10478342 | Dick et al. | Nov 2019 | B2 |
10524656 | Wiltberger et al. | Jan 2020 | B2 |
10617564 | Andersen et al. | Apr 2020 | B1 |
10684449 | Curatu et al. | Jun 2020 | B2 |
10702416 | Belkin et al. | Jul 2020 | B2 |
10849789 | Dewey et al. | Dec 2020 | B2 |
10925768 | Charles | Feb 2021 | B2 |
20010027314 | Peyman | Oct 2001 | A1 |
20020013573 | Telfair et al. | Jan 2002 | A1 |
20040059321 | Knopp et al. | Mar 2004 | A1 |
20050096639 | Slatkine et al. | May 2005 | A1 |
20050107774 | Lin | May 2005 | A1 |
20050254009 | Baker et al. | Nov 2005 | A1 |
20050286019 | Wiltberger et al. | Dec 2005 | A1 |
20050288745 | Andersen et al. | Dec 2005 | A1 |
20060176913 | Souhaite et al. | Aug 2006 | A1 |
20060195076 | Blumenkranz et al. | Aug 2006 | A1 |
20060265030 | McDaniel | Nov 2006 | A1 |
20070081166 | Brown et al. | Apr 2007 | A1 |
20070129709 | Andersen et al. | Jun 2007 | A1 |
20070159600 | Gil et al. | Jul 2007 | A1 |
20070213693 | Plunkett | Sep 2007 | A1 |
20080089481 | Gertner | Apr 2008 | A1 |
20080108934 | Berlin et al. | May 2008 | A1 |
20080161781 | McArdle et al. | Jul 2008 | A1 |
20080204658 | Van Saarloos | Aug 2008 | A1 |
20090137993 | Kurtz | May 2009 | A1 |
20090157062 | Hauger et al. | Jun 2009 | A1 |
20090161827 | Gertner | Jun 2009 | A1 |
20100002837 | Gertner et al. | Jan 2010 | A1 |
20100076419 | Chew et al. | Mar 2010 | A1 |
20100142767 | Fleming | Jun 2010 | A1 |
20100324543 | Kurtz et al. | Dec 2010 | A1 |
20110144627 | Smith et al. | Jun 2011 | A1 |
20110172649 | Schuele et al. | Jul 2011 | A1 |
20110190741 | Deisinger et al. | Aug 2011 | A1 |
20120016349 | Brownell | Jan 2012 | A1 |
20120050308 | Nakano et al. | Mar 2012 | A1 |
20120083772 | Rubinfeld et al. | Apr 2012 | A1 |
20120089134 | Horvath et al. | Apr 2012 | A1 |
20120283557 | Berlin | Nov 2012 | A1 |
20130103011 | Grant et al. | Apr 2013 | A1 |
20130123761 | Belkin et al. | May 2013 | A1 |
20130204236 | Awdeh | Aug 2013 | A1 |
20130218145 | Belkin et al. | Aug 2013 | A1 |
20130289450 | Homer | Oct 2013 | A1 |
20140094785 | Charles | Apr 2014 | A1 |
20140114297 | Woodley et al. | Apr 2014 | A1 |
20140128731 | Gonzalez et al. | May 2014 | A1 |
20140128851 | Wysopal | May 2014 | A1 |
20140128852 | Gooding et al. | May 2014 | A1 |
20140135747 | Donitzky et al. | May 2014 | A1 |
20140135753 | Feklistov et al. | May 2014 | A1 |
20140276681 | Schuele et al. | Sep 2014 | A1 |
20140307077 | Prabhakar | Oct 2014 | A1 |
20150164635 | Renke | Jun 2015 | A1 |
20150223683 | Davidovics et al. | Aug 2015 | A1 |
20150266706 | Hashimoto | Sep 2015 | A1 |
20150272782 | Schuele et al. | Oct 2015 | A1 |
20150313759 | Vera et al. | Nov 2015 | A1 |
20160008172 | Kahook et al. | Jan 2016 | A1 |
20160067087 | Tedford et al. | Mar 2016 | A1 |
20160089269 | Horvath et al. | Mar 2016 | A1 |
20160095752 | Srinivasan | Apr 2016 | A1 |
20160113816 | Herekar et al. | Apr 2016 | A1 |
20160346126 | Luttrull et al. | Dec 2016 | A1 |
20160354241 | Mordaunt et al. | Dec 2016 | A1 |
20160367399 | Goldshleger et al. | Dec 2016 | A1 |
20170038284 | Nemati | Feb 2017 | A1 |
20170087014 | Potter, Jr. et al. | Mar 2017 | A1 |
20170246033 | Bor et al. | Aug 2017 | A1 |
20170340483 | Rill et al. | Nov 2017 | A1 |
20170360604 | Bach et al. | Dec 2017 | A1 |
20180085257 | Horvath et al. | Mar 2018 | A1 |
20180125708 | Bohme et al. | May 2018 | A1 |
20180168737 | Ren et al. | Jun 2018 | A1 |
20180207029 | Herekar et al. | Jul 2018 | A1 |
20180214305 | Schuele et al. | Aug 2018 | A1 |
20180344527 | Palanker et al. | Dec 2018 | A1 |
20190078073 | Streeter et al. | Mar 2019 | A1 |
20190099291 | Herekar et al. | Apr 2019 | A1 |
20190105519 | Herekar et al. | Apr 2019 | A1 |
20190142636 | Tedford et al. | May 2019 | A1 |
20190151146 | Kim | May 2019 | A1 |
20190247225 | Stobrawa et al. | Aug 2019 | A1 |
20190269554 | Goldshleger et al. | Sep 2019 | A1 |
20190343680 | Belkin et al. | Nov 2019 | A1 |
20190344076 | Irazoqui et al. | Nov 2019 | A1 |
20190358085 | Fu et al. | Nov 2019 | A1 |
20200107724 | Wiltberger et al. | Apr 2020 | A1 |
20200146887 | Horvath et al. | May 2020 | A1 |
20200306080 | Herekar et al. | Oct 2020 | A1 |
20200345546 | Belkin et al. | Nov 2020 | A1 |
20200360187 | Schuele et al. | Nov 2020 | A1 |
20200379216 | Curatu et al. | Dec 2020 | A1 |
Number | Date | Country |
---|---|---|
2015210430 | Sep 2015 | AU |
2015315113 | Mar 2016 | AU |
2017258835 | Nov 2017 | AU |
2640203 | Aug 2007 | CA |
1579351 | Feb 2005 | CN |
101411607 | Apr 2009 | CN |
201537172 | Aug 2010 | CN |
202016006265 | Mar 2017 | DE |
0224322 | Jun 1987 | EP |
0651982 | May 1995 | EP |
0689811 | Jan 1996 | EP |
1602321 | Dec 2005 | EP |
2301421 | Mar 2011 | EP |
2301424 | Mar 2011 | EP |
2301425 | Mar 2011 | EP |
2602005 | Jun 2013 | EP |
1856774 | Jun 2016 | EP |
2695016 | Mar 2017 | EP |
2992931 | Aug 2017 | EP |
2391318 | Dec 2017 | EP |
2729099 | Nov 2019 | EP |
3191040 | Jul 2020 | EP |
3517081 | Nov 2020 | EP |
2854729 | Mar 2021 | EP |
2655837 | Jun 1991 | FR |
2010148635 | Jul 2010 | JP |
6083823 | Feb 2017 | JP |
20180106113 | Oct 2018 | KR |
20190022216 | Mar 2019 | KR |
2499582 | Nov 2013 | RU |
2553507 | Jun 2015 | RU |
9216259 | Oct 1992 | WO |
1993012727 | Jul 1993 | WO |
9316631 | Sep 1993 | WO |
9412092 | Jun 1994 | WO |
1998022016 | May 1998 | WO |
9918868 | Apr 1999 | WO |
02064031 | Aug 2002 | WO |
2014018104 | Jan 2004 | WO |
2004027487 | Apr 2004 | WO |
2006119584 | Nov 2006 | WO |
2007103349 | Sep 2007 | WO |
2008112236 | Sep 2008 | WO |
2010113193 | Oct 2010 | WO |
2011163508 | Jun 2011 | WO |
2011085274 | Jul 2011 | WO |
2011151812 | Dec 2011 | WO |
2013004255 | Jan 2013 | WO |
2013035091 | Mar 2013 | WO |
2013059481 | Apr 2013 | WO |
2013122711 | Aug 2013 | WO |
2013165689 | Nov 2013 | WO |
2014132162 | Sep 2014 | WO |
2015069197 | May 2015 | WO |
2016018864 | Feb 2016 | WO |
2016207739 | Dec 2016 | WO |
2017031570 | Mar 2017 | WO |
2017069819 | Apr 2017 | WO |
2018005796 | Jan 2018 | WO |
2018021780 | Feb 2018 | WO |
2018049246 | Mar 2018 | WO |
2018232397 | Dec 2018 | WO |
2019109125 | Jun 2019 | WO |
2020008323 | Jan 2020 | WO |
2020012841 | Jan 2020 | WO |
2020018436 | Jan 2020 | WO |
2020050308 | Mar 2020 | WO |
202093060 | May 2020 | WO |
2020089737 | May 2020 | WO |
2020183342 | Sep 2020 | WO |
2021048723 | Mar 2021 | WO |
Entry |
---|
Katta et al., “Optical Coherence Tomography Image-Guided Smart Laser Knife for Surgery,” Lasers in Surgery and Medicine, Wiley Online Library, pp. 1-11, Jul. 2017. |
Barnes et al., “Control of Intraocular Pressure Elevations after Argon Laser Trabeculoplasty: Comparison of Brimonidine 0.2% to Apraclonidine 1.0%,” Opthalmology, vol. 106, No. 10, pp. 2033-2037, year 1999. |
Yakopson et al., “Brimonidine 0.1% vs. Apraclonidine 0.5% for Prevention of Intraocular Pressure Elevation after Selective Laser Trabeculoplasty,” Investigative Ophthalmology & Visual Science, vol. 49, p. 1234, May 2008. |
Kim et at., “Effect of Prophylactic Topical Brimonidine (0.15%) Administration on the Development of Subconjunctival Hemorrhage after Intravitreal Injection,” Retina, The Journal for Retinal and Vitreous Diseases, vol. 31, No. 2, pp. 389-392, year 2011. |
Hong et al., “Effect of Prophylactic Brimonidine Instillation on Bleeding during Strabismus Surgery in Adults,” American Journal of Ophthalmology, vol. 144, No. 3, pp. 469-470, Sep. 2007. |
Goldsmith et al., “Anterior Chamber Width Measurement by High-Speed Optical Coherence Tomography,” Ophthalmology, vol. 112, No. 2, pp. 238-244, year 2005. |
Norden, “Effect of Prophilactic Brimonidine on Bleeding Complications and Flap Adherence After Laser in situ Keratomileusis,” Journal of Refractive Surgery, vol. 18, No. 4, pp. 468-471, Jul./Aug. 2002. |
Kohnen et al., “Internal Anterior Chamber Diameter using Optical Coherence Tomography Compared with White-to-White Distances using Automated Measurements,” Journal of Cataract & Refractive Surgery, vol. 32, pp. 1809-1813, Nov. 2006. |
Zhang et al., “Perioperative Medications for Preventing Temporarily Increased Intraocular Pressure after Laser Trabeculoplasty (Review),” Cochrane Database of Systematic Reviews 2017, issue 2, pp. 1-117, year 2017. |
AU Application # 2019297135 Office Action dated Sep. 30, 2021. |
International Application # PCT/IB2021/054187 Search Report dated Jul. 30, 2021. |
Nagar et al., “A randomised, prospective study comparing selective laser trabeculoplasty with latanoprost for the control of intraocular pressure in ocular hypertension and open angle glaucoma,” British Journal of Ophthalmology, vol. 89, pp. 1413-1417, year 2005. |
Hong et al., “Repeat Selective Laser Trabeculoplasty,” Journal of Glaucoma, vol. 18, issue 3, pp. 180-183, Mar. 2009. |
Goyal et al., “Effect of primary selective laser trabeculoplasty on tonographic outflow facility—a randomised clinical trial,” British Journal of Ophthalmology, BMJ Publishing Group, vol. 94, issue 11, pp. 1-22, year 2010. |
Franco et al., “Effect of Second SLT on IOP,” Investigative Ophthalmology & Visual Science, vol. 48, pp. 1-2 May 2007. |
Chen et al., “A Comparison between 90 degrees and 180 degrees Selective Laser Trabeculoplasty,” Journal of Glaucoma, vol. 13, issue 1, p. 1, Feb. 2004. |
Mequio et al., “Efficacy of Repeat Selective Laser Trabeculoplasty,” Investigative Ophthalmology & Visual Science, vol. 48, p. 1, year 2007. |
Grulkowski et al., “Anterior segment imaging with Spectral OCT system using a high-speed CMOS camera,” Optics Express, vol. 17, No. 6, p. 4842-4858, year 2009. |
Shields et al., “Noncontact Transscleral ND:YAG Cyclophotocoagulation: A Long-Term Follow-Up of 500 Patients,” Transactions of the American Ophthalmological Society, vol. XCII, pp. 271-287, year 1994. |
Liu et al., “Real-time visual analysis of microvascular blood flow for critical care,” CVPR2015 paper as Open Access Version, provided by the Computer Vision Foundation, pp. 2217-2225, year 2015. |
Desco et al., “Effect of prophylactic brimonidine on bleeding complications after cataract surgery,” European Journal of Ophthalmology, vol. 15, pp. 228-232, year 2005. |
Pasquali et al., “Dilute brimonidine to improve patient comfort and subconjunctival hemorrhage after LASIK,” Journal of Refractive Surgery, vol. 29, pp. 469-475, year 2013. |
Sacks et al., “Non-contact direct selective laser trabeculoplasty: light propagation analysis,” Biomedical Optics Express, vol. 11, pp. 2889-2904, year 2020. |
Kasuga et al., “Trabecular Meshwork Length in Men and Women by Histological Assessment,” Current Eye Research, Early Online, pp. 1-5, Jun. 2012. |
International Application # PCT/IB2020/058300 Search Report dated Dec. 27, 2020. |
SensoMotoric Instruments GmbH (SMI), “SG 3000”, Product Flyer, pp. 1-2, year 2010. |
Ashik et al., “The precision of ophthalmic biometry using calipers,” Canadian Journal of Ophthalmology, vol. 48, issue 6, pp. 1-13, Dec. 2013. |
Balalzsi, “Noncontact Thermal Mode Nd:YAG Laser Transscleral Cyclocoagulation in the Treatment of Glaucoma,” Ophthalmology, vol. 98, pp. 1858-1863, year 1991. |
Leung et al., “Anterior chamber angle imaging with optical coherence tomography,” Eye, vol. 25, pp. 261-267, year 2011. |
Tasman et al., “The Wills Eye Hospital Atlas of Clinical Ophthalmology,” Lippincott Williams & Wilkins, p. 158, year 2001. |
Gaasterland, “Laser Therapies: Iridotomy, Iridoplasty, and Trabeculoplasty,” as appears in “The Glaucoma Book: A Practical Evidence-Based Approach to Patient Care,” Springer, p. 722, year 2010. |
Kara, “Bleeding in Retinal Images Using Image Processing”, A Thesis submitted to the graduate school of applied sciences of Near East University, pp. 1-79, Nicosia, Larnaca, year 2019. |
Navilas Operator Manual, Document Version 2.10, 2012 OD-OS GmbH, pp. 1-94, Sep. 2012. |
EP Application # 20201567.3 Search Report dated Jun. 22, 2021. |
U.S. Appl. No. 16/420,194 Office Action dated Jul. 22, 2021. |
Vogel et al., “Optical properties of human sclera, and their consequences for transscleral laser applications”, Lasers in Surgery and Medicine , vol. 11, pp. 331-340, 1991. |
Geffen et al., “Transscleral Selective Laser Trabeculoplasty Without a Gonioscopy Lens”, Journal of Glaucoma, Inc, vol. 26, No. 3, pp. 201-207, Mar. 2017. |
Das et al., “Sclera Recognition—A Survey”, 2nd IAPR Asian Conference on Pattern Recognition, pp. 1-5, Naha, Japan, Nov. 5-8, 2013. |
Kaya et al., “Designing a Pattern Stabilization Method Using Scleral Blood Vessels for Laser Eye Surgery”, International Conference on Pattern Recognition, pp. 698-701, Istanbul, Turkey, Aug. 23-26, 2010. |
International Application # PCT/IB2019/055564 search report dated Oct. 10, 2019. |
Arany, “Photobiomodulation therapy: Easy to do, but difficult to get right”, LaserFocusWorld, pp. 1-6, Jul. 31, 2019 downloaded from www.laserfocusworld.com/lasers-sources/article/14037967/photobiomodulation-therapyeasy-to-do-but-difficult-to-get-right, pp. 22-24, year 2019. |
Borzabadi-Farahani, “Effect of low-level laser irradiation on proliferation of human dental mesenchymal stem cells; a systemic review”, Journal of Photochemistry and Photobiology B: Biology, vol. 162, pp. 577-582, Sep. 2016. |
Ascott et al., “Trabecular Repopulation by Anterior Trabecular Meshwork Cells After Laser Trabeculoplasty”, American Journal of Ophthalmology, vol. 107, issue 1, pp. 1-6, Jan. 1989. |
Cao et al., “Peripheral Iridotomy,” Medscape 25, pp. 1-12, Jun. 15, 2020. |
Husain, “Laser Peripheral Iridotomy—Practical Points”, YouTube presentation, p. 1, Sep. 28, 2016, downloaded from https://www.youtube.com/watch?=Azxzsv31yls. |
Sacks et al., U.S. Appl. No. 17/136,052, filed Dec. 29, 2020. |
Smith et al., “Light scatter from the central human cornea”, Journal “Eye”, issue 4, pp. 584-588, year 1990. |
Turati et al., “Patterned Laser Trabeculoplasty”, Ophthalmic Surgery, Lasers and Imaging , vol. 41, No. 5, pp. 538-545, 2010. |
Nozaki et al.,“Patterned Laser Trabeculoplasty with PASCAL streamline 577”, Investigative Ophthalmology & Visual Science, vol. 54, p. 1867, Jun. 2013. |
Barkana et al., “Selective Laser Trabeculoplasty”, Survey of Ophthalmology, vol. 52, No. 6, pp. 634-653, year 2007. |
Ivandic et al., “Early Diagnosis of Ocular Hypertension Using a Low-Intensity Laser Irradiation Test”, Photomedicine and Laser Surgey, vol. 00, No. 00, pp. 1-5, year 2009. |
EP Application No. 19830473.5 Search Report dated Feb. 28, 2022. |
AU Application No. 2019297135 Office Action dated Jan. 5, 2022. |
U.S. Appl. No. 16/935,236 Office Action dated Jan. 6, 2022. |
IN Application No. 202147003401 Office Action dated Jan. 13, 2022. |
CN Application No. 201980043641.6 Office Action dated Feb. 18, 2022. |
IN Application No. 201948052117 Office Actio ndated Feb. 16, 2022. |
International Application # PCT/IB2021/059821 Search Report dated Apr. 7, 2022. |
Number | Date | Country | |
---|---|---|---|
20210267800 A1 | Sep 2021 | US |
Number | Date | Country | |
---|---|---|---|
62692868 | Jul 2018 | US | |
62739238 | Sep 2018 | US | |
62748461 | Oct 2018 | US |