Image stabilization techniques and methods

Abstract
Direct optical imaging of anatomical features and structures from within a biological organ in a dynamic environment (where the tissue being imaged is in motion due to cardiac rhythms, respiration, etc) presents certain image stability issues due (and/or related) to the motion of the target structure and may limit the ability of the user to visually interpret the image for the purposes of diagnostics and therapeutics. Systems and mechanisms for the purpose of actively stabilizing the image or for compiling and re-displaying the image information in a manner that is more suitable to interpretation by the user are disclosed.
Description
FIELD OF THE INVENTION

The present invention relates generally to catheter control systems and methods for stabilizing images of moving tissue regions such as a heart which are captured when intravascularly accessing and/or treating regions of the body.


BACKGROUND OF THE INVENTION

Conventional devices for accessing and visualizing interior regions of a body lumen are known. For example, various catheter devices are typically advanced within a patient's body, e.g., intravascularly, and advanced into a desirable position within the body. Other conventional methods have utilized catheters or probes having position sensors deployed within the body lumen, such as the interior of a cardiac chamber. These types of positional sensors are typically used to determine the movement of a cardiac tissue surface or the electrical activity within the cardiac tissue. When a sufficient number of points have been sampled by the sensors, a “map” of the cardiac tissue may be generated.


Another conventional device utilizes an inflatable balloon which is typically introduced intravascularly in a deflated state and then inflated against the tissue region to be examined. Imaging is typically accomplished by an optical fiber or other apparatus such as electronic chips for viewing the tissue through the membrane(s) of the inflated balloon. Moreover, the balloon must generally be inflated for imaging. Other conventional balloons utilize a cavity or depression formed at a distal end of the inflated balloon. This cavity or depression is pressed against the tissue to be examined and is flushed with a clear fluid to provide a clear pathway through the blood.


However, many of the conventional catheter imaging systems lack the capability to provide therapeutic treatments or are difficult to manipulate in providing effective therapies. For instance, the treatment in a patient's heart for atrial fibrillation is generally made difficult by a number of factors, such as visualization of the target tissue, access to the target tissue, and instrument articulation and management, amongst others.


Conventional catheter techniques and devices, for example such as those described in U.S. Pat. Nos. 5,895,417; 5,941,845; and 6,129,724, used on the epicardial surface of the heart may be difficult in assuring a transmural lesion or complete blockage of electrical signals. In addition, current devices may have difficulty dealing with varying thickness of tissue through which a transmural lesion is desired.


Conventional accompanying imaging devices, such as fluoroscopy, are unable to detect perpendicular electrode orientation, catheter movement during the cardiac cycle, and image catheter position throughout lesion formation. The absence of real-time visualization also poses the risk of incorrect placement and ablation of structures such as sinus node tissue which can lead to fatal consequences.


Moreover, because of the tortuous nature of intravascular access, devices or mechanisms at the distal end of a catheter positioned within the patient's body, e.g., within a chamber of the heart, are typically no longer aligned with the handle. Steering or manipulation of the distal end of the catheter via control or articulation mechanisms on the handle is easily disorienting to the user as manipulation of a control on the handle in a first direction may articulate the catheter distal end in an unexpected direction depending upon the resulting catheter configuration leaving the user to adjust accordingly. However, this results in reduced efficiency and longer procedure times as well as increased risks to the patient. Accordingly, there is a need for improved catheter control systems which facilitate the manipulation and articulation of a catheter.


SUMMARY OF THE INVENTION

Accordingly, various methods and techniques may be effected to stabilize the images of the tissue when directly visualizing moving regions of tissue, such as the tissue which moves in a beating heart with an imaging assembly positioned in proximity to or against the tissue. Systems and mechanisms are described that can capture and process video images in order to provide a “stabilized” output image and/or create a larger composite image generated from a series of images for the purposes of simplifying the output image for user interpretation during diagnostic and therapeutic procedures.


Typically, images can be captured/recorded by a video camera at a rate of, e.g., 10-100 fps (frames per second), based on the system hardware and software configurations. Much higher video capture rates are also possible in additional variations. The images can then be captured and processed with customizable and/or configurable DSP (digital signal processing) hardware and software at much higher computational speeds (e.g., 1.5-3 kHz as well as relatively slower or faster rates) in order to provide real-time or near real-time analysis of the image data. Additionally, analog signal processing hardware may also be incorporated. A variety of algorithms, e.g., optical flow, image pattern matching, etc. can be used to identify, track and monitor the movement of whole images or features, elements, patterns, and/or structures within the image(s) in order to generate velocity and/or displacement fields that can be utilized by further algorithmic processing to render a more stabilized image. For the imaging assembly, examples of various algorithms which may be utilized may include, e.g., optical flow estimation to compute an approximation to the motion field from time-varying image intensity. Additionally, methods for evaluating motion estimation may also include, e.g., correlation, block matching, feature tracking, energy-based algorithms, as well as, gradient-based approaches, among others.


In some cases, the image frames may be shifted by simple translation and/or rotation and may not contain a significant degree of distortion or other artifacts to greatly simplify the image processing methods and increase overall speed. Alternatively, the hardware and software system can also create a composite image that is comprised (or a combination) of multiple frames during a motion cycle by employing a variety of image stitching algorithms. A graphical feature, e.g., a circle, square, dotted-lines, etc, can be superimposed or overlaid on the composite image in order to indicate the actual position of the camera (image) based on the information obtained from the image tracking software as the camera/hood undergoes a certain excursion, displacement, etc., relative to the target tissue of the organ structure.


An estimate of motion and pixel shifts may also be utilized. For example, a fibrillating heart can achieve 300 bpm (beats per minute), which equals 5 beats per second. Given a video capture rate of 30 fps (frames per second) there would then be roughly 6 frames captured during each beat. Given a typical displacement of, e.g., 1 cm of the camera/hood relative to the plane of the surface of the target tissue per beat, each image may record a displacement of about 1.6 mm per frame. With a field of view (FOV), e.g., of about 7 mm, then each frame may represent an image shift of about 23%. Given an image sensor size of, e.g., 220 pixels×224 pixels, the number of pixels displaced per frame is, e.g., 50 pixels.


Image processing and analysis algorithms may be extremely sensitive to instabilities in, e.g., image intensity, lighting conditions and to variability/instability in the lighting (or image intensity) over the sequence of image frames, as this can interfere with the analysis and/or interpretation of movement within the image. Therefore, mechanisms and methods of carefully controlling the consistency of the lighting conditions may be utilized for ensuring accurate and robust image analysis. Furthermore, mechanisms and methods for highlighting surface features, structures, textures, and/or roughness may also be utilized. For example, a plurality of peripheral light sources, e.g., from flexible light fiber(s), can create even symmetrical illumination or can be tailored to have one or all illuminating sources active or by activating sources near each other in order to provide focused lighting from one edge or possibly alternate the light sources in order to best represent, depict, characterize, highlight features of the tissue, etc. The light source can be configured such that all light sources are from one origin of a given wavelength or the wavelength can be adjusted for each light element. Also, the light bundles can be used to multiplex the light to other different sources so that a given wavelength can be provided at one or more light sources and can be controlled to provide the best feature detection (illumination) and also to provide the most suitable image for feature detection or pattern matching.


As further described herein, light fibers can be located at the periphery of the hood or they can be configured within the hood member. The incidence angle can be tailored such that the reflected light is controlled to minimize glare and other lighting artifacts that could falsely appear as surface features of interest and therefore possibly interfere with the image tracking system. The lighting requirements that provide optimal visual views of the target tissue for the user may vary from the lighting requirements utilized by the software to effectively track features on the target tissue in an automated manner. The lighting conditions can be changed accordingly for different conditions (e.g., direct viewing by the user or under software control) and can be automatically (e.g., software controlled) or manually configurable. Lighting sources could include, e.g., light emitting diodes, lasers, incandescent lights, etc., with a broad spectrum from near-infrared (>760 nm) through the visible light spectrum.


As the camera actively tracks its position relative to the target tissue, the power delivered by the RF generator during ablation may also be controlled as a function of the position of the hood in order to deliver energy to the tissue at a consistent level. In situations where the excursions of the hood/camera occur with varying velocity, the power level may be increased during periods of rapid movements and/or decreased during periods of slower movements such that the average delivery of energy per region/area (per unit time) is roughly constant to minimize regions of incomplete or excessive ablation thus potentially reducing or eliminating damage to surrounding tissue, structures or organs. Alternatively, the tracking of the target tissue may be utilized such that only particular regions in the moving field receive energy whereas other areas in the field receive none (or relatively less energy) by modulating the output power accordingly by effectively gating the power delivery to a location(s) on the target tissue. This technique could ultimately provide higher specificity and focal delivery of ablative energy despite a moving RF electrode system relative to the target tissue.


Active or dynamic control of the hood using control wires, etc., may also be used in order to match/synchronize the excursion of the device with that of the tissue by utilizing surface sensors and/or optical video image to provide feedback to motion control.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A shows a side view of one variation of a tissue imaging apparatus during deployment from a sheath or delivery catheter.



FIG. 1B shows the deployed tissue imaging apparatus of FIG. 1A having an optionally expandable hood or sheath attached to an imaging and/or diagnostic catheter.



FIG. 1C shows an end view of a deployed imaging apparatus.



FIGS. 2A and 2B show one example of a deployed tissue imager positioned against or adjacent to the tissue to be imaged and a flow of fluid, such as saline, displacing blood from within the expandable hood.



FIGS. 3A and 3B show examples of various visualization imagers which may be utilized within or along the imaging hood.



FIGS. 4A and 4B show perspective and end views, respectively, of an imaging hood having at least one layer of a transparent elastomeric membrane over the distal opening of the hood.



FIGS. 5A and 5B show perspective and end views, respectively, of an imaging hood which includes a membrane with an aperture defined therethrough and a plurality of additional openings defined over the membrane surrounding the aperture.



FIGS. 6A and 6B show perspective views of a targeted tissue region which is actively moving and a visualization assembly which is positioned against the tissue while also moving in a corresponding manner to effect a stable image of the moving tissue.



FIG. 7 illustrates a schematic diagram that represents an example of how the imaging output captured from the imager may be transmitted to a processor for processing the captured images as well as other data.



FIG. 8 shows a representative view of a visualization assembly electronically coupled to a video processing unit and an optional image processing unit for providing multiple images of the tissue region.



FIG. 9 depicts a series of images which may be captured during a total excursion of the tissue displacement of the hood relative to the moving tissue.



FIG. 10A shows an example of how an entire video image taken by the imager may be processed to identify one or more sub-sample regions at discrete locations.



FIGS. 10B and 10C show examples of how the identified sub-sample regions may be located in various patterns.



FIG. 11 illustrates how multiple captured images of the underlying tissue region may overlap when taken while the tissue moves relative to the hood over a predetermined time sequence.



FIG. 12 illustrates multiple images captured over an excursion length relative to the tissue region and processed to create a composite image.



FIG. 13 illustrates how a graphical feature such as a positional indicator can be superimposed over the composite image.



FIG. 14 illustrate an example of how the tissue image may be displayed in alternative ways such as the unprocessed image on a first monitor and the stabilized composite image on a second optional monitor.



FIGS. 15A and 15B show examples of alternative views of the tissue images which may be toggled via a switch on a single monitor.



FIGS. 16A and 16B show examples of end views of the hood having multiple sensors positioned upon the membrane.



FIGS. 17A and 17B show partial cross-sectional side views of examples of an optical tracking sensor which may be positioned upon the hood.



FIGS. 18A and 18B show partial cross-sectional side views of another example of an optical tracking sensor having integrated light sources for tracking the tissue images.



FIGS. 19A and 19B illustrate representations of a local coordinate system provided by the sensor and a global coordinate system relative to the tissue region of interest.



FIG. 20 shows an end view of an example of sensors which may be positioned along the hood over or upon the membrane for providing an estimated average of tissue displacement.



FIG. 21 shows another variation where an accelerometer may be mounted to the hood or along the catheter.



FIGS. 22A to 22D show various examples of various configurations for angling light incident upon the tissue surface.



FIGS. 23A and 23B show perspective assembly and end views, respectively, of a hood having multiple optical fibers positioned longitudinally along the hood to provide for angled lighting of the underlying tissue surface from multiple points of emitted light.



FIGS. 24A and 24B side and cross-sectional side views of the hood with a plurality of light fibers illuminating the surface of the target tissue.



FIGS. 25A and 25B show examples of light fibers having its ends further retracted for providing another lighting angle.



FIG. 26 shows illustrates yet another variation where a dark-field optical pathway may be created which highlights surface features of tissue but keeps the background relatively dark.





DETAILED DESCRIPTION OF THE INVENTION

A tissue-imaging and manipulation apparatus described herein is able to provide real-time images in vivo of tissue regions within a body lumen such as a heart, which is filled with blood flowing dynamically therethrough and is also able to provide intravascular tools and instruments for performing various procedures upon the imaged tissue regions. Such an apparatus may be utilized for many procedures, e.g., facilitating transseptal access to the left atrium, cannulating the coronary sinus, diagnosis of valve regurgitation/stenosis, valvuloplasty, atrial appendage closure, arrhythmogenic focus ablation, among other procedures. Although intravascular applications are described, other extravascular approaches or applications may be utilized with the devices and methods herein.


One variation of a tissue access and imaging apparatus is shown in the detail perspective views of FIGS. 1A to 1C. As shown in FIG. 1A, tissue imaging and manipulation assembly 10 may be delivered intravascularly through the patient's body in a low-profile configuration via a delivery catheter or sheath 14. In the case of treating tissue, it is generally desirable to enter or access the left atrium while minimizing trauma to the patient. To non-operatively effect such access, one conventional approach involves puncturing the intra-atrial septum from the right atrial chamber to the left atrial chamber in a procedure commonly called a transseptal procedure or septostomy. For procedures such as percutaneous valve repair and replacement, transseptal access to the left atrial chamber of the heart may allow for larger devices to be introduced into the venous system than can generally be introduced percutaneously into the arterial system.


When the imaging and manipulation assembly 10 is ready to be utilized for imaging tissue, imaging hood 12 may be advanced relative to catheter 14 and deployed from a distal opening of catheter 14, as shown by the arrow. Upon deployment, imaging hood 12 may be unconstrained to expand or open into a deployed imaging configuration, as shown in FIG. 1B. Imaging hood 12 may be fabricated from a variety of pliable or conformable biocompatible material including but not limited to, e.g., polymeric, plastic, or woven materials. One example of a woven material is Kevlar® (E. I. du Pont de Nemours, Wilmington, Del.), which is an aramid and which can be made into thin, e.g., less than 0.001 in., materials which maintain enough integrity for such applications described herein. Moreover, the imaging hood 12 may be fabricated from a translucent or opaque material and in a variety of different colors to optimize or attenuate any reflected lighting from surrounding fluids or structures, i.e., anatomical or mechanical structures or instruments. In either case, imaging hood 12 may be fabricated into a uniform structure or a scaffold-supported structure, in which case a scaffold made of a shape memory alloy, such as Nitinol, or a spring steel, or plastic, etc., may be fabricated and covered with the polymeric, plastic, or woven material. Hence, imaging hood 12 may comprise any of a wide variety of barriers or membrane structures, as may generally be used to localize displacement of blood or the like from a selected volume of a body lumen or heart chamber. In exemplary embodiments, a volume within an inner surface 13 of imaging hood 12 will be significantly less than a volume of the hood 12 between inner surface 13 and outer surface 11.


Imaging hood 12 may be attached at interface 24 to a deployment catheter 16 which may be translated independently of deployment catheter or sheath 14. Attachment of interface 24 may be accomplished through any number of conventional methods. Deployment catheter 16 may define a fluid delivery lumen 18 as well as an imaging lumen 20 within which an optical imaging fiber or assembly may be disposed for imaging tissue. When deployed, imaging hood 12 may expand into any number of shapes, e.g., cylindrical, conical as shown, semi-spherical, etc., provided that an open area or field 26 is defined by imaging hood 12. The open area 26 is the area within which the tissue region of interest may be imaged. Imaging hood 12 may also define an atraumatic contact lip or edge 22 for placement or abutment against the tissue region of interest. Moreover, the diameter of imaging hood 12 at its maximum fully deployed diameter, e.g., at contact lip or edge 22, is typically greater relative to a diameter of the deployment catheter 16 (although a diameter of contact lip or edge 22 may be made to have a smaller or equal diameter of deployment catheter 16). For instance, the contact edge diameter may range anywhere from 1 to 5 times (or even greater, as practicable) a diameter of deployment catheter 16. FIG. 1C shows an end view of the imaging hood 12 in its deployed configuration. Also shown are the contact lip or edge 22 and fluid delivery lumen 18 and imaging lumen 20.


As seen in the example of FIGS. 2A and 2B, deployment catheter 16 may be manipulated to position deployed imaging hood 12 against or near the underlying tissue region of interest to be imaged, in this example a portion of annulus A of mitral valve MV within the left atrial chamber. As the surrounding blood 30 flows around imaging hood 12 and within open area 26 defined within imaging hood 12, as seen in FIG. 2A, the underlying annulus A is obstructed by the opaque blood 30 and is difficult to view through the imaging lumen 20. The translucent fluid 28, such as saline, may then be pumped through fluid delivery lumen 18, intermittently or continuously, until the blood 30 is at least partially, and preferably completely, displaced from within open area 26 by fluid 28, as shown in FIG. 2B.


Although contact edge 22 need not directly contact the underlying tissue, it is at least preferably brought into close proximity to the tissue such that the flow of clear fluid 28 from open area 26 may be maintained to inhibit significant backflow of blood 30 back into open area 26. Contact edge 22 may also be made of a soft elastomeric material such as certain soft grades of silicone or polyurethane, as typically known, to help contact edge 22 conform to an uneven or rough underlying anatomical tissue surface. Once the blood 30 has been displaced from imaging hood 12, an image may then be viewed of the underlying tissue through the clear fluid 30. This image may then be recorded or available for real-time viewing for performing a therapeutic procedure. The positive flow of fluid 28 may be maintained continuously to provide for clear viewing of the underlying tissue. Alternatively, the fluid 28 may be pumped temporarily or sporadically only until a clear view of the tissue is available to be imaged and recorded, at which point the fluid flow 28 may cease and blood 30 may be allowed to seep or flow back into imaging hood 12. This process may be repeated a number of times at the same tissue region or at multiple tissue regions.



FIG. 3A shows a partial cross-sectional view of an example where one or more optical fiber bundles 32 may be positioned within the catheter and within imaging hood 12 to provide direct in-line imaging of the open area within hood 12. FIG. 3B shows another example where an imaging element 34 (e.g., CCD or CMOS electronic imager) may be placed along an interior surface of imaging hood 12 to provide imaging of the open area such that the imaging element 34 is off-axis relative to a longitudinal axis of the hood 12, as described in further detail below. The off-axis position of element 34 may provide for direct visualization and uninhibited access by instruments from the catheter to the underlying tissue during treatment.


In utilizing the imaging hood 12 in any one of the procedures described herein, the hood 12 may have an open field which is uncovered and clear to provide direct tissue contact between the hood interior and the underlying tissue to effect any number of treatments upon the tissue, as described above. Yet in additional variations, imaging hood 12 may utilize other configurations. An additional variation of the imaging hood 12 is shown in the perspective and end views, respectively, of FIGS. 4A and 4B, where imaging hood 12 includes at least one layer of a transparent elastomeric membrane 40 over the distal opening of hood 12. An aperture 42 having a diameter which is less than a diameter of the outer lip of imaging hood 12 may be defined over the center of membrane 40 where a longitudinal axis of the hood intersects the membrane such that the interior of hood 12 remains open and in fluid communication with the environment external to hood 12. Furthermore, aperture 42 may be sized, e.g., between 1 to 2 mm or more in diameter and membrane 40 can be made from any number of transparent elastomers such as silicone, polyurethane, latex, etc. such that contacted tissue may also be visualized through membrane 40 as well as through aperture 42.


Aperture 42 may function generally as a restricting passageway to reduce the rate of fluid out-flow from the hood 12 when the interior of the hood 12 is infused with the clear fluid through which underlying tissue regions may be visualized. Aside from restricting out-flow of clear fluid from within hood 12, aperture 42 may also restrict external surrounding fluids from entering hood 12 too rapidly. The reduction in the rate of fluid out-flow from the hood and blood in-flow into the hood may improve visualization conditions as hood 12 may be more readily filled with transparent fluid rather than being filled by opaque blood which may obstruct direct visualization by the visualization instruments.


Moreover, aperture 42 may be aligned with catheter 16 such that any instruments (e.g., piercing instruments, guidewires, tissue engagers, etc.) that are advanced into the hood interior may directly access the underlying tissue uninhibited or unrestricted for treatment through aperture 42. In other variations wherein aperture 42 may not be aligned with catheter 16, instruments passed through catheter 16 may still access the underlying tissue by simply piercing through membrane 40.


In an additional variation, FIGS. 5A and 5B show perspective and end views, respectively, of imaging hood 12 which includes membrane 40 with aperture 42 defined therethrough, as described above. This variation includes a plurality of additional openings 44 defined over membrane 40 surrounding aperture 42. Additional openings 44 may be uniformly sized, e.g., each less than 1 mm in diameter, to allow for the out-flow of the translucent fluid therethrough when in contact against the tissue surface. Moreover, although openings 44 are illustrated as uniform in size, the openings may be varied in size and their placement may also be non-uniform or random over membrane 40 rather than uniformly positioned about aperture 42 in FIG. 5B. Furthermore, there are eight openings 44 shown in the figures although fewer than eight or more than eight openings 44 may also be utilized over membrane 40.


Additional details of tissue imaging and manipulation systems and methods which may be utilized with apparatus and methods described herein are further described, for example, in U.S. patent application Ser. No. 11/259,498 filed Oct. 25, 2005 (U.S. Pat. No. 7,860,555); Ser. No. 11/763,399 filed Jun. 14, 2007 (U.S. Pub. 2007/0293724); and Ser. No. 12/118,439 filed May 9, 2008 (U.S. Pub. 2009/0030412), each of which is incorporated herein by reference in its entirety.


In utilizing the devices and methods above, various procedures may be accomplished. One example of such a procedure is crossing a tissue region such as in a transseptal procedure where a septal wall is pierced and traversed, e.g., crossing from a right atrial chamber to a left atrial chamber in a heart of a subject. Generally, in piercing and traversing a septal wall, the visualization and treatment devices described herein may be utilized for visualizing the tissue region to be pierced as well as monitoring the piercing and access through the tissue. Details of transseptal visualization catheters and methods for transseptal access which may be utilized with the apparatus and methods described herein are described in U.S. patent application Ser. No. 11/763,399 filed Jun. 14, 2007 (U.S. Pat. Pub. 2007/0293724 A1), incorporated herein by reference above. Additionally, details of tissue visualization and manipulation catheter which may be utilized with apparatus and methods described herein are described in U.S. patent application Ser. No. 11/259,498 filed Oct. 25, 2005 (U.S. Pat. Pub. 2006/0184048 A1), also incorporated herein above.


When visualizing moving regions of tissue, such as the tissue which moves in a beating heart, the relative movement between the imaging assembly in the device and the tissue region may result in tissue images which are difficult to capture accurately. Accordingly, various methods and techniques may be effected to stabilize the images of the tissue, e.g., (1) mechanical stabilization of the hood-camera assembly under dynamic control using imaging from the camera or surface sensors or both; (2) software stabilization (algorithmic based stabilization via pattern matching, feature detection, optical flow, etc); (3) sensor-based tracking of excursion of the assembly relative to the tissue (surface optical sensor, accelerometer, EnSite NavX® (St. Jude Medical, Minn.), Carto® Navigation System (Biosense Webster, Calif.), etc.) as feedback into an algorithm; and/or (4) signal feedback of biological functions into an algorithm (EKG, respiration, etc).


Systems and mechanisms are described that can capture and process video images in order to provide a “stabilized” output image and/or create a larger composite image generated from a series of images for the purposes of simplifying the output image for user interpretation during diagnostic and therapeutic procedures.


Typically, images can be captured/recorded by a video camera at a rate of, e.g., 10-100 fps (frames per second), based on the system hardware and software configurations. Much higher video capture rates are also possible in additional variations. The images can then be captured and processed with customizable and/or configurable DSP (digital signal processing) hardware and software at much higher computational speeds (e.g., 1.5-3 kHz as well as relatively slower or faster rates) in order to provide real-time or near real-time analysis of the image data. Additionally, analog signal processing hardware may also be incorporated.


A variety of algorithms, e.g., optical flow, image pattern matching, etc. can be used to identify, track and monitor the movement of whole images or features, elements, patterns, and/or structures within the image(s) in order to generate velocity and/or displacement fields that can be utilized by further algorithmic processing to render a more stabilized image. For example, see B. Horn and B. Schunck. Determining optical flow. Artificial Intelligence, 16(1-3):185-203, August 1981. and B. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In IJCAI81, pages 674-679, 1981. and J. Shin, S. Kim, S. Kang, S.-W. Lee, J. Paik, B. Abidi, and M. Abidi. Optical flow-based real-time object tracking using non-prior training active feature model. Real-Time Imaging, 11(3):204-218, June 2005 and J. Barron, D. Fleet, S. Beauchemin. Performance of Optical Flow Techniques. International Journal of Computer Vision, 12 (1):43-77, 1994, D. Fleet, Y. Weiss. Optical Flow Estimation. Handbook of Mathematical Models in Computer Vision. (Editors: N. Paragios, et al.). Pages. 239-258, 2005.). Each of these references is incorporated herein by reference in its entirety. For the imaging assembly, examples of various algorithms which may be utilized may include, e.g., optical flow estimation to compute an approximation to the motion field from time-varying image intensity. Additionally, methods for evaluating motion estimation may also include, e.g., correlation, block matching, feature tracking, energy-based algorithms, as well as, gradient-based approaches, among others.


In some cases, the image frames may be shifted by simple translation and/or rotation and may not contain a significant degree of distortion or other artifacts to greatly simplify the image processing methods and increase overall speed. Alternatively, the hardware and software system can also create a composite image that is comprised (or a combination) of multiple frames during a motion cycle by employing a variety of image stitching algorithms, also known in the art (see e.g., R. Szeliski. Image Alignment and Stitching. A Tutorial. Handbook of Mathematical Models in Computer Vision. (Editors: N. Paragios, et al.). Pages 273-292, 2005), which is incorporated herein by reference in its entirety. A graphical feature, e.g., a circle, square, dotted-lines, etc, can be superimposed or overlaid on the composite image in order to indicate the actual position of the camera (image) based on the information obtained from the image tracking software as the camera/hood undergoes a certain excursion, displacement, etc., relative to the target tissue of the organ structure.


An estimate of motion and pixel shifts may also be utilized. For example, a fibrillating heart can achieve 300 bpm (beats per minute), which equals 5 beats per second. Given a video capture rate of 30 fps (frames per second) there would then be roughly 6 frames captured during each beat. Given a typical displacement of, e.g., 1 cm of the camera/hood relative to the plane of the surface of the target tissue per beat, each image may record a displacement of about 1.6 mm per frame. With a field of view (FOV), e.g., of about 7 mm, then each frame may represent an image shift of about 23%. Given an image sensor size of, e.g., 220 pixels×224 pixels, the number of pixels displaced per frame is, e.g., 50 pixels.


Image processing and analysis algorithms may be extremely sensitive to instabilities in, e.g., image intensity, lighting conditions and to variability/instability in the lighting (or image intensity) over the sequence of image frames, as this can interfere with the analysis and/or interpretation of movement within the image. Therefore, mechanisms and methods of carefully controlling the consistency of the lighting conditions may be utilized for ensuring accurate and robust image analysis. Furthermore, mechanisms and methods for highlighting surface features, structures, textures, and/or roughness may also be utilized. For example, a plurality of peripheral light sources, e.g., from flexible light fiber(s), or individual light emitting diodes (LEDs) can create even symmetrical illumination or can be tailored to have one or all illuminating sources active or by activating sources near each other in order to provide focused lighting from one edge or possibly alternate the light sources in order to best represent, depict, characterize, highlight features of the tissue, etc. The light source can be configured such that all light sources are from one origin and of a given wavelength or the wavelength can be adjusted for each light element. Also, the light bundles can be used to multiplex the light to other different sources so that a given wavelength can be provided at one or more light sources and can be controlled to provide the best feature detection (illumination) and also to provide the most suitable image for feature detection or pattern matching.


As further described herein, light fibers can be located at the periphery of the hood or they can be configured within the hood member. The incidence angle can be tailored such that the reflected light is controlled to minimize glare and other lighting artifacts that could falsely appear as surface features of interest and therefore possibly interfere with the image tracking system. The lighting requirements that provide optimal visual views of the target tissue for the user may vary from the lighting requirements utilized by the software to effectively track features on the target tissue in an automated manner. The lighting conditions can be changed accordingly for different conditions (e.g., direct viewing by the user or under software control) and can be automatically (e.g., software controlled) or manually configurable. Lighting sources could include, e.g., light emitting diodes, lasers, incandescent lights, etc., with a broad spectrum from near-infrared (>650 nm) through the visible light spectrum.


As the camera actively tracks its position relative to the target tissue, the power delivered by the RF generator during ablation may also be controlled as a function of the position of the hood in order to deliver energy to the tissue at a consistent level. In situations where the excursions of the hood/camera occur with varying velocity, the power level may be increased during periods of rapid movements and/or decreased during periods of slower movements such that the average delivery of energy per region/area (per unit time) is roughly constant to minimize regions of incomplete or excessive ablation thus potentially reducing or eliminating damage to surrounding tissue, structures or organs. Alternatively, the tracking of the target tissue may be utilized such that only particular regions in the moving field receive energy whereas other areas in the field receive none (or relatively less energy) by modulating the output power accordingly by effectively gating the power delivery to a location(s) on the target tissue. This technique could ultimately provide higher specificity and focal delivery of ablative energy despite a moving RF electrode system relative to the target tissue.


Active or dynamic control of the hood using control wires, etc., may also be used in order to match/synchronize the excursion of the device with that of the tissue by utilizing surface sensors and/or optical video image to provide feedback to motion control.


Turning now to FIG. 6A, a perspective view of a typical section of target tissue T from an organ that is actively moving, either due to contractility (such as in a heart), respiration, peristalsis, or other tissue motion, as depicted by arrows in the plane of the tissue, which can include lateral tissue movement 52 or rotational tissue movement 54. The target tissue T can also experience displacement along an axis normal to the plane of the tissue. The hood 12 of the visualization assembly may be seen in proximity to the tissue region T where the hood 12 may be positioned upon an articulatable section 50 of the catheter 16.


As shown in the perspective view of FIG. 6B, any movement of the underlying tissue region T may be matched by the catheter assembly by moving the hood 12 in a manner which corresponds to the movement of the tissue T. For instance, any axial displacement 58 of the hood 12 corresponding to any out-of-plane movement of the tissue region T may be achieved by advancing and retracting the catheter 16. Similarly, any corresponding lateral catheter movement 56 of the hood 12 may be accomplished by articulating the steerable section 50 the catheter 16 to match correspondingly to the lateral movements 52 of the tissue T. Steerable section 50 may be manually articulated to match movements of the hood 12 with that of the tissue. Alternatively, computer control of the steerable section 50 may be utilized to move the hood 12 accordingly. Examples of steering and control mechanisms which may be utilized with the devices and methods disclosed herein may be seen in U.S. patent application Ser. No. 11/848,429 filed Aug. 31, 2007 (U.S. Pub. 2008/0097476) which shows and describes computer-controlled articulation and steering. Other examples include U.S. patent application Ser. No. 12/108,812 filed Apr. 24, 2008 (U.S. Pub. 2008/0275300) which shows and describes multiple independently articulatable sections; U.S. patent application Ser. No. 12/117,655 filed May 8, 2008 (U.S. Pub. 2008/0281293) which shows and describes alternative steering mechanisms; U.S. patent application Ser. No. 12/499,011 filed Jul. 7, 2009 (U.S. Pub. 2010/0004633); and Ser. No. 12/967,288 filed Dec. 14, 2010 which show and describe steering control mechanisms and handles. Each of these references are incorporated herein by reference in its entirety.



FIG. 7 illustrates a schematic diagram that represents an example of how the imaging output captured from the imager 60 (as described above) within or adjacent to the hood 12 may be transmitted to a processor 62 that may comprise imaging processing hardware as well as software algorithms for processing the captured images as well as other data. For instance, processor 62 may also receive signals from one or more sensors 64 located along or in proximity to the hood 12 which sense and detect positional information of the hood 12 and/or catheter 16. Such sensors may include, e.g., surface optical tracking sensors, positional information received from a NavX® system, Carto® Navigation System (Biosense Webster, Calif.), accelerometers, etc. Processor 62 may also receive biological signals or physiological data 66 detected by one or more sensors also located along or in proximity to the hood 12 or from other internal or external inputs, e.g., electrocardiogram data, respiration, etc.


With the processor 62 programmed to receive and process both the positional information from one or more sensor signals 64 as well as physiological information of the subject from one or more biological signals 66, the processor 62 may optionally display one or more types of images. For example, a composite image 70 may be processed and displayed where the image represents a composite image that is combined, stitched-together, or comprised of multiple images taken over the excursion distance captured by the imager 60 during relative movement between the tissue and hood 12. Additionally and/or alternatively, a stabilized composite image 68 may be displayed which represents an image where the motion due to tissue displacement is reduced, minimized, or eliminated and a single view of an “average” image. Additionally and/or alternatively, a raw video image 72 may be displayed as well which shows the unprocessed image captured by the imager 60. Each of these different types of images may be displayed individually or simultaneously on different screens or different portions of a screen if so desired.



FIG. 8 depicts the hood 12 and catheter 16 imaging a tissue region T which may undergo tissue displacement, e.g., within three degrees of freedom (linear displacements 52 and rotational displacement 54). The captured images may be transmitted to a video processing unit 80 (VPU) which may process the raw video images 72 for display upon a monitor. The VPU 80 may further transmit its images to an image processing unit 82 which may contain the processor 62 and algorithm for stabilizing and/or stitching together a composite image. The stabilized composite or averaged image 68 captured from tissue regions A, B, C during tissue movement can be viewed on a second monitor. Optionally, the processing function of image processing unit 82 can also be turned off or toggled, e.g., via a switch 84, for use on one or more monitors to enable the physician to select the type of image for viewing. Additionally, although a typical target tissue T is described, the imaging angle of the tissue may change relative to the hood 12 during image capture of the tissue region T due to the relative tissue movement. Because of this changing image, images of the tissue T at various stages at a particular given time interval may be displayed if so desired for comparison.


As an example of the range of images the imager within or adjacent to the hood 12 may capture during relative tissue movement, FIG. 9 depicts a series of images which may be captured during the total excursion L of the tissue displacement of the hood 12 relative to the moving tissue. The image represented by the static field of view 90 through hood 12 shows an initial position of hood 12 relative to the tissue. As the tissue moves relative to the hood 12, the sampling rate or frame rate of the imager may be sufficiently high enough such that as the tissue moves relative to the hood, each subsequent imaged tissue regions in a first direction 92′, 92″ may be captured at increments of a distance d which may allow for overlapping regions of the tissue to be captured. Likewise, the imaged tissue regions in a second direction 94′, 94″ may be captured as well at overlapping increments of distance d between each captured image.


In processing the captured images to provide a stabilized or composite tissue image for display, one example is illustrated in FIG. 10A which shows an example of how an entire video image 100 (e.g., 220×224) can be taken by the imager and then processed by the processor 62 to identify one or more sub-sample regions 102 (e.g., 10×10 pixel subsets up to 18×18 pixel subsets) at discrete locations. These identified sub-sample regions 102 may be identified consistently between each subsequent image taken for mapping and/or stitching purposes between the subsequent images. The sub-sample regions 102 may range between, e.g., 1-20 subsets, and may be oriented in various patterns to best match or be removed from obstructions (e.g., various hood features, etc.) for the most robust image quality and processing. FIG. 10B shows an example of how the identified sub-sample regions 102′ may be located in a first staggered pattern while FIG. 10C shows another example of how the sub-sampled regions 102″ can be identified along the corners of the image 100. The location of the sub-sampled regions are illustrated for exemplary purposes and other locations over the image 100 may be identified as needed or desired.


With the sub-sample regions identified, FIG. 11 depicts how the multiple captured images 110 (e.g., F1, F2, F3, etc.) of the underlying tissue region may overlap when taken while the tissue moves relative to the hood 12, e.g., at 1 cm increments, over a predetermined time sequence. Subsequent images may be automatically compared 112 by the processor 62 for comparison from one image to the other utilizing the identified sub-sample regions between each image. An example of such a comparison between two subsequent images is shown, e.g., between frame F1 to frame F2 and between frame F2 and frame F3, etc. In this manner, multiple images may be compared between each sampled frame.


Another example is shown in FIG. 12 depicting multiple images (e.g., F1 to Fn) which are captured over an excursion length L of the hood 12 relative to the tissue region. While the images are captured and tracked over a time period, the imaging processing unit 82 may receive these multiple images and process the images to create a composite image 120 stitched from each individual captured frame in order to simplify viewing by the user. With this composite image 120, which can be continuously updated and maintained as the tissue T and/or hood 12 moves relative to one another, a graphical feature such as a positional indicator 122, e.g. a circle, can be superimposed over the composite image 120 at a first location for display to the user to depict the actual position of the hood 12 in real time relative to the tissue T, as shown in FIG. 13. Positional indicator 122′ is also illustrated at a second location along the tissue T. Use of the positional indicator for display may be incorporated to depict where the hood 12 is in real time relative to the tissue T, or more specifically, where hood 12 is relative to anatomical markers or features of interest, especially when using ablation energy.


As previously described and as shown in FIG. 14, the tissue image may be displayed in alternative ways, such as the unprocessed image 72 on a first monitor 130 and the stabilized composite image 68 on a second optional monitor 132. Alternatively, a single monitor 130 may be used with a switch 134 electrically coupled to the imaging processing unit 82. Switch 134 may allow the user to compare one image with the other and decide which one is best suited for the task at hand, e.g., gaining access to the target region, exploring within the target region, or to stabilize and track the target tissue in order to commence with ablation or other therapy. Thus, the user may use the switch 134 to toggle between the unprocessed image view 72, as shown in FIG. 15A, and the stabilized composite image 68 on a single monitor 130, as shown in FIG. 15B.


In an alternative variation, rather than using the imager within or adjacent to the hood 12 for tracking and sampling the images of the underlying tissue, one or more individual sensors (as previously mentioned) may be positioned along or upon the hood 12 such as along the membrane 40 in proximity to aperture 42. As shown in the end view of hood 12 in FIG. 16A, a first optical displacement sensor 140 is shown positioned upon membrane 40 with a second sensor 142 positioned on an opposite side of membrane 40 relative to first sensor 140. Another variation is shown in the end view of FIG. 16B which shows membrane 40 with the addition of a third sensor 144 and fourth sensor 146 positioned along membrane 40 uniformly relative to one another around aperture 42. In other variations, a single sensor may be used or any number of sensors greater than one may be utilized as practicable.


In use, multiple sensors may provide for multiple readings to increase the accuracy of the sensors and the displacements of the hood 12 may be directly tracked with the sensor-based modality like an optical displacement sensor such as those found in optical computer mice directly mounted to the face of the hood 12 (or some feature/portion of the hood 12). Additional sensors may be able to provide a more robust reading, especially if one sensor is reading incorrectly due to poor contact or interference from blood. Because the hood 12 is deformable, the relative position of each sensor relative to each other may be independent of one another, thus, the detected values may be averaged or any accumulated errors may be limited.



FIG. 17A shows a partial cross-sectional side view of one example of an optical tracking sensor 150 which may be positioned upon the hood 12 as described above. Sensor 150 may generally comprise a light source such as an optical fiber 152 optically coupled to a light source and having a distal end which may be angled with respect to the sensor 150 from which transmitted light 154 may be emitted such that the transmitted light 154 is incident upon the tissue T at an angle to create a side-lighting effect which may highlight or exaggerate surface features. FIG. 17B shows a partial cross-sectional side view of another variation where sensor 150 may also include a detector 156 having various sizes (e.g., 16×16 pixels, 18×18 pixels, or any other detector size) as well as a lens 158. Moreover, although a single optical fiber 152 is shown, multiple fibers may be utilized in other variations.



FIG. 18A shows a cross-sectional side view of another variation of an optical displacement sensor 160 having an integrated light source 162, e.g., light emitting diode (LED), laser, etc., rather than using a light source removed from the sensor 160. The integrated light source 162 may be angled (or angled via a lens or reflector) such that the emitted light 164 incident upon the tissue T emerges at an angle relative to the tissue surface. An example of an integrated light source 162 (such as an LED) similar to a configuration of an optical mouse sensor is shown in the cross-sectional side view of FIG. 18B for integration into sensor 160. As shown, the integrated light source 162 may be positioned within the sensor and a light pipe 166 may be positioned in proximity to the light source 162 to direct the emitted light through the sensor such that the emitted light 164 emerge at an angle with respect to the imaged tissue T to highlight/exaggerate surface features. The light reflected from the tissue may be reflected through lens 158 which may direct the reflected light onto detector 156. The detected image may then be transmitted to the processor 62, as previously described.


With respect to the use of one or more positional sensors, such sensors may be mounted upon or along the hood 12 and/or catheter 16 to calculate a position of the hood 12. Additional examples of positional sensors which may be utilized with the systems and methods described herein are shown and described in further detail in U.S. patent application Ser. No. 11/848,532 filed Aug. 31, 2007 (U.S. Pub. 2009/0054803), which is incorporated herein by reference in its entirety. An example is shown FIG. 19A which illustrates a representation of a local coordinate system 170 for providing three degrees-of-freedom (e.g., x, y, θ) and a global coordinate system 172 referenced with respect to a predetermined grid, marker, pattern, etc. such as from tissue fiber images, tissue texture, etc. The sensor 140 positioned upon or along the hood 12 and/or catheter 16 may provide for proper sensing via local information generated relative to the sensor 140 itself (rather than global information) which may be calculated such that in the event the tracking is lost, the system can simply wait for a subsequent cardiac cycle (or tissue movement) to recalculate the excursion again as the tracking would only be active when the image is stabilized and the catheter control is well maintained. The local coordinate system 174 may for provide two degrees-of-freedom (e.g., x, y) referenced within the sensor 140 since the light source and sensing occurs from the same reference point.


Alternatively, the global coordinate system 172 may be utilized relative to the tissue region. If imaging of the tissue surface does not provide sufficient “markers” to track then alternative methods of providing fiducial markers may be utilized, e.g., sensor may calculate axial displacements (two degrees-of-freedom) and possibly at least one rotational degree-of-freedom to provide a rotational component. FIG. 19B shows another variation where only axial displacements are detected to provide a more robust sensing due to the relatively fewer pixel representations of the surface.


An example of sensors 140, 142 which may be positioned along the hood 12 over or upon the membrane 40 in proximity to aperture 42 is shown in the end view of FIG. 20. These positional sensors 140, 142 may be independent or integrated with the sensors used for imaging. As shown, the one or more sensors may be used individually for detecting a linear translation of the hood 12 where the displacement of the entire hood 12 may be estimated based upon an averaging of the translation of each individual sensor where (y1+y2)/2=yhood and (x1+x2)/2=xhood. An angle of rotation may also be estimated by utilizing the rotation of each individual sensor where the overall rotation of the hood 12 is a function of the displacements of each sensor 140, 142. In this example, the overall estimated rotation of the hood, Θhood=f ([x1, y1], [x2, y2]) and the overall displacement may be determined as an average between the displacements of each sensor 140, 142 where yhood=(y1+y2)/2 and xhood=(x1+x2)/2. The use of the two sensors is illustrated as an example and more than two sensors may optionally be utilized. Moreover, the positioning of the sensors may be varied as well and are not limited to positioning upon the membrane 40.



FIG. 21 shows another variation where an accelerometer 180 may be mounted to the hood 12 or along catheter 16 and may be used to provide a mechanism of gating or timing the excursions or at least the change in directions of detected movements 182. This variation may be incorporated as another input to the tracking algorithm by the processor.


Turning back to the emission of an angled light incident upon the tissue surface, as previously mentioned the emitted light for surface detection may be angled relative to the sensor as well as relative to the tissue. An example is illustrated in the representative assembly of FIG. 22A, which shows an optical fiber 152 positioned at an angle such that the emitted light 154 is incident upon the tissue T at an angle, Θ. The angled light may illuminate surface features or details along the tissue surface that can be used for tracking. Low-angled (relative to the surface) side lighting provided by optical fiber 152 coupled to light source 194 may be used or an LED, laser, or other lighting elements could also be utilized along the hood 12 in proximity to the tissue surface to achieve similar low angled lighting to provide improved detail to small surface features. Alternately, the light source can be comprised of one or more LEDs mounted directly on or in the hood. The reflected light may be captured within the field of view 192 of imaging element 190 positioned within or adjacent to hood 12, as shown.



FIG. 22B shows a variation where direct axial illumination, as shown, may be used where one or more optical fibers 152, LEDs, or other light source may be positioned so as to emit light axially relative to the imaging element 190. In this variation, the light is incident upon the tissue surface perpendicularly. FIG. 22C depicts another variation where a shallower angle of illumination can help highlight the features of the tissue surface in the captured images. FIG. 22D also illustrates another variation where multiple light sources 150, 150′ may be used for emitting light 154, 154′ at multiple angled positions for depicting tissue surface features at multiple angles.



FIG. 23A shows a perspective view of an assembly having a hood 12 with multiple optical fibers positioned longitudinally along the hood to provide for angled lighting of the underlying tissue surface from multiple points of emitted light. As illustrated also in the end view of hood 12 in FIG. 23B, the terminal emitting ends 202 of multiple optical fibers 200 (which may be each optically coupled to a common light source 194 or each individually to separate light sources) may be positioned circumferentially around the hood 12 such that the ends 202 are directed towards the aperture 42. In this manner, the emitted light may converge on the aperture 42 (or any region of the tissue surface being visualized) to provide shallow angle edge lighting. The number of fibers can range anywhere from, e.g., 1 to 50 or more, and can be symmetrically or asymmetrically distribute about the hood face. Alternatively, optical fibers 200 may be replaced by individual LEDs.



FIG. 24A illustrates a side view of hood 12 with a plurality of light fibers 200 on the surface of the target tissue T and FIG. 24B shows a cross-sectional side view of FIG. 24A illustrating the illumination angle, Θ, of the incident light 204 which may penetrate the tissue T to a certain depth depending on the light intensity and wavelength.


Another variation is shown in the cross-sectional side view of FIG. 25A showing optical fibers 200, LEDs, or other light sources along hood 12 with a relatively more retracted position of its ends 202 relative to the tissue surface to provide an alternative lighting scheme and to also enable to the light fibers to be shorter and not be subjected to the sharp bend and motion of the hood edge. As shown, the light fiber may have a relatively shallower tip angle, Θ, to redirect the light inward. FIG. 25B similarly depicts a hood 12 with light fibers 200 having its ends 202 which are further retracted for providing another lighting angle, Θ, as well as providing shorter light fibers that may be less prone to damage during hood manipulation or repeated deployments and retraction from within the delivery sheath.



FIG. 26 illustrates yet another variation where a dark-field optical pathway may be created which highlights surface features of tissue T but keeps the background relatively dark. Segmentation of the image may be effected to provide better detail in the raw image for further processing rather than trying to process an image (using thresholding, segmenting, etc.) with poorly defined detail and features. Typically, a higher quality image can be easier to process rather than applying powerful image processing algorithms to a poor quality image. In particular, the substrate may be relatively translucent or dark. In the case of tissue, it may be that the incident light spectrum can “penetrate” the tissue with little reflection and scatter, yet help highlight surface features or texture. This may be utilized with relatively thin or somewhat translucent tissue types or thin membranes.


The applications of the disclosed invention discussed above are not limited to certain treatments or regions of the body, but may include any number of other applications as well. Modification of the above-described methods and devices for carrying out the invention, and variations of aspects of the invention that are obvious to those of skill in the arts are intended to be within the scope of this disclosure. Moreover, various combinations of aspects between examples are also contemplated and are considered to be within the scope of this disclosure as well.

Claims
  • 1. A method for imaging a tissue region in motion, comprising: intravascularly advancing a deployment catheter towards the tissue region;expanding a non-inflatable barrier or membrane defining an open area and projecting distally from the deployment catheter against the tissue region;purging the tissue region of blood contained within the open area via a transparent fluid while the tissue region is moving, wherein the open area is in fluid communication through an opening with a blood-filled environment external to the barrier or membrane;visualizing the tissue region via a visualization element through the transparent fluid;sensing a displacement of the barrier or membrane relative to the tissue region with a plurality of sensors during tissue movement;capturing via the visualization element a plurality of images of a tissue region over a plurality of corresponding time intervals while the tissue region is moving; and,displaying a composite image of the plurality of images.
  • 2. The method of claim 1 wherein purging comprises introducing saline through the deployment catheter.
  • 3. The method of claim 1 wherein sensing a displacement comprises detecting lateral and/or rotational tissue movement relative to the barrier or membrane.
  • 4. The method of claim 1 wherein capturing via the visualization element comprises imaging the tissue region via an electronic imager positioned within or adjacent to the open area.
  • 5. The method of claim 1 wherein capturing via the visualization element comprises receiving positional data of the barrier or membrane from one or more sensors.
  • 6. The method of claim 1 further comprising receiving physiological data of a subject while capturing the plurality of images.
  • 7. The method of claim 1 wherein displaying a composite image comprises displaying an image stitched from the plurality of images captured during movement of the barrier or membrane relative to the tissue region.
  • 8. The method of claim 7 further comprising displaying a stabilized image of the tissue region averaged from the plurality of images.
  • 9. The method of claim 1 further comprising superimposing a graphical feature upon the composite image where the graphical feature is indicative of a position of the barrier or membrane relative to the tissue region.
  • 10. The method of claim 1 wherein capturing a plurality of images comprises illuminating the tissue region via at least one light source angled with respect to the tissue region.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of priority to U.S. Provisional Application 61/304,235 filed Feb. 12, 2010, which is incorporated herein by reference in its entirety.

US Referenced Citations (549)
Number Name Date Kind
623022 Johnson Apr 1899 A
2305462 Wolf Dec 1942 A
2453862 Peter Nov 1948 A
3559651 Moss Feb 1971 A
3874388 King et al. Apr 1975 A
3903877 Terada Sep 1975 A
4175545 Termanini Nov 1979 A
4326529 Doss et al. Apr 1982 A
4445892 Hussein et al. May 1984 A
4470407 Hussein et al. Sep 1984 A
4517976 Murakoshi et al. May 1985 A
4569335 Tsuno Feb 1986 A
4576146 Kawazoe et al. Mar 1986 A
4615333 Taguchi Oct 1986 A
4619247 Inoue et al. Oct 1986 A
4676258 Inokuchi et al. Jun 1987 A
4681093 Ono et al. Jul 1987 A
4709698 Johnston et al. Dec 1987 A
4710192 Liotta et al. Dec 1987 A
4727418 Kato et al. Feb 1988 A
4784133 Mackin Nov 1988 A
4838246 Hahn et al. Jun 1989 A
4848323 Marijnissen et al. Jul 1989 A
4911148 Sosnowski et al. Mar 1990 A
4914521 Adair Apr 1990 A
4943290 Rexroth et al. Jul 1990 A
4950285 Wilk Aug 1990 A
4957484 Murtfeldt Sep 1990 A
4961738 Mackin Oct 1990 A
4976710 Mackin Dec 1990 A
4991578 Cohen Feb 1991 A
4994069 Ritchart et al. Feb 1991 A
4998916 Hammerslag et al. Mar 1991 A
4998972 Chin et al. Mar 1991 A
5047028 Qian Sep 1991 A
5057106 Kasevich et al. Oct 1991 A
5090959 Samson et al. Feb 1992 A
5123428 Schwarz Jun 1992 A
RE34002 Adair Jul 1992 E
5156141 Krebs et al. Oct 1992 A
5171259 Inoue Dec 1992 A
5281238 Chin et al. Jan 1994 A
5282827 Kensey et al. Feb 1994 A
5306234 Johnson Apr 1994 A
5313943 Houser et al. May 1994 A
5330496 Alferness Jul 1994 A
5334159 Turkel Aug 1994 A
5334193 Nardella Aug 1994 A
5336252 Cohen Aug 1994 A
5339800 Wilta et al. Aug 1994 A
5348554 Imran et al. Sep 1994 A
5353792 Lubbers et al. Oct 1994 A
5370647 Graber et al. Dec 1994 A
5373840 Knighton Dec 1994 A
5375612 Cottenceau et al. Dec 1994 A
5385148 Lesh et al. Jan 1995 A
5391182 Chin Feb 1995 A
5403326 Harrison et al. Apr 1995 A
5405376 Mulier et al. Apr 1995 A
5421338 Crowley et al. Jun 1995 A
5431649 Mulier et al. Jul 1995 A
5453785 Lenhardt et al. Sep 1995 A
5462521 Brucker et al. Oct 1995 A
5471515 Fossum et al. Nov 1995 A
5498230 Adair Mar 1996 A
5505730 Edwards Apr 1996 A
5515853 Smith et al. May 1996 A
5527338 Purdy Jun 1996 A
5549603 Feiring Aug 1996 A
5558619 Kami et al. Sep 1996 A
5571088 Lennox et al. Nov 1996 A
5575756 Karasawa et al. Nov 1996 A
5575810 Swanson et al. Nov 1996 A
5584872 LaFontaine et al. Dec 1996 A
5591119 Adair Jan 1997 A
5593422 Muijs Van de Moer et al. Jan 1997 A
5593424 Northrup, III Jan 1997 A
5672153 Lax et al. Sep 1997 A
5676693 LaFontaine Oct 1997 A
5681308 Edwards et al. Oct 1997 A
5695448 Kimura et al. Dec 1997 A
5697281 Eggers et al. Dec 1997 A
5697882 Eggers et al. Dec 1997 A
5709224 Behl et al. Jan 1998 A
5713907 Hogendijk et al. Feb 1998 A
5713946 Ben-Haim Feb 1998 A
5716321 Kerin et al. Feb 1998 A
5722403 McGee et al. Mar 1998 A
5725523 Mueller Mar 1998 A
5746747 McKeating May 1998 A
5749846 Edwards et al. May 1998 A
5749890 Shaknovich May 1998 A
5754313 Pelchy et al. May 1998 A
5766137 Omata Jun 1998 A
5769846 Edwards et al. Jun 1998 A
5792045 Adair Aug 1998 A
5797903 Swanson et al. Aug 1998 A
5823947 Yoon et al. Oct 1998 A
5827268 Laufer Oct 1998 A
5829447 Stevens et al. Nov 1998 A
5842973 Bullard Dec 1998 A
5843118 Sepetka et al. Dec 1998 A
5848969 Panescu et al. Dec 1998 A
5860974 Abele Jan 1999 A
5860991 Klein et al. Jan 1999 A
5865791 Whayne et al. Feb 1999 A
5873815 Kerin et al. Feb 1999 A
5879366 Shaw et al. Mar 1999 A
5895417 Pomeranz et al. Apr 1999 A
5897487 Ouchi Apr 1999 A
5897553 Mulier et al. Apr 1999 A
5902328 LaFontaine et al. May 1999 A
5904651 Swanson et al. May 1999 A
5908445 Whayne et al. Jun 1999 A
5925038 Panescu et al. Jul 1999 A
5928250 Koike et al. Jul 1999 A
5929901 Adair et al. Jul 1999 A
5941845 Tu et al. Aug 1999 A
5944690 Falwell et al. Aug 1999 A
5964755 Edwards Oct 1999 A
5968053 Revelas Oct 1999 A
5971983 Lesh Oct 1999 A
5986693 Adair et al. Nov 1999 A
5997571 Farr et al. Dec 1999 A
6004269 Crowley et al. Dec 1999 A
6012457 Lesh Jan 2000 A
6024740 Lesh et al. Feb 2000 A
6027501 Goble et al. Feb 2000 A
6036685 Mueller Mar 2000 A
6043839 Adair et al. Mar 2000 A
6047218 Whayne et al. Apr 2000 A
6063077 Schaer May 2000 A
6063081 Mulier et al. May 2000 A
6068653 LaFontaine May 2000 A
6071279 Whayne et al. Jun 2000 A
6071302 Sinofsky et al. Jun 2000 A
6081740 Gombrich et al. Jun 2000 A
6086528 Adair Jul 2000 A
6086534 Kesten Jul 2000 A
6099498 Addis Aug 2000 A
6099514 Sharkey et al. Aug 2000 A
6102905 Baxter et al. Aug 2000 A
6112123 Kelleher et al. Aug 2000 A
6115626 Whayne et al. Sep 2000 A
6123703 Tu et al. Sep 2000 A
6123718 Tu et al. Sep 2000 A
6129724 Fleischman et al. Oct 2000 A
6139508 Simpson et al. Oct 2000 A
6142993 Whayne et al. Nov 2000 A
6152144 Lesh et al. Nov 2000 A
6156350 Constantz Dec 2000 A
6159203 Sinofsky Dec 2000 A
6161543 Cox et al. Dec 2000 A
6164283 Lesh Dec 2000 A
6167297 Benaron Dec 2000 A
6168591 Sinofsky Jan 2001 B1
6168594 LaFontaine et al. Jan 2001 B1
6174307 Daniel et al. Jan 2001 B1
6178346 Amundson et al. Jan 2001 B1
6190381 Olsen et al. Feb 2001 B1
6211904 Adair et al. Apr 2001 B1
6224553 Nevo May 2001 B1
6231561 Frazier et al. May 2001 B1
6234995 Peacock, III May 2001 B1
6235044 Root et al. May 2001 B1
6237605 Vaska et al. May 2001 B1
6238393 Mulier et al. May 2001 B1
6240312 Alfano et al. May 2001 B1
6254598 Edwards et al. Jul 2001 B1
6258083 Daniel et al. Jul 2001 B1
6263224 West Jul 2001 B1
6270492 Sinofsky Aug 2001 B1
6275255 Adair et al. Aug 2001 B1
6290689 Delaney et al. Sep 2001 B1
6306081 Ishikawa et al. Oct 2001 B1
6310642 Adair et al. Oct 2001 B1
6311692 Vaska et al. Nov 2001 B1
6314962 Vaska et al. Nov 2001 B1
6314963 Vaska et al. Nov 2001 B1
6315777 Comben Nov 2001 B1
6315778 Gambale et al. Nov 2001 B1
6322536 Rosengart et al. Nov 2001 B1
6325797 Stewart et al. Dec 2001 B1
6328727 Frazier et al. Dec 2001 B1
6358247 Altman et al. Mar 2002 B1
6358248 Mulier et al. Mar 2002 B1
6375654 McIntyre Apr 2002 B1
6379345 Constantz Apr 2002 B1
6385476 Osadchy et al. May 2002 B1
6387043 Yoon May 2002 B1
6387071 Constantz May 2002 B1
6394096 Constantz May 2002 B1
6396873 Goldstein et al. May 2002 B1
6398780 Farley et al. Jun 2002 B1
6401719 Farley et al. Jun 2002 B1
6409722 Hoey et al. Jun 2002 B1
6416511 Lesh et al. Jul 2002 B1
6419669 Frazier et al. Jul 2002 B1
6423051 Kaplan et al. Jul 2002 B1
6423055 Farr et al. Jul 2002 B1
6423058 Edwards et al. Jul 2002 B1
6428536 Panescu et al. Aug 2002 B2
6436118 Kayan Aug 2002 B1
6440061 Wenner et al. Aug 2002 B1
6440119 Nakada et al. Aug 2002 B1
6458151 Saltiel Oct 2002 B1
6461327 Addis et al. Oct 2002 B1
6464697 Edwards et al. Oct 2002 B1
6474340 Vaska et al. Nov 2002 B1
6475223 Werp et al. Nov 2002 B1
6478769 Parker Nov 2002 B1
6482162 Moore Nov 2002 B1
6484727 Vaska et al. Nov 2002 B1
6485489 Teirstein et al. Nov 2002 B2
6488671 Constantz et al. Dec 2002 B1
6494902 Hoey et al. Dec 2002 B2
6497705 Comben Dec 2002 B2
6500174 Maguire et al. Dec 2002 B1
6502576 Lesh Jan 2003 B1
6514249 Maguire et al. Feb 2003 B1
6517533 Swaminathan Feb 2003 B1
6527979 Constantz et al. Mar 2003 B2
6532380 Close et al. Mar 2003 B1
6533767 Johansson et al. Mar 2003 B2
6537272 Christopherson et al. Mar 2003 B2
6540733 Constantz et al. Apr 2003 B2
6540744 Hassett et al. Apr 2003 B2
6544195 Wilson et al. Apr 2003 B2
6547780 Sinofsky Apr 2003 B1
6558375 Sinofsky et al. May 2003 B1
6558382 Jahns et al. May 2003 B2
6562020 Constantz et al. May 2003 B1
6572609 Farr et al. Jun 2003 B1
6579285 Sinofsky Jun 2003 B2
6585732 Mulier et al. Jul 2003 B2
6587709 Solf et al. Jul 2003 B2
6593884 Gilboa et al. Jul 2003 B1
6605055 Sinofsky et al. Aug 2003 B1
6613062 Leckrone et al. Sep 2003 B1
6622732 Constantz Sep 2003 B2
6626855 Weng et al. Sep 2003 B1
6626899 Houser et al. Sep 2003 B2
6626900 Sinofsky et al. Sep 2003 B1
6635070 Evans et al. Oct 2003 B2
6645202 Pless et al. Nov 2003 B1
6650923 Lesh et al. Nov 2003 B1
6658279 Swanson et al. Dec 2003 B2
6659940 Adler Dec 2003 B2
6673090 Root et al. Jan 2004 B2
6676656 Sinofsky Jan 2004 B2
6679836 Couvillon, Jr. Jan 2004 B2
6682526 Parker et al. Jan 2004 B1
6689128 Sliwa, Jr. et al. Feb 2004 B2
6692430 Adler Feb 2004 B2
6701581 Senovich et al. Mar 2004 B2
6701931 Sliwa, Jr. et al. Mar 2004 B2
6702780 Gilboa et al. Mar 2004 B1
6704043 Goldstein et al. Mar 2004 B2
6706039 Mulier et al. Mar 2004 B2
6712798 Constantz Mar 2004 B2
6719747 Constantz et al. Apr 2004 B2
6719755 Sliwa, Jr. et al. Apr 2004 B2
6730063 Delaney et al. May 2004 B2
6736810 Hoey et al. May 2004 B2
6751492 Ben-Haim Jun 2004 B2
6755790 Stewart et al. Jun 2004 B2
6755811 Constantz Jun 2004 B1
6764487 Mulier et al. Jul 2004 B2
6771996 Bowe et al. Aug 2004 B2
6773402 Govari et al. Aug 2004 B2
6780151 Grabover et al. Aug 2004 B2
6805128 Pless et al. Oct 2004 B1
6805129 Pless et al. Oct 2004 B1
6811562 Pless Nov 2004 B1
6833814 Gilboa et al. Dec 2004 B2
6840923 Lapcevic Jan 2005 B1
6840936 Sliwa, Jr. et al. Jan 2005 B2
6849073 Hoey et al. Feb 2005 B2
6858005 Ohline et al. Feb 2005 B2
6858026 Sliwa, Jr. et al. Feb 2005 B2
6863668 Gillespie et al. Mar 2005 B2
6866651 Constantz Mar 2005 B2
6887237 McGaffigan May 2005 B2
6892091 Ben-Haim et al. May 2005 B1
6896690 Lambrecht et al. May 2005 B1
6899672 Chin et al. May 2005 B2
6915154 Docherty et al. Jul 2005 B1
6916284 Moriyama Jul 2005 B2
6923805 LaFontaine et al. Aug 2005 B1
6929010 Vaska et al. Aug 2005 B2
6932809 Sinofsky Aug 2005 B2
6939348 Malecki et al. Sep 2005 B2
6942657 Sinofsky et al. Sep 2005 B2
6949095 Vaska et al. Sep 2005 B2
6953457 Farr et al. Oct 2005 B2
6955173 Lesh Oct 2005 B2
6962589 Mulier et al. Nov 2005 B2
6971394 Sliwa, Jr. et al. Dec 2005 B2
6974464 Quijano et al. Dec 2005 B2
6979290 Mourlas et al. Dec 2005 B2
6982740 Adair et al. Jan 2006 B2
6984232 Vanney et al. Jan 2006 B2
6994094 Schwartz Feb 2006 B2
7019610 Creighton, IV et al. Mar 2006 B2
7025746 Tal Apr 2006 B2
7030904 Adair et al. Apr 2006 B2
7041098 Farley et al. May 2006 B2
7042487 Nakashima May 2006 B2
7044135 Lesh May 2006 B2
7052493 Vaska et al. May 2006 B2
7090683 Brock et al. Aug 2006 B2
7118566 Jahns Oct 2006 B2
7156845 Mulier et al. Jan 2007 B2
7163534 Brucker et al. Jan 2007 B2
7166537 Jacobsen et al. Jan 2007 B2
7169144 Hoey et al. Jan 2007 B2
7186214 Ness Mar 2007 B2
7207984 Farr et al. Apr 2007 B2
7217268 Eggers et al. May 2007 B2
7242832 Carlin et al. Jul 2007 B2
7247155 Hoey et al. Jul 2007 B2
7261711 Mulier et al. Aug 2007 B2
7263397 Hauck et al. Aug 2007 B2
7276061 Schaer et al. Oct 2007 B2
7309328 Kaplan et al. Dec 2007 B2
7416552 Paul et al. Aug 2008 B2
7435248 Taimisto et al. Oct 2008 B2
7527625 Knight et al. May 2009 B2
7534204 Starksen et al. May 2009 B2
7569052 Phan et al. Aug 2009 B2
7736347 Kaplan et al. Jun 2010 B2
7758499 Adler Jul 2010 B2
7860555 Saadat Dec 2010 B2
7860556 Saadat Dec 2010 B2
7901348 Soper et al. Mar 2011 B2
8050746 Saadat et al. Nov 2011 B2
8078266 Saadat et al. Dec 2011 B2
8131350 Saadat et al. Mar 2012 B2
8137333 Saadat et al. Mar 2012 B2
8187189 Jung et al. May 2012 B2
8333012 Rothe et al. Dec 2012 B2
8382662 Soper et al. Feb 2013 B2
20010005789 Root et al. Jun 2001 A1
20010020126 Swanson et al. Sep 2001 A1
20010031912 Adler Oct 2001 A1
20010039416 Moorman et al. Nov 2001 A1
20010047136 Domanik et al. Nov 2001 A1
20010047184 Connors Nov 2001 A1
20010052930 Adair et al. Dec 2001 A1
20020004644 Koblish Jan 2002 A1
20020026145 Bagaoisan et al. Feb 2002 A1
20020054852 Cate May 2002 A1
20020065455 Ben-Haim et al. May 2002 A1
20020068853 Adler et al. Jun 2002 A1
20020080248 Adair et al. Jun 2002 A1
20020087166 Brock et al. Jul 2002 A1
20020087169 Brock et al. Jul 2002 A1
20020091304 Ogura et al. Jul 2002 A1
20020138088 Nash et al. Sep 2002 A1
20020165598 Wahr et al. Nov 2002 A1
20020169377 Khairkhahan et al. Nov 2002 A1
20030009085 Arai et al. Jan 2003 A1
20030035156 Cooper Feb 2003 A1
20030036698 Kohler et al. Feb 2003 A1
20030069593 Tremulis et al. Apr 2003 A1
20030120142 Dubuc et al. Jun 2003 A1
20030130572 Phan et al. Jul 2003 A1
20030144657 Bowe et al. Jul 2003 A1
20030171741 Ziebol et al. Sep 2003 A1
20030181939 Bonutti Sep 2003 A1
20030208222 Zadno-Azizi Nov 2003 A1
20030212394 Pearson et al. Nov 2003 A1
20030216720 Sinofsky et al. Nov 2003 A1
20030220574 Markus et al. Nov 2003 A1
20030222325 Jacobsen et al. Dec 2003 A1
20040006333 Arnold et al. Jan 2004 A1
20040049211 Tremulis et al. Mar 2004 A1
20040054335 Lesh et al. Mar 2004 A1
20040054389 Osypka Mar 2004 A1
20040082833 Adler Apr 2004 A1
20040097788 Mourlas et al. May 2004 A1
20040117032 Roth Jun 2004 A1
20040133113 Krishnan Jul 2004 A1
20040138707 Greenhalgh Jul 2004 A1
20040147806 Adler Jul 2004 A1
20040147911 Sinofsky Jul 2004 A1
20040147912 Sinofsky Jul 2004 A1
20040147913 Sinofsky Jul 2004 A1
20040158143 Flaherty et al. Aug 2004 A1
20040158289 Girouard et al. Aug 2004 A1
20040167503 Sinofsky Aug 2004 A1
20040181237 Forde et al. Sep 2004 A1
20040199052 Banik et al. Oct 2004 A1
20040210239 Nash et al. Oct 2004 A1
20040215180 Starkebaum et al. Oct 2004 A1
20040215183 Hoey et al. Oct 2004 A1
20040220471 Schwartz Nov 2004 A1
20040230131 Kassab et al. Nov 2004 A1
20040248837 Raz et al. Dec 2004 A1
20040249367 Saadat et al. Dec 2004 A1
20040254523 Fitzgerald et al. Dec 2004 A1
20040260182 Zuluaga et al. Dec 2004 A1
20050014995 Amundson et al. Jan 2005 A1
20050015048 Chiu et al. Jan 2005 A1
20050020914 Amundson et al. Jan 2005 A1
20050027163 Chin et al. Feb 2005 A1
20050038419 Arnold et al. Feb 2005 A9
20050059862 Phan Mar 2005 A1
20050059954 Constantz Mar 2005 A1
20050059965 Eberl et al. Mar 2005 A1
20050065504 Melsky et al. Mar 2005 A1
20050090818 Pike, Jr. et al. Apr 2005 A1
20050096643 Brucker et al. May 2005 A1
20050101984 Chanduszko et al. May 2005 A1
20050107736 Landman et al. May 2005 A1
20050119523 Starksen et al. Jun 2005 A1
20050124969 Fitzgerald et al. Jun 2005 A1
20050131401 Malecki et al. Jun 2005 A1
20050154252 Sharkey et al. Jul 2005 A1
20050158899 Jacobsen et al. Jul 2005 A1
20050159702 Sekiguchi et al. Jul 2005 A1
20050165279 Adler et al. Jul 2005 A1
20050165391 Maguire et al. Jul 2005 A1
20050165466 Morris et al. Jul 2005 A1
20050182295 Soper et al. Aug 2005 A1
20050182465 Ness Aug 2005 A1
20050197530 Wallace et al. Sep 2005 A1
20050197623 Leeflang et al. Sep 2005 A1
20050215895 Popp et al. Sep 2005 A1
20050222557 Baxter et al. Oct 2005 A1
20050222558 Baxter et al. Oct 2005 A1
20050228452 Mourlas et al. Oct 2005 A1
20050234436 Baxter et al. Oct 2005 A1
20050234437 Baxter et al. Oct 2005 A1
20050267328 Blumzvig et al. Dec 2005 A1
20050267452 Farr et al. Dec 2005 A1
20060009715 Khairkhahan et al. Jan 2006 A1
20060009737 Whiting et al. Jan 2006 A1
20060015096 Hauck et al. Jan 2006 A1
20060022234 Adair et al. Feb 2006 A1
20060025651 Adler et al. Feb 2006 A1
20060025787 Morales et al. Feb 2006 A1
20060030844 Knight et al. Feb 2006 A1
20060069303 Couvillon et al. Mar 2006 A1
20060074398 Whiting et al. Apr 2006 A1
20060084839 Mourlas et al. Apr 2006 A1
20060084945 Moll et al. Apr 2006 A1
20060089637 Werneth et al. Apr 2006 A1
20060111614 Saadat et al. May 2006 A1
20060122587 Sharareh Jun 2006 A1
20060146172 Jacobsen et al. Jul 2006 A1
20060149134 Soper et al. Jul 2006 A1
20060149331 Mann et al. Jul 2006 A1
20060155242 Constantz Jul 2006 A1
20060161133 Laird et al. Jul 2006 A1
20060167439 Kalser et al. Jul 2006 A1
20060183992 Kawashima Aug 2006 A1
20060184048 Saadat Aug 2006 A1
20060217755 Eversull et al. Sep 2006 A1
20060224167 Weisenburgh et al. Oct 2006 A1
20060253113 Arnold et al. Nov 2006 A1
20060258909 Saadat et al. Nov 2006 A1
20060271032 Chin et al. Nov 2006 A1
20070005019 Okishige Jan 2007 A1
20070015964 Eversull et al. Jan 2007 A1
20070016130 Leeflang et al. Jan 2007 A1
20070043338 Moll et al. Feb 2007 A1
20070043413 Eversull et al. Feb 2007 A1
20070049923 Jahns Mar 2007 A1
20070055142 Webler Mar 2007 A1
20070078451 Arnold et al. Apr 2007 A1
20070083187 Eversull et al. Apr 2007 A1
20070083217 Eversull et al. Apr 2007 A1
20070093808 Mulier et al. Apr 2007 A1
20070100241 Adler May 2007 A1
20070100324 Tempel et al. May 2007 A1
20070106146 Altmann et al. May 2007 A1
20070106214 Gray et al. May 2007 A1
20070106287 O'Sullivan May 2007 A1
20070135826 Zaver et al. Jun 2007 A1
20070167801 Webler et al. Jul 2007 A1
20070167828 Saadat Jul 2007 A1
20070255097 Jung et al. Nov 2007 A1
20070265609 Thapliyal et al. Nov 2007 A1
20070265610 Thapliyal et al. Nov 2007 A1
20070270686 Ritter et al. Nov 2007 A1
20070282371 Lee et al. Dec 2007 A1
20070287886 Saadat Dec 2007 A1
20070293724 Saadat et al. Dec 2007 A1
20080009747 Saadat et al. Jan 2008 A1
20080009859 Auth et al. Jan 2008 A1
20080015445 Saadat et al. Jan 2008 A1
20080015563 Hoey et al. Jan 2008 A1
20080015569 Saadat et al. Jan 2008 A1
20080027464 Moll et al. Jan 2008 A1
20080033241 Peh et al. Feb 2008 A1
20080033290 Saadat et al. Feb 2008 A1
20080045827 Rongen et al. Feb 2008 A1
20080057106 Erickson et al. Mar 2008 A1
20080058590 Saadat et al. Mar 2008 A1
20080058591 Saadat et al. Mar 2008 A1
20080058650 Saadat et al. Mar 2008 A1
20080058836 Moll et al. Mar 2008 A1
20080097476 Peh et al. Apr 2008 A1
20080183081 Lys et al. Jul 2008 A1
20080188759 Saadat et al. Aug 2008 A1
20080214889 Saadat et al. Sep 2008 A1
20080228032 Starksen et al. Sep 2008 A1
20080275300 Rothe et al. Nov 2008 A1
20080281293 Peh et al. Nov 2008 A1
20080287790 Li Nov 2008 A1
20080287805 Li Nov 2008 A1
20090030276 Saadat et al. Jan 2009 A1
20090030412 Willis et al. Jan 2009 A1
20090054803 Saadat et al. Feb 2009 A1
20090062790 Malchano et al. Mar 2009 A1
20090076489 Welches et al. Mar 2009 A1
20090076498 Saadat et al. Mar 2009 A1
20090082623 Rothe et al. Mar 2009 A1
20090125022 Saadat et al. May 2009 A1
20090143640 Saadat et al. Jun 2009 A1
20090187074 Saadat et al. Jul 2009 A1
20090203962 Miller et al. Aug 2009 A1
20090221871 Peh et al. Sep 2009 A1
20090227999 Willis et al. Sep 2009 A1
20090264727 Markowitz et al. Oct 2009 A1
20090267773 Markowitz et al. Oct 2009 A1
20090275799 Saadat et al. Nov 2009 A1
20090275842 Saadat et al. Nov 2009 A1
20090299363 Saadat et al. Dec 2009 A1
20090326572 Peh et al. Dec 2009 A1
20100004506 Saadat Jan 2010 A1
20100004633 Rothe et al. Jan 2010 A1
20100004661 Verin et al. Jan 2010 A1
20100010311 Miller et al. Jan 2010 A1
20100016662 Salsman et al. Jan 2010 A1
20100094081 Rothe et al. Apr 2010 A1
20100130836 Malchano et al. May 2010 A1
20100198081 Hanlin et al. Aug 2010 A1
20100292558 Saadat et al. Nov 2010 A1
20110060227 Saadat Mar 2011 A1
20110060298 Saadat Mar 2011 A1
20110144576 Rothe et al. Jun 2011 A1
20110301418 Gharib et al. Dec 2011 A1
20110306833 Saadat et al. Dec 2011 A1
20120004544 Saadat et al. Jan 2012 A9
20120004577 Saadat et al. Jan 2012 A1
20120059366 Drews et al. Mar 2012 A1
20120277596 Jung et al. Nov 2012 A1
Foreign Referenced Citations (39)
Number Date Country
10028155 Dec 2000 DE
0283661 Sep 1988 EP
0301288 Feb 1999 EP
59093413 May 1984 JP
59-181315 Oct 1984 JP
01-221133 Sep 1989 JP
03-284265 Dec 1991 JP
05-103746 Apr 1993 JP
09-051897 Feb 1997 JP
11-299725 Nov 1999 JP
2001-258822 Sep 2001 JP
WO 9221292 Dec 1992 WO
WO 9407413 Apr 1994 WO
WO 9503843 Feb 1995 WO
WO 9818388 May 1998 WO
WO 03039350 May 2003 WO
WO 03053491 Jul 2003 WO
WO 03101287 Dec 2003 WO
WO 2004043272 May 2004 WO
WO 2004080508 Sep 2004 WO
WO 2005070330 Aug 2005 WO
WO 2005077435 Aug 2005 WO
WO 2005081202 Sep 2005 WO
WO 2006017517 Feb 2006 WO
WO 2006024015 Mar 2006 WO
WO 2006083794 Aug 2006 WO
WO 2006091597 Aug 2006 WO
WO 2006126979 Nov 2006 WO
WO 2007067323 Jun 2007 WO
WO 2007079268 Jul 2007 WO
WO 2007133845 Nov 2007 WO
WO 2007134258 Nov 2007 WO
WO 2008015625 Feb 2008 WO
WO 2008021994 Feb 2008 WO
WO 2008021997 Feb 2008 WO
WO 2008021998 Feb 2008 WO
WO 2008024261 Feb 2008 WO
WO 2008079828 Jul 2008 WO
WO 2009112262 Sep 2009 WO
Non-Patent Literature Citations (77)
Entry
Avitall, “A Catheter System to Ablate Atrial Fibrillation in a Sterile Pericarditis Dog Model”, PACE, vol. 17, p. 774, 1994.
Avitall, “Right-Sided Driven Atrial Fibrillation in a Sterile Pericarditis Dog Model”, PACE, vol. 17, p. 774, 1994.
Avitall, “Vagally Mediated Atrial Fibrillation in a Dog Model can be Ablated by Placing Linear Radiofrequency Lesions at the Junction of the Right Atrial Appendage and the Superior Vena Cava”, PACE, vol. 18, p. 857, 1995.
Baker, “Nonpharmacologic Approaches to the Treatment of Atrial Fibrillation and Atrial Flutter”, J. Cardiovasc. Electrophysiol., vol. 6, pp. 972-978, 1995.
Bhakta, “Principles of Electroanatomic Mapping”, Indian Pacing & Electrophysiol J., vol. 8, No. 1, pp. 32-50, 2008.
Bidoggia, “Transseptal Left Heart Catheterization: Usefulness of the Intracavitary Electrocardiogram in the Localization of the Fossa Ovalis”, Cathet Cardiovasc Diagn., vol. 24, No. 3, pp. 221-225, 1991.
Bredikis, “Surgery of Tachyarrhythmia: Intracardiac Closed Heart Cryoablation”, PACE, vol. 13, pp. 1980-1984, 1990.
Cox, “Cardiac Surgery for Arrhythmias”, J. Cardiovasc. Electrophysiol., vol. 15, pp. 250-262, 2004.
Cox, “Five-Year Experience With the Maze Procedure for Atrial Fibrillation”, The Annals Thoracic Surgery, vol. 56, pp. 814-824, 1993.
Cox, “Modification of the Maze Procedure for Atrial Flutter and Atrial Fibrillation”, The Journal of Thoracic and Cardiovascular Surgery, vol. 110, pp. 473-484, 1995.
Cox, “The Status of Surgery for Cardiac Arrhythmias”, Circulation, vol. 71, pp. 413-417, 1985.
Cox, “The Surgical Treatment of Atrial Fibrillation”, The Journal of Thoracic and Cardiovascular Surgery, vol. 101, pp. 584-592, 1991.
Elvan, Replication of the “Maze” Procedure by Radiofrequency Catheter Ablation Reduces the Ability to Induce Atrial Fibrillation, PACE, vol. 17, p. 774, 1994.
Elvan, Radiofrequency Catheter Ablation (RFCA) of the Atria Effectively Abolishes Pacing Induced Chronic Atrial Fibrillation, PACE, vol. 18, p. 856, 1995.
Elvan, Radiofrequency Catheter Ablation of the Atria Reduces Inducibility and Duration of Atrial Fibrillation in Dogs, Circulation, vol. 91, pp. 2235-2244, 1995.
European Patent Application No. 06734083.6 filed Jan. 30, 2006 in the name of Saadat et al., Examination Communication mailed May 18, 2010.
European Patent Application No. 06734083.6 filed Jan. 30, 2006 in the name of Saadat et al., extended European Search Report mailed Jul. 1, 2009.
European Patent Application No. 06734083.6 filed Jan. 30, 2006 in the name of Saadat et al., office action mailed Oct. 23, 2009.
European Patent Application No. 07841754.0 filed Aug. 31, 2007 in the name of Saadat et al., Supplemental European Search Report mailed Jun. 30, 2010.
European Patent Application No. 08746822.9 filed Apr. 24, 2008 in the name of Rothe et al., European Search Report mailed Mar. 29, 2010.
European Patent Application No. 08746822.9 filed Apr. 24, 2008 in the name of Rothe et al., Office Action mailed Jul. 13, 2010.
Fieguth, Inhibition of Atrial Fibrillation by Pulmonary Vein Isolation and Auricular Resection—Experimental Study in a Sheep Model, European J. Cardiothorac. Surg., vol. 11, pp. 714-721, 1997.
Hoey, Intramural Ablation Using Radiofrequency Energy via Screw-Tip Catheter and Saline Electrode, PACE, vol. 18, p. 487, 1995.
Huang, Increase in the Lesion Size and Decrease in the Impedance Rise with a Saline Infusion Electrode Catheter for Radiofrequency, Circulation, vol. 80, No. 4, pp. II-324, 1989.
Moser, Angioscopic Visualization of Pulmonary Emboli, CHEST, vol. 77, No. 2, pp. 198-201, 1980.
Nakamura, Percutaneous Intracardiac Surgery With Cardioscopic Guidance, SPIE, vol. 1652, pp. 214-216, 1992.
Pappone, Circumferential Radiofrequency Ablation of Pulmonary Vein Ostia, Circulation, vol. 102, pp. 2619-2628, 2000.
Sethi, Transseptal Catheterization for the Electrophysiologist: Modification with a “View”, J. Interv. Card. Electrophysiol., vol. 5, pp. 97-99, 2001, Kluwer Academic Publishers, Netherlands.
Thiagalingam, Cooled Needle Catheter Ablation Creates Deeper and Wider Lesions than Irrigated Tip Catheter Ablation, J. Cardiovasc. Electrophysiol., vol. 16, pp. 1-8, 2005.
U.S. Appl. No. 11/259,498, filed Oct. 25, 2005 in the name of Saadat et al., Non-final Office Action mailed Feb. 25, 2010.
U.S. Appl. No. 11/560,742, filed Nov. 16, 2006 in the name of Saadat, Non-final Office Action mailed Jun. 10, 2010.
U.S. Appl. No. 11/687,597, filed Mar. 16, 2007 in the name of Saadat et al., Non-final Office Action mailed Jul. 21, 2010.
U.S. Appl. No. 11/828,267, filed Jul. 25, 2007 in the name of Saadat et al., Non-final Office Action mailed Jan. 14, 2010.
U.S. Appl. No. 12/117,655, filed May 8, 2008 in the name of Peh et al., Final Office Action mailed Mar. 1, 2010.
U.S. Appl. No. 12/117,655, filed May 8, 2008 in the name of Saadat et al., Non-final Office Action mailed Jun. 8, 2009.
U.S. Appl. No. 61/286,283, filed Dec. 14, 2009 in the name of Rothe et al.
U.S. Appl. No. 61/297,462, filed Jan. 22, 2010 in the name of Rothe et al.
Uchida, Developmental History of Cardioscopes, Coronary Angioscopy, pp. 187-197, 2001, Futura Publishing Co., Armonk, NY.
Willkampf, Radiofrequency Ablation with a Cooled Porous Electrode Catheter, JACC, vol. 11, No. 2, p. 17A, 1988.
U.S. Appl. No. 11/775,771, filed Jul. 10, 2007 in the name of Saadat et al., Non-final Office Action mailed Aug. 27, 2010.
U.S. Appl. No. 11/828,267, filed Jul. 25, 2007 in the name of Saadat et al., final Office Action mailed Sep. 16, 2010.
U.S. Appl. No. 11/259,498, filed Oct. 25, 2005 in the name of Saadat, Notice of Allowance mailed Nov. 15, 2010.
U.S. Appl. No. 11/560,742, filed Nov. 16, 2006 in the name of Saadat, Notice of Allowance mailed Nov. 15, 2010.
U.S. Appl. No. 12/464,800, filed May 12, 2009 in the name of Peh et al., non-final Office Action mailed Nov. 24, 2010.
U.S. Appl. No. 11/848,429, filed Aug. 31, 2007 in the name of Peh et al., non-final Office Action mailed Nov. 24, 2010.
European Patent Application No. 06734083.6 filed Jan. 30, 2006 in the name of Voyage Medical, Inc., Office Action mailed Nov. 12, 2010.
European Patent Application No. 07812146.4 filed Jun. 14, 2007 in the name of Voyage Medical, Inc., European Search Report mailed Nov. 18, 2010.
European Patent Application No. 07799466.3 filed Jul. 10, 2007 in the name of Voyage Medical, Inc., European Search Report mailed Nov. 18, 2010.
U.S. Appl. No. 12/117,655, filed May 8, 2008 in the name of Peh et al., non-final Office Action mailed Dec. 16, 2010.
U.S. Appl. No. 12/026,455, filed Feb. 5, 2008 in the name of Saadat et al., non-final Office Action mailed Dec. 27, 2010.
U.S. Appl. No. 12/947,198, filed Nov. 16, 2010 in the name of Saadat, non-final Office Action mailed Feb. 18, 2011.
U.S. Appl. No. 11/560,732, filed Nov. 16, 2006 in the name of Saadat, Notice of Allowance mailed Feb. 3, 2011.
U.S. Appl. No. 12/947,246, filed Nov. 16, 2006 in the name of Saadat, Notice of Allowance mailed Feb. 18, 2011.
U.S. Appl. No. 11/687,597, filed Mar. 16, 2007 in the name of Saadat, Notice of Allowance mailed Feb. 24, 2011.
U.S. Appl. No. 11/560,732, filed Mar. 16, 2007 in the name of Saadat, Notice of Allowance mailed Feb. 24, 2011.
U.S. Appl. No. 11/848,207, filed Aug. 30, 2007 in the name of Saadat et al., non-final Office Action mailed Feb. 25, 2011.
Japanese Patent Application No. 2007-554156 filed Jan. 30, 2006 in the name of Voyage Medical, Inc., Office Action mailed Feb. 15, 2011.
European Patent Application No. 07758716.0 filed Mar. 16, 2007 in the name of Voyage Medical, Inc., Supplemental European Search Report mailed Feb. 28, 2011.
U.S. Appl. No. 11/848,202, filed Aug. 30, 2007 in the name of Saadat et al., non-final Office Action mailed Mar. 11, 2011.
U.S. Appl. No. 11/763,399, filed Jun. 14, 2007 in the name of Saadat et al., non-final Office Action mailed Apr. 11, 2011.
U.S. Appl. No. 12/499,011, filed Jul. 7, 2009 in the name of Rothe et al., non-final Office Action mailed Apr. 12, 2011.
U.S. Appl. No. 12/367,019, filed Feb. 6, 2009 in the name of Miller et al., non-final Office Action mailed Apr. 22, 2011.
U.S. Appl. No. 11/959,158, filed Dec. 18, 2007 in the name of Saadat et al., non-final Office Action mailed Apr. 25, 2011.
U.S. Appl. No. 11/848,532, filed Aug. 31, 2007 in the name of Saadat et al., non-final Office Action mailed Apr. 26, 2011.
U.S. Appl. No. 11/828,281, filed Jul. 25, 2007 in the name of Peh et al., non-final Office Action mailed Apr. 27, 2011.
U.S. Appl. No. 11/961,950, filed Dec. 20, 2007 in the name of Saadat et al., non-final Office Action mailed May 9, 2011.
U.S. Appl. No. 11/961,995, filed Dec. 20, 2007 in the name of Saadat et al., non-final Office Action mailed May 9, 2011.
U.S. Appl. No. 11/962,029, filed Dec. 20, 2007 in the name of Saadat et al., non-final Office Action mailed May 9, 2011.
U.S. Appl. No. 11/828,267, filed Jul. 25, 2007 in the name of Saadat et al., non-final Office Action mailed May 11, 2011.
Japanese Patent Application No. 2009-500630 filed Mar. 16, 2007 in the name of Voyage Medical, Inc., Office Action mailed Apr. 27, 2011.
U.S. Appl. No. 11/775,771, filed Jul. 10, 2007 in the name of Saadat et al., final Office Action mailed May 12, 2011.
U.S. Appl. No. 11/877,386, filed Oct. 23, 2007 in the name of Saadat et al., non-final Office Action mailed May 20, 2011.
U.S. Appl. No. 11/775,819, filed Jul. 10, 2007 in the name of Saadat et al., non-final Office Action mailed May 20, 2011.
U.S. Appl. No. 11/775,837, filed Jul. 10, 2007 in the name of Saadat et al., non-final Office Action mailed May 23, 2011.
U.S. Appl. No. 12/117,655, filed May 8, 2008 in the name of Peh et al., final Office Action mailed Jun. 2, 2011.
U.S. Appl. No. 12/323,281, filed Nov. 25, 2008 in the name of Saadat et al., non-final Office Action mailed Jun. 7, 2011.
Japanese Patent Application No. 2007-554156 filed Jan. 30, 2006 in the name of Voyage Medical, Inc., Notice of Allowance mailed Jun. 13, 2011.
Related Publications (1)
Number Date Country
20120016221 A1 Jan 2012 US
Provisional Applications (1)
Number Date Country
61304235 Feb 2010 US