System and method for image-based alignment of an endoscope

Information

  • Patent Grant
  • 10321803
  • Patent Number
    10,321,803
  • Date Filed
    Thursday, April 23, 2015
    9 years ago
  • Date Issued
    Tuesday, June 18, 2019
    5 years ago
Abstract
Systems and methods for endoscopic procedures employ a first technique to ensure initial correct alignment of an endoscope (100) with a desired target (10). A reference image (51) is then acquired from an imaging arrangement associated with the endoscope. During a subsequent stage of the procedure, tracking of the endoscope position relative to the target is performed partially or entirely by image-based tracking by comparing features in real-time video image (52) produced by imaging arrangement with features in the reference image (51). The feature comparison may be performed visually by a user, or may be automated to offer more specific corrective suggestions to the user.
Description
FIELD AND BACKGROUND OF THE INVENTION

The present invention relates to endoscopy techniques and, in particular, it concerns a system and method for image-based alignment of an endoscope during at least part of an endoscopic procedure.


The present invention will be exemplified in the context of a system as described in the co-assigned PCT application published as WO 03/086498 entitled “Endoscope Structure and Techniques for Navigation in Branched Structure” to Gilboa, which is hereby incorporated by reference in its entirety. The aforementioned patent application discloses a method and apparatus in which a thin locatable guide, enveloped by a sheath, is used to navigate a bronchoscopic tool to a target location within the lung, aimed in particular to deliver treatments to the lung periphery beyond the bronchoscope's own reach. The coordinates of the target are predetermined based upon three-dimensional CT data. A location sensor is incorporated at the locatable guide's tip. The enveloped guide is inserted into the lung via the working channel of a bronchoscope. First, the bronchoscope's tip is directed to the furthest reachable location in the direction of the target. Next, the guide is advanced beyond the tip of the bronchoscope towards the designated target, based on the combination of the CT data and the position of the guide's tip as measured in body coordinates. When the guide's tip at the target, the guide is withdrawn, freeing the enveloping sheath for insertion a bronchoscopic tool. In order to prevent the distal end portion of the sheath from sliding away from the target, the sheath is locked to the bronchoscope's body and the bronchoscope itself is held steadily to prevent it from slipping further into the lungs or outwards. Because the airways in the periphery of the lung are narrow, approximately in the same dimensions as the sheath, sideways movements are extremely limited.


The above system may also be used to navigate the tip of the bronchoscope to a target located inside the main bronchus and not only to targets in the periphery of the lungs. Although for such centrally-located target the physician has direct visualization of the scene in front of the bronchoscope, it is not always sufficient for visually identifying the designated targets, since many of these targets are hidden in the tissue outside the airways. Hence, it is a benefit to combine the CT data into the navigational aids also for targets inside the main bronchus, where the bronchoscope can reach and direct vision exists, but yet the target itself is hidden.


When using the navigation system for navigating the tip of the bronchoscope itself, many of the mechanical features of the locatable guide described in WO 03/086498 are not needed. Specifically, the steerability of the guide is not needed, and the enveloping sheath is also not needed. However the principle of using a separate locatable guide having a location sensor at its tip and being inserted into the working channel of a regular bronchoscope actually changes the bronchoscope from a non-locatable bronchoscope to a locatable bronchoscope, thereby offering major advantages as will become clear.


As in the prior art apparatus, the locatable guide can be inserted into and withdrawn from the bronchoscope's working channel as needed. Unlike the periphery of the lung, the central airways are much wider than the bronchoscope. As a consequence, when the tip of the bronchoscope is on target, it can move sideways in addition to sliding in and out. Therefore stabilizing the bronchoscope's tip during treatment is a three dimensional task, involving the operation of the steering ability of the bronchoscope. An example for the importance for maintaining the location of the bronchoscope's tip at the designated target during the insertion of the bronchoscopic tool is the use of the Transbronchial Histology Needle, by which a needle is guided towards a target such as a lymph node which neighbors the main bronchus from the outside and thus is invisible to the bronchoscope image but its coordinates are known from the CT data. Any mistake in directing the needle may result not only in failure of the procedure, but worse, in causing damage to vital organs such as the aorta or other major blood vessels.


In principle, the same methods as presented in WO 03/086498 may be used in the context of the major airways. Specifically, by using the location of the tip of the bronchoscope as measured by the location measurement sensor, a directing display is produced corresponding to a simulation or schematic diagram of the view from the distal tip of the guide, which is based on the relative location of the target versus the position of the tip of the guide in six degrees of freedom. In the central airways, this view is supplemented by the direct video image from the bronchoscope imaging arrangement. Based on these two displays, the physician brings the tip of the bronchoscope to the target. When the tip of the bronchoscope is correctly aligned with and adjacent to the target (FIG. 7), the guide with the location sensor is withdrawn (as shown in FIG. 8), thereby freeing the bronchoscope's working channel for insertion a bronchoscopic tool FIG. 9a). Once the locatable guide is released, the directing display can no longer function for directing the tip to target. Instead, the physician has to hold the bronchoscope as steadily as possible during withdrawal of the guide and the insertion of the tool. If the bronchoscope slips from the target location (for example, as shown in FIG. 9b), the physician may notice the chance of position in the video image, but has no effective tool available to help him return the tip of the bronchoscope reliably to the desired target (other than reinserting the guide and repeating the navigation process).


Hence, it would be of benefit to have a method and corresponding system for confirming correct alignment of the tip of an endoscope after removal of a locatable guide used to achieve initial alignment, particularly for procedures involving a target which is obscured from view.


SUMMARY OF THE INVENTION

The present invention is a system and method for image-based alignment of an endoscope.


According to the teachings of the present invention there is provided, a method for confirming correct alignment of a distal end of an endoscope including an imaging arrangement during an endoscopic procedure, the method comprising: (a) positioning the distal end of the endoscope adjacent to a target location and capturing a reference image using the imaging arrangement; (b) sensing a real-time video image using the imaging arrangement; and (c) comparing features of the real-time video image with the reference image to confirm correct alignment of the endoscope.


According to a further feature of the present invention, the step of positioning employs a target location identified in three-dimensional image data of a region of a body to be treated.


According to a further feature of the present invention, the three-dimensional image data is derived from an imaging technique selected from: computerized tomography; magnetic resonance imaging; positron emission tomography; and ultrasound.


According to a further feature of the present invention, the step of positioning employs a position sensor associated with the distal end of the endoscope, the position sensor being part of a position measuring system.


According to a further feature of the present invention, the step of positioning is performed by comparing the position of the distal end of the endoscope as measured by the position measuring system and the target location as identified in the image data.


According to a further feature of the present invention, the position sensor is part of an elongated element deployed within a working channel of the endoscope, and wherein the elongated element is withdrawn from the working channel prior to the comparing.


According to a further feature of the present invention, the target location is not visible in the reference image.


According to a further feature of the present invention, the reference image and the real-time video image are displayed simultaneously to facilitate performance of the comparing features visually by a user.


According to a further feature of the present invention, the comparing includes co-processing the reference image and at least one frame from the real-time video to determine a measure of mismatch, the method further comprising generating an alarm signal if the measure of mismatch exceeds a predefined value.


According to a further feature of the present invention, the comparing includes co-processing the reference image and at least one frame from the real-time video to determine a displacement correction required to compensate for erroneous movement of the endoscope, the method further comprising generating a display indicative to a user of the displacement correction required to compensate for the erroneous movement of the endo scope.


According to a further feature of the present invention, the comparing includes co-processing the reference image and at least one frame from the real-time video to determine a transformation relating the real-time video frame to the reference image, the method further comprising generating a display corresponding to the real-time video with addition of an indication of a target location, position of the indication being derived at least in part by use of the transformation.


According to a further feature of the present invention, the endoscope is a bronchoscope.


There is also provided according to the teachings of the present invention, a system for ensuring correct alignment of an endoscope during performance of an endoscopic procedure, the system comprising: (a) an endoscope having a distal end for insertion into a body; (b) an imaging arrangement associated with the endoscope and configured to generate a real-time video image of a region beyond the distal end; and (c) a processing system associated with the imaging arrangement and configured to: (i) in an initial state of alignment with a target location, derive from the imaging arrangement a reference image corresponding to correct alignment with the target location, (ii) derive from the imaging arrangement real-time images of the region beyond the distal end, and (iii) co-process the reference image and the real-time images to determine a current alignment status of the endoscope with the target location.


According to a further feature of the present invention, the processing system is configured to co-process the reference image and the real-time images to determine a measure of mismatch, the processing system further generating an alarm signal if the measure of mismatch exceeds a predefined value.


According to a further feature of the present invention, there is also provided a display for displaying at least the real-tire images to a user, wherein the processing system is configured to co-process the reference image and the real-time images to determine a displacement correction required to compensate for erroneous movement of the endoscope, the processing system further generating an indication on the display indicative to a user of the displacement correction required to compensate for the erroneous movement of the endoscope.


According to a further feature of the present invention, there is also provided a display for displaying at least the real-time images to a user, wherein the processing system is configured to co-process the reference image and the real-time images to determine a transformation relating the real-time video frame to the reference image) the processing system further generating on the display an indication of a target location, position of the indication being derived at least in part by use of the transformation.


According to a further feature of the present invention, there is also provided a position measuring system including a position sensor carried by an elongated element removably deployable along a working channel of the endoscope.


There is also provided according to the teachings of the present invention, a method for facilitating performance of an endoscopic procedure on a target which is obscured from view by an imaging arrangement of an endoscope, the method comprising: (a) generating real-time video from the imaging arrangement of the endoscope; (b) determining a position of the target and a position of a distal end of the endoscope; (c) generating a display including the real-time video and a simulated view of the target correctly positioned within the real-time video; and (d) adjusting the display so as to maintain the target correctly positioned within the real-time video when the endoscope is moved.


According to a further feature of the present invention, the adjusting includes comparing features from at least one frame of the real-time video with features from a reference image derived from the imaging arrangement during the step of determining, thereby deriving a transformation relating the real-time video frame to the reference image.


According to a further feature of the present invention, the determining a position of a distal end of the endoscope is performed using position measuring system including a position sensor carried by an elongated element removably deployable along a working channel of the endoscope.





BRIEF DESCRIPTION OF THE DRAWINGS

The invention is herein described, by way of example only, with reference to the accompanying, drawings, wherein:



FIG. 1 is a video display of the target zone inside the bronchial tree when the tip of the bronchoscope is off target.



FIG. 2 is a simulated tip view corresponding to the position of FIG. 1.



FIG. 3 is a video display of the target zone inside the bronchial tree when the tip of the bronchoscope is on the target.



FIG. 4 is a simulated tip view corresponding to the position of FIG. 3.



FIG. 5 is a display of a combination of a stored image and live video in the context of first preferred embodiment of the invention.



FIG. 6 is the display of a combination of a stored image and live video in the context of a second preferred embodiment of the invention.



FIG. 7 is a schematic side cross-sectional view showing the bronchoscope having been correctly aligned by use of a position measurement sensor with a target which is obscured from view.



FIG. 8 is a view similar to FIG. 7 after removal of the position measurement sensor to free a working lumen of the bronchoscope.



FIG. 9a is a view similar to FIG. 8 after insertion of a tool along the working lumen.



FIG. 9b is a view similar to FIG. 9a after erroneous movement has disrupted alignment of the tool with the obscured target.



FIG. 10 is a schematic illustration of the components of a system, constructed and operative according to the teachings of the present invention, for ensuring correct alignment of an endoscope during performance of an endoscopic procedure.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

The present invention is a system and method for image-based alignment of an endoscope.


The principles and operation of systems and methods according to the present invention may be better understood with reference to the drawings and the accompanying description.


First in general terms, the present invention provides systems and methods for endoscopic procedures (exemplified herein with reference to bronchoscopic procedures) wherein a first technique is used to ensure initial correct alignment of an endoscope with a desired target and a reference image is acquired from an imaging arrangement associated with the endoscope. Then, during a subsequent stage of the procedure, tracking of the endoscope position relative to the target is performed partially or entirely by image-based tracking by comparing features in the realtime video image produced by the imaging arrangement with features in the reference image.


Thus, according to a first aspect of the present invention, a method for confirming correct alignment of a distal end of an endoscope during an endoscopic procedure includes: positioning the distal end of the endoscope adjacent to a target location and capturing, a reference image using the imaging arrangement; sensing a real-time video image using the imaging arrangement; and comparing features of the real-time video image with the reference image to confirm correct alignment of the endoscope.


It will immediately be appreciated that the present invention offers profound advantages, particularly for cases where the desired target is obscured from view (such as behind other tissue) or is not readily identifiable directly by visual imaging. In such cases, navigation under video imaging alone is insufficient. Nevertheless, after use of a primary tracking system (such as that of the aforementioned WO 03/086498) to achieve initial alignment, use of feature-based optical tracking based on features not necessarily belonging to the target frees the system from subsequent dependence on the primary tracking system, thereby allowing removal of the position measurement probe and/or rendering navigation more robust and reliable in the face of disturbances such as movement of the patient's body or the like. These and other advantages of the present invention will become clearer from the subsequent description.


Referring now to the drawings, FIG. 10 shows schematically a preferred implementation of a system, constructed and operative according to the teachings of the present invention, for implementing the methods of the present invention. For one set of preferred implementations of the method of the present invention, the system is substantially similar to that described in WO 03/086498 with certain changes to the display and/or image processing systems, as will be described below. Thus, a position measurement sensor 101 and video sensor 102 are incorporated in the distal tip of bronchoscope 100. An electro-magnetic tracking system induces electro-magnetic fields from antennae 107, senses the signals from the location sensor 101 and determines the position of the tip of the bronchoscope in six degrees of freedom. A processing system 108 gathers that position information together with the video image from the tip of the bronchoscope as produced by the video camera 104. The processing system may display to the physician live video, captured images and simulated views on a display screen 110.


Further details of a particularly preferred position measuring system for measuring position in six degrees-of-freedom may be found in U.S. Pat. No. 6,188,355 and PCT Application Publication Nos. WO 00/10456 and WO 01/67035 Most preferably, at least one, and preferably three, reference sensors (not shown) are also attached to the chest of the patient and their 6 DOF coordinates sent to processing system 108 where they are used to calculate the patient coordinate frame of reference.


It should be noted in this context that the term “position sensor” is used herein in the description and claims to refer to any element which can be associated permanently or temporarily with an object and functions together with other components of a position measuring system to determine the position and/or attitude of the object. It should be appreciated that the terminology does not necessarily imply that the position sensor itself is capable of any measurement alone. Nor does this terminology imply any particular function of the position sensor, such that the “sensor” may be a transmitter, a receiver or any other element which functions as part of a position measuring system, depending upon the technology employed. In all such cases, the element is referred to as a “position sensor” since its presence associated with the object allows sensing by the system of the object's position.


Although described herein with reference to a non-limiting preferred implementation employing a bronchoscope, it should be noted that the present invention is equally applicable to substantially any intra-body endoscopic procedure.


As in the aforementioned WO 03/086498, the location of the desired target within the body is preferably determined in an offline preparation session prior to the procedure in which the target is identified in three-dimensional image data of a region of a body to be treated. The three-dimensional image data is preferably derived from an imaging technique selected from: computerized tomography; magnetic resonance imaging; positron emission tomography; and ultrasound. Most commonly, computerized tomography (“CT”) data is used. Then, after suitable calibration to register the position measurement system coordinates with the CT data, a simulated tip view or other visual navigation aids as described in WO 03/086498 are used to guide the bronchoscope into alignment with the target. These navigation aids are based on comparing the position of the distal end of the endoscope as measured by the position measuring system and the target location as identified in the image data. Then, according to one particularly preferred set of embodiments, the position sensor is withdrawn from a working channel of the endoscope as part of an elongated element.



FIGS. 1-4 illustrate schematically examples the displays which are preferably available to the physician during initial alignment of the bronchoscope. Specifically, FIG. 1 shows an example of a target area. The target is obscured from view, being located behind the tissue of a bifurcation inside the bronchus. The target 10, marked in a broken line, is not visible in the video image. In one embodiment of the invention, it is not marked in the video display at all. In another preferred embodiment it's the target's location, as calculated by the processing system 108, is displayed in the video display by an artificial mark such as a line, a point, a broken line, a colored area, a three dimensional entity or a combination of any of the above.


In FIG. 1, the tip of the bronchoscope is shown to be positioned off the direction of the target. FIG. 2 is the tip view used for directing to the target. The direction of the target relative to the tip is presented by dot 20 marking the target and arrow 22 aiming from the tip to the target. The target may also be presented in a simulated view of the boundary of the actual target lesion as calculated from the CT data or by a colored area or by three-dimensional entity or by any combination of the above. According to the example, the tip should be deflecting in the 12 o'clock direction in order to be on target FIGS. 3 and 4 show the same scenario when the tip is on target. According to the prior art described above, after achieving alignment with the target as shown, the physician has to try to hold the bronchoscope steady while withdrawing the locatable guide and inserting a tool along the lumen. According to the present invention, before the guide is withdrawn, the image as shown in FIG. 3 is captured and stored in a memory device of processing system 108. Now the system has two sources of images to control the location of the bronchoscope's distal tip, a real-time live video image 52 and a captured video image 51 where the tip was located at the desired target location, as shown in FIG. 5.


The present invention may be implemented in a number of different embodiments with different degrees of sophistication as to how the comparison between features of the real-time video and the reference image is performed. According, to a first basic embodiment, the reference image 51 and the real-time video image 52 are displayed simultaneously on display device 110 as illustrated in FIG. 5, thereby facilitating visual comparison of the image features by a user. In this case, the physician himself compares the two images and decides whether the bronchoscope is located in the required location, and if not, in what direction the tip of the bronchoscope should be deflected.


In more sophisticated embodiments, the system preferably co-processes the reference image and the real-time images to determine a current alignment status of the endoscope with the target location. Thus, processing system 108 is configured to: derive from the imaging arrangement of the endo scope, in an initial state of alignment with a target location, a reference image corresponding to correct alignment with the target location; derive from the imaging, arrangement real-time images of the region beyond the distal end, and co-process the reference image and the real-time images to determine a current alignment status of the endoscope with the target location.


Here too, the co-processing may be implemented at various different levels of sophistication. In a simplest case, a correlation between the reference image and the current video image may offer a measure of mismatch. The user can then empirically adjust the position of the bronchoscope tip to maximize the correlation (minimize the mismatch), thereby returning to the correct position. Application of a threshold to the measure of mismatch may be used to activate an alarm signal.


In more preferred implementations, the system tracks features or regions from the reference image in the video image to provide more specific indications to the user of the required correction for any erroneous movement of the bronchoscope off target. For small-scale lateral displacements, this may be implemented simply by correlating a central sub-window 56 of reference image 51 centered on target location 55 with a corresponding sized sliding window (i.e., at multiple different positions) in the real-time video to find the best match, thereby identifying the position of the target sub-window in the real-time video image.


At a next level of sophistication the tracking may also allow for scaling and/or rotation of the sub-window. This allows the system to maintain target tracking during rotation, as well as small-scale advancing or withdrawal, of the bronchoscope. A further level of sophistication may employ planar transformations such as affine transformations which approximate the distortions caused by viewing a surface from different viewing angles.


At the top end of the range of sophistication in the tracking algorithms are tracking techniques based on three-dimensional modeling of the viewed scene and reconstruction of the camera path. Such techniques, often referred to as “Structure from Motion”, are well developed in the field of optical tracking and computer vision, and allow reconstruction of three-dimensional models from a single moving camera. Details of processing techniques for implementing structure from motion may be found in papers from the Robotics Research Group in the Department of Engineering Science, Oxford University (UK) such as “Automatic Camera Tracking” by Andrew W. Fitzgibbon et al. Video Registration (2003) and “Feature Based Methods for Structure and Motion Estimation” by P. H. S. Torr et al. Vision Algorithms: Theory and Practice (2000), both available from http://www.robots.ox.ac.uk/.


In the present application, structure-from-motion processing can be greatly simplified by the use of model data based on CT data or the like. Thus, for example, given that the initial reference image is taken from a known position as established by the primary alignment system, a “depth” (i.e., camera-to-surface distance) associated with each pixel of the reference image can be derived directly from CT data, thereby providing an initial three-dimensional model from which processing can begin. This approach has advantages of robustness under significant changes of view, and even where there is no overlap between the current real-time video field of view and the reference image.


In any or all of the above-mentioned tracking techniques, corrections are preferably made for geometrical distortions introduced by the optics of the imaging arrangement, as is known in the art. These corrections may be performed on the source images prior to implementing the tracking techniques, or may be incorporated into the tracking calculations themselves.


The output from the system (and method) of the present invention may take a number of forms. In a simplest case mentioned above, an alarm may be sounded if a measure of mismatch between the current video and the reference image indicates that the bronchoscope has wandered off target, and the measure of mismatch (or the correlation) may be displayed to the user or indicated by an audio signal to provide feedback indicative of “getting hotter” or “getting colder” with regard to alignment with the target.


In more preferred implementations where features of the reference image are positively tracked within the real-time video, the processing system may generate an indication on the display indicative to a user of the displacement correction required to compensate for the erroneous movement of the endoscope. This may take the form of all arrow or vector such as line 62 in FIG. 6 which indicates the movement of the bronchoscope required to bring the center of the field of view into alignment with the target position illustrated as 61, Thus, FIG. 6 corresponds to the display when the bronchoscope is in the position of FIG. 9b. After performing the required corrective motion, the bronchoscope returns to the position of FIG. 9a and the live video 52 of FIG. 6 would again appear similar to the reference image 51.


Alternatively, or additionally, a transformation (2 or 3 dimensional) calculated by the processing system for relating the real-time video frame to the reference image may be used to determine the position of the obscured target within the real-time video image. The target location can then be designated in the real-time video image, for example, by the sub-window frame 58 or the marker 57 as shown in FIG. 5.


According to a most preferred option, which is believed to be of patentable significance in its own right, the present invention provides an augmented reality endoscopic display in which a simulated view of an obscured target is displayed in the context of the real-time video so that the target appears correctly positioned within the video image and moves so as to maintain the correct positioning of the target within the real-time video when the endoscope is moved. This augmented reality display allows the user to operate the endoscope in a fully intuitive manner as if the target were directly viewable via the video imaging arrangement of the endoscope. Thus, the user will see obscured target 10 of FIGS. 1 and 3 as if the tissue in front of the target was semi-transparent.


In practical terms, the simulated view of the target used for the augmented reality display is preferably derived from three-dimensional imaging data such as CT in which the target tissue has been designated prior to the procedure. The target tissue volume is then preferably exported as a three-dimensional graphic object, or a closed body geometrical approximation to the tissue volume is generated. Then, during the procedure, information regarding the relative positions and orientations of the endoscope tip and the target tissue is used to determine the position, viewing angle and scaling factors which should be used to represent the target correctly in the real-time video image. The target is preferably indicated as a semi-transparent video overlay so that it appears as a ghost image without completely hiding the tissue actually viewed in the video image. Alternatively, a dashed outline or any other suitable indication may be used.


The augmented reality display is advantageous both during initial alignment of the endoscope with the target and during subsequent performance of a procedure. Most preferably, during a procedure, real-time adjustment of the simulated target image within the video image is performed on the basis of the optical tracking of the present invention. Where optical tracking is performed in two dimensions only, the adjustment of the target appearance will correspondingly be reduced to a two-dimensional manipulation. Where three-dimensional model based tracking is used, full three-dimensional augmented reality functionality is preferably maintained.


In each case, an alarm is preferably activated if the location error, i.e., the misalignment of the real-time video from the reference image, exceeds a predefined value, for example, the size of the target. The alarm may be an audio alarm and/or a visual alarm indication such as a flashing symbol or a color change of part or all of the display. An alarm (which may be distinct from the off-target alarm) is preferably also generated if the tracking algorithm fails to maintain reliable tracking between the reference image and the current image.


As mentioned earlier, the optical tracking of the present invention does not require the target location to be visible in the reference image. In fact, it should be noted that the reference image need not even include the direction to the target in its field of view. For example, if an endoscope is to be used in a procedure with a tool which generates images, takes samples or otherwise treats a region of tissue located laterally next to the distal end of the endoscope, the imaging arrangement of the endoscope will typically not offer a view of the tissue of the target, nor of any tissue which overlies the target. Nevertheless, once correct alignment of the endoscope has been achieved using the primary tracking system, the optical tracking of the present invention based on a reference image of the scene beyond the tip of the endoscope is effective to ensure correct alignment with the target, despite the fact that the target is outside the field of view.


It will be appreciated that the above descriptions are intended only to serve as examples, and that many other embodiments are possible within the scope of the present invention as defined in the appended claims.

Claims
  • 1. A method for confirming correct alignment of a distal end portion of an endoscope including an imaging arrangement, the method comprising: positioning a distal end portion of an endoscope adjacent to a target location to capture a reference image using an imaging arrangement of the endoscope;generating a real-time video image;comparing features of the real-time video image with the reference image to determine whether the endoscope is correctly aligned;measuring, if the endoscope is not correctly aligned, a mismatch between the reference image and the real-time video image;correlating a sub-window of at least a portion of the reference image provided on a display, the sub-window centered on the target location, with a corresponding sliding window of the real-time video image provided on the display;displaying the sub-window superimposed over the sliding window;designating a marker corresponding to a center of the target location; andupdating a location of the marker in the sliding window when a view of the real-time video changes.
  • 2. The method of claim 1, further comprising providing a visual indication of the mismatch, the visual indication including directionality and magnitude information.
  • 3. The method of claim 2, further comprising adjusting a position of the endoscope based on the directionality and magnitude information derived from the mismatch.
  • 4. The method of claim 1, further comprising rotating the sub-window.
  • 5. The method of claim 4, further comprising tracking the target location during rotation of the sub-window.
  • 6. The method of claim 5, further comprising employing planar transformations to approximate distortions caused by different viewing angles of the target location.
  • 7. The method of claim 1, further comprising employing the target location in three-dimensional image data of a region of a body to be treated.
  • 8. The method of claim 7, wherein the three-dimensional image data is derived from an imaging technique selected from: computerized tomography, magnetic resonance imaging, positron emission tomography, and ultrasound.
  • 9. A system for confirming correct alignment of an endoscope, the system comprising: an endoscope having a distal end portion for insertion into a body;an imaging arrangement configured to generate a real-time video image of a region beyond the distal end portion; anda processing system associated with the imaging arrangement and configured to: (i) derive, when the imaging arrangement is in an initial state of alignment with a target location, a reference image,(ii) derive from the imaging arrangement real-time images of the region beyond the distal end portion,(iii) co-process the reference image and the real-time images to determine a current alignment status of the endoscope with the target location, wherein a mismatch is measured, if the endoscope is not correctly aligned, between the reference image and the real-time video images,(iv) correlate a sub-window of the reference image provided on a display, the sub-window centered on the target location, with a corresponding sliding window of the real-time video image provided on the display;(v) designate a marker corresponding to a center of the target location; and(vi) update a location of the marker in the sliding window when a view of the real-time video changes.
  • 10. The system of claim 9, wherein a visual indication of the mismatch is provided, the visual indication including directionality and magnitude information.
  • 11. The system of claim 10, wherein a position of the endoscope is adjusted based on the directionality and magnitude information derived from the mismatch.
  • 12. The system of claim 9, wherein the sub-window is rotated.
  • 13. The system of claim 12, wherein the target location is tracked during rotation of the sub-window.
  • 14. The system of claim 13, wherein planar transformations are employed to approximate distortions caused by different viewing angles of the target location.
  • 15. The system of claim 9, wherein the target location is employed in three-dimensional image data of the region of the body to be treated.
  • 16. The system of claim 15, wherein the three-dimensional image data is derived from an imaging technique selected from: computerized tomography, magnetic resonance imaging, positron emission tomography, and ultrasound.
  • 17. A method for confirming correct alignment of a distal end portion of an endoscope including an imaging arrangement, the method comprising: positioning a distal end portion of an endoscope adjacent to a target location to capture a reference image using an imaging arrangement of the endoscope;generating a real-time video image;comparing features of the real-time video image with the reference image to determine whether the endoscope is correctly aligned;correlating a sub-window of at least a portion of the reference image provided on a display, the sub-window centered on the target location, with a corresponding sliding window of the real-time video image provided on the display;displaying the sub-window superimposed over the sliding window;rotating the sub-window;designating a marker corresponding to a center of the target location; andupdating a location of the marker in the sliding window when a view of the real-time video changes.
  • 18. A system for confirming correct alignment of an endoscope, the system comprising: an endoscope having a distal end portion for insertion into a body;an imaging arrangement configured to generate a real-time video image of a region beyond the distal end portion; anda processing system associated with the imaging arrangement and configured to: (i) derive, when the imaging arrangement is in an initial state of alignment with a target location, a reference image,(ii) derive from the imaging arrangement real-time images of the region beyond the distal end portion,(iii) co-process the reference image and the real-time images to determine a current alignment status of the endoscope with the target location,(iv) correlate a sub-window of the reference image provided on a display, the sub-window centered on the target location, with a corresponding sliding window of the real-time video image provided on the display, wherein the sub-window is rotated;(v) designate a marker corresponding to a center of the target location; and(vi) update a location of the marker in the sliding window when a view of the real-time video changes.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation application of U.S. patent application Ser. No. 11/568,260, filed on Apr. 22, 2008, the entire contents of which are hereby incorporated herein by reference, which is a continuation of PCT international patent application number PCT/IL2005/000452, filed May 1, 2005, the entire contents of which are hereby incorporated herein by reference.

US Referenced Citations (807)
Number Name Date Kind
1576781 Phillips Mar 1926 A
1735726 Bornhardt Nov 1929 A
2407845 Nemeyer Sep 1946 A
2650588 Drew Sep 1953 A
2697433 Sehnder Dec 1954 A
3016899 Stenvall Jan 1962 A
3017887 Heyer Jan 1962 A
3061936 Dobbeleer Nov 1962 A
3073310 Mocarski Jan 1963 A
3109588 Polhemus et al. Nov 1963 A
3191652 Benson et al. Jun 1965 A
3294083 Alderson Dec 1966 A
3310264 Appleton Mar 1967 A
3367326 Frazier Feb 1968 A
3439256 Kahne et al. Apr 1969 A
3577160 White May 1971 A
3614950 Rabey Oct 1971 A
3644825 Davis, Jr. et al. Feb 1972 A
3674014 Tillander Jul 1972 A
3702935 Carey et al. Nov 1972 A
3704707 Halloran Dec 1972 A
3747166 Eross Jul 1973 A
3821469 Whetstone et al. Jun 1974 A
3868565 Kuipers Feb 1975 A
3941127 Froning Mar 1976 A
3983474 Kuipers Sep 1976 A
4017858 Kuipers Apr 1977 A
4024997 Kolpin May 1977 A
4037592 Kronner Jul 1977 A
4052620 Brunnett Oct 1977 A
4054881 Raab Oct 1977 A
4117337 Staats Sep 1978 A
4173228 Van Steenwyk et al. Nov 1979 A
4182312 Mushabac Jan 1980 A
4202349 Jones May 1980 A
4228799 Anichkov et al. Oct 1980 A
4256112 Kopf et al. Mar 1981 A
4262306 Renner Apr 1981 A
4287809 Egli et al. Sep 1981 A
4298874 Kuipers Nov 1981 A
4314251 Raab Feb 1982 A
4317078 Weed et al. Feb 1982 A
4319136 Jinkins Mar 1982 A
4328548 Crow et al. May 1982 A
4328813 Ray May 1982 A
4339953 Iwasaki Jul 1982 A
4341220 Perry Jul 1982 A
4346384 Raab Aug 1982 A
4358856 Stivender et al. Nov 1982 A
4368536 Pfeiler Jan 1983 A
4396885 Constant Aug 1983 A
4396945 DiMatteo et al. Aug 1983 A
4403321 Kruger Sep 1983 A
4418422 Richter et al. Nov 1983 A
4419012 Stephenson et al. Dec 1983 A
4422041 Lienau Dec 1983 A
4431005 McCormick Feb 1984 A
4447224 DeCant, Jr. et al. May 1984 A
4447462 Tafuri et al. May 1984 A
4485815 Amplatz et al. Dec 1984 A
4506676 Duska Mar 1985 A
4543959 Sepponen Oct 1985 A
4548208 Niemi Oct 1985 A
4571834 Fraser et al. Feb 1986 A
4572198 Codrington Feb 1986 A
4583538 Onik et al. Apr 1986 A
4584577 Temple Apr 1986 A
4586491 Carpenter May 1986 A
4607619 Seike et al. Aug 1986 A
4608977 Brown Sep 1986 A
4613866 Blood Sep 1986 A
4617925 Laitinen Oct 1986 A
4618978 Cosman Oct 1986 A
4621628 Brudermann Nov 1986 A
4625718 Olerud et al. Dec 1986 A
4638798 Shelden et al. Jan 1987 A
4642786 Hansen Feb 1987 A
4645343 Stockdale et al. Feb 1987 A
4649504 Krouglicof et al. Mar 1987 A
4651732 Frederick Mar 1987 A
4653509 Oloff et al. Mar 1987 A
4659971 Suzuki et al. Apr 1987 A
4660970 Ferrano Apr 1987 A
4669172 Petruzzi Jun 1987 A
4673352 Hansen Jun 1987 A
4685583 Noon Aug 1987 A
4688037 Krieg Aug 1987 A
4701049 Beckman et al. Oct 1987 A
4705395 Hageniers Nov 1987 A
4705401 Addleman et al. Nov 1987 A
4706665 Gouda Nov 1987 A
4709156 Murphy et al. Nov 1987 A
4710708 Rorden et al. Dec 1987 A
4719419 Dawley Jan 1988 A
4722056 Roberts et al. Jan 1988 A
4722336 Kim et al. Feb 1988 A
4723544 Moore et al. Feb 1988 A
4727565 Ericson Feb 1988 A
RE32619 Damadian Mar 1988 E
4733969 Case et al. Mar 1988 A
4737032 Addleman et al. Apr 1988 A
4737794 Jones Apr 1988 A
4737921 Goldwasser et al. Apr 1988 A
4742356 Kuipers May 1988 A
4742815 Ninan et al. May 1988 A
4743770 Lee May 1988 A
4743771 Sacks et al. May 1988 A
4745290 Frankel et al. May 1988 A
4750487 Zanetti Jun 1988 A
4753528 Hines et al. Jun 1988 A
4761072 Pryor Aug 1988 A
4764016 Johansson Aug 1988 A
4771787 Wurster et al. Sep 1988 A
4779212 Levy Oct 1988 A
4782239 Hirose et al. Nov 1988 A
4787591 Villacorta Nov 1988 A
4788481 Niwa Nov 1988 A
4791934 Brunnett Dec 1988 A
4793355 Crum et al. Dec 1988 A
4794262 Sato et al. Dec 1988 A
4797907 Anderton Jan 1989 A
4803976 Frigg et al. Feb 1989 A
4804261 Kirschen Feb 1989 A
4805615 Carol Feb 1989 A
4809679 Shimonaka et al. Mar 1989 A
4809694 Ferrara Mar 1989 A
4821200 Oberg Apr 1989 A
4821206 Arora Apr 1989 A
4821731 Martinelli et al. Apr 1989 A
4822163 Schmidt Apr 1989 A
4825091 Breyer et al. Apr 1989 A
4829373 Leberl et al. May 1989 A
4836778 Baumrind et al. Jun 1989 A
4838265 Cosman et al. Jun 1989 A
4841967 Chang et al. Jun 1989 A
4845771 Wislocki et al. Jul 1989 A
4849692 Blood Jul 1989 A
4860331 Williams et al. Aug 1989 A
4862893 Martinelli Sep 1989 A
4869247 Howard, III et al. Sep 1989 A
4875165 Fencil et al. Oct 1989 A
4875478 Chen Oct 1989 A
4884566 Mountz et al. Dec 1989 A
4889526 Rauscher et al. Dec 1989 A
4896673 Rose et al. Jan 1990 A
4905698 Strohl, Jr. et al. Mar 1990 A
4923459 Nambu May 1990 A
4931056 Ghajar et al. Jun 1990 A
4945305 Blood Jul 1990 A
4945914 Allen Aug 1990 A
4951653 Fry et al. Aug 1990 A
4951677 Crowley et al. Aug 1990 A
4955891 Carol Sep 1990 A
4961422 Marchosky et al. Oct 1990 A
4977655 Martinelli Dec 1990 A
4989608 Ratner Feb 1991 A
4991579 Allen Feb 1991 A
4994069 Ritchart et al. Feb 1991 A
5002058 Martinelli Mar 1991 A
5005592 Cartmell Apr 1991 A
5013317 Cole et al. May 1991 A
5016639 Allen May 1991 A
5017139 Mushabac May 1991 A
5021888 Kondou et al. Jun 1991 A
5027818 Bova et al. Jul 1991 A
5030196 Inoue Jul 1991 A
5030222 Calandruccio et al. Jul 1991 A
5031203 Trecha Jul 1991 A
5042486 Pfeiler et al. Aug 1991 A
5047036 Koutrouvelis Sep 1991 A
5050608 Watanabe et al. Sep 1991 A
5054492 Scribner et al. Oct 1991 A
5057095 Fabian Oct 1991 A
5059186 Yamamoto et al. Oct 1991 A
5059789 Salcudean Oct 1991 A
5078140 Kwoh Jan 1992 A
5079699 Tuy et al. Jan 1992 A
5086401 Glassman et al. Feb 1992 A
5094241 Allen Mar 1992 A
5097839 Allen Mar 1992 A
5098426 Sklar et al. Mar 1992 A
5099845 Besz et al. Mar 1992 A
5099846 Hardy Mar 1992 A
5104393 Isner et al. Apr 1992 A
5105829 Fabian et al. Apr 1992 A
5107839 Houdek et al. Apr 1992 A
5107843 Aarnio et al. Apr 1992 A
5107862 Fabian et al. Apr 1992 A
5109194 Cantaloube Apr 1992 A
5119817 Allen Jun 1992 A
5142930 Allen et al. Sep 1992 A
5143076 Hardy et al. Sep 1992 A
5152288 Hoenig et al. Oct 1992 A
5160337 Cosman Nov 1992 A
5161536 Vilkomerson et al. Nov 1992 A
5171245 Cezana Dec 1992 A
5178164 Allen Jan 1993 A
5178621 Cook et al. Jan 1993 A
5186174 Schlondorff et al. Feb 1993 A
5187475 Wagener et al. Feb 1993 A
5188126 Fabian et al. Feb 1993 A
5190059 Fabian et al. Mar 1993 A
5193106 DeSena Mar 1993 A
5197476 Nowacki et al. Mar 1993 A
5197965 Cherry et al. Mar 1993 A
5198768 Keren Mar 1993 A
5198877 Schulz Mar 1993 A
5207688 Carol May 1993 A
5211164 Allen May 1993 A
5211165 Dumoulin et al. May 1993 A
5211176 Ishiguro et al. May 1993 A
5212720 Landi et al. May 1993 A
5214615 Bauer May 1993 A
5219351 Teubner et al. Jun 1993 A
5222499 Allen et al. Jun 1993 A
5224049 Mushabac Jun 1993 A
5228442 Imran Jul 1993 A
5230338 Allen et al. Jul 1993 A
5230622 Brossoit Jul 1993 A
5230623 Guthrie et al. Jul 1993 A
5233990 Barnea Aug 1993 A
5237996 Waldman et al. Aug 1993 A
5249581 Horbal et al. Oct 1993 A
5251127 Raab Oct 1993 A
5251635 Dumoulin et al. Oct 1993 A
5253647 Takahashi et al. Oct 1993 A
5255680 Darrow et al. Oct 1993 A
5257636 White Nov 1993 A
5257998 Ota et al. Nov 1993 A
5261404 Mick et al. Nov 1993 A
5265610 Darrow et al. Nov 1993 A
5265611 Hoenig et al. Nov 1993 A
5269759 Hernandez et al. Dec 1993 A
5271400 Dumoulin et al. Dec 1993 A
5273025 Sakiyama et al. Dec 1993 A
5274551 Corby, Jr. Dec 1993 A
5279309 Taylor et al. Jan 1994 A
5285787 Machida Feb 1994 A
5291199 Overman et al. Mar 1994 A
5291889 Kenet et al. Mar 1994 A
5295483 Nowacki et al. Mar 1994 A
5297549 Beatty et al. Mar 1994 A
5299253 Wessels Mar 1994 A
5299254 Dancer et al. Mar 1994 A
5299288 Glassman et al. Mar 1994 A
5300080 Clayman et al. Apr 1994 A
5301061 Nakada et al. Apr 1994 A
5305091 Gelbart et al. Apr 1994 A
5305203 Raab Apr 1994 A
5306271 Zinreich et al. Apr 1994 A
5307072 Jones, Jr. Apr 1994 A
5309913 Kormos et al. May 1994 A
5315630 Sturm et al. May 1994 A
5316024 Hirschi et al. May 1994 A
5318025 Dumoulin et al. Jun 1994 A
5320111 Livingston Jun 1994 A
5320249 Strech Jun 1994 A
5325728 Zimmerman et al. Jul 1994 A
5325873 Hirschi et al. Jul 1994 A
5329944 Fabian et al. Jul 1994 A
5330485 Clayman et al. Jul 1994 A
5333168 Fernandes et al. Jul 1994 A
5341807 Nardella Aug 1994 A
5353795 Souza et al. Oct 1994 A
5353800 Pohndorf et al. Oct 1994 A
5353807 DeMarco Oct 1994 A
5359417 Muller et al. Oct 1994 A
5368030 Zinreich et al. Nov 1994 A
5371778 Yanof et al. Dec 1994 A
5375596 Twiss et al. Dec 1994 A
5377678 Dumoulin et al. Jan 1995 A
5380302 Orth Jan 1995 A
5383454 Bucholz Jan 1995 A
5383852 Stevens-Wright Jan 1995 A
5385146 Goldreyer Jan 1995 A
5385148 Lesh et al. Jan 1995 A
5386828 Owens et al. Feb 1995 A
5389101 Heilbrun et al. Feb 1995 A
5391199 Ben-Haim Feb 1995 A
5394457 Leibinger et al. Feb 1995 A
5394875 Lewis et al. Mar 1995 A
5397329 Allen Mar 1995 A
5398684 Hardy Mar 1995 A
5399146 Nowacki et al. Mar 1995 A
5400384 Fernandes et al. Mar 1995 A
5400771 Pirak et al. Mar 1995 A
5402801 Taylor Apr 1995 A
5408409 Glassman et al. Apr 1995 A
5413573 Koivukangas May 1995 A
5417210 Funda et al. May 1995 A
5419325 Dumoulin et al. May 1995 A
5423334 Jordan Jun 1995 A
5425367 Shapiro et al. Jun 1995 A
5425382 Golden et al. Jun 1995 A
5426683 O'Farrell, Jr. et al. Jun 1995 A
5426687 Goodall et al. Jun 1995 A
5427097 Depp Jun 1995 A
5429132 Guy et al. Jul 1995 A
5432543 Hasegawa Jul 1995 A
5433198 Desai Jul 1995 A
RE35025 Anderton Aug 1995 E
5437277 Dumoulin et al. Aug 1995 A
5443066 Dumoulin et al. Aug 1995 A
5443489 Ben-Haim Aug 1995 A
5444756 Pai et al. Aug 1995 A
5445144 Wodicka et al. Aug 1995 A
5445150 Dumoulin et al. Aug 1995 A
5445166 Taylor Aug 1995 A
5446548 Gerig et al. Aug 1995 A
5447154 Cinquin et al. Sep 1995 A
5448610 Yamamoto et al. Sep 1995 A
5453686 Anderson Sep 1995 A
5456718 Szymaitis Oct 1995 A
5457641 Zimmer et al. Oct 1995 A
5458718 Venkitachalam Oct 1995 A
5464446 Dreessen et al. Nov 1995 A
5469847 Zinreich et al. Nov 1995 A
5474075 Goldberg et al. Dec 1995 A
5478341 Cook et al. Dec 1995 A
5478343 Ritter Dec 1995 A
5480422 Ben-Haim Jan 1996 A
5480439 Bisek et al. Jan 1996 A
5483961 Kelly et al. Jan 1996 A
5485849 Panescu et al. Jan 1996 A
5487391 Panescu Jan 1996 A
5487729 Avellanet et al. Jan 1996 A
5487757 Truckai et al. Jan 1996 A
5489256 Adair Feb 1996 A
5490196 Rudich et al. Feb 1996 A
5494034 Schlondorff et al. Feb 1996 A
5503416 Aoki et al. Apr 1996 A
5506102 McDonnell Apr 1996 A
5513637 Twiss et al. May 1996 A
5514146 Lam et al. May 1996 A
5515160 Schulz et al. May 1996 A
5515641 D'Alessandro May 1996 A
5517990 Kalfas et al. May 1996 A
5522815 Durgin, Jr. et al. Jun 1996 A
5531227 Schneider Jul 1996 A
5531520 Grimson et al. Jul 1996 A
5531686 Lundquist et al. Jul 1996 A
5535973 Bailey et al. Jul 1996 A
5542938 Avellanet et al. Aug 1996 A
5543951 Moehrmann Aug 1996 A
5546940 Panescu et al. Aug 1996 A
5546949 Frazin et al. Aug 1996 A
5546951 Ben-Haim Aug 1996 A
5551429 Fitzpatrick et al. Sep 1996 A
5555883 Avitall Sep 1996 A
5558091 Acker et al. Sep 1996 A
5566681 Manwaring et al. Oct 1996 A
5568384 Robb et al. Oct 1996 A
5568809 Ben-haim Oct 1996 A
5571083 Lemelson Nov 1996 A
5572999 Funda et al. Nov 1996 A
5573493 Sauer et al. Nov 1996 A
5573533 Strul Nov 1996 A
5575794 Walus et al. Nov 1996 A
5575798 Koutrouvelis Nov 1996 A
5577991 Akui et al. Nov 1996 A
5583909 Hanover Dec 1996 A
5588430 Bova et al. Dec 1996 A
5590215 Allen Dec 1996 A
5592939 Martinelli Jan 1997 A
5595193 Walus et al. Jan 1997 A
5596228 Anderton et al. Jan 1997 A
5600330 Blood Feb 1997 A
5603318 Heilbrun et al. Feb 1997 A
5611025 Lorensen et al. Mar 1997 A
5617462 Spratt Apr 1997 A
5617857 Chader et al. Apr 1997 A
5619261 Anderton Apr 1997 A
5622169 Golden et al. Apr 1997 A
5622170 Schulz Apr 1997 A
5627873 Hanover et al. May 1997 A
5628315 Vilsmeier et al. May 1997 A
5630431 Taylor May 1997 A
5636644 Hart et al. Jun 1997 A
5638819 Manwaring et al. Jun 1997 A
5640170 Anderson Jun 1997 A
5642395 Anderton et al. Jun 1997 A
5643175 Adair Jul 1997 A
5643268 Vilsmeier et al. Jul 1997 A
5645065 Shapiro et al. Jul 1997 A
5646524 Gilboa Jul 1997 A
5647361 Damadian Jul 1997 A
5662111 Cosman Sep 1997 A
5664001 Tachibana et al. Sep 1997 A
5674296 Bryan et al. Oct 1997 A
5676673 Ferre et al. Oct 1997 A
5681260 Ueda et al. Oct 1997 A
5682886 Delp et al. Nov 1997 A
5682890 Kormos et al. Nov 1997 A
5690108 Chakeres Nov 1997 A
5694945 Ben-Haim Dec 1997 A
5695500 Taylor et al. Dec 1997 A
5695501 Carol et al. Dec 1997 A
5696500 Diem Dec 1997 A
5697377 Wittkampf Dec 1997 A
5702406 Vilsmeier et al. Dec 1997 A
5711299 Manwaring et al. Jan 1998 A
5713853 Clark et al. Feb 1998 A
5713946 Ben-Haim Feb 1998 A
5715822 Watkins et al. Feb 1998 A
5715836 Kliegis et al. Feb 1998 A
5718241 Ben-Haim et al. Feb 1998 A
5727552 Ryan Mar 1998 A
5727553 Saad Mar 1998 A
5728047 Edoga Mar 1998 A
5729129 Acker Mar 1998 A
5730129 Darrow et al. Mar 1998 A
5730130 Fitzpatrick et al. Mar 1998 A
5732703 Kalfas et al. Mar 1998 A
5735278 Hoult et al. Apr 1998 A
5738096 Ben-Haim Apr 1998 A
5740802 Nafis et al. Apr 1998 A
5741214 Ouchi et al. Apr 1998 A
5742394 Hansen Apr 1998 A
5744953 Hansen Apr 1998 A
5748767 Raab May 1998 A
5749362 Funda et al. May 1998 A
5749835 Glantz May 1998 A
5752513 Acker et al. May 1998 A
5755725 Druais May 1998 A
RE35816 Schulz Jun 1998 E
5758667 Slettenmark Jun 1998 A
5762064 Polvani Jun 1998 A
5767960 Orman Jun 1998 A
5769789 Wang et al. Jun 1998 A
5769843 Abela et al. Jun 1998 A
5769861 Vilsmeier Jun 1998 A
5772594 Barrick Jun 1998 A
5775322 Silverstein et al. Jul 1998 A
5776064 Kalfas et al. Jul 1998 A
5782765 Jonkman Jul 1998 A
5787886 Kelly et al. Aug 1998 A
5792055 McKinnon Aug 1998 A
5795294 Luber et al. Aug 1998 A
5797849 Vesely et al. Aug 1998 A
5799055 Peshkin et al. Aug 1998 A
5799099 Wang et al. Aug 1998 A
5800352 Ferre et al. Sep 1998 A
5800535 Howard, III Sep 1998 A
5802719 O'Farrell, Jr. et al. Sep 1998 A
5803089 Ferre et al. Sep 1998 A
5807252 Hassfeld et al. Sep 1998 A
5810008 Dekel et al. Sep 1998 A
5810728 Kuhn Sep 1998 A
5810735 Halperin et al. Sep 1998 A
5820553 Hughes Oct 1998 A
5820591 Thompson et al. Oct 1998 A
5823192 Kalend et al. Oct 1998 A
5823958 Truppe Oct 1998 A
5828725 Levinson Oct 1998 A
5828770 Leis et al. Oct 1998 A
5829444 Ferre et al. Nov 1998 A
5831260 Hansen Nov 1998 A
5833608 Acker Nov 1998 A
5834759 Glossop Nov 1998 A
5836869 Kudo Nov 1998 A
5836954 Heilbrun et al. Nov 1998 A
5840024 Taniguchi et al. Nov 1998 A
5840025 Ben-Haim Nov 1998 A
5842984 Avitall Dec 1998 A
5843051 Adams et al. Dec 1998 A
5843076 Webster, Jr. et al. Dec 1998 A
5846183 Chilcoat Dec 1998 A
5848967 Cosman Dec 1998 A
5851183 Bucholz Dec 1998 A
5865726 Katsurada et al. Feb 1999 A
5865846 Bryan et al. Feb 1999 A
5868674 Glowinski et al. Feb 1999 A
5868675 Henrion et al. Feb 1999 A
5871445 Bucholz Feb 1999 A
5871455 Ueno Feb 1999 A
5871487 Warner et al. Feb 1999 A
5873822 Ferre et al. Feb 1999 A
5879499 Corvi Mar 1999 A
5882304 Ehnholm et al. Mar 1999 A
5884410 Prinz Mar 1999 A
5889834 Vilsmeier et al. Mar 1999 A
5891034 Bucholz Apr 1999 A
5891157 Day et al. Apr 1999 A
5893885 Webster, Jr. Apr 1999 A
5904691 Barnett et al. May 1999 A
5907395 Schulz et al. May 1999 A
5913820 Bladen et al. Jun 1999 A
5920395 Schulz Jul 1999 A
5921992 Costales et al. Jul 1999 A
5923727 Navab Jul 1999 A
5928248 Acker Jul 1999 A
5935160 Auricchio et al. Aug 1999 A
5938603 Ponzi Aug 1999 A
5938694 Jaraczewski et al. Aug 1999 A
5947925 Ashiya et al. Sep 1999 A
5947980 Jensen et al. Sep 1999 A
5947981 Cosman Sep 1999 A
5950629 Taylor et al. Sep 1999 A
5951475 Gueziec et al. Sep 1999 A
5951571 Audette Sep 1999 A
5954647 Bova et al. Sep 1999 A
5954796 McCarty et al. Sep 1999 A
5957844 Dekel et al. Sep 1999 A
5967980 Ferre et al. Oct 1999 A
5967982 Barnett Oct 1999 A
5968047 Reed Oct 1999 A
5971997 Guthrie et al. Oct 1999 A
5976156 Taylor et al. Nov 1999 A
5980535 Barnett et al. Nov 1999 A
5983126 Wittkampf Nov 1999 A
5987349 Schulz Nov 1999 A
5987960 Messner et al. Nov 1999 A
5999837 Messner et al. Dec 1999 A
5999840 Grimson et al. Dec 1999 A
6001130 Bryan et al. Dec 1999 A
6004269 Crowley et al. Dec 1999 A
6006126 Cosman Dec 1999 A
6006127 Van Der Brug et al. Dec 1999 A
6013087 Adams et al. Jan 2000 A
6014580 Blume et al. Jan 2000 A
6016439 Acker Jan 2000 A
6019725 Vesely et al. Feb 2000 A
6024695 Taylor et al. Feb 2000 A
6035229 Silverstein et al. Mar 2000 A
6050724 Schmitz et al. Apr 2000 A
6059718 Taniguchi et al. May 2000 A
6061588 Thornton et al. May 2000 A
6063022 Ben-Haim May 2000 A
6071288 Carol et al. Jun 2000 A
6073043 Schneider Jun 2000 A
6076008 Bucholz Jun 2000 A
6086529 Arndt Jul 2000 A
6096050 Audette Aug 2000 A
6104944 Martinelli Aug 2000 A
6112111 Glantz Aug 2000 A
6117070 Akiba Sep 2000 A
6118845 Simon et al. Sep 2000 A
6122538 Sliwa, Jr. et al. Sep 2000 A
6122541 Cosman et al. Sep 2000 A
6131396 Duerr et al. Oct 2000 A
6134003 Tearney et al. Oct 2000 A
6139183 Graumann Oct 2000 A
6147480 Osadchy et al. Nov 2000 A
6149592 Yanof et al. Nov 2000 A
6156067 Bryan et al. Dec 2000 A
6161032 Acker Dec 2000 A
6165181 Heilbrun et al. Dec 2000 A
6167296 Shahidi Dec 2000 A
6171303 Ben-Haim et al. Jan 2001 B1
6172499 Ashe Jan 2001 B1
6175756 Ferre et al. Jan 2001 B1
6178345 Vilsmeier et al. Jan 2001 B1
6183444 Glines et al. Feb 2001 B1
6188355 Gilboa Feb 2001 B1
6192280 Sommer et al. Feb 2001 B1
6194639 Botella et al. Feb 2001 B1
6200262 Ouchi Mar 2001 B1
6201387 Govari Mar 2001 B1
6203493 Ben-Haim Mar 2001 B1
6203497 Dekel et al. Mar 2001 B1
6210362 Ponzi Apr 2001 B1
6210378 Ouchi Apr 2001 B1
6211666 Acker Apr 2001 B1
6213995 Steen et al. Apr 2001 B1
6216027 Willis et al. Apr 2001 B1
6223067 Vilsmeier et al. Apr 2001 B1
6226543 Gilboa et al. May 2001 B1
6233476 Strommer et al. May 2001 B1
6236875 Bucholz et al. May 2001 B1
6246231 Ashe Jun 2001 B1
6246784 Summers et al. Jun 2001 B1
6253770 Acker et al. Jul 2001 B1
6259942 Westermann et al. Jul 2001 B1
6273896 Franck et al. Aug 2001 B1
6285902 Kienzle, III et al. Sep 2001 B1
6286798 Chun Sep 2001 B1
6298262 Franck et al. Oct 2001 B1
6306097 Park et al. Oct 2001 B1
6314310 Ben-Haim et al. Nov 2001 B1
6319250 Falwell et al. Nov 2001 B1
6332089 Acker et al. Dec 2001 B1
6335617 Osadchy et al. Jan 2002 B1
6341231 Ferre et al. Jan 2002 B1
6343728 Carbone Feb 2002 B1
6345112 Summers et al. Feb 2002 B1
6346940 Fukunaga Feb 2002 B1
6351659 Vilsmeier Feb 2002 B1
6355027 Le et al. Mar 2002 B1
6366799 Acker et al. Apr 2002 B1
6368316 Jansen et al. Apr 2002 B1
6373240 Govari Apr 2002 B1
6381485 Hunter et al. Apr 2002 B1
6381490 Ostrovsky Apr 2002 B1
6422994 Kaneko et al. Jul 2002 B1
6423009 Downey et al. Jul 2002 B1
6424856 Vilsmeier et al. Jul 2002 B1
6427314 Acker Aug 2002 B1
6428547 Vilsmeier et al. Aug 2002 B1
6434415 Foley et al. Aug 2002 B1
6437567 Schenck et al. Aug 2002 B1
6443894 Sumanaweera et al. Sep 2002 B1
6445943 Ferre et al. Sep 2002 B1
6447504 Ben-Haim et al. Sep 2002 B1
6453190 Acker et al. Sep 2002 B1
6468265 Evans et al. Oct 2002 B1
6470207 Simon et al. Oct 2002 B1
6474341 Hunter et al. Nov 2002 B1
6478802 Kienzle, III et al. Nov 2002 B2
6484049 Seeley et al. Nov 2002 B1
6484118 Govari Nov 2002 B1
6485413 Boppart et al. Nov 2002 B1
6490475 Seeley et al. Dec 2002 B1
6493573 Martinelli et al. Dec 2002 B1
6498944 Ben-Haim et al. Dec 2002 B1
6499488 Hunter et al. Dec 2002 B1
6503195 Keller et al. Jan 2003 B1
6516046 Frohlich et al. Feb 2003 B1
6527443 Vilsmeier et al. Mar 2003 B1
6547722 Higuma et al. Apr 2003 B1
6551325 Neubauer et al. Apr 2003 B2
6579277 Rabiner et al. Jun 2003 B1
6580938 Acker Jun 2003 B1
6584174 Schubert et al. Jun 2003 B2
6585639 Kotmel et al. Jul 2003 B1
6591129 Ben-Haim et al. Jul 2003 B1
6593884 Gilboa et al. Jul 2003 B1
6609022 Vilsmeier et al. Aug 2003 B2
6611700 Vilsmeier et al. Aug 2003 B1
6612485 Lackner et al. Sep 2003 B2
6618612 Acker et al. Sep 2003 B1
6626339 Gates et al. Sep 2003 B2
6628980 Atalar et al. Sep 2003 B2
6631876 Phillips Oct 2003 B1
6640128 Vilsmeier et al. Oct 2003 B2
6650927 Keidar Nov 2003 B1
6666864 Bencini et al. Dec 2003 B2
6676659 Hutchins et al. Jan 2004 B2
6690963 Ben-Haim et al. Feb 2004 B2
6694162 Hartlep Feb 2004 B2
6701179 Martinelli et al. Mar 2004 B1
6706041 Costantino Mar 2004 B1
6711429 Gilboa et al. Mar 2004 B1
6712842 Gifford, III et al. Mar 2004 B1
6751492 Ben-Haim Jun 2004 B2
6770027 Banik et al. Aug 2004 B2
6788967 Ben-Haim et al. Sep 2004 B2
6796963 Carpenter et al. Sep 2004 B2
6850794 Shahidi Feb 2005 B2
6887236 Gilboa May 2005 B2
6947788 Gilboa et al. Sep 2005 B2
6960161 Amling et al. Nov 2005 B2
6995729 Govari et al. Feb 2006 B2
7022066 Yokoi et al. Apr 2006 B2
7101380 Khachin et al. Sep 2006 B2
7182756 Saeed et al. Feb 2007 B2
7197354 Sobe Mar 2007 B2
7233820 Gilboa Jun 2007 B2
7236567 Sandkamp et al. Jun 2007 B2
7286868 Govari Oct 2007 B2
7301332 Govari et al. Nov 2007 B2
7321228 Govari Jan 2008 B2
7324915 Altmann et al. Jan 2008 B2
7343195 Strommer et al. Mar 2008 B2
7353125 Nieminen et al. Apr 2008 B2
7357795 Kaji et al. Apr 2008 B2
7366562 Dukesherer et al. Apr 2008 B2
7370656 Gleich et al. May 2008 B2
7373271 Schneider May 2008 B1
7386339 Strommer et al. Jun 2008 B2
7397364 Govari Jul 2008 B2
7399296 Poole et al. Jul 2008 B2
7497029 Plassky et al. Mar 2009 B2
7505809 Strommer et al. Mar 2009 B2
7536218 Govari et al. May 2009 B2
RE40852 Martinelli et al. Jul 2009 E
7570987 Raabe et al. Aug 2009 B2
7577474 Vilsmeier Aug 2009 B2
7579837 Fath et al. Aug 2009 B2
7587235 Wist et al. Sep 2009 B2
7597296 Conway Oct 2009 B2
7599535 Kiraly et al. Oct 2009 B2
7599810 Yamazaki Oct 2009 B2
7630753 Simon et al. Dec 2009 B2
7634122 Bertram et al. Dec 2009 B2
7636595 Marquart et al. Dec 2009 B2
7641609 Ohnishi et al. Jan 2010 B2
7648458 Niwa et al. Jan 2010 B2
7652468 Kruger et al. Jan 2010 B2
7657300 Hunter et al. Feb 2010 B2
7659912 Akimoto et al. Feb 2010 B2
7660623 Hunter et al. Feb 2010 B2
7680528 Pfister et al. Mar 2010 B2
7684849 Wright et al. Mar 2010 B2
7686767 Maschke Mar 2010 B2
7688064 Shalgi et al. Mar 2010 B2
7696899 Immerz et al. Apr 2010 B2
7697972 Verard et al. Apr 2010 B2
7697973 Strommer et al. Apr 2010 B2
7697974 Jenkins et al. Apr 2010 B2
7720517 Drysen May 2010 B2
7722565 Wood et al. May 2010 B2
7725154 Beck et al. May 2010 B2
7725164 Suurmond et al. May 2010 B2
7727269 Abraham-Fuchs et al. Jun 2010 B2
7729742 Govari Jun 2010 B2
7744605 Vilsmeier et al. Jun 2010 B2
7747307 Wright et al. Jun 2010 B2
7751865 Jascob et al. Jul 2010 B2
8083432 Limpert Dec 2011 B2
8317149 Greenburg et al. Nov 2012 B2
8663088 Greenburg et al. Mar 2014 B2
9055881 Gilboa et al. Jun 2015 B2
20010004676 Ouchi Jun 2001 A1
20010007918 Vilsmeier et al. Jul 2001 A1
20010007925 Ritchart et al. Jul 2001 A1
20010031919 Strommer et al. Oct 2001 A1
20010034530 Malackowski et al. Oct 2001 A1
20010036245 Kienzle et al. Nov 2001 A1
20010037064 Shahidi Nov 2001 A1
20020026097 Akiba Feb 2002 A1
20020035864 Paltieli et al. Mar 2002 A1
20020067408 Adair et al. Jun 2002 A1
20020077544 Shahidi Jun 2002 A1
20020087100 Onuki et al. Jul 2002 A1
20020095081 Vilsmeier Jul 2002 A1
20020128565 Rudy Sep 2002 A1
20020137014 Anderson et al. Sep 2002 A1
20020162555 West et al. Nov 2002 A1
20020165503 Morris et al. Nov 2002 A1
20020173689 Kaplan Nov 2002 A1
20020193686 Gilboa Dec 2002 A1
20030018251 Solomon Jan 2003 A1
20030028096 Niwa et al. Feb 2003 A1
20030040657 Yamaya et al. Feb 2003 A1
20030069474 Couvillon Apr 2003 A1
20030074011 Gilboa et al. Apr 2003 A1
20030086599 Armato et al. May 2003 A1
20030114742 Lewkowicz et al. Jun 2003 A1
20030135119 Lee et al. Jul 2003 A1
20030142753 Gunday Jul 2003 A1
20030160721 Gilboa et al. Aug 2003 A1
20030171653 Yokoi et al. Sep 2003 A1
20030227547 Iddan Dec 2003 A1
20040015049 Zaar Jan 2004 A1
20040024309 Ferre et al. Feb 2004 A1
20040086161 Sivaramakrishna et al. May 2004 A1
20040097804 Sobe May 2004 A1
20040138548 Strommer et al. Jul 2004 A1
20040169509 Czipott et al. Sep 2004 A1
20040249267 Gilboa Dec 2004 A1
20040260201 Mueller Dec 2004 A1
20050011786 Wood et al. Jan 2005 A1
20050033149 Strommer et al. Feb 2005 A1
20050054895 Hoeg et al. Mar 2005 A1
20050059890 Deal et al. Mar 2005 A1
20050080342 Gilreath et al. Apr 2005 A1
20050085715 Dukesherer et al. Apr 2005 A1
20050085718 Shahidi Apr 2005 A1
20050090818 Pike et al. Apr 2005 A1
20050107688 Strommer May 2005 A1
20050119527 Banik et al. Jun 2005 A1
20050182292 Suzuki Aug 2005 A1
20050182295 Soper et al. Aug 2005 A1
20050197566 Strommer et al. Sep 2005 A1
20050229934 Willeford Oct 2005 A1
20060015126 Sher Jan 2006 A1
20060058647 Strommer et al. Mar 2006 A1
20060064006 Strommer et al. Mar 2006 A1
20060069313 Couvillon et al. Mar 2006 A1
20060149134 Soper Jul 2006 A1
20060184016 Glossop Aug 2006 A1
20070083107 Ferre et al. Apr 2007 A1
20070163597 Mikkaichi et al. Jul 2007 A1
20070167714 Kiraly et al. Jul 2007 A1
20070167738 Timinger et al. Jul 2007 A1
20070167743 Honda et al. Jul 2007 A1
20070167804 Park et al. Jul 2007 A1
20070167806 Wood et al. Jul 2007 A1
20070225553 Shahidi Sep 2007 A1
20070225559 Clerc et al. Sep 2007 A1
20070265639 Danek et al. Nov 2007 A1
20070276180 Greenburg et al. Nov 2007 A1
20070287901 Strommer et al. Dec 2007 A1
20070293721 Gilboa Dec 2007 A1
20080086051 Voegele Apr 2008 A1
20080097154 Makower et al. Apr 2008 A1
20080097187 Gielen et al. Apr 2008 A1
20080118135 Averbuch et al. May 2008 A1
20080125760 Gilboa May 2008 A1
20080132757 Tgavalekos Jun 2008 A1
20080132909 Jascob et al. Jun 2008 A1
20080132911 Sobe Jun 2008 A1
20080139886 Tatsuyama Jun 2008 A1
20080139915 Dolan et al. Jun 2008 A1
20080144909 Wiemker et al. Jun 2008 A1
20080147000 Seibel et al. Jun 2008 A1
20080154172 Mauch Jun 2008 A1
20080157755 Kruger et al. Jul 2008 A1
20080161682 Kendrick et al. Jul 2008 A1
20080162074 Schneider Jul 2008 A1
20080183071 Strommer et al. Jul 2008 A1
20080188749 Rasche et al. Aug 2008 A1
20090182224 Shmarak et al. Jul 2009 A1
20090234223 Onoda et al. Sep 2009 A1
20090318797 Hadani Dec 2009 A1
20100016757 Greenburg et al. Jan 2010 A1
20130158346 Soper et al. Jun 2013 A1
Foreign Referenced Citations (104)
Number Date Country
964149 Mar 1975 CA
3042343 Jun 1982 DE
3508730 Sep 1986 DE
3520782 Dec 1986 DE
3717871 Dec 1988 DE
3831278 Mar 1989 DE
3838011 Jul 1989 DE
4213426 Oct 1992 DE
4225112 Dec 1993 DE
4233978 Apr 1994 DE
19610984 Sep 1997 DE
19715202 Oct 1998 DE
19751761 Oct 1998 DE
19832296 Feb 1999 DE
19747427 May 1999 DE
10085137 Nov 2002 DE
0062941 Oct 1982 EP
0119660 Sep 1984 EP
0155857 Sep 1985 EP
0319844 Jun 1989 EP
0326768 Aug 1989 EP
0350996 Jan 1990 EP
0419729 Apr 1991 EP
0427358 May 1991 EP
0456103 Nov 1991 EP
0581704 Feb 1994 EP
0600610 Jun 1994 EP
0651968 May 1995 EP
0655138 May 1995 EP
0796633 Sep 1997 EP
0857461 Aug 1998 EP
0894473 Feb 1999 EP
0908146 Apr 1999 EP
0930046 Jul 1999 EP
1078644 Feb 2001 EP
1255113 Nov 2002 EP
1543765 Jun 2005 EP
1667749 Jun 2006 EP
2096523 Sep 2009 EP
2417970 Sep 1979 FR
2618211 Jan 1989 FR
2094590 Sep 1982 GB
2164856 Apr 1986 GB
63-240851 Oct 1988 JP
03-267054 Nov 1991 JP
06-125869 May 1994 JP
06194639 Jul 1994 JP
07-043619 Feb 1995 JP
09-253038 Sep 1997 JP
10-197807 Jul 1998 JP
2000-075218 Mar 2000 JP
2000-279379 Oct 2000 JP
2001-231743 Aug 2001 JP
2001-275942 Oct 2001 JP
8809151 Dec 1988 WO
8905123 Jun 1989 WO
9005494 May 1990 WO
9103982 Apr 1991 WO
9104711 Apr 1991 WO
9107726 May 1991 WO
9203090 Mar 1992 WO
9206645 Apr 1992 WO
9404938 Mar 1994 WO
9423647 Oct 1994 WO
9424933 Nov 1994 WO
9507055 Mar 1995 WO
9611624 Apr 1996 WO
9632059 Oct 1996 WO
9729682 Aug 1997 WO
9729684 Aug 1997 WO
9736192 Oct 1997 WO
9749453 Dec 1997 WO
9808554 Mar 1998 WO
9838908 Sep 1998 WO
9915097 Apr 1999 WO
9921498 May 1999 WO
9923956 May 1999 WO
9926549 Jun 1999 WO
9927839 Jun 1999 WO
9929253 Jun 1999 WO
9933406 Jul 1999 WO
9937208 Jul 1999 WO
9938449 Aug 1999 WO
9952094 Oct 1999 WO
9960939 Dec 1999 WO
0006701 Feb 2000 WO
0014056 Mar 2000 WO
0016684 Mar 2000 WO
0035531 Jun 2000 WO
0119235 Mar 2001 WO
0130437 May 2001 WO
0167035 Sep 2001 WO
0187136 Nov 2001 WO
0187398 Nov 2001 WO
0191842 Dec 2001 WO
0224054 Mar 2002 WO
02064011 Aug 2002 WO
02070047 Sep 2002 WO
03086498 Oct 2003 WO
2004023986 Mar 2004 WO
2005025635 Mar 2005 WO
2005074380 Aug 2005 WO
2006116597 Nov 2006 WO
2007109418 Sep 2007 WO
Non-Patent Literature Citations (59)
Entry
European Patent Office, Supplementary European Search Report dated Sep. 18, 2008, 4 pages.
United States Patent and Trademark Office, Final Office Action dated May 1, 2012 in U.S. Appl. No. 12/476,976, 6 pages.
European Patent Office, Decision to Grant dated Apr. 13, 2012 in European Patent Application No. 10191689, 1 page.
United States Patent and Trademark Office, Office Action dated Feb. 22, 2012 in U.S. Appl. No. 12/233,933, 10 pages.
European Patent Office, Extended European Search Report dated Feb. 20, 2012 in European Patent Application No. 06701745, 9 pages.
United States Patent and Trademark Office, Final Office Action dated Dec. 19, 2011 in U.S. Appl. No. 10/571,793, 8 pages.
European Patent Office, Extended European Search Report dated Nov. 22, 2011 in European Patent Application No. 11182823, 5 pages.
European Patent Office, Extended European Search Report dated Nov. 21, 2011 in European Patent Application No. 11182823, 5 pages.
United States Patent and Trademark Office, Office Action dated Nov. 18, 2011 in U.S. Appl. No. 12/476,976, 8 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Oct. 7, 2011 in International Patent Application No. PCT/US2011/040579, 8 pages.
European Patent Office, Extended European Search Report dated Sep. 6, 2011 in European Patent Application No. 03719056, 6 pages.
European Patent Office, Extended European Search Report dated Sep. 6, 2011 in European Patent Application No. 11174666, 6 pages.
Japanese Patent Office, Official Action dated Aug. 23, 2011 in Japanese Patent Application No. 2007-552806, 7 pages.
Japanese Patent Office, Examiner's Report dated Aug. 19, 2011 in Japanese Patent Application No. JP2007-552806, 7 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Jun. 30, 2011 in International Patent Application No. PCT/US2009/069073, 6 pages.
United States Patent and Trademark Office, Office Action dated May 24, 2011 in U.S. Appl. No. 10/571,793, 8 pages.
United States Patent and Trademark Office, Office Action dated Mar. 31, 2011 in U.S. Appl. No. 12/643,917, 10 pages.
European Patent Office, Extended European Search Report dated Mar. 8, 2011 in European Patent Application No. 10191689, 4 pages.
United States Patent and Trademark Office, Office Action dated Dec. 23, 2010 in U.S. Appl. No. 10/571,793, 11 pages.
European Patent Office, Supplementary European Search Report dated Nov. 15, 2010 in European Patent Application No. EP10159373.9, 12 pages.
United States Patent and Trademark Office, Office Action dated Oct. 4. 2010 in U.S. Appl. No. 12/271,175, 11 pages.
European Patent Office, Examination Report dated Sep. 11, 2010 in European Patent Application No. 3719056, 4 pages.
United States Patent and Trademark Office, Final Office Action dated Jun. 23, 2010 in U.S. Appl. No. 10/571,793, 10 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Apr. 23, 2010 in International Patent Application No. PCT/US2009/069073, 8 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Apr. 8, 2010 in International Patent Application No. PCT/IB2008/002543, 7 pages.
European Patent Office, Examination Report dated Mar. 30, 2010 in European Patent Application No. EP05737664.2, 5 pages.
Japanese Patent Office, Official Action dated Mar. 12, 2010 in Japanese Patent Application No. 2006-526007, 5 pages.
European Patent Office, Extended European Search Report dated Dec. 1, 2009 in European Patent Application No. 09157586, 7 pages.
United States Patent and Trademark Office, Office Action dated Nov. 27, 2009 in U.S. Appl. No. 10/571,793, 11 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Nov. 12, 2009 in International Patent Application No. PCT/IL2009/000697, 9 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Sep. 22, 2009 in International Patent Application No. PCT/IL2009/000553, 7 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Jul. 28, 2009 in International Patent Application No. PCT/IL2005/000159, 6 pages.
European Patent Office, Examination Report dated Jul. 14, 2009 in European Patent Application No. 03719056, 6 pages.
United States Patent and Trademark Office, Office Action dated Jun. 24, 2009 in U.S. Appl. No. 10/571,695, 11 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Mar. 30, 2009 in International Patent Application No. PCT/IL2006/000113, 6 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Mar. 16, 2009 in International Patent Application No. PCT/IB2008/002543, 9 pages.
United States Patent and Trademark Office, Final Office Action dated Mar. 12, 2009 in U.S. Appl. No. 10/597,747, 7 pages.
European Patent Office, Supplementary European Search Report dated Feb. 27, 2009 in European Patent Application No. 03719056, 6 pages.
European Patent Office, Decision to Grant dated Feb. 20, 2009 in European Patent Application No. 04770514, 24 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Dec. 15, 2008 in International Patent Application No. PCT/IL2006/000113, 6 pages.
Japanese Patent Office, Official Action dated Dec. 12, 2008 in Japanese Patent Application No. 2008-583508, 9 pages.
European Patent Office, Supplementary European Search Report dated Oct. 7, 2008 in European Patent Application No. 04770514, 4 pages.
United States Patent and Trademark Office, Office Action dated Sep. 11, 2008 in U.S. Appl. No. 10/597,747, 9 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Jul. 11, 2008 in International Patent Application No. PCT-IL2005/000159, 12 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Oct. 9, 2007 in International Patent Application No. PCT/IL2004/000843, 4 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Sep. 24, 2007 in International Patent Application No. PCT/IL2004/000843, 4 pages.
WIPO, U.S. International Search Authority, International Search Report and Written Opinion dated Jul. 11, 2007 in International Patent Application No. PCT/IL2005/000159, 6 pages.
United States Patent and Trademark Office, Notice of Allowance dated Oct. 6, 2006 in U.S. Appl. No. 10/491,099, 7 pages.
China Patent and Trademark Office, Office Action dated Jun. 19, 2006 in Chinese Patent Application No. 038135485, 5 pages.
United States Patent and Trademark Office, Office Action dated Dec. 30, 2005 in U.S. Appl. No. 10/491,099, 15 pages.
United States Patent and Trademark Office, Office Action dated Apr. 22, 2005 in U.S. Appl. No. 10/491,099, 5 pages.
United States Patent and Trademark Office, Notice of Allowance dated Jan. 3, 2005 in U.S. Appl. No. 10/137,415, 9 pages.
Shmarak, I. et al., U.S. Appl. No. 10/986,567, filed Nov. 2004 (abandoned, unpublished), 84 pages.
United States Patent and Trademark Office, Office Action dated Jul. 1, 2004 in U.S. Appl. No. 10/10,137,415, 14 pages.
WIPO, U.S. International Preliminary Examining Authority, International Preliminary Report on Patentability dated Jan. 24, 2004 in International Patent Application No. PCT/IL2003/000323, 3 pages.
WIPO, U.S. International Search Authority, International Search Report dated Dec. 8, 2003 in International Patent Application No. PCT/IL2003/000323, 1 page.
Stenoien, D.L. et al., “Ligand-Mediated Assembly and Real-Time Cellular Dynamics of Estrogen Receptor .alpha.—Coactivator Complexes in Living Cells,” Molecular and Cellular Biology, Jul. 2001, pp. 4404-4412, 9 pages.
McKenna, N.J. et al., “Nuclear Receptor Coregulators: Cellular and Molecular Biology,” Endocrine Reviews 20(3):321-344, Jun. 1, 1999, 24 pages.
Ding, X.F. et al., “Nuclear Receptor-Binding Sites of Coactivators Glucocorticoid Receptor Interacting Protein 1 (GRIP1) and Steroid Receptor Coactivator 1 (SRC-1): Multiple Motifs with Different Binding Specificities,” Molecular Endocrinology12:302-313, Feb. 1, 1998 (9 pages).
Related Publications (1)
Number Date Country
20150223668 A1 Aug 2015 US
Continuations (1)
Number Date Country
Parent 11568260 US
Child 14694181 US