Luminal background cleaning

Abstract
Apparatus and methods are described for use with an input angiogram image of a device inserted inside a portion of subject's body, the angiogram image being acquired in the presence of contrast agent. At least one processor (11) includes background-image-generation functionality (13) configured to generate a background image in which a relative value is assigned to a first pixel with respect to a second pixel, at least partially based upon relative values of surroundings of the first pixel and the surroundings of the second pixel in the input image. Cleaned-image-generation functionality (14) generates a cleaned image in which visibility of the radiopaque portions of the device is increased relative to the input image, by dividing the input image by the background image. Output-generation functionality (15) drives a display (16) to display an output based upon the cleaned image. Other applications are also described.
Description
FIELD OF EMBODIMENTS OF THE INVENTION

Applications of the present invention generally relate to medical image processing. Specifically, applications of the present invention relate to background cleaning in images of body lumens and body cavities.


BACKGROUND OF THE INVENTION

Vascular catheterizations, such as coronary catheterizations, are frequently-performed medical interventions. Such interventions are typically performed in order to diagnose the blood vessels for potential disease, and/or to treat diseased blood vessels. Typically, in order to facilitate visualization of blood vessels, the catheterization is performed under extraluminal imaging. Typically, and in order to highlight the vasculature during such imaging, a contrast agent is periodically injected into the applicable vasculature. The contrast agent typically remains in the vasculature only momentarily. During the time that the contrast agent is present in the applicable vasculature, the contrast agent typically hides, in full or in part, or obscures, devices positioned or deployed within that vasculature.


The following articles do not necessarily pertain to medical procedures or body organs, but yet serve as a useful technical background.


An article entitled “Nonlocal linear image regularization and supervised segmentation,” by Gilboa and Osher (SIAM Multiscale Modeling & Simulation, volume 6, issue 2, pp. 595-630, 2007), which is incorporated herein by reference, describes how a nonlocal quadratic functional of weighted differences is examined. The weights are based on image features and represent the affinity between different pixels in the image. By prescribing different formulas for the weights, one can generalize many local and nonlocal linear de-noising algorithms, including the nonlocal means filter and the bilateral filter. In this framework one can show that continuous iterations of the generalized filter obey certain global characteristics and converge to a constant solution. The linear operator associated with the Euler-Lagrange equation of the functional is closely related to the graph Laplacian. Thus, the steepest descent for minimizing the functional as a nonlocal diffusion process may be determined. This formulation allows a convenient framework for nonlocal variational minimizations, including variational denoising, Bregman iterations and the recently-proposed inverse-scale-space. The authors demonstrate how the steepest descent flow can be used for segmentation. Following kernel based methods in machine learning, the generalized diffusion process is used to propagate sporadic initial user's information to the entire image. The process is not explicitly based on a curve length energy and thus can cope well with highly non-convex shapes and corners. Reasonable robustness to noise is achieved.


An article entitled “Nonlocal Operators with Applications to Image Processing,” by Gilboa and Osher (SIAM Multiscale Modeling & Simulation, volume 7, issue 3, pp. 1005-1028, 2008), which is incorporated herein by reference, describes the use of nonlocal operators to define types of flows and functionals for image processing and other applications. The authors describe a main advantage of the technique over classical Partial-Differential-Equation-based (PDE-based) algorithms as being the ability to handle better textures and repetitive structures. This topic can be viewed as an extension of spectral graph theory and the diffusion geometry framework to functional analysis and PDE-like evolutions. Some possible applications and numerical examples of the technique are provided, as is a general framework for approximating Hamilton-Jacobi equations on arbitrary grids in high dimensions, e.g., for control theory.


An article entitled “Non-local regularization of inverse problem,” by Peyre, Bougleux, and Cohenin (Lecture Notes in Computer Science, 2008, Volume 5304/2008, pp. 57-68), which is incorporated herein by reference, proposes a new framework to regularize linear inverse problems using the total variation on non-local graphs. A nonlocal graph allows adaptation of the penalization to the geometry of the underlying function to recover. A fast algorithm computes, iteratively, both the solution of the regularization process and the non-local graph adapted to this solution.


An article entitled “The split Bregman method for L1 regularized problems,” by Goldstein and Osher (SIAM Journal on Imaging Sciences, Volume 2, Issue 2, pp. 323-343), which is incorporated herein by reference, notes that the class of 11-regularized optimization problems has received much attention recently because of the introduction of “compressed sensing,” which allows images and signals to be reconstructed from small amounts of data. Despite this recent attention, many 11-regularized problems still remain difficult to solve, or require techniques that are very problem-specific. The authors show that Bregman iteration can be used to solve a wide variety of constrained optimization problems. Using this technique, the authors propose a “Split Bregman” method, which can solve a very broad class of 11-regularized problems.


In an article entitled “Bregmanized nonlocal regularization for reconvolution and sparse reconstruction,” by Zhang, Burgery, Bresson, and Osher (SIAM Journal on Imaging Sciences, Volume 3, Issue 3, July 2010), which is incorporated herein by reference, the authors propose two algorithms based on Bregman iteration and operator splitting technique for nonlocal TV regularization problems. The convergence of the algorithms is analyzed and applications to deconvolution and sparse reconstruction are presented.


SUMMARY OF EMBODIMENTS

Some applications of the present invention are applied to medical procedures performed, in whole or in part, on or within luminal body structures or body cavities. For some applications, apparatus and methods are provided for facilitating the visualization of devices positioned or deployed within a lumen or cavity at a time when the lumen or cavity is injected with contrast agent.


It should be appreciated that while using coronary catheterization as a primary example, applications of the current invention may be applied to any medical procedure in which a medical device is positioned or deployed within a body lumen or cavity, while the lumen or cavity is injected with a substance for the purpose of better discerning that lumen or cavity by means of medical imaging. Such lumens or cavities include, without limitation, any lumen or cavity of the cardiovascular system, the gastro-intestinal tract, the respiratory tract, the urinary tract, the nasal cavities, and/or any other bodily lumen or cavity.


There is therefore provided, in accordance with some applications of the present invention, apparatus for use with an input angiogram image of a device inserted inside a portion of a body of a subject, the device including radiopaque portions thereof, the angiogram image being acquired in the presence of contrast agent within the body portion, and for use with a display, the apparatus including:


at least one processor that includes:

    • background-image-generation functionality configured to generate a background image in which a relative value is assigned to a first pixel with respect to a second pixel, at least partially based upon relative values of surroundings of the first pixel and the surroundings of the second pixel in the input image;
    • cleaned-image-generation functionality configured to generate a cleaned image in which visibility of the radiopaque portions of the device is increased relative to the input image, by dividing the input image by the background image; and
    • output-generation functionality configured to drive the display to display an output based upon the cleaned image.


For some applications, the background-image-generation functionality is configured to generate the background image by:


in the background image, assigning pixel values to the first pixel and the second pixel that are more similar to one another, than the similarity of pixel values that are assigned to a third pixel and a fourth pixel,


based upon the first pixel and the second pixel having more similar surroundings to one another in the input image, than a similarity of surroundings of the third pixel and the fourth pixel to one another in the input image.


For some applications, the background-image-generation functionality is configured to generate the background image by assigning values to the first and second pixels based upon values of the first pixel and the second pixel in the input image.


For some applications, the output-generation functionality is configured to drive the display to display the cleaned image.


For some applications:


the input angiogram image of the device includes a plurality of input angiogram images of the device,


the cleaned-image-generation functionality is configured to generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images, and


the output-generation functionality is configured to:

    • generate a stabilized image stream by image tracking the cleaned images with respect to each other, based upon locations of the radiopaque portions of the device in the cleaned images, and
    • drive the display to display the stabilized image stream.


For some applications:


the input angiogram image of the device includes a plurality of input angiogram images of the device,


the cleaned-image-generation functionality is configured to generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images, and


the output-generation functionality is configured to:

    • generate a stabilized image stream by image tracking the input images with respect to each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, and
    • drive the display to display the stabilized image stream.


For some applications:


the input angiogram image of the device includes a plurality of input angiogram images of the device,


the cleaned-image-generation functionality is configured to generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images, and


the output-generation functionality is configured to:

    • generate an enhanced image frame by:
      • aligning the cleaned images with each other, based upon locations of the radiopaque portions of the device in the cleaned images, and
      • generating an averaged image frame based upon the aligned images, and
    • drive the display to display the enhanced image frame.


For some applications:


the input angiogram image of the device includes a plurality of input angiogram images of the device,


the cleaned-image-generation functionality is configured to generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images, and


the output-generation functionality is configured to:

    • generate an enhanced image frame by:
      • aligning the input images with each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, and
      • generating an averaged image frame based upon the aligned images, and
    • drive the display to display the enhanced image frame.


For some applications:


the portion of the subject's body includes a lumen of the subject's body,


the device includes an endoluminal data-acquisition device configured to acquire endoluminal data points while the device is at respective locations within the lumen, and


the output-generation functionality is configured:

    • based upon locations of the radiopaque portions of the device in the cleaned image, to determine that a given one of the endoluminal data points corresponds to a given location within the lumen, and
    • to drive the display to display an output, in response to the determining.


For some applications:


the portion of the subject's body includes a lumen of the subject's body,


the device includes an endoluminal data-acquisition device configured to acquire endoluminal data points while the device is at respective locations within the lumen, and


the output-generation functionality is configured:


based upon locations of the radiopaque portions of the device in the cleaned image, to determine that the endoluminal device is at a given location within the lumen, and


in response to the determining that the endoluminal device is at the given location within the lumen, to drive the display to display an endoluminal image of the lumen corresponding to the location.


There is further provided, in accordance with some applications of the present invention, a method for use with an input angiogram image of a device inserted inside a portion of a body of a subject, the device including radiopaque portions thereof, the angiogram image being acquired in the presence of contrast agent within the body portion, the method including:


generating, with a processor, a background image in which a relative value is assigned to a first pixel with respect to a second pixel, at least partially based upon relative values of surroundings of the first pixel and surroundings of the second pixel in the input image;


generating, with the processor, a cleaned image in which visibility of the radiopaque portions of the device is increased relative to the input image, by dividing the input image by the background image; and


generating an output on a display, based upon the cleaned image.


For some applications, generating the background image includes:


in the background image, assigning pixel values to the first pixel and the second pixel that are more similar to one another, than the similarity of pixel values that are assigned to a third pixel and a fourth pixel,


based upon the first pixel and the second pixel having more similar surroundings to one another in the input image, than a similarity of surroundings of the third pixel and the fourth pixel to one another in the input image.


For some applications, generating the background image further includes assigning values to the first and second pixels based upon values of the first pixel and the second pixel in the input image.


For some applications, generating the output includes displaying the cleaned image.


For some applications,


the input angiogram image of the device includes a plurality of input angiogram images of the device,


generating the cleaned image includes generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,


the method further includes generating a stabilized image stream by image tracking the cleaned images with respect to each other, based upon locations of the radiopaque portions of the device in the cleaned images, and


generating the output includes displaying the stabilized image stream.


For some applications,


the input angiogram image of the device includes a plurality of input angiogram images of the device,


generating the cleaned image includes generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,


the method further includes generating a stabilized image stream by image tracking the input images with respect to each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, and


generating the output includes displaying the stabilized image stream.


For some applications,


the input angiogram image of the device includes a plurality of input angiogram images of the device,


generating the cleaned image includes generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,


the method further includes generating an enhanced image frame by:

    • aligning the cleaned images with each other, based upon locations of the radiopaque portions of the device in the cleaned images, and
    • generating an averaged image frame based upon the aligned cleaned images, and


generating the output includes displaying the enhanced image frame.


For some applications,


the input angiogram image of the device includes a plurality of input angiogram images of the device,


generating the cleaned image includes generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,


the method further includes generating an enhanced image frame by:

    • aligning the input images with each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, and
    • generating an averaged image frame based upon the aligned input images, and


generating the output includes displaying the enhanced image frame.


For some applications,


the portion of the subject's body includes a lumen of the subject's body,


the device includes an endoluminal data-acquisition device configured to acquire endoluminal data points while the device is at respective locations within the lumen,


the method further includes, based upon locations of the radiopaque portions of the device in the cleaned image, determining that a given one of the endoluminal data points corresponds to a given location within the lumen, and


generating the output includes generating the output in response to the determining.


For some applications,


the portion of the subject's body includes a lumen of the subject's body,


the device includes an endoluminal device configured to be moved through the lumen,


the method further includes, based upon locations of the radiopaque portions of the device in the cleaned image, determining that the endoluminal device is at a given location within the lumen, and


generating the output includes, in response to the determining that the endoluminal device is at the given location within the lumen, generating an endoluminal image of the lumen corresponding to the location.


The present invention will be more fully understood from the following detailed description of embodiments thereof, taken together with the drawings, in which:





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A is a flow chart, at least some of the steps of which are used to generate a cleaned luminal image from an input luminal image, in accordance with some applications of the present invention;



FIG. 1B is a block diagram showing components of an image-processing system, in accordance with some applications of the present invention;



FIG. 2 shows an input luminal image, to which background cleaning may be subsequently applied, in accordance with some applications of the present invention;



FIG. 3 shows a background luminal image, generated in accordance with some applications of the present invention; and



FIG. 4 shows a clean luminal image, generated in accordance with some applications of the present invention.





DETAILED DESCRIPTION OF EMBODIMENTS
Terminology





    • The terms “medical tool,” “tool”, “device,” and “probe” refer to any type of a diagnostic or therapeutic or other functional tool including, but not limited to, a cardiovascular catheter, a stent delivery and/or placement and/or retrieval tool, a balloon delivery and/or placement and/or retrieval tool, a valve delivery and/or repair and/or placement and/or retrieval tool, a graft delivery and/or placement and/or retrieval tool, a tool for the delivery and/or placement and/or retrieval of an implantable device or of parts of such a device, an implantable device or parts thereof, a tool for closing a gap, a tool for closing a septal defect, a guide wire, a marker wire, a suturing tool, a clipping tool (such as a valve-leaflet-clipping tool), a biopsy tool, an aspiration tool, a navigational tool, a localization tool, a probe comprising one or more location sensors, a tissue characterization probe, a probe for the analysis of fluid, a measurement probe, an electrophysiological probe, a stimulation probe, an ablation tool, a tool for penetrating or opening partial or total occlusions in blood vessels, a drug or substance delivery tool, a chemotherapy tool, a photodynamic therapy tool, a brachytherapy tool, a local irradiation tool, a laser device, a tool for delivering energy, a tool for delivering markers or biomarkers, a tool for delivering biological glue, an irrigation device, a suction device, a ventilation device, a device for delivering and/or placing and/or retrieving a lead of an electrophysiological device, a lead of an electrophysiological device, a pacing device, a coronary sinus device, an imaging device, a sensing probe, a probe comprising an optical fiber, a robotic tool, a tool that is controlled remotely, or any combination thereof.

    • The terms “image” and “imaging” refer to any type of medical imaging, typically presented as a sequence of images and including, but not limited to, imaging using ionizing radiation, imaging using non-ionizing radiation, video, fluoroscopy, angiography, ultrasound, CT, MR, PET, PET-CT, CT angiography, SPECT, Gamma camera imaging, Optical Coherence Tomography (OCT), Near-Infra-Red Spectroscopy (NIRS), Vibration Response Imaging (VRI), Optical Imaging, infrared imaging, electrical mapping imaging, other forms of Functional Imaging, or any combination or fusion thereof. Examples of ultrasound imaging include Endo-Bronchial Ultrasound (EBUS), Trans-Thoracic Echo (TTE), Trans-Esophageal Echo (TEE), Intra-Vascular Ultrasound (IVUS), Intra-Cardiac Ultrasound (ICE), or any combination thereof.

    • The term “contrast agent,” when used in reference to its application in conjunction with imaging, refers to any substance that is used to highlight, and/or enhance in another manner, the anatomical structure, functioning, and/or composition of a bodily organ while the organ is being imaged.

    • The terms “lumen” and “cavity”, refer to any lumen or cavity of the cardiovascular system, the gastro-intestinal tract, the respiratory tract, the urinary tract, the nasal cavities, and any other bodily lumen or cavity.

    • The term “stabilized,” or “stable” when used in the context of displayed images, means a display of a series of images in a manner such that periodic, cyclical, and/or other motion of the body organ(s) being imaged, and/or of a medical tool being observed, is partially or fully reduced, with respect to the entire image frame, or at least a portion thereof.

    • The term “automatic,” when used for describing background cleaning, means “without necessitating user intervention or interaction.” (Such interaction or intervention may still however be optional in some cases.)

    • The term “real time” means without a noticeable delay.

    • The term “near real time” means with a short noticeable delay (such as approximately one or two motion cycles of an applicable organ, and, in the case of procedures relating to organs or vessels the motion of which are primarily as a result of the cardiac cycle, less than two seconds).

    • The term “on-line,” when used in reference to image processing, or to measurements being made on images, means that the image processing is performed, and/or the measurements are made, intra-procedurally, in real time or near real time.





References is now made to FIGS. 1A-1B, which are respectively (A) a flow chart, at least some of the steps of which are used to generate a cleaned luminal image from an input luminal image, in accordance with some applications of the present invention, and (B) a block diagram showing components of a processor of a an image-processing system, in accordance with some applications of the present invention. Applications of the present invention are typically used during medical procedures that are performed, in whole or in part, within body lumens or cavities. For some applications, apparatus and methods are provided for facilitating the visualization of a device positioned or deployed within such a lumen or cavity, at a time when the lumen or cavity is injected with contrast agent.


In accordance with some applications of the present invention, an input image acquired by an imaging device 10 (FIG. 1B) is input into a processor 11 of an image-processing system 12. Typically, within the input image, radiopaque portions of a tool inside a body lumen or a body cavity (and/or other are artifacts inside the lumen or the cavity) are hidden, in full or in part (e.g., are obscured), by contrast agent. The processor generates a cleaned image, in which visibility of the radiopaque portions of the tool (and/or the other artifacts) is improved relative to the input image. Typically, background-image-generation functionality 13 of the processor computes a background image. Subsequently, cleaned-image-generation functionality 14 of the processor divides the input image by the background image, such as to generate the cleaned image, in which visibility of the radiopaque portions of the tool (and/or the other artifacts) is improved relative to the input image. Typically, the cleaned image is generated from the input image automatically by the processor. Further typically, the generation of the cleaned image is performed by the processor on line, with respect to the inputting of the input image into the processor, and/or with respect to acquisition of the input image by an imaging device. Output generation functionality 15 of the processor typically drives a display 16 to display an output based on the cleaned image. For example, the display may display the cleaned image, or the display may display a stabilized image stream, the stabilization of the image stream being performed by performing image tracking based on locations of the radiopaque portions of the tool in the cleaned image.


Typically, in a first step, an input image u0 (FIG. 1A) is inputted to processor 11, e.g., by the image being selected by the processor, or by a user. For some applications, the input image is a coronary angiogram, or a frame from an angiogram, that is acquired while a device that includes radiopaque sections thereof is within the coronary artery. For example, the device may be a balloon, a stent, a guide wire, an imaging probe, or any combination thereof. Typically, during injection of the contrast agent, the radiopaque sections of the device being inserted are hidden, in full or in part (e.g., obscured), by contrast agent.


Subsequent to the input image being inputted, a background image is generated by background-image-generation functionality 13, in accordance with the techniques described hereinbelow.


In the input image, pixels that are near to each other and that lie on the same object, are expected to have approximately the same value. For example, pixels lying on the ribs are expected to have approximately the same value as one another, and pixels lying inside a blood vessel are expected to have approximately the same value as one another. Thus, portions (i.e., pieces) of the input image are expected to have generally homogenous pixel values. However, the input image is not expected to be totally homogenous, since not all pixels lie on the same object. For example, pixels that lie on a rib are expected to have different values from pixels lying on a blood vessel. Thus, the values of the pixels in the input image can be assumed to be generally piecewise homogenous.


The assumption of piecewise homogeneity generally holds for the majority of the image pixels. However, the assumption fails to hold with respect to a portion of the pixels. An example of such pixels is that of the pixels that correspond to the radiopaque portions of an inserted device (e.g., radiopaque markers of a catheter). The values of such pixels are typically different from the values of their surrounding pixels, the surrounding pixels corresponding to the surrounding anatomy (e.g., a blood vessel in which the catheter is placed). Thus, these pixels are non-homogenous with respect to surrounding pixels.


For the purpose of the present invention, the non-homogeneous pixels are considered to be the foreground of the input image. An image that does not contain the non-homogeneous pixels (or in which the visibility of the non-homogeneous pixels is reduced), but which contains the piecewise homogenous pixels is considered to the background image. Thus, in accordance with the present invention, a background image is computed in which large contiguous piecewise homogenous image parts are enhanced relative to the input image, while the non-homogeneous pixels are made less visible relative to the input image. The background image is typically more homogeneous than the input image, and, in the background image, features of the subject's anatomy typically are enhanced relative to the input image, while the visibility of features of the tool is typically reduced relative to the input image.


The background image is typically computed by assigning a relative value to a first pixel in the background image with respect to a second pixel in the background image, based upon the relative values of the surroundings of the first pixel (e.g., a patch of pixels surrounding the first pixel) and the surroundings of the second pixel (e.g., a patch of pixels surrounding the first pixel) in the input image. Thus, pixels that have more similarly appearing surroundings in the input image are assigned more similar values to one another in the background image than pixels that have less similarly appearing surroundings in the input image. Typically, in computing the background image, the aforementioned method for assigning pixel values is traded-off against keeping the value of any given pixel in the background image similar to the value of the pixel in the input image.


For some applications, the background image is generated by computing a background image that is such as to reduce (e.g., that is such as to minimize) the cost of the following function:











x







[



λ
2




(


b


(
x
)


-


u
0



(
x
)



)

2


+




y









(


b


(
y
)


-

b


(
x
)



)

2



w


(

x
,
y

)



dy




]





[

Function





1

]







in which:

    • b(x) is the value of pixel x is the background image;
    • u0(x) is the value of pixel x in the input image;
    • b(y) is the value of pixel y is the background image;
    • w(x,y) is a weight measure that is the inverse of the weighted sum of the squared difference between px (i.e., a patch taken around pixel x) and py (i.e., a patch taken around pixel y).


The first term of Function 1 (i.e., the term in the brackets that appears before the “plus” sign) favors a background image whose pixel values are close to the pixel values of the input image. The second term of Function 1 (i.e., the term in the brackets that appears after the “plus” sign) favors a background image in which pixels the surroundings of which have similar values in the input image, have similar values in the background image. Thus, the resulting background image is typically similar to the input image, and at the same time is typically more homogenous than the input image, giving similar values to similar pixels. λ is a constant that represents the value of the trade-off between the first term and the second term of Function 1, i.e., the trade-off between (a) generating a background image in which the pixels have similar values to the values of the pixels in the input image and (b) generating an image in which the relative values of the pixels in the background image is based upon similarities between patches surrounding respective pixels in the input image. For some applications, the value of λ is set empirically, by testing different values on a wide range of benchmark input images.


For some applications, the background image is generated in accordance with minimization techniques described in “Nonlocal linear image regularization and supervised segmentation,” by Guy Gilboa and Stanley Osher (SIAM Multiscale Modeling & Simulation, volume 6, issue 2, pp. 595630, 2007), and/or in “Nonlocal Operators with Applications to Image Processing,” by Guy Gilboa and Stanley Osher (SIAM Multiscale Modeling & Simulation, volume 7, issue 3, pp. 1005-1028, 2008), both of which articles are incorporated herein by reference. Alternatively or additionally, other minimization techniques are used, such those described in “Non-local regularization of inverse problem,” by Gabriel Peyre, Sebastien Bougleux, and Laurent Cohenin (Lecture Notes in Computer Science, 2008, Volume 5304/2008, pp. 57-68), “The split Bregman method for L1 regularized problems,” by Tom Goldstein and Stanley Osher (SIAM Journal on Imaging Sciences, Volume 2, Issue 2, pp. 323-343), and/or in “Bregmanized nonlocal regularization for reconvolution and sparse reconstruction,” by Xiaoqun Zhang, Martin Burgery, Xavier Bresson, and Stanley Osher (SIAM Journal on Imaging Sciences, Volume 3, Issue 3, July 2010), all of which articles are incorporated herein by reference.


For some applications, an algorithm as described in “Bregmanized nonlocal regularization for reconvolution and sparse reconstruction,” by Zhang et al., which is incorporated herein by reference, is used to generate the background image. For example, algorithm 1 described on Page 17 of the aforementioned article may be used to generate the background image.


For some applications, the weight measure that is used to compute the background image (e.g., in accordance with Function 1, described hereinabove) is computed using the following technique. Given an input image, the processor of the system computes a weight measure between each pixel x in the image, and pixels in the vicinity of pixel x. The weight measure measures the similarity between small image patches centered on respective pixels x. For example, the similarity may be measured by an inverse of the weighted sum of squared differences between these patches. For some applications, the inverse of the weighted sum of squared differences is weighted by a Guassian, e.g., in accordance with techniques described in “A non-local algorithm for image denoising,” by Buades, Coll and Morell (IEEE CVPR 2005, volume 2, pages 60-65), which is incorporated herein by reference.


Typically, subsequent to the generation of the background image of the lumen, a cleaned image is generated by cleaned-image-generation functionality 14, by dividing the input image by the background image. (It should be noted that, throughout the description of the present invention, “dividing the input image by the background image” should be interpreted as being synonymous with “subtracting the background image from the input image” with respect to the mathematical operations that are performed by processor 11.) Typically, in the resulting cleaned image, image elements which are not homogeneous inside the lumen (such as the radiopaque markers of the device inserted in the vessel) remain visible while the vessel itself appears, in whole or in part, “clean” of contrast agent, at least relative to the input image.


Although some of the techniques described herein have been described with reference to an angiogram of a device that in inserted the into the coronary artery, the scope of the present invention includes applying the techniques to images of other body lumen and/or body cavities, mutatis mutandis. For example, the techniques described herein may be applied to an angiogram of the aorta that has an aortic replacement valve placed therein. Alternatively or additionally, the techniques described herein may be applied to an angiogram of a heart chamber that has a replacement valve placed therein. Further alternatively or additionally, the techniques described herein may be applied to an angiogram of a heart ventricle that has a ventricular assist device placed therein.


Reference is now made to FIG. 2, which shows an example of input luminal image. Specifically, the image shown in FIG. 2 is a frame of an angiogram of a coronary artery in which a catheter carrying a stent is inserted. Two radiopaque markers 20 that are disposed on the catheter may be observed at the proximal and distal edges of a balloon on which the pre-deployed stent is mounted.


Reference is now made to FIG. 3, which is a background luminal image generated from the input image, typically automatically and typically on line, in accordance with techniques described hereinabove. In FIG. 3 the contrast-filled coronary artery is clearly visible, while the radiopaque elements of the catheter carrying the stent are not visible.


Reference is now made to FIG. 4, which is a cleaned image that was produced by dividing the input image by the background image, in accordance with the techniques described hereinabove. The cleaned image was generated automatically and on line with respect to the inputting of the input image to the system. It may be observed that in the cleaned image, the contrast-filled vessel is largely invisible, while radiopaque markers 20 of the catheter carrying the stent are seen more clearly than in the input image.


For some applications, a series of cleaned image frames are used to create a stabilized image stream of an angiographic sequence, in which, typically, radiopaque elements of a device (e.g., markers of a catheter carrying a stent) appear relatively stable. Typically, the stabilized image stream of the angiographic sequence is generated in accordance with techniques described in U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556) and/or in U.S. patent application Ser. No. 12/075,244 to Tolkowsky (published as US 2008/0221442), both of which applications are incorporated herein by reference. The stabilized image stream is typically displayed on display 16. For some applications, a plurality of cleaned images are generated, the cleaned images corresponding to respective input images. The input images are stabilized such as to generate a stabilized image stream, based upon the locations of the radiopaque elements of the device in the corresponding cleaned images. Alternatively or additionally, a cleaned, stabilized image stream is generated by stabilizing the cleaned images with respect to each other, based upon the locations of the radiopaque elements of the device in the cleaned images.


For some applications, a stent is deployed within a lumen, and a catheter and/or a balloon carrying radiopaque markers remains within the luminal section in which the stent is deployed. A series of image frames are cleaned in accordance with the techniques described herein, and the cleaned image frames are used to create an enhanced image stream of an angiographic sequence, in which the stent appears more visible than in a native angiogram of the stent. Typically, the enhanced image stream is generated in accordance with techniques described herein, in combination with techniques described in U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556), which is incorporated herein by reference. The enhanced image stream is typically displayed on display 16. For some applications, a plurality of cleaned images are generated, the cleaned images corresponding to respective input images. An enhanced image frame is generated by aligning the input images with each other based upon locations of the radiopaque portions of the device in the corresponding cleaned images, and generating an averaged image frame based upon the aligned input images. Alternatively or additionally, a cleaned, enhanced image frame is generated, by aligning the cleaned images with each other, based upon locations of the radiopaque portions of the device in the cleaned images, and generating an averaged image frame based upon the aligned cleaned images.


For some applications, a series of cleaned image frames are used to create an image stream that is both stabilized and enhanced. Typically, the stabilized, enhanced image stream of the angiographic sequence is generated in accordance with techniques described herein, in combination with techniques described in U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556) and/or in U.S. patent application Ser. No. 12/075,244 to Tolkowsky (published as US 2008/0221442), both of which applications are incorporated herein by reference. The stabilized, enhanced image stream is typically displayed on display 16.


For some applications, a series of input image frames are divided by the respective corresponding background images such as to produce cleaned image frames. The cleaned image frames are used to create an image stream that is stabilized, enhanced and cleaned, by stabilizing and enhancing the cleaned image frames. In an embodiment, such an image stream is produced in accordance with techniques described herein, in combination with techniques described in U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556) and/or in U.S. patent application Ser. No. 12/075,244 to Tolkowsky (published as US 2008/0221442), both of which applications are incorporated herein by reference. The stabilized, enhanced, cleaned image stream is typically displayed on display 16.


For some applications, the visibility of the radiopaque markers on an endoluminal device is increased in an image stream, by cleaning image frames belonging to the image stream, in accordance with the techniques described herein. The increased visibility of the markers is used to facilitate tracking the device comprising those markers, typically automatically and typically on line. Typically, the tracking is performed in accordance with techniques described in U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556), which is incorporated herein by reference. For some applications, the endoluminal device comprising the radiopaque markers is an endoluminal data-acquisition device (e.g., an endoluminal imaging probe), and the increased visibility of the radiopaque markers in the resulting image stream is utilized for co-registering, typically automatically and typically on line, endoluminal data points (e.g., endoluminal images) with the extraluminal images (e.g., extraluminal x-ray images). The endoluminal imaging probe may be ultrasound, optical coherence, infrared, MRI, or any combination thereof. Typically, the co-registration is performed in accordance with techniques described in International Patent Application PCT/IL2011/000612 (published as WO 12/014,212), which is incorporated herein by reference.


For example, based upon locations of radiopaque portions of an endoluminal data-acquisition device in the cleaned image the output-generation functionality of the processor may determine that a given one of the endoluminal data points corresponds to a given location within the lumen, and an output may be generated in response thereto. Alternatively or additionally, based upon locations of the radiopaque portions of an endoluminal device in the cleaned image, the output-generation functionality may determine that the endoluminal device is at a given location within the lumen. In response to determining that the endoluminal device is at the given location within the lumen, the output-generation functionality may drive the display to display an endoluminal image of the lumen corresponding to the location.


It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove. Rather, the scope of the present invention includes both combinations and subcombinations of the various features described hereinabove, as well as variations and modifications thereof that are not in the prior art, which would occur to persons skilled in the art upon reading the foregoing description.

Claims
  • 1. Apparatus for use with at least one input angiogram image of a device inserted inside a portion of a body of a subject, the device including radiopaque portions thereof, the angiogram image being acquired in the presence of contrast agent within the body portion, the apparatus comprising: a display; andat least one processor configured: to generate a background image in which a relative pixel value is assigned to a first pixel with respect to a second pixel, at least partially based upon relative pixel values of a patch of pixels that surround a first pixel of the input angiogram image that is disposed at the same location as the first pixel of the background image and a patch of pixels that surround a second pixel of the input angiogram image that is disposed at the same location as the second pixel of the background image;to generate a cleaned image in which visibility of the radiopaque portions of the device is increased relative to the input angiogram image, by dividing the input angiogram image by the background image; andto drive the display to display an output based upon the cleaned image.
  • 2. The apparatus according to claim 1, wherein the processor is configured to generate the background image by: in the background image, assigning pixel values to the first pixel and the second pixel that are more similar to one another, than the similarity of pixel values that are assigned to a third pixel and a fourth pixel,based upon the patch of pixels that surround the first pixel of the input angiogram image and the patch of pixels that surround the second pixel of the input angiogram image having more similar pixel values to one another, than a similarity, to one another, of pixel values of a patch of pixels that surround a third pixel of the input angiogram image that is disposed at the same location as the third pixel of the background image and a patch of pixels that surround a fourth pixel of the input angiogram image that is disposed at the same location as the fourth pixel of the background image.
  • 3. The apparatus according to claim 1, wherein the processor is further configured to generate the background image by assigning pixel values to the first and second pixels of the background image based upon pixel values of the first pixel and the second pixel of the input angiogram image.
  • 4. The apparatus according to claim 1, wherein the processor is configured to drive the display to display the cleaned image.
  • 5. The apparatus according to claim 1, wherein: the input angiogram image of the device includes a plurality of input angiogram images of the device, andthe processor is configured to: generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,generate a stabilized image stream by image tracking the cleaned images with respect to each other, based upon locations of the radiopaque portions of the device in the cleaned images, anddrive the display to display the stabilized image stream.
  • 6. The apparatus according to claim 1, wherein: the input angiogram image of the device includes a plurality of input angiogram images of the device, andthe processor is configured to: generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,generate a stabilized image stream by image tracking the input images with respect to each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, anddrive the display to display the stabilized image stream.
  • 7. The apparatus according to claim 1, wherein: the input angiogram image of the device includes a plurality of input angiogram images of the device, andthe processor is configured to: generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,generate an enhanced image frame by: aligning the cleaned images with each other, based upon locations of the radiopaque portions of the device in the cleaned images, andgenerating an averaged image frame based upon the aligned images, anddrive the display to display the enhanced image frame.
  • 8. The apparatus according to claim 1, wherein: the input angiogram image of the device includes a plurality of input angiogram images of the device, andthe processor is configured to: generate a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,generate an enhanced image frame by: aligning the input images with each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, andgenerating an averaged image frame based upon the aligned images, anddrive the display to display the enhanced image frame.
  • 9. The apparatus according to claim 1, wherein: the portion of the subject's body includes a lumen of the subject's body,the device includes an endoluminal data-acquisition device configured to acquire endoluminal data points while the device is at respective locations within the lumen, andthe processor is configured: based upon locations of the radiopaque portions of the device in the cleaned image, to determine that a given one of the endoluminal data points corresponds to a given location within the lumen, andto drive the display to display an output, in response to the determining.
  • 10. The apparatus according to claim 1, wherein: the portion of the subject's body includes a lumen of the subject's body,the device includes an endoluminal device configured to be moved through the lumen, andthe processor is configured: based upon locations of the radiopaque portions of the device in the cleaned image, to determine that the endoluminal device is at a given location within the lumen, andin response to the determining that the endoluminal device is at the given location within the lumen, to drive the display to display an endoluminal image of the lumen corresponding to the location.
  • 11. A method for use with at least one input angiogram image of a device inserted inside a portion of a body of a subject, the device including radiopaque portions thereof, the angiogram image being acquired in the presence of contrast agent within the body portion, the method comprising: generating, with a processor, a background image in which a relative pixel value is assigned to a first pixel with respect to a second pixel, at least partially based upon relative pixel values of a patch of pixels that surround a first pixel of the input angiogram image that is disposed at the same location as the first pixel of the background image and a patch of pixels that surround a second pixel of the input angiogram image that is disposed at the same location as the second pixel of the background image;generating, with the processor, a cleaned image in which visibility of the radiopaque portions of the device is increased relative to the input angiogram image, by dividing the input angiogram image by the background image; andgenerating an output on a display, based upon the cleaned image.
  • 12. The method according to claim 11, wherein generating the background image comprises: in the background image, assigning pixel values to the first pixel and the second pixel that are more similar to one another, than the similarity of pixel values that are assigned to a third pixel and a fourth pixel,based upon the patch of pixels that surround the first pixel of the input angiogram image and the patch of pixels that surround the second pixel of the input angiogram image having more similar pixel values to one another, than a similarity, to one another, of pixel values of a patch of pixels that surround a third pixel of the input angiogram image that is disposed at the same location as the third pixel of the background image and a patch of pixels that surround a fourth pixel of the input angiogram image that is disposed at the same location as the fourth pixel of the background image.
  • 13. The method according to claim 11, wherein generating the background image further comprises assigning pixel values to the first and second pixels of the background image based upon pixel values of the first pixel and the second pixel of the input angiogram image.
  • 14. The method according to claim 11, wherein generating the output comprises displaying the cleaned image.
  • 15. The method according to claim 11, wherein the input angiogram image of the device includes a plurality of input angiogram images of the device,wherein generating the cleaned image comprises generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,the method further comprising generating a stabilized image stream by image tracking the cleaned images with respect to each other, based upon locations of the radiopaque portions of the device in the cleaned images, andwherein generating the output comprises displaying the stabilized image stream.
  • 16. The method according to claim 11, wherein the input angiogram image of the device includes a plurality of input angiogram images of the device,wherein generating the cleaned image comprises generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,the method further comprising generating a stabilized image stream by image tracking the input images with respect to each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, andwherein generating the output comprises displaying the stabilized image stream.
  • 17. The method according to claim 11, wherein the input angiogram image of the device includes a plurality of input angiogram images of the device,wherein generating the cleaned image comprises generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,the method further comprising generating an enhanced image frame by: aligning the cleaned images with each other, based upon locations of the radiopaque portions of the device in the cleaned images, andgenerating an averaged image frame based upon the aligned cleaned images, andwherein generating the output comprises displaying the enhanced image frame.
  • 18. The method according to claim 11, wherein the input angiogram image of the device includes a plurality of input angiogram images of the device,wherein generating the cleaned image comprises generating a plurality of cleaned images, the cleaned images corresponding to respective input angiogram images,the method further comprising generating an enhanced image frame by: aligning the input images with each other, based upon locations of the radiopaque portions of the device in the corresponding cleaned images, andgenerating an averaged image frame based upon the aligned input images, andwherein generating the output comprises displaying the enhanced image frame.
  • 19. The method according to claim 11, wherein the portion of the subject's body includes a lumen of the subject's body,wherein the device includes an endoluminal data-acquisition device configured to acquire endoluminal data points while the device is at respective locations within the lumen,the method further comprising, based upon locations of the radiopaque portions of the device in the cleaned image, determining that a given one of the endoluminal data points corresponds to a given location within the lumen, andwherein generating the output comprises generating the output in response to the determining.
  • 20. The method according to claim 11, wherein the portion of the subject's body includes a lumen of the subject's body,wherein the device includes an endoluminal device configured to be moved through the lumen,the method further comprising, based upon locations of the radiopaque portions of the device in the cleaned image, determining that the endoluminal device is at a given location within the lumen, andwherein generating the output comprises, in response to the determining that the endoluminal device is at the given location within the lumen, generating an endoluminal image of the lumen corresponding to the location.
CROSS REFERENCES TO RELATED APPLICATIONS

The present application is a U.S. national phase of PCT Application no. PCT/IL2012/000246 to Barzelay (published as WO 12/176191), filed Jun. 21, 2012, which claims priority from U.S. Provisional Patent Application 61/457,866 to Barzelay, filed Jun. 23, 2011. The present application is related to International Patent Application PCT/IL2011/000612 (published as WO 12/014,212), entitled “Co-use of endoluminal data and extraluminal imaging,” filed Jul. 28, 2011, which: (a) claims the benefit of: U.S. Provisional Patent Application 61/344,464, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 29 Jul. 2010; U.S. Provisional Patent Application 61/344,875, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 1 Nov. 2010; U.S. Provisional Patent Application 61/457,339, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 3 Mar. 2011; U.S. Provisional Patent Application 61/457,455, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 1 Apr. 2011; U.S. Provisional Patent Application 61/457,780, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 2 Jun. 2011; and U.S. Provisional Patent Application 61/457,951, entitled “Co-use of endoluminal data and extraluminal imaging,” filed 15 Jul. 2011; and (b) is a continuation-in-part of U.S. patent application Ser. No. 12/650,605 to Cohen (published as US 2010/0172556), filed Dec. 31, 2009, which: (i) is a continuation of U.S. patent application Ser. No. 12/666,879 to Steinberg (issued as U.S. Pat. No. 8,781,193), which is the US national phase of PCT Application No. PCT/IL2009/001089 to Cohen (published as WO 10/058,398), filed Nov. 18, 2009, which claims priority from the following patent applications: U.S. Provisional Patent Application 61/193,329, entitled “Apparatuses and methods for the automatic generation of a road map from angiographic images of a cyclically-moving organ,” to Steinberg, filed Nov. 18, 2008U.S. Provisional Patent Application 61/193,915, entitled “Image processing and tool actuation for medical procedures,” to Steinberg, filed Jan. 8, 2009U.S. Provisional Patent Application 61/202,181, entitled “Image processing and tool actuation for medical procedures,” to Steinberg, filed Feb. 4, 2009U.S. Provisional Patent Application 61/202,451, entitled “Image processing and tool actuation for medical procedures,” to Steinberg, filed Mar. 2, 2009U.S. Provisional Patent Application 61/213,216, entitled “Image processing and tool actuation for medical procedures,” to Steinberg, filed May 18, 2009U.S. Provisional Patent Application 61/213,534, entitled “Image Processing and Tool Actuation for Medical Procedures,” to Steinberg, filed Jun. 17, 2009U.S. Provisional Patent Application 61/272,210, entitled “Image processing and tool actuation for medical procedures,” to Steinberg, filed Sep. 1, 2009 andU.S. Provisional Patent Application 61/272,356, entitled “Image Processing and Tool Actuation for Medical Procedures” to Steinberg, filed Sep. 16, 2009; and (ii) is a continuation-in-part of U.S. patent application Ser. No. 12/075,244 to Tolkowsky (published as US 2008/0221442, now abandoned), filed Mar. 10, 2008, entitled “Imaging for use with moving organs,” which claims the benefit of U.S. Provisional Patent Application Nos.: 60/906,091 filed on Mar. 8, 2007,60/924,609 filed on May 22, 2007,60/929,165 filed on Jun. 15, 2007,60/935,914 filed on Sep. 6, 2007, and60/996,746 filed on Dec. 4, 2007,all entitled “Apparatuses and methods for performing medical procedures on cyclically-moving body organs.” The present application is related to the following patent applications: U.S. patent application Ser. No. 12/075,214 to Iddan (published as 2008/0221439, now abandoned), filed Mar. 10, 2008, entitled “Tools for use with moving organs.”U.S. patent application Ser. No. 12/075,252 to Iddan (published as US 2008/0221440), filed Mar. 10, 2008, entitled “Imaging and tools for use with moving organs.”U.S. patent application Ser. No. 12/781,260 to Blank (published as US 2010/0228076), filed May 17, 2010, entitled “Controlled actuation and deployment of a medical device.”U.S. patent application Ser. No. 12/487,315 to Iddan (issued as U.S. Pat. No. 8,700,130), filed Jun. 18, 2009, entitled “Stepwise advancement of a medical tool,” which claims the benefit of U.S. Provisional Patent Application No. 61/129,331 to Iddan, filed on Jun. 19, 2008, entitled “Stepwise advancement of a medical tool.” All of the above-mentioned applications are incorporated herein by reference.

PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/IL2012/000246 6/21/2012 WO 00 2/7/2014
Publishing Document Publishing Date Country Kind
WO2012/176191 12/27/2012 WO A
US Referenced Citations (466)
Number Name Date Kind
3871360 Van Horn et al. Mar 1975 A
3954098 Dick et al. May 1976 A
4016871 Schiff Apr 1977 A
4031884 Henzel Jun 1977 A
4245647 Randall Jan 1981 A
4270143 Morris May 1981 A
4316218 Gay Feb 1982 A
4382184 Wernikoff May 1983 A
4545390 Leary Oct 1985 A
4709385 Pfeiler Nov 1987 A
4712560 Schaefer et al. Dec 1987 A
4723938 Goodin et al. Feb 1988 A
4741328 Gabbay May 1988 A
4758223 Rydell Jul 1988 A
4770184 Greene, Jr. et al. Sep 1988 A
4849906 Chodos et al. Jul 1989 A
4865043 Shimoni Sep 1989 A
4878115 Elion Oct 1989 A
4920413 Nakamura Apr 1990 A
4991589 Hongo et al. Feb 1991 A
4994965 Crawford et al. Feb 1991 A
5020516 Biondi Jun 1991 A
5054045 Whiting et al. Oct 1991 A
5054492 Scribner Oct 1991 A
5056524 Oe Oct 1991 A
5062056 Lo et al. Oct 1991 A
5150292 Hoffmann et al. Sep 1992 A
5176619 Segalowitz Jan 1993 A
5177796 Feig et al. Jan 1993 A
5293574 Roehm et al. Mar 1994 A
5295486 Wollschlager et al. Mar 1994 A
5330496 Alferness Jul 1994 A
5357550 Asahina et al. Oct 1994 A
5429144 Wilk Jul 1995 A
5457728 Whiting et al. Oct 1995 A
5457754 Han et al. Oct 1995 A
5486192 Walinsky et al. Jan 1996 A
5537490 Yukawa Jul 1996 A
5538494 Matsuda Jul 1996 A
5577502 Darrow et al. Nov 1996 A
5586201 Whiting et al. Dec 1996 A
5596990 Yock Jan 1997 A
5613492 Feinberg Mar 1997 A
5619995 Lobodzinski Apr 1997 A
5630414 Horbaschek May 1997 A
5674217 Wahlstrom et al. Oct 1997 A
5724977 Yock Mar 1998 A
5764723 Weinberger Jun 1998 A
5766208 McEwan Jun 1998 A
5792157 Mische et al. Aug 1998 A
5807327 Green et al. Sep 1998 A
5809105 Roehm et al. Sep 1998 A
5822391 Whiting et al. Oct 1998 A
5830222 Makower Nov 1998 A
5879305 Yock Mar 1999 A
5885218 Teo Mar 1999 A
5885244 Leone et al. Mar 1999 A
5916194 Jacobsen et al. Jun 1999 A
5921934 Teo Jul 1999 A
5971976 Wang et al. Oct 1999 A
6088488 Hardy et al. Jul 2000 A
6095976 Nachtomy Aug 2000 A
6120455 Teo Sep 2000 A
6120523 Crocker et al. Sep 2000 A
6126608 Kemme et al. Oct 2000 A
6148095 Prause et al. Nov 2000 A
6152878 Nachtomy Nov 2000 A
6195445 Dubuisson-Jolly et al. Feb 2001 B1
6233478 Liu May 2001 B1
6246898 Vesely et al. Jun 2001 B1
6254541 Teo Jul 2001 B1
6267727 Teo Jul 2001 B1
6278767 Hsieh Aug 2001 B1
6331181 Tierney et al. Dec 2001 B1
6370417 Horbaschek et al. Apr 2002 B1
6377011 Ben-Ur Apr 2002 B1
6442415 Bis et al. Aug 2002 B1
6454715 Teo Sep 2002 B2
6454776 Tajima et al. Sep 2002 B1
6473635 Rasche Oct 2002 B1
6491636 Chenal Dec 2002 B2
6493575 Kesten et al. Dec 2002 B1
6496716 Langer et al. Dec 2002 B1
6532380 Close et al. Mar 2003 B1
6538634 Chui et al. Mar 2003 B1
6546271 Reisfeld Apr 2003 B1
6576007 Dehdashtian et al. Jun 2003 B2
6589176 Jago Jul 2003 B2
6616596 Milbocker Sep 2003 B1
6643533 Knoplioch Nov 2003 B2
6659953 Sumanaweera et al. Dec 2003 B1
6666863 Wentzel et al. Dec 2003 B2
6704593 Stainsby Mar 2004 B2
6708052 Mao et al. Mar 2004 B1
6711436 Duhaylongsod Mar 2004 B1
6718055 Suri Apr 2004 B1
6726675 Beyar Apr 2004 B1
6728566 Subramanyan Apr 2004 B1
6731973 Voith May 2004 B2
6786896 Madhani et al. Sep 2004 B1
6788827 Makram-Ebeid Sep 2004 B1
6796972 Sinofsky et al. Sep 2004 B1
6835177 Fritz et al. Dec 2004 B2
6858003 Evans et al. Feb 2005 B2
6912471 Heigl Jun 2005 B2
6937696 Mostafavi Aug 2005 B1
6959266 Mostafavi Oct 2005 B1
6973202 Mostafavi Dec 2005 B2
6980675 Evron et al. Dec 2005 B2
6996430 Gilboa et al. Feb 2006 B1
6999852 Green Feb 2006 B2
7031504 Argiro et al. Apr 2006 B1
7070555 Siess Jul 2006 B2
7085342 Younis et al. Aug 2006 B2
7134994 Alpert Nov 2006 B2
7155046 Aben et al. Dec 2006 B2
7155315 Niemeyer et al. Dec 2006 B2
7180976 Wink et al. Feb 2007 B2
7191100 Mostafavi Mar 2007 B2
7209779 Kaufman Apr 2007 B2
7215802 Klingensmith May 2007 B2
7221973 Nitz May 2007 B2
7269457 Shafer Sep 2007 B2
7289652 Florent et al. Oct 2007 B2
7321677 Evron et al. Jan 2008 B2
7339585 Verstraelen et al. Mar 2008 B2
7343032 Oakley et al. Mar 2008 B2
7343195 Strommer et al. Mar 2008 B2
7359554 Klingensmith Apr 2008 B2
7369691 Kondo et al. May 2008 B2
7397935 Kimmel Jul 2008 B2
7398116 Edwards Jul 2008 B2
7517318 Altmann Apr 2009 B2
7546154 Hornegger et al. Jun 2009 B2
7587074 Zarkh et al. Sep 2009 B2
7599730 Hunter Oct 2009 B2
7604601 Altmann Oct 2009 B2
7650179 Redel et al. Jan 2010 B2
7653426 Yatsuo et al. Jan 2010 B2
7668362 Olson et al. Feb 2010 B2
7693349 Gering Apr 2010 B2
7697974 Jenkins Apr 2010 B2
7713210 Byrd May 2010 B2
7729743 Sabczynski et al. Jun 2010 B2
7729746 Redel et al. Jun 2010 B2
7740584 Donaldson Jun 2010 B2
7742629 Zarkh et al. Jun 2010 B2
7773787 Tek et al. Aug 2010 B2
7773792 Kimmel Aug 2010 B2
7778488 Nord Aug 2010 B2
7778688 Strommer Aug 2010 B2
7822291 Guetter Oct 2010 B2
7831076 Altmann Nov 2010 B2
7844126 Mory et al. Nov 2010 B2
7848553 Hertel Dec 2010 B2
7877132 Rongen Jan 2011 B2
7889905 Higgins et al. Feb 2011 B2
7892177 Rold et al. Feb 2011 B2
7914442 Gazdzinski Mar 2011 B1
7916912 Abramov et al. Mar 2011 B2
7925064 Cloutier et al. Apr 2011 B2
7925069 Ortyn et al. Apr 2011 B2
7925327 Weese Apr 2011 B2
7927275 Kuban Apr 2011 B2
7930014 Huennekens et al. Apr 2011 B2
7935055 Burckhardt May 2011 B2
7961926 Viswanathan Jun 2011 B2
7965905 Allon et al. Jun 2011 B2
7970187 Puts Jun 2011 B2
7978916 Klingensmith Jul 2011 B2
7992100 Lundstrom Aug 2011 B2
8025622 Rold et al. Sep 2011 B2
8029447 Kanz Oct 2011 B2
8052605 Muller Nov 2011 B2
8055327 Strommer et al. Nov 2011 B2
8077939 Le Bezet et al. Dec 2011 B2
8080474 Chen Dec 2011 B2
8086000 Weijers Dec 2011 B2
8126241 Zarkh et al. Feb 2012 B2
8155411 Hof Apr 2012 B2
8157742 Taylor Apr 2012 B2
8165361 Li Apr 2012 B2
8172763 Nelson May 2012 B2
8189886 Huo et al. May 2012 B2
8199981 Koptenko et al. Jun 2012 B2
8200040 Pfister Jun 2012 B2
8208995 Tearney et al. Jun 2012 B2
8213676 Bendall Jul 2012 B2
8233718 Klingensmith Jul 2012 B2
8260395 Markowitz et al. Sep 2012 B2
8271068 Khamene Sep 2012 B2
8275201 Rangwala et al. Sep 2012 B2
8289284 Glynn Oct 2012 B2
8295577 Zarkh et al. Oct 2012 B2
8298147 Huennekens et al. Oct 2012 B2
8303503 Nair Nov 2012 B2
8364242 Li Jan 2013 B2
8396276 Gatta Mar 2013 B2
8396533 Barbu et al. Mar 2013 B2
8409098 Olson Apr 2013 B2
8411927 Chang et al. Apr 2013 B2
8428318 Zhuo Apr 2013 B2
8428691 Byrd Apr 2013 B2
8433115 Chen Apr 2013 B2
8457374 Lendl Jun 2013 B2
8478387 Xu Jul 2013 B2
8483488 Richter Jul 2013 B2
8515146 Zhu et al. Aug 2013 B2
8565859 Wang et al. Oct 2013 B2
8605976 Diamant et al. Dec 2013 B2
8625865 Zarkh et al. Jan 2014 B2
8700128 Assis et al. Apr 2014 B2
8731642 Zarkh et al. May 2014 B2
8861830 Brada et al. Oct 2014 B2
20010031919 Strommer et al. Oct 2001 A1
20010055418 Nakamura Dec 2001 A1
20020049375 Strommer et al. Apr 2002 A1
20020058869 Axelsson et al. May 2002 A1
20020090119 Saito et al. Jul 2002 A1
20020114497 Wetzel et al. Aug 2002 A1
20020188307 Pintor et al. Dec 2002 A1
20020193686 Gilboa Dec 2002 A1
20030014100 Maria Meens et al. Jan 2003 A1
20030018251 Solomon Jan 2003 A1
20030021381 Koppe et al. Jan 2003 A1
20030023141 Stelzer et al. Jan 2003 A1
20030069499 Lienard et al. Apr 2003 A1
20030088179 Seeley et al. May 2003 A1
20030095710 Tessadro May 2003 A1
20030139772 Fisher et al. Jul 2003 A1
20030157073 Peritt Aug 2003 A1
20040034380 Woolfson et al. Feb 2004 A1
20040077941 Reddy et al. Apr 2004 A1
20040097805 Verard et al. May 2004 A1
20040133129 Harari et al. Jul 2004 A1
20040165756 Mielekamp Aug 2004 A1
20040176681 Mao et al. Sep 2004 A1
20040215235 Jackson et al. Oct 2004 A1
20040249270 Kondo et al. Dec 2004 A1
20040254570 Hadjicostis et al. Dec 2004 A1
20040267113 Thomson Dec 2004 A1
20050004503 Samson et al. Jan 2005 A1
20050008210 Evron et al. Jan 2005 A1
20050015009 Mourad et al. Jan 2005 A1
20050031176 Hertel Feb 2005 A1
20050033199 van der Steen Feb 2005 A1
20050043614 Huizenga et al. Feb 2005 A1
20050054916 Mostafavi Mar 2005 A1
20050067568 Harding et al. Mar 2005 A1
20050080336 Byrd et al. Apr 2005 A1
20050090737 Burrel et al. Apr 2005 A1
20050096589 Shachar May 2005 A1
20050107688 Strommer May 2005 A1
20050107808 Evans et al. May 2005 A1
20050111719 Pesatore et al. May 2005 A1
20050113685 Maschke et al. May 2005 A1
20050137661 Sra Jun 2005 A1
20050141766 Nagahashi et al. Jun 2005 A1
20050143777 Sra Jun 2005 A1
20050154281 Xue et al. Jul 2005 A1
20050182295 Soper et al. Aug 2005 A1
20050197557 Strommer et al. Sep 2005 A1
20050197559 Boese et al. Sep 2005 A1
20050197566 Strommer et al. Sep 2005 A1
20050201510 Mostafavi Sep 2005 A1
20050228359 Doyle Oct 2005 A1
20050234331 Sendai Oct 2005 A1
20050273050 Yokoyama et al. Dec 2005 A1
20050288577 Weese Dec 2005 A1
20060007188 Reiner Jan 2006 A1
20060058647 Strommer et al. Mar 2006 A1
20060074285 Zarkh et al. Apr 2006 A1
20060106318 Davidson May 2006 A1
20060120581 Eck et al. Jun 2006 A1
20060129142 Reynolds Jun 2006 A1
20060147897 Grinvald Jul 2006 A1
20060149134 Soper et al. Jul 2006 A1
20060155327 Briganti et al. Jul 2006 A1
20060159318 Alyassin et al. Jul 2006 A1
20060165270 Borgert et al. Jul 2006 A1
20060173287 Sabczynski et al. Aug 2006 A1
20060184016 Glossop Aug 2006 A1
20060188135 Zarkh et al. Aug 2006 A1
20060193505 Glukhovsky et al. Aug 2006 A1
20060224188 Libbus et al. Oct 2006 A1
20060224232 Chobotov Oct 2006 A1
20060241369 Lienard et al. Oct 2006 A1
20060241445 Altmann Oct 2006 A1
20060241465 Huennekens et al. Oct 2006 A1
20060241478 Lewis Oct 2006 A1
20060253024 Altmann Nov 2006 A1
20060253029 Altmann Nov 2006 A1
20060253031 Altmann Nov 2006 A1
20060257006 Bredno et al. Nov 2006 A1
20060259137 Artof et al. Nov 2006 A1
20060269108 Viswanathan Nov 2006 A1
20060287595 Maschke Dec 2006 A1
20070021816 Rudin Jan 2007 A1
20070038061 Huennekens et al. Feb 2007 A1
20070038081 Eck et al. Feb 2007 A1
20070043292 Camus Feb 2007 A1
20070053558 Puts et al. Mar 2007 A1
20070055128 Glossop Mar 2007 A1
20070060798 Krupnik et al. Mar 2007 A1
20070088431 Bourang et al. Apr 2007 A1
20070106146 Altmann et al. May 2007 A1
20070116342 Zarkh et al. May 2007 A1
20070123771 Redel et al. May 2007 A1
20070142907 Moaddeb et al. Jun 2007 A1
20070147706 Sasaki et al. Jun 2007 A1
20070165916 Cloutier et al. Jul 2007 A1
20070173861 Strommer Jul 2007 A1
20070208388 Jahns Sep 2007 A1
20070219630 Chu Sep 2007 A1
20070232896 Gilboa et al. Oct 2007 A1
20070248253 Manzke et al. Oct 2007 A1
20070255139 Deschinger Nov 2007 A1
20070269135 Ono Nov 2007 A1
20070276216 Beyar et al. Nov 2007 A1
20080008366 Desh Jan 2008 A1
20080015677 Glossop et al. Jan 2008 A1
20080021331 Grinvald Jan 2008 A1
20080082049 Evans et al. Apr 2008 A1
20080089566 Node-Langlois Apr 2008 A1
20080114238 Lloyd May 2008 A1
20080118117 Gauldie et al. May 2008 A1
20080119922 Alkhatib May 2008 A1
20080137923 Spahn Jun 2008 A1
20080137935 Spahn Jun 2008 A1
20080146923 Mejia Jun 2008 A1
20080146942 Dala-Krishna Jun 2008 A1
20080177183 Courtney Jul 2008 A1
20080188739 Rongen et al. Aug 2008 A1
20080221439 Iddan et al. Sep 2008 A1
20080221440 Iddan et al. Sep 2008 A1
20080221442 Tolkowsky et al. Sep 2008 A1
20080247621 Zarkh et al. Oct 2008 A1
20080253686 Bayer Oct 2008 A1
20080262346 Assis et al. Oct 2008 A1
20080267475 Lendl Oct 2008 A1
20080283771 Li Nov 2008 A1
20080300487 Govari Dec 2008 A1
20090016587 Strobel et al. Jan 2009 A1
20090074284 Zeineh et al. Mar 2009 A1
20090093676 Davidson Apr 2009 A1
20090103682 Chen et al. Apr 2009 A1
20090105579 Garibaldi Apr 2009 A1
20090116715 Bredno et al. May 2009 A1
20090171201 Olson Jul 2009 A1
20090177444 Wiemker et al. Jul 2009 A1
20090216112 Assis et al. Aug 2009 A1
20090245601 Cohen et al. Oct 2009 A1
20090257631 Baumgart Oct 2009 A1
20090264752 Markowitz et al. Oct 2009 A1
20090264753 Von Schulthess Oct 2009 A1
20090275831 Hall Nov 2009 A1
20090281418 Ruijters et al. Nov 2009 A1
20090299195 Muller et al. Dec 2009 A1
20090304593 Frinking et al. Dec 2009 A1
20090306547 Iddan et al. Dec 2009 A1
20100041949 Tolkowsky Feb 2010 A1
20100049034 Eck et al. Feb 2010 A1
20100054573 Shekhara Mar 2010 A1
20100067768 Ionasec et al. Mar 2010 A1
20100094124 Schoonenberg et al. Apr 2010 A1
20100094127 Xu Apr 2010 A1
20100099979 Schoonenberg et al. Apr 2010 A1
20100111396 Boucheron May 2010 A1
20100114289 Camus May 2010 A1
20100123715 Hansegard May 2010 A1
20100134517 Saikaly et al. Jun 2010 A1
20100135546 Cziria Jun 2010 A1
20100157041 Klaiman et al. Jun 2010 A1
20100160764 Steinberg et al. Jun 2010 A1
20100160773 Cohen et al. Jun 2010 A1
20100161022 Tolkowsky Jun 2010 A1
20100161023 Cohen et al. Jun 2010 A1
20100171819 Tolkowsky et al. Jul 2010 A1
20100172556 Cohen et al. Jul 2010 A1
20100174192 Azuma Jul 2010 A1
20100191102 Steinberg et al. Jul 2010 A1
20100198063 Huber Aug 2010 A1
20100220917 Steinberg et al. Sep 2010 A1
20100222671 Cohen et al. Sep 2010 A1
20100228076 Blank et al. Sep 2010 A1
20100246910 Wiemker Sep 2010 A1
20100272340 Bar-Aviv et al. Oct 2010 A1
20100290693 Cohen et al. Nov 2010 A1
20100310140 Schneider Dec 2010 A1
20100312100 Zarkh et al. Dec 2010 A1
20100318115 Chanduszko et al. Dec 2010 A1
20100331670 Strommer et al. Dec 2010 A1
20110015520 Meetz et al. Jan 2011 A1
20110026786 Mohamed Feb 2011 A1
20110033094 Zarkh Feb 2011 A1
20110034801 Baumgart Feb 2011 A1
20110052030 Bruder et al. Mar 2011 A1
20110071404 Schmidtt et al. Mar 2011 A1
20110075912 Rieber et al. Mar 2011 A1
20110087104 Moore Apr 2011 A1
20110112398 Zarkh et al. May 2011 A1
20110118825 Hunter et al. May 2011 A1
20110157154 Bernard et al. Jun 2011 A1
20110228992 Wels et al. Sep 2011 A1
20110230758 Eichler Sep 2011 A1
20110235889 Spahn Sep 2011 A1
20110286627 Takacs et al. Nov 2011 A1
20110293163 Kargar et al. Dec 2011 A1
20110319752 Steinberg et al. Dec 2011 A1
20120004529 Tolkowsky et al. Jan 2012 A1
20120004533 Peng Jan 2012 A1
20120004537 Tolkowsky et al. Jan 2012 A1
20120014574 Ferschel et al. Jan 2012 A1
20120029339 Cohen et al. Feb 2012 A1
20120051606 Saikia Mar 2012 A1
20120059220 Holsing Mar 2012 A1
20120059253 Wang et al. Mar 2012 A1
20120065507 Brunke Mar 2012 A1
20120069167 Liu et al. Mar 2012 A1
20120072190 Sharma et al. Mar 2012 A1
20120083696 Kitamura Apr 2012 A1
20120093379 Florent et al. Apr 2012 A1
20120123238 Vaillant et al. May 2012 A1
20120130242 Burgess May 2012 A1
20120140998 Zhu Jun 2012 A1
20120207367 Kneepkens Aug 2012 A1
20120215093 Ji Aug 2012 A1
20120224751 Kemp Sep 2012 A1
20120230565 Steinberg et al. Sep 2012 A1
20120245460 Slomka Sep 2012 A1
20120250974 Miyamoto Oct 2012 A1
20120294498 Popovic Nov 2012 A1
20120300981 Yeh et al. Nov 2012 A1
20120310081 Adler et al. Dec 2012 A1
20130004044 Ross Jan 2013 A1
20130030295 Huennekens Jan 2013 A1
20130046167 Shah Feb 2013 A1
20130053664 Jian et al. Feb 2013 A1
20130109958 Baumgart May 2013 A1
20130109959 Baumgart May 2013 A1
20130116739 Brada et al. May 2013 A1
20130120296 Merritt May 2013 A1
20130120297 Merrit May 2013 A1
20130123616 Merrit May 2013 A1
20130308844 Florent et al. Nov 2013 A1
20130329030 Tolkowsky et al. Dec 2013 A1
20130329977 Tolkowsky et al. Dec 2013 A1
20140094660 Tolkowsky et al. Apr 2014 A1
20140094689 Cohen et al. Apr 2014 A1
20140094690 Tolkowsky et al. Apr 2014 A1
20140094691 Steinberg et al. Apr 2014 A1
20140094692 Tolkowsky et al. Apr 2014 A1
20140094693 Cohen et al. Apr 2014 A1
20140100451 Tolkowsky et al. Apr 2014 A1
20140107479 Klaiman et al. Apr 2014 A1
20140111541 Tolkowsky et al. Apr 2014 A1
20140112566 Steinberg et al. Apr 2014 A1
20140114184 Klaiman et al. Apr 2014 A1
20140114185 Tolkowsky et al. Apr 2014 A1
20140114308 Tolkowsky et al. Apr 2014 A1
20140114333 Tolkowsky et al. Apr 2014 A1
20140121513 Tolkowsky et al. May 2014 A1
20150282737 Tolkowsky et al. Oct 2015 A1
20150282889 Cohen et al. Oct 2015 A1
20150282890 Cohen et al. Oct 2015 A1
20150283319 Tolkowsky et al. Oct 2015 A1
Foreign Referenced Citations (44)
Number Date Country
2 570 079 Mar 2013 EP
9410904 May 1994 WO
9907354 Feb 1999 WO
0033755 Jun 2000 WO
0110313 Feb 2001 WO
0143642 Jun 2001 WO
03043516 May 2003 WO
03096894 Nov 2003 WO
2005026891 Mar 2005 WO
2005051452 Jun 2005 WO
2005124689 Dec 2005 WO
2006027781 Mar 2006 WO
2006066122 Jun 2006 WO
2006066124 Jun 2006 WO
2006076409 Jul 2006 WO
2006114721 Nov 2006 WO
2006121984 Nov 2006 WO
2007002685 Jan 2007 WO
2007014028 Feb 2007 WO
2007015199 Feb 2007 WO
2008007350 Jan 2008 WO
2008062358 May 2008 WO
2008107905 Sep 2008 WO
2009153794 Dec 2009 WO
2010058398 May 2010 WO
2011046903 Apr 2011 WO
2011046904 Apr 2011 WO
2011128797 Oct 2011 WO
2011145094 Nov 2011 WO
2012014212 Feb 2012 WO
2012028190 Mar 2012 WO
2012095755 Jul 2012 WO
2012107857 Aug 2012 WO
2012138872 Oct 2012 WO
2012138874 Oct 2012 WO
2012176191 Dec 2012 WO
2013025602 Feb 2013 WO
2013061225 May 2013 WO
2013084345 Jun 2013 WO
2013128233 Sep 2013 WO
2013169814 Nov 2013 WO
2013175472 Nov 2013 WO
2014002095 Jan 2014 WO
2015155770 Oct 2015 WO
Non-Patent Literature Citations (195)
Entry
Communication dated Sep. 5, 2014 from the USPTO in U.S. Appl. No. 14/143,289.
Communication dated Oct. 24, 2014 from the USPTO in U.S. Appl. No. 12/650,121.
Communication dated Aug. 29, 2014 from the USPTO in U.S. Appl. No. 14/098,140.
Communication dated Nov. 7, 2014 from the USPTO in U.S. Appl. No. 14/096,968.
Communication dated Sep. 5, 2014 from the USPTO in U.S. Appl. No. 14/143,430.
Communication dated Sep. 11, 2014 from the USPTO in U.S. Appl. No. 12/650,152.
Communication dated Oct. 15, 2014 from the USPTO in U.S. Appl. No. 12/781,366.
Communication dated Oct. 8, 2014 from the USPTO in U.S. Appl. No. 14/098,093.
Communication dated Oct. 14, 2014 from the USPTO in U.S. Appl. No. 12/075,252.
Boyle et al., entitled “Assessment of a Novel Angiographic Image Stabilization System for Percutaneous Coronary Intervention” (Journal of Interventional Cardiology, vol. 20 No. 2, 2007.
Timinger et al., entitled “Motion compensated coronary interventional navigation by means of diaphragm tracking and elastic motion models” (Phys Med Biol. Feb. 7, 2005;50(3):491-503.
Timinger et al., entitled “Motion compensation for interventional navigation on 3D static roadmaps based on an affine model and gating” (Phys Med Biol. Mar. 7, 2004;49(5):719-32.
Turski et al., entitled “Digital Subtraction Angiography ‘Road Map’” (American Journal of Roentgenology, 1982.
Iddan et al., entitled “3D imaging in the studio and elsewhere” (SPIE Proceedings vol. 4298, 2001.
“Catheter Insertion Simulation with Combined Visual and Haptic Feedback,” by Zorcolo et al. (Center for Advanced Studies, Research and Development in Sardinia).
“4D-CT imaging of a volume influenced by respiratory motion on multi-slice CT Tinsu Pan,” by Lee et al., (Medical Physics, Feb. 2004, vol. 31, Issue 2, pp. 333-340)—an abstract.
“New 4-D imaging for real-time intraoperative MRI: adaptive 4-D scan,” by Tokuda et al. (Med Image Comput Assist Intent Int Conf. 2006;9(Pt 1):454-61) an abstract.
“Real-time interactive viewing of 4D kinematic MR joint studies,” by Schulz et al. (Med Image Comput Assist Intent Int Conf. 2005;8(Pt 1):467-73.)—an abstract.
“4D smoothing of gated SPECT images using a left-ventricle shape model and a deformable mesh,” by Brankov et al., (Nuclear Science Symposium Conference Record, 2004 IEEE, Oct. 2004, vol. 5, 2845-2848).
“Prospective motion correction of X-ray images for coronary interventions,” by Shechter et al. (IEEE Trans Med Imaging. Apr. 2005;24(4):441-50).
“Cardiac Imaging: We Got the Beat!” by Elizabeth Morgan (Medical Imaging, Mar. 2005).
“Noninvasive Coronary Angiography by Retrospectively ECG-Gated Multislice Spiral CT,” by Achenbach et al., (Circulation. Dec. 5, 2000;102(23):2823-8).
“Spatially-adaptive temporal smoothing for reconstruction of dynamic and gated image sequences,” by Brankov et al., (Nuclear Science Symposium Conference Record, 2000 IEEE, 2000, vol. 2, 15/146-15/150)—an abstract.
“Full-scale clinical implementation of a video based respiratory gating system,” by Ramsey et al., (Engineering in Medicine and Biology Society, 2000. Proceedings of the 22nd Annual International Conference of the IEEE, 2000, vol. 3, 2141-2144)—an abstract.
“Three-Dimensional Respiratory-Gated MR Angiography of the Coronary Arteries: Comparison with Conventional Coronary Angiography,” by Post et al., (AJR, 1996; 166: 1399-1404).
Soffie Mansson, et al., “Dosimetric verification of breathing adapted radiotherapy using polymer gel”, Journal of Physics: Conference series 56 (200) 300-303.
“From 2D to 4D” AXIOM Innovations— Mar. 2008, www.siemens.com/healthcare-magazine.
A Brochure: Impella® 2.5, Percutaneous Circulatory Support System, ABIOMED™, 2007.
Frangi et al., entitled “Multiscale vessel enhancement filtering” (Medical Image Computing and Computer Assisted Intervention—MICCAI 1998—Lecture Notes in Computer Science, vol. 1496, Springer Verlag, Berlin, Germany, pp. 130-137).
Dijkstra, entitled “A Note on Two Problems in Connexion with Graphs” (Numerische Mathematik I, 269-271, 1959).
Zarkh et al., entitled “Guide wire navigation and therapeutic device localization for catheterization procedure” (International Congress Series 1281 (2005) 311-316.
Brochure: At the Transvascular Cardiovascular Therapeutics (TCT) conference held in Washington DC, USA in Oct. 2008, Paieon Inc. demonstrated the CardiOp-THV system for real-time navigation and positioning of a trans-catheter heart valve.
Brochure: At the TCT conference held in San Francisco, USA in Sep. 2009, Paieon Inc. demonstrated the IC-PRO Comprehensive Imaging Workstation for providing assistance in cardiac catheterization procedures.
An International Search Report dated Sep. 8, 2009, which issued during the prosecution of Applicant's PCT/IL09/00610.
An International Search Report dated Jan. 15, 2009, issued during the prosecution of Applicant's PCT Patent Application No. PCT/IL08/000316.
An International Search Report dated May 19 2010 issued during the prosecution of Applicant's PCT Patent Application No. PCT/IL2009/001089.
“A new point matching algorithm for non-rigid registration,” by Chui (Computer Vision and Image Understanding 89 (2003) 114-141).
“Advanced and Challenges in Super-Resolution,” by Farsiu (International Journal of Imaging Systems and Technology, vol. 14, No. 2, pp. 47-57, Special issue on high-resolution image reconstruction, Aug. 2004).
“Image Registration by Minimization of Residual Complexity,” by Myronenko (CVPR 2009).
“Image inpainting,” by Bertalmio (ACM Siggraph 2000, New Orleans, Louisiana, USA, Jul. 2000).
“Nonrigid registration using free-form deformations: application to breast MR images,” by Rueckert, (IEEE Trans. Med. Img, vol. 18, No. 8, 1999).
“Unwarping of unidirectionally distorted EPI images,” by Kybic (IEEE Trans. Med. Img., vol. 19, No. 2, 2000).
“Geometrically Correct 3-D Reconstruction of Intravascular Ultrasound Images by Fusion with Biplane Angiography—Methods and Validation,” Andreas Wahle, IEEE Transactions on Medical Imaging, Final Manuscript #187/98, June 30, 1999.
An International Search Report dated Jan. 6, 2012, which issued during the prosecution of Applicant's PCT Application No. PCT/IL11/00391.
An Official Action dated Nov. 28, 2011, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,252.
An Official Action dated Dec. 8, 2011, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,244.
U.S. Appl. No. 60/845,347 to Strommer et al., filed Sep. 2006.
International Search Report dated Mar. 2, 2012, issued in PCT/IL11/00612.
Office Action dated Mar. 14, 2012, issued in U.S. Appl. No. 12/075,214.
Office Action dated Mar. 15, 2012, issued in U.S. Appl. No. 12/649,944.
Office Action dated Mar. 15, 2012, issued in U.S. Appl. No. 12/650,152.
Office Action dated May 22, 2012, issued in U.S. Appl. No. 12/075,244.
Umeda, H. et al., “Promising efficacy of primary gradual and prolonged balloon angioplasty in small coronary arteries: A randomized comparison with cutting balloon angioplasty and conventional balloon angioplasty”, 2004.
W. Santamore et al., “A microcomputer based automated quantative coronary angiographic analysis system,” Annals of Biomedical Engineering, vol. 16, pp. 367-377, 1988.
V. Duddalwar, “Multislice CT angiography: a practical guide to CT angiography in vascular imaging and intervention,” the British Journal of Radiology, 77 (2004), S27-S38.
Official Action dated Oct. 23, 2012, which issued during the prosecution of JP Application No. 2009-552328.
Official Action dated Nov. 23, 2012, which issued during the prosecution of U.S. Appl. No. 12/649,944.
Official Action dated Aug. 27, 2012, which issued during the prosecution U.S. Appl. No. 12/075,214.
International Search Report dated Oct. 10 2012 , which issued during the prosecution of PCT/IL2012/000246.
Communication dated Sep. 5, 2012 , which issued during the prosecution of EP Application 09 766 329.8-1526.
Communication dated Oct. 29, 2012 , which issued during the prosecution of EP Application 08 719941.0-1265/2129284.
Computer translation of JP 2010-253017 to Takeshi.
G. Mancini et al., “Automated quantitative coronary arteriography: morphologic and physiologic validation in vivo of a rapid digital angiographic method,” Circulation 1987;75:452-460.
I. Kompatsiaris et al., “Deformable Boundary Detection of Stents in Angiographic Images,” IEEE Transactions on Medical Imaging, vol. 19, No. 6, June 2000.
L. Yaneza et al., “ Atherosclerotic Plaque Can Be Quantified Using Multifractal and Wavelet Decomposition Techniques,” Abstracts—Angiography & Interventional Cardiology, JACC Mar. 3, 2004.
Official Action dated Oct. 31, 2012, which issued during the prosecution U.S. Appl. No. 12/075,244.
Official Action dated Sep. 20, 2012, which issued during the prosecution of U.S. Appl. No. 12/649,955.
U.S. Appl. No. 61/359,431.
W. Goodman et al., “Coronary-Artery Calcification in Young Adults With End-Stage Renal Disease Who Are Undergoing Dialysis,” The New England Journal of Medicine, vol. 342 No. 20.
W. Leung et al., “Coronary Artery Quantitation and Data Management System for Paired Cineangiograms,” Catheterization and Cardiovascular Diagnosis 24:121-134 (1991).
A search report dated Nov. 23, 2012, which issued during the prosecution of Applicant's EP Application 09 827264.4-1265/2358269.
An examination report dated Dec. 5, 2012, which issued during the prosecution of Applicant's EP Application 09766329.8.
An Official Action dated Dec. 10, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,156.
An Official Action dated Dec. 11, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,152.
An Official Action dated Jan. 22, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,121.
An Official Action dated Jan. 28, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,366.
An Official Action dated Feb. 4, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,185.
Peng Wang et al.: “Image-Based Device Tracking for the Co-registration of Angiography and Intravascular Ultrasound Images”, MICCAI 2011, Part I, LINCS 6891, pp. 161-168, 2011.
An Official Action dated Jul. 2, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,244.
An Official Action dated Jun. 19, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,214.
An Official Action dated May 31, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,252.
An Official Action dated May 6, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/487,315.
A Notice of Allowance dated Jun. 4, 2013, which issued in Applicant's U.S. Appl. No. 12/649,960.
An Official Action dated Sep. 6, 2013 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,605.
An Official Action dated Aug. 30, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,121.
An Official Action dated Sep. 12, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,260.
A Notice of Allowance in Applicant's U.S. Appl. No. 12/781,414.
An Official Action dated Aug. 3, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,294.
An Official Action dated Jun. 19, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,252.
An Official Action dated Jun. 18, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,366.
An Official Action dated Jun. 7, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,156.
An Official Action dated May 29, 2012, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,121.
Buddy D. Ratner, “Biomaterials Science: An Introduction to Materials in Medicine”, Elsevier, chapter 7, 1996.
Gerald E. Miller, “Fundamentals of Biomedical Transport Processes, Morgan & Claypool Publishers”, chapter 2, 2010.
Gerhard Albert ten Brinke, “Automated coronary flow reserve assessment using planar x-ray angiography”, PhD dissertation, Universiteit Twente, 2011.
Jerry T. Wong et al., “Quantification of fractional flow reserve based on angiographic image data”, Int J Cardiovasc Imaging 28:13-22, Jan. 7, 2011.
Kassab, G. S. et al., “Cross-sectional area and vol. compliance of porcine left coronary arteries,” Am. J. Physiol. Heart Circ. Physiol. 281, H623-H628, Aug. 2011.
Molloi S. et al., “Absolute volumetric coronary blood flow measurement with digital subtraction angiography”. Int J Card Imaging 14:137-145, 1998.
Molloi, S. et al., “Estimation of coronary artery hyperemic blood flow based on arterial lumen volume using angiographic images,” The International Journal of Cardiovascular Imaging, vol. 28, No. 1, 1-11, Jan. 7, 2011.
Molloi, S. et al., “Quantification of coronary artery lumen volume by digital angiography: in vivo validation,” Circulation 104, 2351-2357, Nov. 6, 2001.
Molloi, S. et al., “Quantification of volumetric coronary blood flow with dual-energy digital subtraction angiography,” Circulation 93, 1919-1927, May 15, 1996.
Molloi, S. et al., “Regional volumetric coronary blood flow measurement by digital angiography: in vivo validation,” Acad. Radiol. 11, 7, 757-766, Jul. 2004.
Sian Sen et al., “Development and Validation of a New, Adenosine-Independent Index of Stenosis Severity From Coronary Wave-Intensity Analysis”. Journal of the American College of Cardiology, vol. 59, Apr. 10, 2012.
Yunlong Huo et al., “A validated predictive model of coronary fractional flow reserve,” J. R. Soc. Interface, Nov. 23, 2011.
A Notice of Allowance issued in Applicant's U.S. Appl. No. 13/965,893.
An Official Action dated Nov. 13, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/666,879.
An Official Action dated Oct. 2, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,211.
An Official Action dated Oct. 21, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,214.
An Official Action dated Oct. 23, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/648,913.
An Official Action dated Oct. 25, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,366.
An Official Action dated Oct. 3, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,185.
An Official Action dated Oct. 30, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,335.
An Official Action dated Oct. 4, 2013, which issued during the prosecution of Applicant's U.S. Appl. No. 12/649,955.
Correspondence from the International Searching Authority in Applicant's PCT/IL13/50549.
Correspondence from the International Searching Authority in Applicant's PCT/IL2013/050438.
International Search Report and Written Opinion for International Patent Application PCT/IL2013/050438 mailed Dec. 2, 2013.
Office Action, dated Jan. 7, 2014, issued by the United States Patent and Trademark Office, in counterpart U.S. Appl. No. 12/075,244.
Office Action, dated Feb. 12, 2014, issued by the United States Patent and Trademark Office, in counterpart U.S. Appl. No. 12/781,260.
Office Action, dated Dec. 31, 2013, issued by the United States Patent and Trademark Office, in counterpart U.S. Appl. No. 12/075,252.
Notice of Allowance, dated Jan. 3, 2014, issued by the United States Patent and Trademark Office, in counterpart U.S. Appl. No. 13/965,872.
Search Report, dated Jan. 22, 2014, issued by the International Searching Authority, in counterpart Application No. PCT/IL13/50549.
Written Opinion, dated Jan. 22, 2014, issued by the International Searching Authority, in counterpart Application No. PCT/IL13/50549.
An Official Action dated Feb. 20, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,214.
An Official Action dated May 6, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,252.
A Notice of Allowance issued in Applicant's U.S. Appl. No. 12/666,879.
An Official Action dated Mar. 21, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/648,913.
An Official Action dated Apr. 3, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/649,955.
An Official Action dated Mar. 14, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,605.
An Official Action dated Apr. 25, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,121.
An Official Action dated Apr. 24, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,366.
An Official Action dated Apr. 17, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,211.
An Official Action dated Apr. 28, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,185.
An Official Action dated May 5, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 14/096,968.
An Official Action dated Feb. 14, 2014 , which issued during the prosecution of Applicant's U.S. Appl. No. 14/098,140.
Pyxaras et al., “Quantitative angiography optical coherence tomography for the functional assessment of nonobstructive coronary stenoses” (Medscape), Oct. 2013, 11 pages total.
Tu et al., “In vivo comparison of arterial lumen dimensions assessed by co-registered 3D quantitative coronary angiography intravascular ultrasound and optical coherence tomography.”, Int J Cardiovasc Imaging (2012) 28:1315-1327, Jan. 20, 2012, DOI 10.1007/s10554-012-0016-6, 13 pages total.
Tu et al, “Fusion of 3D QCA and IVUS/OCT”, Int J Cardiovasc Imaging (2011) 27:197-207, Jan. 25, 2011, DOI 10.1007/s10554-011-9809-2, 11 pages total.
Communication dated Mar. 25, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/075,252.
Communication dated Apr. 13, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/649,944.
Communication dated Apr. 22, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 14/142,082.
Communication dated Apr. 10, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/648,913.
Communication dated Apr. 10, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/650,152.
Communication dated Mar. 16, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 13/228,211.
Communication dated Mar. 23, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 14/097,922.
Communication dated Mar. 16, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 13/228,185.
Communication dated Feb. 23, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 14/143,184.
Communication dated May 6, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/781,260.
Communication dated May 11, 2015 from the United States Patent and Trademark Office in counterpart U.S. Appl. No. 12/650,605.
Communication dated May 12, 2015 from the Japanese Patent Office in counterpart application No. 521284/2013.
A Notice of Allowance dated Jun. 24, 2014, issued in Applicant's U.S. Appl. No. 14/097,603.
An Official Action dated Jul. 3, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/648,913.
An Official Action dated Jul. 30, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,214.
An Official Action dated Jul. 31, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/649,944.
An Official Action dated Jun. 18, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,244.
An Official Action dated May 21, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,156.
An Official Action dated May 29, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 14/097,922.
An Official Action dated May 30, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,335.
An Official Action dated Jun. 3, 2014, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,260.
Communication dated Dec. 11, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/648,913.
Communication dated Feb. 4, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/649,955.
Communication dated Nov. 24, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/781,260.
Communication dated Jan. 6, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/650,605.
Communication dated Feb. 6, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/650,121.
Communication dated Nov. 24, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/649,944.
Communication dated Feb. 6, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/781,366.
Communication dated Jan. 16, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 13/228,229.
Communication dated Jan. 6, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 13/228,335.
Communication dated Nov. 28, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 14/097,922.
Communication dated Dec. 4, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 14/098,114.
Communication dated Nov. 24, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 12/650,156.
Communication dated Dec. 19, 2014, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 14/143,430.
Communication dated Jan. 12, 2015, issued by the United States Patent and Trademark Office in corresponding U.S. Appl. No. 14/143,289.
Communication dated Jan. 23, 2015, issued by the European Patent Office in counterpart Application No. 12802046.8.
An Official Action dated Aug. 17, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 12/648,913.
An Official Action dated Aug. 27, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 12/650,121.
An Official Action dated Oct. 22, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 12/649,944.
An Official Action dated Sep. 11, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 12/781,260.
An Official Action dated Sep. 21, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,229.
An Official Action dated Sep. 3, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,211.
An Official Action dated Oct. 7, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,185.
An Official Action dated Aug. 11, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,335.
An Official Action dated Aug. 12, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 14/097,922.
An Official Action dated Oct. 7, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 14/142,082.
An Official Action dated Aug. 25, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 14/143,184.
An Official Action dated Sep. 23, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 14/742,996.
A European Patent report dated Sep. 8, 2015, which issued during the prosecution of Applicant's European Patent Application No. 08719941.0.
An Official Action dated Sep. 4, 2015, which issued during the prosecution of Applicant's Canadian Application No. 2,874,415.
An International Search Report and Written Opinion dated Aug. 25, 2015, which issued during prosecution of Applicant's PCT/IL2015/050372.
An Official Action dated Jul. 28, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 12/075,252.
An Official Action dated Jul. 6, 2015 , which issued during the prosecution of Applicant's U.S. Appl. No. 12/649,955.
An Official Action dated May 19, 2015 which issued during the prosecution of Applicant's U.S. Appl. No. 13/228,229.
An Official Action dated May 21, 2015, which issued during the prosecution of Applicant's U.S. Appl. No. 14/098,140.
An Official Action dated May 21, 2015, which issued during the prosecution of Applicant's Canadian Application No. 2,874,415.
An Official Action dated Jul. 2, 2015, which issued during the prosecution of Applicant's Canadian Application No. 2,875,346.
An English-translation of an Official Action dated Jun. 23, 2015, which issued during the prosecution of Applicant's JP Application No. 2014-164417.
Related Publications (1)
Number Date Country
20140140597 A1 May 2014 US
Provisional Applications (1)
Number Date Country
61457866 Jun 2011 US