This disclosure is directed to methods for compensating motion in digitized fluoroscopy images used to guide catheter ablation procedures.
Atrial fibrillation (AFib) is the most common type of cardiac arrhythmia, and leads to an increased stroke risk for patients. Current treatments approaches include using radio-frequency ablations, in particular when drug therapy fails, and catheter ablation procedures that can be performed in electrophysiology (EP) labs equipped with modern C-arm X-ray systems. These devices often provide 3-D tomographic imaging to facilitate inter-procedural 3-D soft-tissue imaging. Electro-anatomic mapping systems are also available to visualize the catheter position in 3-D within a registered 3-D data set. While these systems can lower an X-ray dose, they add effort and cost to the procedure. In addition, mapping systems are virtual reality systems, and do not allow for instant confirmation of catheter positions under real-time X-ray. In some instances, they may even be off with respect to the underlying anatomy.
Augmented fluoroscopy, overlaying 2-D renderings obtained from either CT, MR, or C-arm CT 3-D data sets onto live fluoroscopic images, can facilitate more precise realtime catheter navigation and also reduce the X-ray dose. However, catheter navigation under augmented fluoroscopy is compromised by cardiac and respiratory motion. A first approach to handle this effect involved tracking commonly used circumferential mapping (CFM) catheters. As atrial fibrillation therapy takes place in the vicinity of the circumferential mapping catheter, tracking this catheter can be assumed to reliably capture the motion of the relevant treatment region if the device has been firmly positioned. Otherwise complete isolation of the pulmonary veins (PVs) may fail due to undetected residual PV-atrial electrical connections. Another previously proposed method involved a 3-D model of the catheter and applied an unconstrained 2-D/3-D registration approach to align the catheter model to biplane fluoroscopy images. An initial registration is performed manually to align the 3-D data to 2-D fluoroscopy with contrast injection showing the target organ. Once the 3-D overlay moves in sync with live fluoroscopic images, catheters can be guided to anatomical structures otherwise not visible under fluoroscopy with more confidence. A yet different approach for monoplane fluoroscopic imaging tracks a catheter only in 2-D and moves the overlay image on the live fluoroscopic images to be in sync with the cardiac and respiratory motion, observed by localizing the 2-D mapping catheter.
In EP labs equipped with biplane C-arm systems, often only one image plane is used at a time to reduce the radiation exposure to the patient. In this case, the above methods are not ideal, as they require re-initialization of the catheter model if the C-arm projection geometry changes during the intervention. Even though model re-initialization is not a very time consuming step, it does interrupt the workflow.
Exemplary embodiments of the invention as described herein generally include methods and systems for 3-D motion compensation even if only monoplane X-ray images are available using a training phase that employs a biplane sequence to estimate a patient-specific motion model for the circumferential mapping catheter positioned at the pulmonary vein considered for ablation to obtain pulmonary vein isolation (PVI). A constrained model-based 2-D/3-D registration method is then used to track a circumferential mapping catheter.
According to an aspect of the invention, there is provided a method for model based motion tracking of a catheter during an ablation procedure including receiving a training series of biplanar fluoroscopic images of a catheter acquired under conditions that will be present during an ablation procedure, segmenting and processing the series of biplanar images to produce a distance transform image for each biplanar image at each acquisition time, minimizing, for each pair of biplanar images at each acquisition time, a first cost function of the distance transform image for each pair of biplanar images to yield a translation parameter that provides a best fit for a model of the catheter to each pair of biplanar images at each acquisition time, and calculating an updated catheter model for each acquisition time from the translation parameter.
According to a further aspect of the invention, the method includes calculating a model center for each updated catheter model, calculating a single motion vector from the model centers of the updated catheter models determining a viewing direction vector from a last row of a projection matrix from homogenous coordinates to a 3-D space, determining a search direction vector perpendicular to the viewing direction vector and the single motion vector, minimizing a second cost function of one distance transform image for each acquisition time, the cost function parameterized by the perpendicular to search direction vector and the single motion vector to determine optimal coefficients of the perpendicular search direction vector and the single motion vector, and using the optimal coefficients with the perpendicular search direction vector and the single motion vector to update the catheter model at each acquisition time.
According to a further aspect of the invention, segmenting a biplanar image comprises using a cascade of weighted combinations of classification and regression trees (CARTs), where each CART is a weak classifier in which each node has a threshold associated with a feature that discriminates between a catheter and background and which partitions the feature space.
According to a further aspect of the invention, processing the series of biplanar images comprises smoothing segmentation results with a median filter, thinning the smoothed segmentation so that a catheter thickness need not be considered, and calculating for every background pixel in each image a distance to a closest object pixel to produce the distance transform image.
According to a further aspect of the invention, minimizing the first cost function comprises a multi-scale grid search that further comprises sub-sampling a search space and using a regions around a smallest cost function to initialize a search at a smaller scale.
According to a further aspect of the invention, the first cost function is
where PA,t and PB,t, are projection matrices for biplanar image planes A and B, IDT, At, IDT, Bt are the distance transform images for each image in the pair of biplane images at time t,
is a transformation matrix for the fluoroscopic images at time t with translation parameters rt=rt,x, rt,y, rt,z)T and superscript u for ‘unconstrained’, respectively, and mit, is the 3-D catheter model at time t, and the sum is over all points i in the catheter model.
According to a further aspect of the invention, the updated catheter model is calculated from mi,t+1=Ttu({circumflex over (r)}t)·mi,t for all points i in the catheter model.
According to a further aspect of the invention, the model center for each updated catheter model is calculated from
where the sum is over all points i in the catheter model.
According to a further aspect of the invention, calculating the single motion vector comprises performing a principal component analysis over all images.
According to a further aspect of the invention, the second cost function is
where P is a projection matrix for one of the two biplanar images acquired for each time t, IDT,t is the distance transform image for the one biplanar image at each time t, where transformation matrix
{circumflex over (λ)}t, {circumflex over (μ)}t are the optimal coefficients of the perpendicular search direction vector vp and the single motion vector vm with vp=(vp,x, Vp,y, vp,z)T, Vm=(vm,x, vm,y, vm,z)T the index c is for ‘constrained’, and mi,t, is the 3-D catheter model at time t, where the sum is over all points i in the catheter model.
According to a further aspect of the invention, the method includes weighting the transformation matrix over successive time steps as
T
c′({circumflex over (λ)}t,{circumflex over (μ)}t)=αTc({circumflex over (λ)}t,{circumflex over (μ)}t)+(1−α)Tc({circumflex over (λ)}t−1,{circumflex over (μ)}t−1) for αε(0,1)
According to another aspect of the invention, there is provided a method for model based motion tracking of a catheter during an ablation procedure including receiving a training series of biplanar fluoroscopic images of a catheter acquired under conditions that will be present during an ablation procedure, providing a series of catheter models updated from a single initial catheter model to correspond to each pair of biplanar images in the training series, calculating a model center for each updated catheter model, calculating a single motion vector from the model centers of the updated catheter models, determining a viewing direction vector from a last row of a projection matrix from homogenous coordinates to a 3-D space, determining a search direction vector perpendicular to the viewing direction vector and the single motion vector, minimizing a first cost function of one distance transform image for each acquisition time, the cost function parameterized by the perpendicular search direction vector and the single motion vector to determine optimal coefficients of the perpendicular search direction vector and the single motion vector, and using the optimal coefficients with the perpendicular search direction vector and the single motion vector to update the catheter model at each acquisition time.
According to a further aspect of the invention, providing a series of catheter models includes segmenting and processing the series of biplanar images to produce a distance transform image for each biplanar image at each acquisition time, minimizing, for each pair of biplanar images at each acquisition time, a first cost function of the distance transform image for each pair of biplanar images to yield a translation parameter that provides a best fit of the initial catheter model 1 to each pair of biplanar images at each acquisition time, and calculating an updated catheter model for each acquisition time from the translation parameter.
According to another aspect of the invention, there is provided a non-transitory program storage device readable by a computer, tangibly embodying a program of instructions executed by the computer to perform the method steps for model based motion tracking of a catheter during an ablation procedure.
a) depicts examples of Haar-like feature prototypes, according to an embodiment of the invention.
b) depicts an example of a classification and regression tree (CART), according to an embodiment of the invention.
c) depicts a classifier cascade of N stages of strong classifiers, according to an embodiment of the invention.
a)-(f) summarize the image segmentation and post-processing steps on a fluoroscopic image, according to an embodiment of the invention.
a) shows the 2-D tracking error for constrained and unconstrained approaches for each frame of a sequence, according to an embodiment of the invention.
b) illustrates 3-D tracking error of the constrained and unconstrained approach for the same sequence shown in
a) illustrates a comparison of the 2-D tracking accuracy of constrained and unconstrained 2-D/3-D registration, according to an embodiment of the invention.
b) compares 3-D error for constrained motion compensation versus no motion compensation, according to an embodiment of the invention.
c) compares 3-D motion compensation error obtained for constrained 2-D/3-D registration versus unconstrained 2-D/3-D registration, according to an embodiment of the invention.
d) illustrates a comparison of motion compensation between the constrained monoplane 2-D/3-D approach and the biplane approach, according to an embodiment of the invention.
a)-(b) illustrate the motion compensation error along the viewing direction of the constrained and unconstrained methods, respectively, according to an embodiment of the invention.
a)-(b) show 2-D and 3D tracking error results, respectively, according to an embodiment of the invention.
a)-(e) shows examples of how segmentation results depend on the number of stages used for classification, according to an embodiment of the invention.
a) is a table of results for the segmentation, median filtering, skeletonization, distance transform, and constrained registration, according to an embodiment of the invention.
b) is a table of comparisons of several motion compensation methods to utilizing 2-D/3-D registration, according to an embodiment of the invention.
a)-(b) illustrate a comparison showing the differences with and without motion compensation applied to the fluoroscopic overlay, according to an embodiment of the invention.
a)-(b) show a fluoroscopic image with a motion-compensated 3-D overlay compared to the original X-ray frame using contrast injection, according to an embodiment of the invention.
Exemplary embodiments of the invention as described herein generally include systems and methods for creating a patient-specific motion model and using the motion model to track a catheter during an ablation procedure. Accordingly, while the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the invention to the particular forms disclosed, but on the contrary, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention.
As used herein, the term “image” refers to multi-dimensional data composed of discrete image elements (e.g., pixels for 2-dimensional images and voxels for 3-dimensional images). The image may be, for example, a medical image of a subject collected by computer tomography, magnetic resonance imaging, ultrasound, or any other medical imaging system known to one of skill in the art. The image may also be provided from non-medical contexts, such as, for example, remote sensing systems, electron microscopy, etc. Although an image can be thought of as a function from R3 to R or R7, the methods of the inventions are not limited to such images, and can be applied to images of any dimension, e.g., a 2-dimensional picture or a 3-dimensional volume. For a 2- or 3-dimensional image, the domain of the image is typically a 2- or 3-dimensional rectangular array, wherein each pixel or voxel can be addressed with reference to a set of 2 or 3 mutually orthogonal axes. The terms “digital” and “digitized” as used herein will refer to images or volumes, as appropriate, in a digital or digitized format acquired via a digital acquisition system or via conversion from an analog image.
According to an embodiment of the invention, a constrained 2-D/3-D registration can be applied to perform motion compensation. An approach according to an embodiment of the invention is based on a 3-D catheter model and an estimated patient-specific motion model learned during a training phase. Set-up of the 3-D catheter model requires only a single biplane X-ray image, and the training phase estimates the motion of the CFM catheter at the PV using a biplane sequence in which the mapping catheter is tracked using an unconstrained 2-D/3-D registration. The principal motion axis is determined from the trajectory established during the tracking training phase. This axis is considered a patient-specific motion model. Since the main axis of the motion by itself is not sufficient to provide a good search space for a constrained registration, another axis is needed. According to an embodiment of the invention, a vector perpendicular to the viewing direction and the main axis is used, because the search region for the constrained registration can then be reduced to a 2-D search space, spanned by the principal axis and a vector parallel to the image plane. This allows tracking motion that is parallel to the image plane, and also capturing depth information with respect to the pre-acquired motion model. A constrained approach according to an embodiment of the invention can be simply stated as dimension reduction of the search space from 3-D to 2-D.
A 3-D catheter model generation from two views using manually selected points as input is now summarized. A 3-D catheter model generation method according to an embodiment of the invention requires that points set on the catheter are displayed in a single biplane X-ray acquisition. Beyond that, this method puts no further restrictions on an operator, i.e., no special parts of the catheter need to be selected. Put differently, 2-D points pA, pBεR2 are selected on the elliptically-shaped part of the catheter in each image plane. The two image planes are denoted by A and B. Two-dimensional ellipses CA, CB E R3×3 in the image planes are then fitted to these points using matrices of the form
with the coefficients of the fitted ellipse a, b, c, d, e, fεR. If the 2-D points were on a perfect ellipse, the matrices would satisfy the following equations
{tilde over (p)}
A
T
C
A
{tilde over (p)}
A=0 (1)
{tilde over (p)}
B
T
C
B
{tilde over (p)}
B=0 (2)
with the 2-D points pA and pB represented in homogeneous coordinates as {tilde over (p)}A=(pAT,1) and {tilde over (p)}B=(pBT, 1). For ellipse fitting, at least six points are used, and ellipse fitting can be performed in a least-squares sense if more than six points are provided. A constraint ac−b2>0 is used to ensure that the solutions for CA and CB are elliptical. After the ellipses in the image planes are fitted, ellipse reconstruction in 3-D can be performed by computing two 3-D cones QA, QBεR4×4 using the projection matrices PA, PB E R3×4. The cones are spanned from the cameras' optical centers to the ellipses on the image planes, as shown in
Q
A
=P
A
T
C
A
P
A, (3)
Q
B
=P
A
T
C
B
P
B, (4)
The quadrics QA, QBεR4×4 are of rank 3. Every intersection of a plane with one of the cones yields a valid solution of an ellipse in 3-D when projected onto the respective image plane. Given two cones, there are two possible solutions by calculating the two intersecting planes of the two 3-D cones. The two intersecting planes can be found by calculating a quadric Q(η) from Q(η)=QA+ηQB such that the resulting Q(η) is of rank 2. The solution is found by calculating the latent roots of Q(η)=QA=ηQB, and the equations of the two intersecting planes are found by an eigenanalysis of the matrix Q(η). Intersecting the 3-D cones with the two 3-D planes yields two possible solutions. The more circular solution of the two ellipses in 3-D is assumed to be the correct one, given the assumption that the pulmonary veins are tubularly shaped and that the circumferential mapping catheter is attached to the PV firmly. The 3-D model is denoted as mi=(mi,x, mi,y, mi,z,1)T E R4 in homogeneous coordinates with iε[1, N] and N is the number of model points. Experiments indicated that N=50 is sufficient to achieve good results. Increasing the number of model points did not lead to improved results. This step needs to be performed only once.
Once a 3-D model of the catheter has been generated, only monoplane fluoroscopic imaging is needed for the remaining tasks of motion estimation and compensation. A method according to an embodiment of the invention does not require manual re-initialization of the catheter model when C-arm angulation changes, minimizing user interaction during interventions.
To segment a catheter in a fluoroscopic image, a 1024×1024 pixel input image is cropped to a 400×400 pixel region-of-interest (ROI) around the projected center of the catheter model. A ROI size of 400×400 was chosen ensure that one can always find the circumferential mapping catheter in all images. Using a different ROI size, or to be more specific, a higher search range, would only affect the runtime. In the first frame, the position is known from the initialization step, and in all the subsequent frames, the tracking result from the previous frame is used.
A catheter segmentation method according to an embodiment of the invention should be reliable and fast enough to allow motion estimation and compensation at the frame rate of the EP procedure, which typically ranges between 1 to 10 frames per second. According to an embodiment of the invention, a combination of Haar-like features and a cascade of boosted classifiers are used. AdaBoost in combination with Haar-like features has proven extremely successful in the field of real-time face detection. Haar-like features can calculate various patterns of intensity differences. Some features detect edges, whereas others focus on line-like structures and are useful for detecting thin objects such as the catheter. Examples of Haar-like feature prototypes are shown in
Even for moderate window sizes, the number of generated features is large, about 40,000 for a 15×15 pixel window. The most suitable features for discriminating between catheter and background are selected by an AdaBoost algorithm and integrated into a classifier. The idea is to combine several weak classifiers, which only have to be slightly better than chance, to form a strong classifier. In the simplest case, a weak classifier amounts to a single feature and threshold. During training, weak classifiers are repeatedly evaluated on samples of a training set in which the catheter has been annotated. The classifier with a minimum classification error is added to a linear combination of weak classifiers until the overall error is below the desired threshold. After each iteration, the importance of individual samples is re-weighted to put more emphasis on misclassified samples for the next evaluation.
According to an embodiment of the invention, instead of single features and thresholds, classification and regression trees (CARTs) are used as weak classifiers. A CART is a small tree of fixed size.
Several strong classifiers, each comprising a weighted combinations of CARTs, are organized into a cascade with N stages.
Different types of circumferential mapping catheters may be used for EP ablation purposes. These mapping catheters can differ in the number of electrodes as well as the thickness of the catheter. To ensure that a method according to an embodiment of the invention can handle a wide variety of catheters, post-processing is performed on the segmentation output from the learning-based catheter classifier. First, the catheter segmentation results are smoothed by a median filter. Experiments indicated that a 5×5 pixel kernel size yielded good results. Second, the smoothed segmentation is thinned so that the catheter thickness can be disregarded in the following registration step. Third, to obtain a smooth representation of the circumferential mapping catheter for subsequent registrations, the distance transform for the skeletonized image is calculated. The distance transform of an image calculates for every background pixel the distance to the closest object pixel. The resulting image is represented as IDT.
a)-(f) summarize the image processing steps on a fluoroscopic image, according to an embodiment of the invention.
Next is presented a method according to an embodiment of the invention for the generation of a patient specific motion model. A short biplane sequence is used to generate 3-D samples of the position of the circumferential mapping catheter recorded during a training phase. The principal axis derived from the sample positions is taken as the main direction of PV motion. The motion model should be acquired in the same state (heart rate in sinus rhythm, arrhythmia) that will be present during the application of the motion model. Next, a constrained model-based 2-D/3-D registration is applied to track the circumferential mapping catheter in 3-D using monoplane fluoroscopy. To this end, the motion model estimated during the training phase limits the allowed motion to two directions. The first motion is parallel to the principal motion axis. The second allowed motion direction is parallel to the image plane, because the underlying motion vector is designed to be perpendicular to the principal axis and the viewing direction.
A motion model according to an embodiment of the invention is created by a 2-D/3-D registration of the previously generated 3-D catheter model to biplane fluoroscopic images acquired during a training phase. A flowchart of a method according to an embodiment of the invention of creating a patient-specific motion model and using the motion model to track a catheter during an ablation procedure is presented in
with translation parameters rt=(rt,x, rt,y, rt,z)T and superscript u for ‘unconstrained’. The cost function according to an embodiment of the invention can be formulated as
with the projection matrices for image planes A and B, PA,t and PB,t, respectively, as well as the 3-D catheter model, mi,t, at time t. Minimization can be performed using a multi-scale grid search approach. The search space was sub-sampled and the region around the smallest cost function was used for a smaller search grid. The projection matrix used during the training phase tracking is not required to be one of the projection matrices used for model generation, i.e., the C-arm can be moved in between catheter model generation and the training phase. The same holds for the actual motion compensation. Given the parameters {circumflex over (r)} found by the nearest-neighbor search, the catheter model can be updated at step 44 to mi,t+1εR4 by
∀i:mi,t+1=Ttu({circumflex over (r)}t)·mi,t. (7)
During the training phase, the same transformation Ttu({circumflex over (r)}t) can be applied to the 3-D volumetric data set that is used for image overlay. This way, a 3-D motion compensation can be shown during the training phase.
A patient specific motion model according to an embodiment of the invention is calculated from the circumferential mapping catheter positions. To this end, the catheter model for every time step is reduced to the center of the model at step 45 by
The principal axis for the catheter centers
Next, motion compensation according to an embodiment of the invention by model-constrained registration is introduced. The assumption of an approach according to an embodiment of the invention is that only monoplane fluoroscopic imaging is available. A constraint according to an embodiment of the invention is the reduction of the 3-D search space to a 2-D search space by introducing a second feasible motion vector that is perpendicular to the viewing direction and the principal motion vector. This results in a 2-D search plane for the catheter model that is semi-parallel to the image plane. The cost function is the distance transform IDT of the post-processed segmentation result. By using the main motion vector vm, the 2-D search space also allows some depth estimation from a single X-ray view. Motion analyses of the left atrium have revealed that the dominant motion is in the anterior-posterior and superior-inferior directions. The degree of rotation is much less, possibly due to deformation of the left atrium. Physicians position their C-arms in standard viewing positions, and usually only angulations in the left-anterior-oblique (LAO), posterior-anterior (PA), or right-anterior-oblique (RAO) directions are used. Angulations towards cranial or caudal directions are, at least to the knowledge of the inventors, not common for EP procedures. If image acquisition is performed with the C-arm in an LAO, PA, or RAO position, most of the motion is captured, as the motion of the left atrium is parallel to the image plane.
Referring again to
v
p
=v
m
×v
v. (9)
Any point on that plane can be represented by a linear combination of these two vectors vp and vm. This translation can be rewritten in matrix notation as
with vp=(vp,x, vp,y, vp,z)T and vm=(vm,x, vm,y, Vm,z)T and the index c for ‘constrained’. An objective function according to an embodiment of the invention for the constrained registration is then defined using the distance transformed images for image planes A or B, IDT,A or IDT,B, respectively In the remainder of this section, the indices A and B are omitted, and P stands either for PA or PB. The same holds for IDT,t. A cost function according to an embodiment of the invention for the constrained registration can then be stated as:
This cost function is minimized for each time step in the training series. Optimization of the cost function to determine the parameters {circumflex over (λ)}t, {circumflex over (μ)}t is performed at step 48 using a nearest-neighbor search, as for the training phase. Given the parameters {circumflex over (λ)}t, {circumflex over (μ)}t, the catheter model can be updated at step 49 to mi,t+1, εR4 by
∀i:mi,t+1=Tc({circumflex over (λ)}t,{circumflex over (μ)}t) (12)
The same transformation Tc({circumflex over (λ)}t, {circumflex over (μ)}t) is then applied to the 3-D volumetric data set that is used to compute the image overlay by 2-D forward projection of the 3-D model based on the known projection geometry. This way, one can achieve 3-D motion compensation for monoplane fluoroscopic images.
According to other embodiments of the invention, additional processing can be performed on the transformation matrices Tc({circumflex over (λ)}t, {circumflex over (μ)}t). Since the parameters {circumflex over (λ)}t, {circumflex over (μ)}t are calculated for each frame in the time series, the matrices Tc({circumflex over (λ)}t, {circumflex over (μ)}t) are time dependent. According to an embodiment of the invention, one can weight the transformations over successive time steps: Tc′({circumflex over (λ)}t, {circumflex over (μ)}t)=αTc({circumflex over (λ)}t, {circumflex over (μ)}t)+(1+α)Tc({circumflex over (λ)}t−1, {circumflex over (μ)}t−1) for αε(0,1) exemplary, non-limiting value of a is 0.90.
According to an embodiment of the invention, the results of the constrained registration are compared to an unconstrained method that uses a full 3-D translation as a motion model. To this end, an unconstrained registration to a monoplane fluoroscopy image is performed. According to an embodiment of the invention, EQ. (6) can be adapted to the monoplane case by rewriting it as:
Motion compensation can be performed using {circumflex over (r)}′ to update the catheter model as in EQ. (7) and applying the same transformation to the 3-D data set used to generate the overlay images.
Next, the performance of a motion-model constrained 2-D/3-D registration algorithm for motion compensation according to an embodiment of the invention is evaluated and the results are presented. The tracking accuracy of the constrained and unconstrained methods were calculated by comparison to a gold-standard segmentation. For evaluation, 13 clinical biplane sequences were available. Two different types of circumferential mapping catheters were encountered. One type was used in 11 biplane sequences, and a second type was used in the two other sequences. The fluoroscopic sequences were acquired during standard electrophysiology procedures. The circumferential mapping catheter was placed at the ostium of the pulmonary vein during image acquisition. The catheter is usually firmly placed to ensure a good wall contact. A suboptimal wall contact may lead to undetected residual PV-atrial electrical connections, and potentially to an incomplete pulmonary vein isolation. One gold-standard segmentation was available for each sequence, i.e., the catheter was segmented by one expert observer in each frame of the whole sequence. Data was acquired from six different patients at one clinical site. All X-ray sequences were recorded on an AXIOM Artis dBC biplane C-arm system (Siemens AG, Healthcare Sector, Forchheim, Germany). The training of the classifier was performed on a two-fold cross validation, i.e., the biplane sequence considered for evaluation was excluded from the training data set. For each sequence, a 3-D model was generated as described above. Then, a constrained method was evaluated by using each image plane of the biplane sequences independently. The frames used for the generation of the motion model were excluded from evaluation. For the unconstrained approach, the same frames were used for evaluation to arrive at comparable results. The constrained method used a training phase of 50% of the sequence. The shortest sequence available comprised 10 frames, and the longest 117. Individual sequences for training of the motion model were not available. To evaluate the influence of the number of frames used during the training phase, the three longest sequences available were used, comprising 79, 95 and 117 frames, respectively. The training phases for this evaluation were chosen to comprise 5, 10, 20, 30, and 40 frames, respectively.
Since motion estimation and compensation is performed in 3-D, and for each case there are biplane sequences to derive the ground truth position in 3-D, a 3-D error can be estimated as well. To this end, the tip of the circumferential mapping catheter was manually localized in 3-D by triangulation from two views. This can only be used as an estimation for the actual 3-D error. An accurate evaluation would require a high resolution 3-D data set for each time instant. Such data is unfortunately not available. The 3-D trajectories of the catheter tip were taken as the gold-standard for the observed 3-D motion.
For the 26 tested sequences, the observed motion was 4.5 mm±2.4 mm. The constrained motion compensation approach yielded a 3-D tracking error of 1.58 mm±0.95 mm. The unconstrained approach performed considerably worse with an average 3-D error of 3.21 mm±1.62 mm. Even though the constrained motion compensation method performed well, a gold-standard biplane method is still superior regarding the 3-D accuracy (0.7 mm±0.4 mm). However, its better accuracy comes at the cost of increased X-ray dose.
As drift is an often discussed issue when evaluating tracking methods, tracking error over time was also considered. This question is especially interesting, as only one previous frame is considered when tracking the current frame. In particular, the tracking result of the previous frame is used for cropping the region-of-interest in the current frame. Apart from that, all frames are treated independently. For example, the 2-D tracking error for constrained 52 and unconstrained 51 approaches for each frame of sequence #19 is given in
The error along the viewing direction was also computed. Since this direction is excluded in the search space for the constrained method, the error along the viewing direction was the largest among all directions, averaging at 1.03 mm±0.94 mm. However, the unconstrained method also had its largest error along the viewing direction with an average of 2.99 mm±0.43 mm
A method according to an embodiment of the invention uses a boosted classifier cascade. To evaluate how many stages in the cascade are needed for achieving good motion compensation results, an experiment using a constrained method according to an embodiment of the invention was performed.
The number of stages used in the cascade can also be addressed by looking at the time required to estimate the motion for one frame. Fast and accurate methods are desired for interventional applications. The time in ms was measured for each frame and the average, the minimum and the maximum were calculated. The results of the computational time of the catheter segmentation versus the number of stages in the boosted classifier cascade are shown in
Besides the runtime of the classification, the runtime of all components of our presented algorithm are also of interest. Results for the segmentation, median filtering, skeletonization, distance transform, and constrained registration are presented in Table I, shown in
The initialization of the catheter model used for motion compensation in a method according to an embodiment of the invention is required only once. A 2-D/3-D registration according to an embodiment of the invention incorporates the projection matrix, so no model reinitialization is required if the viewing direction of a C-arm is changed. Although model re-initialization is usually not time consuming, it does interrupt the workflow because of manual interaction involving the user. In fact, while the catheter model can be calculated in less than 55 ms on a PC platform, user feedback needed for model initialization carries the risk of slowing the workflow.
During the experiments, it was found that a catheter model of 50 points yielded good results. Increasing the number of model points did not improve tracking accuracy.
A method according to an embodiment of the invention relies on a constrained 2-D/3-D registration involving a 3-D catheter model as well as a motion model, and can capture the relevant motion at the site of ablation and take it into account real-time. Since an approach according to an embodiment of the invention uses a 3-D catheter model, re-initialization after repositioning the C-arm can be avoided. However, the reconstruction of the 3-D catheter model requires at least two views in the same cardiac cycle and breathing phase.
The computation of the motion model requires a training phase. In clinical practice, the training phase could be included into the workflow. At the beginning of each AFib ablation procedure, the signals at the PVs are documented and the correct position of the circumferential mapping catheter is verified by contrast injection and, if available, using a short biplane sequence. This sequence could be sufficient to initialize a motion model according to an embodiment of the invention. As four pulmonary veins are typically ablated during a procedure, four individual motion models, one for each PV, may need to be trained. Evaluating the 3-D tracking error with respect to number of frames used during the training phase indicated that a method according to an embodiment of the invention is insensitive to the length of the training phase, as shown in
A method according to an embodiment of the invention can achieve a 3-D accuracy of about 1.6 mm. However, to reduce the 3-D error, one could employ simultaneous biplane imaging, which comes at the cost of a higher dose for the patient and the medical staff.
A method according to an embodiment of the invention cannot take into account motion along the viewing direction because depth information cannot be reliably estimated from monoplane projection images. Using an unconstrained approach, the 3-D error remains high, especially along the viewing direction, as shown in
The distance transform provides the main input for the cost function. As long as only one circumferential mapping catheter appears in the image, there is only one global optimum for the cost function. A multi-scale grid-search approach according to an embodiment of the invention did not detect any local optima. If multiple elliptical shaped catheters were used, local optima would appear and an optimization strategy according to an embodiment of the invention could detect one of these. A visualization of a cost function for one frame in a small area around an optimum is shown in
A method according to an embodiment of the invention is purely image driven, and can be extended to learn the motion difference between the circumferential mapping and the coronary sinus (CS) catheter. The same idea could be applied to using the diaphragm for motion compensation. Embodiments of the invention for motion estimation rely on the assumption that the circumferential mapping catheter is thinly placed at the PV where ablation takes place. If the mapping catheter floated freely within the left atrium, one could not get a reliable motion estimate. In such a case, an additional motion analysis stage could detect the free motion.
To further improve the accuracy, other methods such as probabilistic boosting trees or random forests can be used for catheter segmentation. For a higher efficiency, a different skeletonization method other than the thinning algorithm could be used. Since motion compensation for AFib ablation procedures should be usable within an interventional setup, near real-time performance is desirable. In this context, real-time is regarded as the frame rate that is used for image acquisition during the intervention. AFib ablation procedures are lengthy procedures, and fluoroscopy times often accumulate to more than 30 minutes. This is why physicians try to reduce dose, e.g., by lowering the acquisition frame rates. One reason for the long fluoroscopy times is the complexity encountered when trying to isolate the pulmonary veins. Embodiments of the invention can shorten the procedure time by offering navigation based on fluoroscopy overlay images.
A comparison showing the difference with and without motion compensation applied to the fluoroscopic overlay is presented in
It is to be understood that embodiments of the present invention can be implemented in various forms of hardware, software, firmware, special purpose processes, or a combination thereof. In one embodiment, the present invention can be implemented in software as an application program tangible embodied on a computer readable program storage device. The application program can be uploaded to, and executed by, a machine comprising any suitable architecture.
The computer system 181 also includes an operating system and micro instruction code. The various processes and functions described herein can either be part of the micro instruction code or part of the application program (or combination thereof) which is executed via the operating system. In addition, various other peripheral devices can be connected to the computer platform such as an additional data storage device and a printing device.
It is to be further understood that, because some of the constituent system components and method steps depicted in the accompanying figures can be implemented in software, the actual connections between the systems components (or the process steps) may differ depending upon the manner in which the present invention is programmed. Given the teachings of the present invention provided herein, one of ordinary skill in the related art will be able to contemplate these and similar implementations or configurations of the present invention.
While the present invention has been described in detail with reference to exemplary embodiments, those skilled in the art will appreciate that various modifications and substitutions can be made thereto without departing from the spirit and scope of the invention as set forth in the appended claims.
This application claims priority from “Constrained Registration For Motion Compensation In Atrial Fibrillation Ablation Procedures”, U.S. Provisional Application No. 61/473,313 of Liao, et al., filed Apr. 8, 2011, the contents of which are herein incorporated by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
61473313 | Apr 2011 | US |