The present disclosure relates to a system, apparatus, and method of navigation and position confirmation for surgical procedures. More particularly, the present disclosure relates to a system and method for constructing a fluoroscopic-based three dimensional volume from two dimensional fluoroscopic images captured using a standard fluoroscopic imaging device.
There are several commonly applied methods for treating various maladies affecting organs including the liver, brain, heart, lung and kidney. Often, one or more imaging modalities, such as magnetic resonance imaging, ultrasound imaging, computed tomography (CT), as well as others are employed by clinicians to identify areas of interest within a patient and ultimately targets for treatment.
An endoscopic approach has proven useful in navigating to areas of interest within a patient, and particularly so for areas within luminal networks of the body such as the lungs. To enable the endoscopic, and more particularly the bronchoscopic, approach in the lungs, endobronchial navigation systems have been developed that use previously acquired MRI data or CT image data to generate a three dimensional rendering or volume of the particular body part such as the lungs. In particular, previously acquired images, acquired from an MRI scan or CT scan of the patient, are utilized to generate a three dimensional or volumetric rendering of the patient.
The resulting volume generated from the MRI scan or CT scan is then utilized to create a navigation plan to facilitate the advancement of a navigation catheter (or other suitable device) through a bronchoscope and a branch of the bronchus of a patient to an area of interest. Electromagnetic tracking may be utilized in conjunction with the CT data to facilitate guidance of the navigation catheter through the branch of the bronchus to the area of interest. In certain instances, the navigation catheter may be positioned within one of the airways of the branched luminal networks adjacent to, or within, the area of interest to provide access for one or more medical instruments.
Thus, in order to generate a navigation plan, or in order to even generate a three dimensional or volumetric rendering of the patient's anatomy, such as the lung, a clinician is required to utilize an MRI system or CT system to acquire the necessary image data for construction of the three dimensional volume. An MRI system or CT-based imaging system is extremely costly, and in many cases not available in the same location as the location where a navigation plan is generated or where a navigation procedure is carried out.
A fluoroscopic imaging device is commonly located in the operating room during navigation procedures. The standard fluoroscopic imaging device may be used by a clinician to visualize and confirm the placement of a tool after it has been navigated to a desired location. However, although standard fluoroscopic images display highly dense objects such as metal tools and bones as well as large soft-tissue objects such as the heart, the fluoroscopic images have difficulty resolving small soft-tissue objects of interest such as lesions. Further, the fluoroscope image is only a two dimensional projection. In order to be able to see small soft-tissue objects in three dimensional space, an X-ray volumetric reconstruction is needed. Several solutions exist that provide three dimensional volume reconstruction of soft-tissues such as CT and Cone-beam CT which are extensively used in the medical world. These machines algorithmically combine multiple X-ray projections from known, calibrated X-ray source positions into three dimensional volume in which the soft-tissues are visible.
In order to navigate tools to a remote soft-tissue target for biopsy or treatment, both the tool and the target should be visible in some sort of a three dimensional guidance system. The majority of these systems use some X-ray device to see through the body. For example, a CT machine can be used with iterative scans during procedure to provide guidance through the body until the tools reach the target. This is a tedious procedure as it requires several full CT scans, a dedicated CT room and blind navigation between scans. In addition, each scan requires the staff to leave the room. Another option is a Cone-beam CT machine which is available in some operation rooms and is somewhat easier to operate, but is expensive and like the CT only provides blind navigation between scans, requires multiple iterations for navigation and requires the staff to leave the room.
Accordingly, there is a need for a system that can achieve the benefits of the CT and Cone-beam CT three dimensional image guidance without the underlying costs, preparation requirements, and radiation side effects associated with these systems.
The present disclosure is directed to a system and method for constructing three dimensional volumetric data in which small soft-tissue objects are visible from a video stream (or plurality of images) captured by a standard fluoroscopic imaging device available in most procedure rooms. The fluoroscopic-based constructed three dimensional volumetric data may be used for guidance, navigation planning, improved navigation accuracy, navigation confirmation, and treatment confirmation.
Soft tissue objects are not visible in standard fluoroscopic images and video because they are obscured by dense objects such as dense tissue. The present disclosure is directed to a system and method for creating a three-dimensional pseudo-volume in which the obscuring objects are filtered based on the three dimensional position, and then projected back into two dimensions. In one aspect, multiple fluoroscopic images, each captured at different angles, are utilized to construct the three dimensional pseudo-volume data. The present disclosure describes a system and method which is capable of constructing the three dimensional pseudo-volume data utilizing fluoroscopic images captured from a short range of angles relative to a patient or target region of a patient. Additionally, the present disclosure is also directed to a system and method for improving a previously created three dimensional rendering utilizing two dimensional images, or video, captured by a standard fluoroscopic imaging device.
As described in greater detail below, one aspect of the present disclosure is to determine three dimensional positions of features in the fluoroscopic video, such as three dimensional catheter position, three dimensional target tissue (for example a lesion) position, etc. In order to accomplish this, the pose of the fluoroscopic imaging device for each frame must be determined or known. If an external angle measurement device coupled to the fluoroscopic imaging device is utilized, then the angles and pose of the imaging device is known for each frame. However, when external measurement devices are not utilized other techniques are employed to determine the poses of the fluoroscopic imaging device for each frame, as described in greater detail below. For example, previously acquired CT scan data may be registered to the fluoroscopic video in order to algorithmically find the pose of the fluoroscope for each frame of the captured video. Alternatively, by tracking a few visible markers (two dimensional visible features) in the fluoroscopic video, the pose and three dimensional positions may be solved together by using some structure from motion technique. In some aspects, easier techniques in which the poses are known (by an angle measurement device) may be utilized and only the three dimensional features' positions need to be solved. In order to correct for movements, at least one marker (or a surgical device such as a catheter tip) may be utilized.
Multiple fluoroscope two dimensional images can be processed algorithmically to create pseudo three dimensional volumetric data, similar to Cone-beam CT, but with varying arbitrary angle range and fluoroscope poses, where the fluoroscope is rotated manually. The angle range may be very small (˜30° which may result in poor three dimensional reconstruction quality. The algorithm used is an iterative accelerated projection/back-projection method which unlike the analytic algorithms (Radon transform, FDK . . . ) doesn't assume any predetermined angle range or angle rotation rate. In order to overcome the poor three dimensional reconstruction, instead of displaying the raw three dimensional reconstructed data to the user, the three dimensional reconstruction may be cropped around the area of interest (also referred to herein as the “FluoroCT Blob”). The cropped three dimensional data is then reprojected into two dimensional virtual fluoroscope images in which local soft-tissue features are visible. The reprojected two dimensional images are of good quality (compared to the poor three dimensional reconstruction), especially if the projections are done from the same fluoroscope poses as were seen in the video.
In order to reconstruct the three dimensional data, the pose of the fluoroscopic imaging device must be determined for each two dimensional fluoroscope frame in the video, relative to some fixed coordinate system. The pose of the fluoroscopic imaging device for each frame captured may be determined using any of the methods described below. For example the pose may be determined using an external measurement device coupled to the fluoroscopic imaging device.
Additionally, or alternatively, the pose may be determined using an external measurement device and a single marker or a catheter tip. Specifically, in some cases, the fluoroscope (or C-arm) may jitter during rotation, in which case some stabilization is needed. Unfortunately, due to filtering, the angle measurement device may still report smooth angles throughout the video, ignoring the high-frequencies which are present in the actual camera poses. In this case, a single two dimensional marker can be tracked throughout the video and used to stabilize the camera poses, or to increase their accuracy at the area of the marker. Since the marker will usually be located at the region of interest, this increases the camera pose accuracy in this region, and thus improves the three dimensional reconstruction quality. This method can also be used to compensate for patient body movement such as breathing during the video. Instead of using the camera pose as reported by the angle measurement device, a compensated camera pose is computed using the tracked two dimensional marker, such that all camera poses will be correct relative to it. The single marker used can be the tip of a tool or a catheter which is currently inserted to the patient.
Additionally, or alternatively, the pose may be determined via registration of the fluoroscopic video to previously acquired CT data. Specifically, a previously acquired CT of the patient may be available. In this case, each frame of the fluoroscope video can be registered to a virtual fluoroscopic frame of the CT (camera pose is searched in CT space until the virtual fluoroscopic image, corresponding to the camera pose, matches the one seen in the video). In this way, the camera pose is realized using image-based features matching.
Additionally, or alternatively, the camera pose may be determined using structure from motion techniques. Specifically, if numerous two dimensional features can be tracked throughout the two dimensional video frames, from beginning to end, then these two dimensional features can be used to realize camera poses (together with three dimensional feature positions) for each frame. These features can be artificial markers which were introduced to the patient during procedure.
Filtering the obscuring tissue from the tissue of interest can be done by cropping at a distance from the center of the generated three dimensional pseudo-volume data, cropping at a distance from the tool/catheter, or registering to previously obtained three dimensional volume data (CT) and using it to know which objects to filter.
Once the soft-tissue (such as a target lesion) is visible in the three dimensional reconstructed data (or in the two dimensional enhanced FluoroCT images), all three dimensional information in fluoroscope coordinates can be obtained. When working with the raw three dimensional reconstructed data the three dimensional information is obtained directly from the data. Alternatively, when working with the two dimensional enhanced FluoroCT images, 2-angles markings are needed in order to realize three dimensional positions (triangulation).
The obtained three dimensional data may be utilized for confirmation of tool to soft-tissue target three dimensional relation. For example, the data may be used to determine whether the tool reached the target, the orientation of the tool relative to the target, the distance between the tool and the target, or whether the target falls within an ablation zone of the tool.
Additionally, or alternatively, the three dimensional data may be utilized for correction of navigation. For example, the three dimensional positions can be transformed from fluoroscope coordinates into navigation system coordinates to improve navigation system accuracy at the region of interest. In one aspect, when utilized in an EMN system, fluoroscope coordinates can be transformed to antenna coordinates by assuming that the C-arm is perfectly perpendicular to the antenna and matching the catheter tip, seen in the fluoroscope video, to the catheter position in antenna at the time the video was taken. Fluoroscopic to/from antenna registration can also be achieved by computing the angle between the C-arm and the antenna using the Earth's magnetic field, attaching an EMN sensor to the fluoroscopic imaging device, or aligning a known two dimensional feature of the antenna.
Aspects of the present disclosure are described in detail with reference to the figures wherein like reference numerals identify similar or identical elements. As used herein, the term “distal” refers to the portion that is being described which is further from a user, while the term “proximal” refers to the portion that is being described which is closer to a user.
According to one aspect of the present disclosure, a system for constructing fluoroscopic-based three dimensional volumetric data from two dimensional fluoroscopic images is provided. The system includes a computing device configured to facilitate navigation of a medical device to a target area within a patient and a fluoroscopic imaging device configured to acquire a fluoroscopic video of the target area about a plurality of angles relative to the target area. The computing device is configured to determine a pose of the fluoroscopic imaging device for each frame of the fluoroscopic video and to construct fluoroscopic-based three dimensional volumetric data of the target area in which soft tissue objects are visible using a fast iterative three dimensional construction algorithm. In aspects, the system is configured to determine the pose of the fluoroscopic imaging device by knowing the angle range of movement of the device and computing the relative rotational speed of the device along the range.
The medical device may be a catheter assembly including an extended working channel configured to be positioned within a luminal network of the patient or the medical device may be a radio-opaque marker configured to be placed within the target area. The radio-opaque marker is at least partially visible in the fluoroscopic video acquired.
The computing device may be further configured to create virtual fluoroscopic images of the patient from previously acquired CT volumetric data and register the generated virtual fluoroscopic images with the acquired fluoroscopic video. The pose of the fluoroscopic imaging device for each frame of the fluoroscopic video may be determined based on the registration between the fluoroscopic video and the virtual fluoroscopic images.
The computing device may further be configured to detect frames missing from the fluoroscopic video and supplement the detected missing frames with corresponding virtual fluoroscopic images. The fluoroscopic-based three dimensional volumetric data may be constructed based on the fluoroscopic video and the corresponding virtual fluoroscopic images. In one aspect, the fluoroscopic-based three dimensional volumetric data may be registered with previously acquired CT data using image-based techniques such as “mutual information.” The fluoroscopic-based three dimensional volumetric data may be registered globally to the previously acquired CT data or locally, at the proximity of the target area of interest. A deep-learning based approach may be utilized in which the computing device “sees” many examples of suitable and non-suitable registrations and learns how to register the very two different modalities.
Additionally, the computing device may be further configured to track the two dimensional position or orientation of the medical device navigated to the target region throughout the fluoroscopic video. The computing device may be further configured to reconstruct positions of the medical device throughout the fluoroscopic video using a structure-from-motion technique. The pose of the fluoroscopic imaging device for each frame of the fluoroscopic video may be determined based on the reconstructed positions. Additionally, or alternatively, the pose of the fluoroscopic imaging device for each frame of the fluoroscopic video may be determined based on an external angle measuring device. The external angle measuring device may include an accelerometer, a gyroscope, or a magnetic field sensor coupled to the fluoroscopic imaging device. Additionally, in aspects, the computing device may be configured to synchronize the captured frames of the target area and compensate for shifts in the fluoroscopic imaging device or patient movement to correct construction of the fluoroscopic-based three dimensional volumetric data. Additionally, or alternatively, the computing device may be configured to crop a region of interest from the fluoroscopic-based three dimensional volumetric data, project the cropped region of interest onto the captured frames, and sharpen or intensify at least one of the region of interest or the captured frame to identify soft tissue objects.
In yet another aspect of the present disclosure a method for constructing fluoroscopic-based three dimensional volumetric data from two dimensional fluoroscopic images is provided. The method includes navigating a medical device to a target area within a patient, acquiring a fluoroscopic video of the target area about a plurality of angles relative to the target area using a fluoroscopic imaging device, determining a pose of the fluoroscopic imaging device for each frame of the fluoroscopic video, and constructing fluoroscopic-based three dimensional volumetric data of the target area in which soft tissue objects are visible using a fast iterative three dimensional construction algorithm. The medical device may be a catheter assembly including an extended working channel configured to be positioned within a luminal network of the patient or the medical device may be a radio-opaque marker configured to be placed within the target area. The radio-opaque marker is at least partially visible in the fluoroscopic video acquired.
The method may further include creating virtual fluoroscopic images of the patient from previously acquired CT volumetric data, and registering the fluoroscopic video with the virtual fluoroscopic images, wherein determining the pose of the fluoroscopic imaging device for each frame of the fluoroscopic video is based on the registration between the fluoroscopic video and the virtual fluoroscopic images. The method may further include detecting frames missing from the fluoroscopic video and supplementing the detected missing frames with corresponding virtual fluoroscopic images. Additionally, in aspects of the disclosure, the method may further include tracking the two dimensional position or orientation of the medical device navigated to the target region throughout the fluoroscopic video.
The positions of the medical device throughout the fluoroscopic video may be reconstructed using a structure-from-motion technique. The pose of the fluoroscopic imaging device for each frame of the fluoroscopic video may be determined based on the reconstructed positions. Additionally, or alternatively, the pose of the fluoroscopic imaging device for each frame of the fluoroscopic video may be determined based on an external angle measuring device. The external angle measuring device may include an accelerometer, a gyroscope, or a magnetic field sensor coupled to the fluoroscopic imaging device. Additionally, in aspects, the method may include synchronizing the captured frames of the target area and compensating for shifts in the fluoroscopic imaging device or patient movement to correct construction of the fluoroscopic-based three dimensional volumetric data. Additionally, or alternatively, the method may include cropping a region of interest from the fluoroscopic-based three dimensional volumetric data, projecting the cropped region of interest onto the captured frames, and sharpening or intensifying at least one of the region of interest or the captured frame to identify soft tissue objects.
Various aspects and embodiments of the present disclosure are described hereinbelow with references to the drawings, wherein:
The present disclosure is directed to a system and method for constructing local three dimensional volumetric data, in which small soft-tissue objects are visible, from a video stream captured by a standard fluoroscopic imaging device available in most procedure rooms. The constructed fluoroscopic-based local three dimensional volumetric data may be used for guidance, navigation planning, improved navigation accuracy, navigation confirmation, and treatment confirmation.
As shown in
EMN system 100 generally includes an operating table 20 configured to support a patient “P,” a bronchoscope 30 configured for insertion through the patient's “P's” mouth into the patient's “P's” airways; monitoring equipment 120 coupled to bronchoscope 30 (e.g., a video display, for displaying the video images received from the video imaging system of bronchoscope 30); a tracking system 50 including a tracking module 52, a plurality of reference sensors 54 and a transmitter mat 56; and a computing device 125 including software and/or hardware used to facilitate identification of a target, pathway planning to the target, navigation of a medical instrument to the target, and confirmation of placement of an EWC 12, or a suitable device therethrough, relative to the target.
A fluoroscopic imaging device 110 capable of acquiring fluoroscopic or x-ray images or video of the patient “P” is also included in this particular aspect of system 100. The images, series of images, or video captured by the fluoroscopic imaging device 110 may be stored within the fluoroscopic imaging device 110 or transmitted to computing device 125 for storage, processing, and display. Additionally, the fluoroscopic imaging device 110 may move relative to the patient “P” so that images may be acquired from different angles or perspectives relative to the patient “P” to create a fluoroscopic video. In one aspect of the present disclosure, fluoroscopic imaging device 110 includes an angle measurement device 111 which is configured to measure the angle of the fluoroscopic imaging device 110 relative to the patient “P.” Angle measurement device 111 may be an accelerometer. Fluoroscopic imaging device 110 may include a single imaging device or more than one imaging device. In embodiments including multiple imaging devices, each imaging device may be a different type of imaging device or the same type. Further details regarding the imaging device 110 are described in U.S. Pat. No. 8,565,858, which is incorporated by reference in its entirety herein.
Computing device 125 may be any suitable computing device including a processor and storage medium, wherein the processor is capable of executing instructions stored on the storage medium. The computing device 125 may further include a database configured to store patient data, CT data sets including CT images, fluoroscopic data sets including fluoroscopic images and video, navigation plans, and any other such data. Although not explicitly illustrated, the computing device 125 may include inputs, or may otherwise be configured to receive, CT data sets, fluoroscopic images/video and other data described herein. Additionally, computing device 125 includes a display configured to display graphical user interfaces. Computing device 125 may be connected to one or more networks through which one or more databases may be accessed.
With respect to the planning phase, computing device 125 utilizes previously acquired CT image data for generating and viewing a three dimensional model of the patient's “P's” airways, enables the identification of a target on the three dimensional model (automatically, semi-automatically, or manually), and allows for determining a pathway through the patient's “P's” airways to tissue located at and around the target. More specifically, CT images acquired from previous CT scans are processed and assembled into a three dimensional CT volume, which is then utilized to generate a three dimensional model of the patient's “P's” airways. The three dimensional model may be displayed on a display associated with computing device 125, or in any other suitable fashion. Using computing device 125, various views of the three dimensional model or enhanced two dimensional images generated from the three dimensional model are presented. The enhanced two dimensional images may possess some three dimensional capabilities because they are generated from three dimensional data. The three dimensional model may be manipulated to facilitate identification of target on the three dimensional model or two dimensional images, and selection of a suitable pathway through the patient's “P's” airways to access tissue located at the target can be made. Once selected, the pathway plan, three dimensional model, and images derived therefrom, can be saved and exported to a navigation system for use during the navigation phase(s). One such planning software is the ILOGIC® planning suite currently sold by Medtronic PLC.
With respect to the navigation phase, a six degrees-of-freedom electromagnetic tracking system 50, e.g., similar to those disclosed in U.S. Pat. Nos. 8,467,589, 6,188,355, and published PCT Application Nos. WO 00/10456 and WO 01/67035, the entire contents of each of which are incorporated herein by reference, or other suitable positioning measuring system, is utilized for performing registration of the images and the pathway for navigation, although other configurations are also contemplated. Tracking system 50 includes a tracking module 52, a plurality of reference sensors 54, and a transmitter mat 56. Tracking system 50 is configured for use with a locatable guide 32 and particularly sensor 44. As described above, locatable guide 32 and sensor 44 are configured for insertion through an EWC 12 into a patient's “P's” airways (either with or without bronchoscope 30) and are selectively lockable relative to one another via a locking mechanism.
Transmitter mat 56 is positioned beneath patient “P.” Transmitter mat 56 generates an electromagnetic field around at least a portion of the patient “P” within which the position of a plurality of reference sensors 54 and the sensor element 44 can be determined with use of a tracking module 52. One or more of reference sensors 54 are attached to the chest of the patient “P.” The six degrees of freedom coordinates of reference sensors 54 are sent to computing device 125 (which includes the appropriate software) where they are used to calculate a patient coordinate frame of reference. Registration, as detailed below, is generally performed to coordinate locations of the three dimensional model and two dimensional images from the planning phase with the patient's “P's” airways as observed through the bronchoscope 30, and allow for the navigation phase to be undertaken with precise knowledge of the location of the sensor 44, even in portions of the airway where the bronchoscope 30 cannot reach. Further details of such a registration technique and their implementation in luminal navigation can be found in U.S. Patent Application Pub. No. 2011/0085720, the entire content of which is incorporated herein by reference, although other suitable techniques are also contemplated.
Registration of the patient's “P's” location on the transmitter mat 56 is performed by moving LG 32 through the airways of the patient's “P.” More specifically, data pertaining to locations of sensor 44, while locatable guide 32 is moving through the airways, is recorded using transmitter mat 56, reference sensors 54, and tracking module 52. A shape resulting from this location data is compared to an interior geometry of passages of the three dimensional model generated in the planning phase, and a location correlation between the shape and the three dimensional model based on the comparison is determined, e.g., utilizing the software on computing device 125. In addition, the software identifies non-tissue space (e.g., air filled cavities) in the three dimensional model. The software aligns, or registers, an image representing a location of sensor 44 with a the three dimensional model and two dimensional images generated from the three dimension model, which are based on the recorded location data and an assumption that locatable guide 32 remains located in non-tissue space in the patient's “P's” airways. Alternatively, a manual registration technique may be employed by navigating the bronchoscope 30 with the sensor 44 to pre-specified locations in the lungs of the patient “P”, and manually correlating the images from the bronchoscope to the model data of the three dimensional model.
Following registration of the patient “P” to the image data and pathway plan, a user interface is displayed in the navigation software which sets for the pathway that the clinician is to follow to reach the target. One such navigation software is the ILOGIC® navigation suite currently sold by Medtronic PLC.
Once EWC 12 has been successfully navigated proximate the target as depicted on the user interface, the locatable guide 32 may be unlocked from EWC 12 and removed, leaving EWC 12 in place as a guide channel for guiding medical instruments including without limitation, optical systems, ultrasound probes, marker placement tools, biopsy tools, ablation tools (i.e., microwave ablation devices), laser probes, cryogenic probes, sensor probes, and aspirating needles to the target.
Having described the components of system 100 depicted in
Turning now to
Different fluoroscopes differ primarily by detector size and source-detector distance. In the fluoroscopic data, pixels are not normalized to a specific scale. Their brightness depends on the gain/exposure used by the fluoroscope and on other objects in the scene which the X-rays traverse between the source and the detector such as the table, the surgical device, etc. In the CT data, each voxel is measured in Hounsfield units. Hounsfield units are measured with respect to the brightness observed of water and air using the following formula:
μ are attenuation coefficients. They range from 0 to infinity and measure how difficult it is for an X-ray beam li to traverse through matter of the three dimensional volume 205. “Thicker” matter has a larger μ. HU scales attenuation coefficients such that air is placed at −1000, water at 0 and thicker matter goes up to infinity.
Using the Beer-Lambert law, each two dimensional pixel pi in the fluoroscope's detector 203 is given by:
pi,j=l0 exp(−∫l
μ(x,y,z) is the attenuation coefficient of the three dimensional volume 205 at position (x,y,z). I0 is the X-ray source 201 energy (higher energy produces brighter image). For simplicity, assume I0=1. Taking the logarithm, each two dimensional pixel pi is represented by:
log(pi,j)=−∫l
This is true for each two dimensional detector pixel pi which provides a linear equation on the three dimensional attenuation coefficients. If many such equations are utilized (to solve for each two dimensional pixel pi of the detector 203) then the attenuation coefficients may be solved and the three dimensional volume data of the three dimensional volume 205 can be constructed.
Fluoroscope to CT—Discretization:
The three dimensional volume 205 can be divided into a discrete grid, with voxels sitting at (xk, yk, zk). Thus, the equation for solving each two dimensional pixel pi can then be written as:
log(pi,j)=−Σkwki,jμ(xk,yk,zk).
The left hand side of the equation above is the observed two dimensional pixel pi of the detector 203 and the right hand side of the equation above is a weighted sum of the attenuation coefficients (that is to be solved) with wki,j which are determined by the known fluoroscopic imaging device position of the X-ray source 201.
In order to be able to solve for the volumetric attenuation coefficient values enough linear equations are needed, that is, enough two dimensional observations of different two dimensional pixels p are required. Standard detectors usually have 1024×1024 pixels, where each pixel is represented by a single equation. Therefore, many two dimensional observations are required to be solved for (many two dimensional observed pixels) to reconstruct a three dimensional volume (to solve for many voxels). That is many two dimensional pixels are required to solve for the many voxels. In order for the weights in the equations to be known, the fluoroscopic imaging device X-ray source 201 configuration (position, orientation, field of view) must be known.
In use, the three dimensional volume 205 is a portion of a patient's body. A fluoroscopic video which is comprised of multiple fluoroscope images (as frames of the video) are taken from many different locations relative to the patient, for example a 180° rotation about the patient, to acquire multiple observations of two dimensional pixels p at different positions relative to the patient. As will be described in greater detail below, the position of the fluoroscopic imaging device relative to the patient at a given time can be determined using multiple techniques, including structure-from-motion analysis of radio-opaque markers placed within the patient (
Turning now to
The fluoroscopic-based three dimensional volume generated by any of the methods described below can be incorporated into system 100 for multiple purposes. For example, the fluoroscopic-based three dimensional volume can be registered with the previously generated three dimensional volumetric data that was utilized for navigation of the medical instrument. The system 100 may utilize the registration between the fluoroscopic-based three dimensional volume and the previously acquired three dimensional volumetric data to update the calculated position of the sensor 44 (
Additionally, users may visually confirm that the placement of the navigated medical instrument is positioned in a desired location relative to a target tissue within a target area. Additionally, the fluoroscopic-based three dimensional volume can be utilized to visualize the target area in three dimensions after a procedure is performed. For example, the fluoroscopic-based three dimensional volume can be utilized to visualize the target area after markers are placed within the target area, after a biopsy is taken, or after a target is treated.
With particular reference to
In step 305, with the radio-opaque markers placed in the target area, the fluoroscopic imaging device is positioned such that all of the radio-opaque markers placed in step 303 are visible. That is, step 305 includes aligning the fluoroscopic imaging device such that it can rotate 30° around the markers with all of the markers visible. In step 307, the fluoroscopic imaging device is used to capture a video of about a 30° rotation of the imaging device 110 about the patient, and thus around the markers (rotation from −15° to +15°). By rotating up to 15° (on each side) from the centered angle, it can be ensured that the markers will remain in the images/frames for the entire rotation video and that the imaging device will not hit the patient or the bed.
In step 309, the two-dimensional position of each radio-opaque marker is tracked throughout the entire video. In step 311, the marker positions are constructed in three dimensional using structure-from-motion techniques and the pose of the fluoroscopic imaging device is obtained for each video frame. Structure-from-motion is a method for reconstructing three dimensional positions of points, along with fluoroscopic imaging device locations (camera poses) by tracking these points in a two dimensional continuous video. In step 311, by introducing markers into the patient, the positions and orientations of those markers can be tracked along a continuous fluoroscope rotation video, their three dimensional position in space can be reconstructed, and the corresponding fluoroscopic imaging device locations can be determined. With the fluoroscopic imaging device locations determined through analysis of the video, the fluoroscopic imaging device locations can be used to solve for the three dimensional data.
In step 313, a local three dimensional volume is constructed. Specifically, a fast iterative reconstruction algorithm (
As described above, the fluoroscopic-based three dimensional volume can be incorporated into system 100 for multiple purposes. For example, the fluoroscopic-based three dimensional volume can be registered with the previously generated three dimensional volumetric data that was utilized for navigation of the medical instrument. The system 100 may utilize the registration between the fluoroscopic-based three dimensional volume and the previously acquired three dimensional volumetric data to update the calculated position of the sensor 44 (
Turning now to
Method 400 begins at step 401 where the extended working channel is navigated to a target area utilizing an electromagnetic navigation system, such as the EMN system 100 (
After the EWC or tool is in position, or after the radio-opaque marker is placed, the fluoroscopic imaging device is positioned such that the navigated tip of the EWC or tool (and/or the placed radio-opaque marker) is visible within the field of view of the fluoroscopic imaging device. That is, step 405 includes aligning the fluoroscopic imaging device such that it can rotate 30° around the marker with the marker visible and/or around the tip of the EWC or tool with the tip of the EWC or tool visible. In step 407, the fluoroscopic imaging device is used to capture a video of about a 30° rotation of the imaging device 110 about the patient, and thus around the marker and/or tip of the EWC or tool (rotation from −15° to +15°). By rotating up to 15° (on each side) from the centered angle, it can be ensured that the marker and/or tip of the EWC or tool will remain in the images/frames for the entire rotation video and that the imaging device will not hit the patient or the bed. Step 407 may include capturing a video of about a 30° rotation around the distal portion of the EWC (and the radio-opaque marker, if placed). If a 30° rotation video is captured, then one angle (in the middle of the range) is enough. That is, two projections with 30° between them are enough to confirm or correct three dimensional relation of tools to soft tissue.
In step 409, the two-dimensional position of the distal portion of the EWC or tool (and/or the radio-opaque marker, if placed) is tracked throughout the entire captured video.
In step 411, virtual fluoroscopic images are created from previously acquired CT data. The previously acquired CT data is typically the CT data used during the planning phase to plan a navigation path to the target. In step 411, the CT data is manipulated to create a computer model of fluoroscopic images of the patient. The location of the target in the virtual fluoroscopic images corresponds to the location of the target identified by the clinician during the planning phase. The virtual fluoroscopic images generated by the system, based off of the previously acquired CT data, depict the field of view that would be captured by a fluoroscopic imaging device. Additionally, each of the virtual fluoroscopic images has a virtual fluoroscopic imaging device pose.
In step 413, each video frame of the fluoroscopic video captured in step 407 is registered to the previously acquired CT data by matching each of the fluoroscopic video frames to the virtual fluoroscopic images. In step 415, the fluoroscopic imaging device pose of each video frame of the captured fluoroscopic video is determined based on the registration of step 413. That is, once a fluoroscopic frame is matched to a virtual fluoroscopic image, the virtual fluoroscopic imaging device pose of the virtual fluoroscopic image can be associated with the corresponding fluoroscopic frame.
In step 417, the origin of the fluoroscopic imaging device pose determined in step 415 is corrected by using the tracked position of the distal portion of the EWC or tool (and/or the radio-opaque marker, if placed), which is used to compensate for movement of the patient, such as movement caused by breathing. In step 419, a local three dimensional volume is constructed. Specifically, a fast iterative reconstruction algorithm (
Method 400 may also include the additional step (step 421) of completing the fluoroscopic video captured in step 407 to include virtual fluoroscopic images that are generated by the system, which are representative of fluoroscopic imaging device poses that are outside the range of fluoroscopic imaging device poses captured in the fluoroscopic video. Specifically, in aspects, the previously generated CT volumetric data of the patient which is used to create a navigation plan may also be utilized by system 100 to generate virtual fluoroscopic images of the patient. The generated virtual fluoroscopic images are fluoroscopic-like images which display a view to the user of what a fluoroscopic image of a patient should look like if captured at a given angle by a fluoroscopic imaging device. In step 421, the virtual fluoroscopic images may be used to fill any gaps in the captured fluoroscopic video (captured in step 407). This may include, for example, replacement of images, such as frames, of the captured video that are skewed or damaged. Additionally, or alternatively, this may include, for example, supplementing the captured fluoroscopic video (captured in step 407) with virtual fluoroscopic images that are representative of fluoroscopic images that are outside the range of angles included in the fluoroscopic video. For example, if the fluoroscopic video included a sweep of about a 30° range about the patient, virtual fluoroscopic images that are outside the 30° range could be incorporated into the video to generate a fluoroscopic video that has a range of greater than 30°.
Method 300 (
Method 500 is a method for constructing a three dimensional volume using either a single radio-opaque marker placed proximate the target or the tip of the extended working channel of the catheter assembly positioned proximate the target, in conjunction with a fluoroscope angle measurement device. Method 500 begins at step 501 where fluoroscope calibration is performed using a fluoroscope calibration jig to compute the canonical fluoroscope projection parameters and geometry. This is done once per fluoroscope device, in a setup phase by a technician. The calibration jig is used to determine, in an automated process, both the projection parameters of the fluoroscope (field of view angle) as well as the geometry of the C-arm: position relative to rotation axis. These parameters are sometimes given in technical drawings per fluoroscope device, but may also be found using our calibration jig. In step 503, the previously acquired CT volumetric data is imported into the system along with the previously generated navigation plan.
In step 505, the extended working channel is navigated to a target area utilizing an electromagnetic navigation technique using a system such as the EMN system 100 (
After the EWC or tool is in position, or after the radio-opaque marker is placed, method 500 proceeds to step 509 which includes aligning the fluoroscopic imaging device such that it can rotate 30° around the marker with the marker visible and/or the tip of the EWC or the tool such that the tip of the EWC or tool is visible. In step 511, the fluoroscopic imaging device is used to capture a video of about a 30° rotation of the imaging device 110 about the patient, and thus around the marker or tip of the EWC or tool (rotation from −15° to +15°). By rotating up to 15° (on each side) from the centered angle, it can be ensured that the marker or tip of the EWC or tool will remain in the images/frames for the entire rotation video and that the imaging device will not hit the patient or the bed. If a 30° rotation video is captured, then one angle (in the middle of the range) is enough. That is, two projections with 30° between them are enough to confirm or correct three dimensional relation of tools to soft tissue. In step 513, the two-dimensional position and orientation of the distal end of the EWC (and/or the radio-opaque marker, if placed) is tracked throughout the entire captured video.
In step 515, the calibration data from step 501 is utilized in combination with measurements from an external angle measurement device to compute the fluoroscopic imaging device location (origin-less) in world coordinates. Specifically, the calibration jig is used, as described above, to automatically find the projection parameters and the C-arm geometry of the specific fluoroscope device by using optimization methods. Once these parameters are known, the angle taken from the angle measurement device, that is, the angle of the detector of the fluoroscope, determines a unique three dimensional pose for the detector. It cannot be located in any other place in space other than the once explained by the given angle and the setup parameters.
In step 517, a local three dimensional volume is constructed. Specifically, a fast iterative reconstruction algorithm (
CT reconstruction methods will now be described. CT reconstruction methods can be divided into Analytic methods (Radon, FDK . . . ) and Algebraic methods (ART, SART . . . ). Analytic methods assume a very specific configuration, such as a full 180° rotation, and reconstruct the CT volume in a single iteration (using some exact formulas). Algebraic methods are more flexible but slower and treat the problem as a large equation system which is solved iteratively (using some gradient descent method). All methods use projection (three dimensional to two dimensional) and back-projection (two dimensional to three dimensional).
With respect to projection, since each detector pixel is essentially a weighted sum of three dimensional voxels along a ray, the detector image can be viewed as a two dimensional projection of the three dimensional volume from some fluoroscopic imaging device location. If the three dimensional volume data is already available, then the fluoroscope images can be reproduced by projecting it from the known fluoroscopic imaging device locations. A three dimensional reconstruction is considered good if its two dimensional projections resemble the observed fluoroscope images it was created from.
With respect to back-projection, at each voxel, the back-projection determines which rays traversed through a particular voxel and sums them together. In order to make this determination, the fluoroscopic imaging device locations must be known. If the back-projection operator were applied to the fluoroscope images of the captured video, then a three dimensional volume could be constructed, but the constructed three dimensional volume would be very blurry and inaccurate because while the true center voxel is summed many times, many irrelevant voxels surrounding the true center voxel are also summed many times. After reconstructing a three dimensional volume using one method or another, the quality of the reconstruction is evaluated by taking the reconstructed three dimensional data, projecting it into two dimensional frames (which may be virtual-fluoroscopic frames) and comparing those two dimensional frames to the original fluoroscopic two dimensional frames from which the three dimensional data was created. A goal of the volume reconstruction algorithm is to find three dimensional data which explains the two dimensional observations, such that if the three dimensional data were to be projected back into two dimensional frames, those frames would look like the original, real, fluoroscopic frames. When the product is blurry, it means that the projections are blurry and do not match the real images. In order to address this issue, the method provides for a ramp-filter and correction iteration.
Turning now to
Method 600 begins in step 601 where the equation begins with an initial volume V (can just be zeros). In step 603, the volume V is projected from known fluoroscopic imaging device locations into images Qt. For example, projection may be done, not of the fluoroscopic images, but of the fluoroscopic images filtered by a ramp function. The iterations with residuals, described below, may undergo a ramp-filter before back-projections. In step 605, the residuals Ri=Pi−Qi are computed where Pi are the observed projections from the captured fluoroscope video. In step 606, Ri is convolved with the ramp-filter as in Radon. In step 607, it is determined whether Ri is below a predetermined threshold. If Ri is below the predetermined threshold (yes in step 607), then method 600 is complete. If Ri is not below the predetermined threshold (no in step 607), then method 600 proceeds to step 609. In step 609, Ri is back-projected into E, the correction volume. In step 611, volume V is set to V+E (V=V+E) and method 600 reverts to step 603 where the volume V (now V+E) is projected from known fluoroscopic imaging device locations into images Qi.
Turning now to
From the foregoing and with reference to the various figure drawings, those skilled in the art will appreciate that certain modifications can also be made to the present disclosure without departing from the scope of the same. For example, although the systems and methods are described as usable with an EMN system for navigation through a luminal network such as the lungs, the systems and methods described herein may be utilized with systems that utilize other navigation and treatment devices such as percutaneous devices. Additionally, although the above-described system and method is described as used within a patient's luminal network, it is appreciated that the above-described systems and methods may be utilized in other target regions such as the liver. Further, the above-described systems and methods are also usable for transthoracic needle aspiration procedures.
Detailed embodiments of the present disclosure are disclosed herein. However, the disclosed embodiments are merely examples of the disclosure, which may be embodied in various forms and aspects. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for teaching one skilled in the art to variously employ the present disclosure in virtually any appropriately detailed structure.
As can be appreciated a medical instrument such as a biopsy tool or an energy device, such as a microwave ablation catheter, that is positionable through one or more branched luminal networks of a patient to treat tissue may prove useful in the surgical arena and the present disclosure is directed to systems and methods that are usable with such instruments and tools. Access to luminal networks may be percutaneous or through natural orifice using navigation techniques. Additionally, navigation through a luminal network may be accomplished using image-guidance. These image-guidance systems may be separate or integrated with the energy device or a separate access tool and may include MRI, CT, fluoroscopy, ultrasound, electrical impedance tomography, optical, and/or device tracking systems. Methodologies for locating the access tool include EM, IR, echolocation, optical, and others. Tracking systems may be integrated to an imaging device, where tracking is done in virtual space or fused with preoperative or live images. In some cases the treatment target may be directly accessed from within the lumen, such as for the treatment of the endobronchial wall for COPD, Asthma, lung cancer, etc. In other cases, the energy device and/or an additional access tool may be required to pierce the lumen and extend into other tissues to reach the target, such as for the treatment of disease within the parenchyma. Final localization and confirmation of energy device or tool placement may be performed with imaging and/or navigational guidance using a standard fluoroscopic imaging device incorporated with methods and systems described above.
While several embodiments of the disclosure have been shown in the drawings, it is not intended that the disclosure be limited thereto, as it is intended that the disclosure be as broad in scope as the art will allow and that the specification be read likewise. Therefore, the above description should not be construed as limiting, but merely as exemplifications of particular embodiments. Those skilled in the art will envision other modifications within the scope and spirit of the claims appended hereto.
This application claims priority to U.S. patent application Ser. No. 15/224,812, filed Aug. 1, 2016, now allowed, which claims priority to U.S. Provisional Application Ser. No. 62/201,750, filed on Aug. 6, 2015, the entire contents of which are incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
5057494 | Sheffield | Oct 1991 | A |
5321113 | Cooper et al. | Jun 1994 | A |
5852646 | Klotz et al. | Dec 1998 | A |
5930329 | Navab | Jul 1999 | A |
5963612 | Navab | Oct 1999 | A |
5963613 | Navab | Oct 1999 | A |
6003517 | Sheffield et al. | Dec 1999 | A |
6038282 | Wiesent et al. | Mar 2000 | A |
6049582 | Navab | Apr 2000 | A |
6050724 | Schmitz et al. | Apr 2000 | A |
6055449 | Navab | Apr 2000 | A |
6081577 | Webber | Jun 2000 | A |
6120180 | Graumann | Sep 2000 | A |
6236704 | Navab et al. | May 2001 | B1 |
6317621 | Graumann et al. | Nov 2001 | B1 |
6351513 | Bani-Hashemi et al. | Feb 2002 | B1 |
6373916 | Inoue et al. | Apr 2002 | B1 |
6389104 | Bani-Hashemi et al. | May 2002 | B1 |
6404843 | Vaillant | Jun 2002 | B1 |
6424731 | Launay et al. | Jul 2002 | B1 |
6473634 | Bami | Oct 2002 | B1 |
6484049 | Seeley et al. | Nov 2002 | B1 |
6485422 | Mikus et al. | Nov 2002 | B1 |
6490475 | Seeley et al. | Dec 2002 | B1 |
6491430 | Seissler | Dec 2002 | B1 |
6546068 | Shimura | Apr 2003 | B1 |
6546279 | Bova et al. | Apr 2003 | B1 |
6549607 | Webber | Apr 2003 | B1 |
6697664 | Kienzle et al. | Feb 2004 | B2 |
6707878 | Claus et al. | Mar 2004 | B2 |
6714810 | Grzeszczuk et al. | Mar 2004 | B2 |
6731283 | Navab | May 2004 | B1 |
6731970 | Schlossbauer et al. | May 2004 | B2 |
6768784 | Green et al. | Jul 2004 | B1 |
6782287 | Grzeszczuk et al. | Aug 2004 | B2 |
6785356 | Grass et al. | Aug 2004 | B2 |
6785571 | Glossop | Aug 2004 | B2 |
6801597 | Webber | Oct 2004 | B2 |
6823207 | Jensen et al. | Nov 2004 | B1 |
6856826 | Seeley et al. | Feb 2005 | B2 |
6856827 | Seeley et al. | Feb 2005 | B2 |
6865253 | Blumhofer et al. | Mar 2005 | B2 |
6898263 | Avinash et al. | May 2005 | B2 |
6944260 | Hsieh et al. | Sep 2005 | B2 |
6956927 | Sukeyasu et al. | Oct 2005 | B2 |
7010080 | Mitschke et al. | Mar 2006 | B2 |
7010152 | Bojer et al. | Mar 2006 | B2 |
7035371 | Boese et al. | Apr 2006 | B2 |
7106825 | Gregerson et al. | Sep 2006 | B2 |
7117027 | Zheng et al. | Oct 2006 | B2 |
7129946 | Ditt et al. | Oct 2006 | B2 |
7130676 | Barrick et al. | Oct 2006 | B2 |
7165362 | Jobs et al. | Jan 2007 | B2 |
7251522 | Essenreiter et al. | Jul 2007 | B2 |
7327872 | Vaillant et al. | Feb 2008 | B2 |
7343195 | Strommer et al. | Mar 2008 | B2 |
7369641 | Tsubaki et al. | May 2008 | B2 |
7440538 | Tsujii | Oct 2008 | B2 |
7467007 | Lothert | Dec 2008 | B2 |
7474913 | Durlak | Jan 2009 | B2 |
7502503 | Bojer et al. | Mar 2009 | B2 |
7505549 | Ohishi et al. | Mar 2009 | B2 |
7508388 | Barfuss et al. | Mar 2009 | B2 |
7551759 | Hristov et al. | Jun 2009 | B2 |
7603155 | Jensen et al. | Oct 2009 | B2 |
7620223 | Xu et al. | Nov 2009 | B2 |
7639866 | Pomero et al. | Dec 2009 | B2 |
7664542 | Boese et al. | Feb 2010 | B2 |
7689019 | Boese et al. | Mar 2010 | B2 |
7689042 | Brunner et al. | Mar 2010 | B2 |
7693263 | Bouvier et al. | Apr 2010 | B2 |
7711082 | Fujimoto et al. | May 2010 | B2 |
7711083 | Heigl et al. | May 2010 | B2 |
7711409 | Keppel et al. | May 2010 | B2 |
7720520 | P et al. | May 2010 | B2 |
7725165 | Chen et al. | May 2010 | B2 |
7734329 | Boese et al. | Jun 2010 | B2 |
7742557 | Brunner et al. | Jun 2010 | B2 |
7761135 | Pfister et al. | Jul 2010 | B2 |
7778685 | Evron et al. | Aug 2010 | B2 |
7787932 | Vilsmeier et al. | Aug 2010 | B2 |
7804991 | Abovitz et al. | Sep 2010 | B2 |
7831096 | Williamson et al. | Nov 2010 | B2 |
7835779 | Anderson et al. | Nov 2010 | B2 |
7853061 | Gorges et al. | Dec 2010 | B2 |
7877132 | Rongen et al. | Jan 2011 | B2 |
7899226 | Pescatore et al. | Mar 2011 | B2 |
7907989 | Borgert et al. | Mar 2011 | B2 |
7912180 | Zou et al. | Mar 2011 | B2 |
7912262 | Timmer et al. | Mar 2011 | B2 |
7916918 | Suri et al. | Mar 2011 | B2 |
7949088 | Nishide et al. | May 2011 | B2 |
7991450 | Virtue et al. | Aug 2011 | B2 |
8000436 | Seppi et al. | Aug 2011 | B2 |
8043003 | Vogt et al. | Oct 2011 | B2 |
8045780 | Boese et al. | Oct 2011 | B2 |
8050739 | Eck et al. | Nov 2011 | B2 |
8090168 | Washburn et al. | Jan 2012 | B2 |
8111894 | Haar | Feb 2012 | B2 |
8111895 | Spahn | Feb 2012 | B2 |
8126111 | Uhde et al. | Feb 2012 | B2 |
8126241 | Zarkh et al. | Feb 2012 | B2 |
8150131 | Harer et al. | Apr 2012 | B2 |
8180132 | Gorges et al. | May 2012 | B2 |
8195271 | Rahn | Jun 2012 | B2 |
8200316 | Keppel et al. | Jun 2012 | B2 |
8208708 | Homan et al. | Jun 2012 | B2 |
8229061 | Hanke et al. | Jul 2012 | B2 |
8248413 | Gattani et al. | Aug 2012 | B2 |
8270691 | Xu et al. | Sep 2012 | B2 |
8271068 | Khamene et al. | Sep 2012 | B2 |
8275448 | Camus et al. | Sep 2012 | B2 |
8306303 | Bruder et al. | Nov 2012 | B2 |
8311617 | Keppel et al. | Nov 2012 | B2 |
8320992 | Frenkel et al. | Nov 2012 | B2 |
8335359 | Fidrich et al. | Dec 2012 | B2 |
8340379 | Razzaque et al. | Dec 2012 | B2 |
8345817 | Fuchs et al. | Jan 2013 | B2 |
8374416 | Gagesch et al. | Feb 2013 | B2 |
8374678 | Graumann | Feb 2013 | B2 |
8423117 | Pichon et al. | Apr 2013 | B2 |
8442618 | Strommer et al. | May 2013 | B2 |
8482606 | Razzaque et al. | Jul 2013 | B2 |
8515527 | Vaillant et al. | Aug 2013 | B2 |
8526688 | Groszmann et al. | Sep 2013 | B2 |
8526700 | Isaacs | Sep 2013 | B2 |
8532258 | Bulitta et al. | Sep 2013 | B2 |
8532259 | Shedlock et al. | Sep 2013 | B2 |
8548567 | Maschke et al. | Oct 2013 | B2 |
8625869 | Harder et al. | Jan 2014 | B2 |
8666137 | Nielsen et al. | Mar 2014 | B2 |
8670603 | Tolkowsky et al. | Mar 2014 | B2 |
8675996 | Liao et al. | Mar 2014 | B2 |
8693622 | Graumann et al. | Apr 2014 | B2 |
8693756 | Tolkowsky et al. | Apr 2014 | B2 |
8694075 | Groszmann et al. | Apr 2014 | B2 |
8706184 | Mohr et al. | Apr 2014 | B2 |
8706186 | Fichtinger et al. | Apr 2014 | B2 |
8712129 | Strommer et al. | Apr 2014 | B2 |
8718346 | Isaacs et al. | May 2014 | B2 |
8750582 | Boese et al. | Jun 2014 | B2 |
8755587 | Bender et al. | Jun 2014 | B2 |
8781064 | Fuchs et al. | Jul 2014 | B2 |
8792704 | Isaacs | Jul 2014 | B2 |
8798339 | Mielekamp et al. | Aug 2014 | B2 |
8827934 | Chopra et al. | Sep 2014 | B2 |
8831310 | Razzaque et al. | Sep 2014 | B2 |
8855748 | Keppel et al. | Oct 2014 | B2 |
9001121 | Finlayson et al. | Apr 2015 | B2 |
9001962 | Funk | Apr 2015 | B2 |
9008367 | Tolkowsky et al. | Apr 2015 | B2 |
9031188 | Belcher et al. | May 2015 | B2 |
9036777 | Dhishi et al. | May 2015 | B2 |
9042624 | Dennerlein | May 2015 | B2 |
9044190 | Rubner et al. | Jun 2015 | B2 |
9087404 | Hansis et al. | Jul 2015 | B2 |
9095252 | Popovic | Aug 2015 | B2 |
9104902 | Xu et al. | Aug 2015 | B2 |
9111175 | Strommer et al. | Aug 2015 | B2 |
9135706 | Zagorchev et al. | Sep 2015 | B2 |
9171365 | Mareachen et al. | Oct 2015 | B2 |
9179878 | Jeon | Nov 2015 | B2 |
9216065 | Cohen et al. | Dec 2015 | B2 |
9232924 | Liu et al. | Jan 2016 | B2 |
9262830 | Bakker et al. | Feb 2016 | B2 |
9265468 | Rai et al. | Feb 2016 | B2 |
9277893 | Tsukagoshi et al. | Mar 2016 | B2 |
9280837 | Grass et al. | Mar 2016 | B2 |
9282944 | Fallavollita et al. | Mar 2016 | B2 |
9375268 | Long | Jun 2016 | B2 |
9401047 | Bogoni et al. | Jul 2016 | B2 |
9406134 | Klingenbeck-Regn et al. | Aug 2016 | B2 |
9433390 | Nathaniel et al. | Sep 2016 | B2 |
9445772 | Callaghan et al. | Sep 2016 | B2 |
9445776 | Han et al. | Sep 2016 | B2 |
9466135 | Koehler et al. | Oct 2016 | B2 |
9743896 | Averbuch | Aug 2017 | B2 |
9833167 | Cohen et al. | Dec 2017 | B2 |
9888898 | Imagawa et al. | Feb 2018 | B2 |
9918659 | Chopra et al. | Mar 2018 | B2 |
10004558 | Long | Jun 2018 | B2 |
10127629 | Razzaque et al. | Nov 2018 | B2 |
10130316 | Funabasama et al. | Nov 2018 | B2 |
10194897 | Cedro et al. | Feb 2019 | B2 |
10373719 | Soper et al. | Aug 2019 | B2 |
10376178 | Chopra | Aug 2019 | B2 |
10405753 | Sorger | Sep 2019 | B2 |
10478162 | Barbagli et al. | Nov 2019 | B2 |
10480926 | Froggatt et al. | Nov 2019 | B2 |
10524866 | Srinivasan et al. | Jan 2020 | B2 |
10555788 | Panescu et al. | Feb 2020 | B2 |
10569071 | Harris et al. | Feb 2020 | B2 |
10603106 | Weide et al. | Mar 2020 | B2 |
10610306 | Chopra | Apr 2020 | B2 |
10638953 | Duindam et al. | May 2020 | B2 |
10639114 | Schuh et al. | May 2020 | B2 |
10674970 | Averbuch et al. | Jun 2020 | B2 |
10682070 | Duindam | Jun 2020 | B2 |
10702137 | Deyanov | Jul 2020 | B2 |
10706543 | Donhowe et al. | Jul 2020 | B2 |
10709506 | Coste-Maniere et al. | Jul 2020 | B2 |
10772485 | Schlesinger et al. | Sep 2020 | B2 |
10796432 | Mintz et al. | Oct 2020 | B2 |
10823627 | Sanborn et al. | Nov 2020 | B2 |
10827913 | Ummalaneni et al. | Nov 2020 | B2 |
10835153 | Rafii-Tari et al. | Nov 2020 | B2 |
10885630 | Li et al. | Jan 2021 | B2 |
10896506 | Zhao et al. | Jan 2021 | B2 |
20020147462 | Mair et al. | Oct 2002 | A1 |
20030013972 | Makin | Jan 2003 | A1 |
20030088179 | Seeley | May 2003 | A1 |
20030128801 | Eisenberg et al. | Jul 2003 | A1 |
20040120981 | Nathan | Jun 2004 | A1 |
20060251213 | Bernhardt et al. | Nov 2006 | A1 |
20070015997 | Higgins et al. | Jan 2007 | A1 |
20080045938 | Weide et al. | Feb 2008 | A1 |
20110152676 | Groszmann et al. | Jun 2011 | A1 |
20120155729 | Benson | Jun 2012 | A1 |
20120289825 | Rai | Nov 2012 | A1 |
20130223702 | Holsing | Aug 2013 | A1 |
20130303945 | Blumenkranz et al. | Nov 2013 | A1 |
20140035798 | Kawada et al. | Feb 2014 | A1 |
20140046174 | Ladtkow | Feb 2014 | A1 |
20140340470 | Perez et al. | Nov 2014 | A1 |
20150085979 | Zheng et al. | Mar 2015 | A1 |
20150148690 | Chopra et al. | May 2015 | A1 |
20150227679 | Kamer et al. | Aug 2015 | A1 |
20150265368 | Chopra et al. | Sep 2015 | A1 |
20160005194 | Schretter et al. | Jan 2016 | A1 |
20160157939 | Larkin et al. | Jun 2016 | A1 |
20160183841 | Duindam et al. | Jun 2016 | A1 |
20160192860 | Allenby et al. | Jul 2016 | A1 |
20160206380 | Sparks et al. | Jul 2016 | A1 |
20160287343 | Eichler et al. | Oct 2016 | A1 |
20160287344 | Donhowe et al. | Oct 2016 | A1 |
20170035380 | Barak et al. | Feb 2017 | A1 |
20170112571 | Thiel et al. | Apr 2017 | A1 |
20170112576 | Coste-Maniere et al. | Apr 2017 | A1 |
20170209071 | Zhao et al. | Jul 2017 | A1 |
20170265952 | Donhowe et al. | Sep 2017 | A1 |
20170311844 | Zhao et al. | Nov 2017 | A1 |
20170319165 | Averbuch | Nov 2017 | A1 |
20180078318 | Barbagli et al. | Mar 2018 | A1 |
20180144092 | Flitsch et al. | May 2018 | A1 |
20180153621 | Duindam et al. | Jun 2018 | A1 |
20180235709 | Donhowe et al. | Aug 2018 | A1 |
20180240237 | Donhowe et al. | Aug 2018 | A1 |
20180256262 | Duindam et al. | Sep 2018 | A1 |
20180263706 | Averbuch | Sep 2018 | A1 |
20180279852 | Rafii-Tari et al. | Oct 2018 | A1 |
20180325419 | Zhao et al. | Nov 2018 | A1 |
20190000559 | Berman et al. | Jan 2019 | A1 |
20190000560 | Berman et al. | Jan 2019 | A1 |
20190008413 | Duindam et al. | Jan 2019 | A1 |
20190038365 | Soper et al. | Feb 2019 | A1 |
20190065209 | Mishra et al. | Feb 2019 | A1 |
20190110839 | Rafii-Tari et al. | Apr 2019 | A1 |
20190175062 | Rafii-Tari et al. | Jun 2019 | A1 |
20190175799 | Hsu et al. | Jun 2019 | A1 |
20190183318 | Froggatt et al. | Jun 2019 | A1 |
20190183585 | Rafii-Tari et al. | Jun 2019 | A1 |
20190183587 | Rafii-Tari et al. | Jun 2019 | A1 |
20190192234 | Gadda et al. | Jun 2019 | A1 |
20190209016 | Herzlinger et al. | Jul 2019 | A1 |
20190209043 | Zhao et al. | Jul 2019 | A1 |
20190216548 | Ummalaneni | Jul 2019 | A1 |
20190239723 | Duindam et al. | Aug 2019 | A1 |
20190239831 | Chopra | Aug 2019 | A1 |
20190250050 | Sanborn et al. | Aug 2019 | A1 |
20190254649 | Walters et al. | Aug 2019 | A1 |
20190269470 | Barbagli | Sep 2019 | A1 |
20190269818 | Dhanaraj et al. | Sep 2019 | A1 |
20190269819 | Dhanaraj et al. | Sep 2019 | A1 |
20190272634 | Li et al. | Sep 2019 | A1 |
20190298160 | Ummalaneni et al. | Oct 2019 | A1 |
20190298451 | Wong et al. | Oct 2019 | A1 |
20190320878 | Duindam et al. | Oct 2019 | A1 |
20190320937 | Duindam et al. | Oct 2019 | A1 |
20190336238 | Yu et al. | Nov 2019 | A1 |
20190343424 | Blumenkranz et al. | Nov 2019 | A1 |
20190350659 | Wang et al. | Nov 2019 | A1 |
20190365199 | Zhao et al. | Dec 2019 | A1 |
20190365479 | Rafii-Tari | Dec 2019 | A1 |
20190365486 | Srinivasan et al. | Dec 2019 | A1 |
20190380787 | Ye et al. | Dec 2019 | A1 |
20200000319 | Saadat et al. | Jan 2020 | A1 |
20200000526 | Zhao | Jan 2020 | A1 |
20200008655 | Schlesinger et al. | Jan 2020 | A1 |
20200030044 | Wang et al. | Jan 2020 | A1 |
20200030461 | Sorger | Jan 2020 | A1 |
20200038750 | Kojima | Feb 2020 | A1 |
20200043207 | Lo et al. | Feb 2020 | A1 |
20200046431 | Soper et al. | Feb 2020 | A1 |
20200046436 | Tzeisler et al. | Feb 2020 | A1 |
20200054399 | Duindam et al. | Feb 2020 | A1 |
20200054408 | Schuh et al. | Feb 2020 | A1 |
20200060771 | Lo et al. | Feb 2020 | A1 |
20200069192 | Sanborn et al. | Mar 2020 | A1 |
20200077870 | Dicarlo et al. | Mar 2020 | A1 |
20200078023 | Cedro et al. | Mar 2020 | A1 |
20200078095 | Chopra et al. | Mar 2020 | A1 |
20200078103 | Duindam et al. | Mar 2020 | A1 |
20200085514 | Blumenkranz | Mar 2020 | A1 |
20200109124 | Pomper et al. | Apr 2020 | A1 |
20200129045 | Prisco | Apr 2020 | A1 |
20200129239 | Bianchi et al. | Apr 2020 | A1 |
20200138514 | Blumenkranz et al. | May 2020 | A1 |
20200138515 | Wong | May 2020 | A1 |
20200142013 | Wong | May 2020 | A1 |
20200155116 | Donhowe et al. | May 2020 | A1 |
20200155232 | Wong | May 2020 | A1 |
20200170623 | Averbuch | Jun 2020 | A1 |
20200170720 | Ummalaneni | Jun 2020 | A1 |
20200179058 | Barbagli et al. | Jun 2020 | A1 |
20200188021 | Wong et al. | Jun 2020 | A1 |
20200188038 | Donhowe et al. | Jun 2020 | A1 |
20200205903 | Srinivasan et al. | Jul 2020 | A1 |
20200205904 | Chopra | Jul 2020 | A1 |
20200214664 | Zhao et al. | Jul 2020 | A1 |
20200229679 | Zhao et al. | Jul 2020 | A1 |
20200242767 | Zhao et al. | Jul 2020 | A1 |
20200275860 | Duindam | Sep 2020 | A1 |
20200297442 | Adebar et al. | Sep 2020 | A1 |
20200315554 | Averbuch et al. | Oct 2020 | A1 |
20200330795 | Sawant et al. | Oct 2020 | A1 |
20200352427 | Deyanov | Nov 2020 | A1 |
20200364865 | Donhowe et al. | Nov 2020 | A1 |
20200383750 | Kemp et al. | Dec 2020 | A1 |
20210000524 | Barry et al. | Jan 2021 | A1 |
Number | Date | Country |
---|---|---|
0013237 | Jul 2003 | BR |
0116004 | Jun 2004 | BR |
0307259 | Dec 2004 | BR |
0412298 | Sep 2006 | BR |
112018003862 | Oct 2018 | BR |
101190149 | Jun 2008 | CN |
103065322 | Apr 2013 | CN |
105232152 | Jan 2016 | CN |
106236257 | Dec 2016 | CN |
106447780 | Feb 2017 | CN |
1644519 | Dec 2008 | CZ |
486540 | Sep 2016 | CZ |
2709512 | Aug 2017 | CZ |
2884879 | Jan 2020 | CZ |
1644519 | Dec 2008 | EP |
2141497 | Jan 2010 | EP |
3127485 | Feb 2017 | EP |
3413830 | Sep 2019 | EP |
3478161 | Feb 2020 | EP |
3641686 | Apr 2020 | EP |
3644885 | May 2020 | EP |
3644886 | May 2020 | EP |
3749239 | Dec 2020 | EP |
H11197259 | Jul 1999 | JP |
PA03005028 | Jan 2004 | MX |
PA03000137 | Sep 2004 | MX |
PA03006874 | Sep 2004 | MX |
225663 | Jan 2005 | MX |
226292 | Feb 2005 | MX |
PA03010507 | Jul 2005 | MX |
PA05011725 | May 2006 | MX |
06011286 | Mar 2007 | MX |
246862 | Jun 2007 | MX |
2007006441 | Aug 2007 | MX |
265247 | Mar 2009 | MX |
284569 | Mar 2011 | MX |
2008038283 | Apr 2008 | WO |
2009081297 | Jul 2009 | WO |
2014186715 | Nov 2014 | WO |
2015101948 | Jul 2015 | WO |
2019006258 | Jan 2019 | WO |
Entry |
---|
Examination Report No. 1 issued in Australian Patent application No. 2020210140 dated Mar. 22, 2021, 4 pages. |
Second Office Action issued in Chinese Patent Application No. 201910105949.0 dated Dec. 17, 2021, with English translation. |
Extended European Search Report issued in European Application No. 19202514.6 dated Jul. 8, 2019, 9 pages. |
Extended European Search Report issued in European Application No. 20184865.2 dated Apr. 7, 2021, 9 pages. |
First Office Action issued in Chinese Patent Application No. 201910105949 dated Jul. 12, 2021 with English translation, 19 pages. |
Australian Examination Report issued in corresponding Appl. No. AU 2019200594 dated Aug. 20, 2019 (5 pages). |
Australian Examination Report No. 2 issued in Appl. No. AU 2016210747 dated Oct. 18, 2017 (4 pages). |
Canadian Office Action issued in Appl. No. 2,937,825 dated Mar. 26, 2018 (4 pages). |
CT scan—Wikipedia, the free encyclopedia [retrieved from internet on Mar. 30, 2017] published on Jun. 30, 2015 as per Wayback Machine. |
Examination Report issued in corresponding Appl. No. AU 2018204631 dated Aug. 12, 2019 (3 pages). |
Examination Report issued in corresponding Application No. AU 2018204631 dated May 24, 2019 (4 pages). |
Extended European Search Report from Appl. No. EP 16182953.6-1666 dated Jan. 2, 2017. |
Extended European Search Report issued in corresponding Appl. No. EP 19156045.7 dated Jul. 15, 2019 (9 pages). |
Japanese Office Action issued in corresponding Appl. No. JP 2019-021423, together with English language translation, dated Jan. 8, 2020 (7 pages). |
Office Action issued in Chinese Appl. No. 201610635896.X dated Jul. 23, 2018, together with English language translation (16 pages). |
Yingchao Li, et al., “Distortion Correction and Geometric Calibration for X-Ray Angiography System”, IEEE Transactions on Nuclear Science, vol. 56, No. 3, (2009), pp. 608-619. |
Communication pursuant to Article 94(3) EPC issued in European Patent Application No. 19156045.7 dated Jul. 14, 2022. |
Chinese Office Action issued in Chinese Patent Application No. 201910416164.5 dated Sep. 9, 2022. English translation not available. |
Sami Sebastian Brandt et al., “Structure-From-Motion Without Correspondence From Tomographic Projections by Bayesian Inversion Theory”, pp. 238-248, Feb. 1, 2007. |
Ulrich, Muller et al.,“Correction of C-Arm Projection Matrices by 3D-2D Rigid Registration of CT-Images Using Mutual Information”, pp. 162-170, Jan. 1, 2003. |
Number | Date | Country | |
---|---|---|---|
20200253571 A1 | Aug 2020 | US |
Number | Date | Country | |
---|---|---|---|
62201750 | Aug 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15224812 | Aug 2016 | US |
Child | 16862467 | US |