This disclosure relates to the field of image-guided navigation of medical devices, and particularly to image-guided navigation of percutaneously-inserted tools from a percutaneous entry point to a target (e.g., a tumor) to perform a procedure, for example, localization (e.g., using a dye, guide wire, or fiducials), biopsy, or ablation of the target.
Computed tomography (CT)-guided needle biopsy is a common method for obtaining tissue samples from lung nodules for lung cancer diagnosis. A patient lies on a CT table and receives a local anesthesia injection to numb the needle path. A target lung nodule is located through a pre-operative CT scan, which is used by a clinician to plan the safest needle path to the target nodule. Using intraoperative CT scans to confirm the positions of the target nodule and the needle, the clinician inserts the needle through the skin, advances it towards and into the target nodule, and removes samples of the target nodule. While CT provides high spatial resolution and good tissue contrast, which enables precise and safe placement of biopsy needles, CT does not provide real-time imaging, tracking, and movement perception.
In aspects, this disclosure features methods of performing image-guided medical procedures using a percutaneously-inserted device. In aspects, the percutaneously-inserted medical device may be used for performing localization, biopsy, ablation, or any other suitable image-guided medical procedure. In one general aspect, this disclosure features a method of performing an image-guided medical procedure. The method includes performing a fluoroscopic sweep of at least a portion of a patient's body that includes a target area to obtain fluoroscopic images; determining a fluoroscopic pose for each of the fluoroscopic images; generating a fluoroscopic 3D reconstruction based on the fluoroscopic images and the fluoroscopic poses, and applying a trajectory of a percutaneously-inserted device to the fluoroscopic 3D reconstruction such that the fluoroscopic 3D reconstruction includes a planned trajectory. The method also includes displaying a live fluoroscopic image and overlaying the planned trajectory of the percutaneously-inserted device on the live fluoroscopic image. The method also includes displaying advancement of a percutaneously-inserted device a first distance in the live fluoroscopic image and determining at two angles that the advancement of the percutaneously-inserted device the first distance is following the planned trajectory. The method also includes displaying advancement of the percutaneously-inserted device a second distance in response to determining at two angles that the advancement of the percutaneously-inserted device the first distance is following the planned trajectory, and determining at two angles that the advancement of the percutaneously-inserted device the second distance is following the planned trajectory.
In aspects, implementations of this disclosure may include one or more of the following features. The method may include applying a mark indicating a critical structure to avoid to the fluoroscopic 3D reconstruction, and overlaying the mark indicating the critical structure to avoid on the live fluoroscopic image. The planned trajectory may include a target, an insertion point, and a line between the insertion point and the target. Performing the fluoroscopic sweep may include performing the fluoroscopic sweep to obtain fluoroscopic images of a radiopaque object. The method may include determining the position of the radiopaque object relative to the target and determining the position of an insertion point based on the position of the radiopaque object relative to the target. The fluoroscopic sweep may be performed after insertion of the percutaneously-inserted device at the insertion point.
The method may include determining that the percutaneously-inserted device is located at the target and performing one or more additional fluoroscopic sweeps to obtain one or more additional fluoroscopic images in response to determining that the percutaneously-inserted device is not located at the target.
The method may include receiving preoperative computed tomography (CT) images including markings indicating the planned trajectory. The method may include registering the fluoroscopic 3D reconstruction to the preoperative CT images. Applying a planned trajectory of the percutaneously-inserted device to the fluoroscopic 3D reconstruction may include transferring the markings on the preoperative CT images to the fluoroscopic 3D reconstruction based on the registering.
Registering may include determining one or more anatomical features that may be in both the fluoroscopic 3D reconstruction and the preoperative CT images. Registering may include aligning the fluoroscopic 3D reconstruction and the preoperative CT images based on the determined one or more anatomical features. The one or more anatomical features may include the target, a lesion, a tumor, or a rib.
In another general aspect, this disclosure features a method of performing an image-guided medical procedure. The method includes performing a first fluoroscopic sweep of at least a portion of a patient's body that includes a target area to obtain first fluoroscopic images, determining a pose for each of the first fluoroscopic images, and generating and displaying a first fluoroscopic 3D reconstruction based on the first fluoroscopic images and the poses. The method also includes marking a planned trajectory in the first fluoroscopic 3D reconstruction, displaying a live fluoroscopic image, and overlaying the planned trajectory on the live fluoroscopic image. The method also includes adjusting the live fluoroscopic image such that the trajectory appears as a point and advancing the percutaneously-inserted device along the planned trajectory based on the adjusted live fluoroscopic image.
In aspects, implementations of this disclosure may include one or more of the following features. The method may include displaying the length of the planned trajectory. The percutaneously-inserted device may be advanced based on the length of the planned trajectory and length markers on the percutaneously-inserted device. The planned trajectory may include an insertion point, a target, and a line between the insertion point and the target. The method may include adjusting the live fluoroscopic image to a second angle to verify the depth of the percutaneously-inserted device in the patient's body. The method may include adjusting the live fluoroscopic image to a third angle to verify the direction of the percutaneously-inserted device.
In another general aspect, this disclosure features a method of performing an image-guided medical procedure. The method includes performing a first fluoroscopic sweep of at least a portion of a patient's body that includes a target area to obtain first fluoroscopic images, determining a first fluoroscopic pose for each of the first fluoroscopic images, generating and displaying a first fluoroscopic 3D reconstruction based on the first fluoroscopic images and the first fluoroscopic poses. The method also includes applying a trajectory of a percutaneously-inserted device to the first fluoroscopic 3D reconstruction such that the first fluoroscopic 3D reconstruction includes a planned trajectory, displaying a live fluoroscopic image, and overlaying the planned trajectory of the percutaneously-inserted device on the live fluoroscopic image. The method also includes performing a second fluoroscopic sweep to obtain second fluoroscopic images of an insertion point, determining a second pose for each of the second fluoroscopic images, and generating and displaying a second fluoroscopic 3D reconstruction based on the second fluoroscopic images and the second poses. The method also includes displaying insertion of a percutaneously-inserted device in the live fluoroscopic image.
In aspects, implementations of this disclosure may include one or more of the following features. The method may include registering the second fluoroscopic 3D reconstruction to the first fluoroscopic 3D reconstruction and transferring the planned trajectory applied to the first fluoroscopic 3D reconstruction to the second fluoroscopic 3D reconstruction based on the registering. The method may include displaying the live fluoroscopic image in at least one fluoroscopic angle after advancement of the inserted percutaneously-inserted device a first distance, and determining that the advancement of the inserted percutaneously-inserted device the first distance is following the planned trajectory based on the live fluoroscopic image in the at least one fluoroscopic angle.
The at least one fluoroscopic angle may include two different fluoroscopic angles at which a direction of the inserted percutaneously-inserted device can be determined from the live fluoroscopic image.
The method may include displaying the live fluoroscopic image in at least one fluoroscopic angle after advancement of the inserted percutaneously-inserted device a second distance and determining that the advancement of the inserted percutaneously-inserted device the second distance is following the planned trajectory based on the live fluoroscopic image in the at least one fluoroscopic angle. The method may include performing a third fluoroscopic sweep to obtain third fluoroscopic images after advancement of the inserted percutaneously-inserted device the second distance, determining a third pose for each of the third fluoroscopic images; and generating and displaying a third fluoroscopic 3D reconstruction based on the third fluoroscopic images and the third poses.
Various exemplary embodiments are illustrated in the accompanying figures with the intent that these examples are not restrictive. It will be appreciated that for simplicity and clarity of the illustration, elements shown in the figures referenced below are not necessarily drawn to scale. Also, where considered appropriate, reference numerals may be repeated among the figures to indicate like, corresponding or analogous elements. The figures are listed below.
This disclosure relates to systems and methods for image-guided, e.g., fluoroscopy- and/or (electromagnetic) EM-guided, navigation of percutaneously-inserted devices, e.g., biopsy tools, treatment tools (e.g., ablation tools), or needles. An advantage of a fluoroscope is that a fluoroscope can provide a live view of the medical procedure, e.g., a biopsy procedure. However, a fluoroscope may provide images having a lower quality than computed tomography (CT) images, which provide good visualization via axial and coronal views. One way to solve this issue is to fuse or merge the live fluoroscopic images with the CT images by, for example, overlaying portions of the CT images on the fluoroscopic images or vice versa. For example, the fusion may show the CT images around the tool in a fluoroscopic 3D reconstruction because CT images show the pleura better than the fluoroscopic 3D reconstruction. Alternatively, the fusion may involve displaying the tool in the fluoroscopic 3D reconstruction in the CT images. In other aspects, once the fluoroscopic 3D reconstruction is aligned with the CT images, the target, e.g., lesion, from either the fluoroscopic 3D reconstruction or the CT images may be displayed in the fused or merged images.
Systems and methods according to aspects of the disclosure involve centering a fluoroscope on an estimated target location, performing a fluoroscopic sweep to obtain fluoroscopic images, e.g., coronal slices, and determining poses of the fluoroscopic images. A fluoroscopic 3D reconstruction may then be generated based on the fluoroscopic images and the poses and presented to a clinician. The fluoroscopic 3D reconstruction may be presented to the clinician as a series of fluoroscopic two-dimensional (2D) reconstruction images corresponding to slices of the fluoroscopic reconstruction. The clinician may plan a trajectory by placing one or more marks on the fluoroscopic 3D reconstruction or on the fluoroscopic 2D reconstruction images including an entry point mark (i.e., the location where the percutaneously-inserted device should be placed), a target mark, and/or a trajectory mark or line indicating a trajectory between the entry point and the target. The entry point, the target, and/or the trajectory may then be displayed in a live 2D fluoroscopy view based on the trajectory planned by the clinician. Alternatively, a navigation system may automatically plan a trajectory including the entry point based on the identification of the target. In aspects, the target, the entry point, the trajectory, and/or a structure to avoid may be selected or identified by the clinician, or automatically identified or segmented by the navigation system. Then, the target, the entry point, the trajectory, and/or a structure to avoid may be marked on a fluoroscopic 3D reconstruction and overlaid on the live 2D fluoroscopy image or in the EM navigation view.
As the clinician inserts and advances a percutaneously-inserted device towards the target while viewing the live 2D fluoroscopy image and the planned trajectory overlaid on the live 2D fluoroscopy image, additional fluoroscopic sweeps are performed to generate additional fluoroscopic 3D reconstructions. In the first sweep, one or more radiopaque objects (e.g., needles) may optionally be placed or laid on the patient such that the one or more radiopaque objects appear in the corresponding first fluoroscopic 3D reconstruction to understand the location of the entry point on the patient's body. Then, a needle is slightly inserted at the entry point based on the first fluoroscopic 3D reconstruction. In second or additional sweeps, after the needle is already introduced into the patient's body, the clinician checks the needle direction versus the trajectory to the target in the respective second or additional fluoroscopic 3D reconstructions that are displayed to the clinician. The clinician adjusts the needle direction each time and advances the needle further. In some aspects, if the clinician confirms that the direction of the needle is good after insertion, the distance to the target is calculated and displayed to the clinician so she can advance the needle the displayed distance in one action using length markers on the needle, and reach the target after the second sweep. In other aspects, the clinician confirms that the direction of the needle is good at different locations along the trajectory towards the target based on respective additional sweeps.
In aspects, a marked fluoroscopic 3D reconstruction is used in live two-dimensional (2D) fluoroscopy with an overlay. First, a planned trajectory, which is on a fluoroscopic 3D reconstruction and may include a target, an insertion point, and a line in between, is overlaid on the live 2D fluoroscopy view or image. Second, the clinician makes sure in two angles that the needle is inserted to follow the trajectory. For example, the needle is advanced in baby steps making sure in two angles that the direction of the needle is good, and then the needle is advanced in larger steps until the second angle verification.
Alternatively, the fluoroscope is adjusted to make the overlaid trajectory appear as a point (e.g., a bullseye view) in the live 2D fluoroscopy image, and then the needle is inserted to follow the trajectory while the clinician checks the direction of the needle using the live 2D fluoroscopy image at a second angle different from the initial or first angle of the live 2D fluoroscopy image. The distance from the needle tip to the target may be determined from trajectory markings on the first fluoroscopic 3D reconstruction based on the first sweep. Then, the clinician can insert and/or advance the needle the determined distance based on length markers disposed on the needle. Optionally, the clinician can check the depth of the needle by using the live 2D fluoroscopy image at a third angle. In aspects, the methods described above can be combined in various combinations, e.g., additional sweeps while inserting and/or advancing the needle that is guided by respective fluoroscopic 3D reconstructions overlaid on the live 2D fluoroscopy view or image.
In each fluoroscopic sweep, the percutaneously-inserted device may be visible in the live 2D fluoroscopic images. As such, the percutaneously-inserted device may be either marked by the clinician or segmented automatically. Also, the actual trajectory of the percutaneously-inserted device may be shown in the live fluoroscopy view. The actual trajectory of the percutaneously-inserted device may be shown together with the planned trajectory in the live fluoroscopy view. In some aspects, the clinician may start the tool at the bullseye (i.e., where the crosshairs of a needle insertion point are centered in a circle marking the target) and follow the trajectory in projections of the fluoroscopic 3D reconstruction that are convenient to the clinician.
In aspects, an electromagnetic (EM) sensor is disposed on the percutaneously-inserted device and EM navigation is used to guide the percutaneously-inserted device to the target. In this aspect, the fluoroscopic 3D reconstruction including marks indicating an insertion point, a target, and a path to the target is registered to the navigation 3D views. Next, the marks in the fluoroscopic 3D reconstruction are transferred to the navigation 3D views based on the registering. Then, the clinician uses the navigation 3D views to guide the insertion and advancement of the percutaneously-inserted device.
In aspects, the methods involve marking an entry point, a target, a trajectory, and/or a structure to avoid on preoperative CT images and registering the fluoroscopic 3D reconstruction to the preoperative CT images so that the markings may be transferred to and displayed in the fluoroscopic 3D reconstruction. A clinician and/or a software application may mark an entry point, a target, a trajectory, and/or a structure to avoid on the preoperative CT images. The clinician may center the live fluoroscopic view on an estimated target location and perform a fluoroscopic sweep. Then, a fluoroscopic 3D reconstruction is generated from the fluoroscopic sweep and registered to the preoperative CT images. The registration of the fluoroscopic 3D reconstruction to the preoperative CT images may be based either on intra-modality registration methods (e.g., the mutual information method) or on markers (e.g., the target or the tool may serve as a marker and may be identified by the clinician or automatically segmented or recognized).
The live fluoroscopy view may show where a percutaneously-inserted device should be inserted and the tool's trajectory towards the target. One or more additional sweeps are performed to capture additional fluoroscopic images, which are used to generate a fluoroscopic 3D reconstruction, in which the position of the tool is either marked by the clinician or segmented automatically. The clinician follows the planned trajectory in projections convenient to the clinician. The actual trajectory of the tool is then shown in the preoperative CT images and/or in the fluoroscopic 3D reconstruction.
In aspects, the navigation system may determine where to percutaneously insert the tool and the trajectory to the target based on preoperative CT images, on a fluoroscopic 3D reconstruction, and/or on the live fluoroscopic image. Then, the clinician navigates the percutaneously-inserted device to the target in 3D, e.g., using the location of the tool obtained from a location sensor (e.g., EM sensor) of the tool or automatically determining the location of the tool based on the fluoroscopic 3D reconstruction from a grayscale calculation or any other suitable image processing method. Optionally, additional sweeps may be performed. After each fluoroscopic sweep, a new fluoroscopic 3D reconstruction may be generated and registered to the previous fluoroscopic 3D reconstruction and the navigation system may be updated. After navigation of the percutaneously-inserted device to the target, the additional fluoroscopic sweeps may be used to confirm whether the percutaneously-inserted device or the tip of the percutaneously-inserted device has reached the target.
The navigation and image system 100 generally includes an operating table or bed 20 configured to support a patient “P;” a tracking system 50 including a tracking module 52, reference sensors 54 and a transmitter mat 56; a biopsy tool 27 or other percutaneously-inserted device, a location sensor 28 disposed on the biopsy tool 27, and a computing device 125 including software and/or hardware used to facilitate identification of a target, pathway planning to the target, navigation of the biopsy tool 27 to the target, and confirmation of the placement of the biopsy tool 27 relative to the target. The computing device 125 may include a video display for displaying a live fluoroscopic view including a live fluoroscopic image captured by the fluoroscopic imaging device 110, a fluoroscopic 3D reconstruction, fluoroscopic 2D reconstruction images, and/or preoperative CT images.
A fluoroscopic imaging device 110 capable of acquiring fluoroscopic or x-ray images or video of the patient “P” is also included in aspects of the navigation and image system 100. The images, series of images, or video captured may be stored within the fluoroscopic imaging device 110 or transmitted to computing device 125 for storage, processing, and display. Additionally, the fluoroscopic imaging device 110 may move relative to the patient “P” so that images may be acquired from different angles or perspectives relative to the patient “P” to create a fluoroscopic video. Fluoroscopic imaging device 110 may include an angle measurement device 111 which is configured to measure the angle of the fluoroscopic imaging device 110 relative to the patient “P.” Angle measurement device 111 may be an accelerometer. Fluoroscopic imaging device 110 may include a single imaging device or more than one imaging device. In aspects including multiple imaging devices, each imaging device may be a different type of imaging device or the same type.
Computing device 125 may be any suitable computing device including a processor and storage medium, wherein the processor is capable of executing instructions stored on the storage medium. The computing device 125 is operably coupled to some or all of the components of the navigation and image system 100 including tracking system 50. The computing device 125 may include a database configured to store patient data, CT data sets including CT images and volumetric renderings, fluoroscopic data sets including fluoroscopic images and video, navigation plans, and any other such data. Although not explicitly illustrated, the computing device 125 may include inputs, or may otherwise be configured to receive, CT data sets, fluoroscopic images/video and other data described herein. Additionally, computing device 125 includes a display configured to display graphical user interfaces described herein. Computing device 125 may be connected to one or more networks through which one or more databases may be accessed.
With respect to the planning phase, computing device 125 utilizes fluoroscopy images and/or previously acquired CT images to enable the identification of an entry point at which the biopsy tool 27 enters patient P's body percutaneously and a target on either or both of those images (automatically, semi-automatically, or manually), and allows for determining a pathway or trajectory from the entry point to tissue located at and/or around the target. The fluoroscopic and/or CT images may be displayed on a display associated with computing device 125, or in any other suitable fashion.
With respect to the navigation phase according to one aspect, a five degrees-of-freedom or a six degrees-of-freedom electromagnetic tracking system 50 or other suitable positioning measuring system may be utilized for performing navigation of the biopsy tool 27, although other configurations are also contemplated. Tracking system 50 includes a tracking module 52, reference sensors 54, a transmitter mat 56, and the location sensor 28 disposed on the biopsy tool 27. Length markers 29 may also be disposed on the biopsy tool 27 to assist with navigation as described herein. For example, systems of this disclosure may determine a distance from the tip of the biopsy tool 27 to the target based on fluoroscopic images. Then, the biopsy tool 27 may be advanced the determined distance using the length markers 29.
Transmitter mat 56 is positioned beneath patient “P.” Transmitter mat 56 generates an electromagnetic field around at least a portion of the patient “P” within which the position of the reference sensors 54 and the location sensor 28 can be determined with use of a tracking module 52. One or more of reference sensors 54 are attached to the chest of the patient “P.” The six degrees of freedom coordinates of reference sensors 54 are sent to computing device 125 (which may be a computing device storing and executing appropriate software) where they are used to calculate a patient coordinate frame of reference. Registration, as detailed herein, is generally performed to align the fluoroscopic 3D reconstruction with the previously acquired CT images.
Previously acquired CT scans, may not provide a basis sufficient for accurate guiding of the biopsy needle 27 or other percutaneously-inserted device to a target during a biopsy or other medical procedure. The inaccuracy may be caused by CT-to-Body divergence (deformation of the patient's lungs during the procedure relative to the lungs at the time of the acquisition of the previously acquired CT data). Thus, another imaging modality is needed to visualize targets and confirm placement of the biopsy needle 27 during a biopsy procedure. For this purpose, the system described herein processes image data captured by the imaging device 110 and/or CT image data as is described herein. This fluoroscopic image data may be utilized to identify targets and be incorporated into, and used to update, the data from the CT scans with, among other things, the actual trajectory of the biopsy needle 27.
Reference is now made to
Memory 102 may store an application 81 and image data 114. Application 81 may include instructions executable by processor 104 for, among other functions, executing the methods of
Memory 102 may include any non-transitory computer-readable storage media for storing data and/or software including instructions that are executable by processor 104 and which control the operation of computing device 125 and in some embodiments, may also control the operation of imaging device 110. Imaging device 110 may be used to capture a sequence of fluoroscopic images based on which the fluoroscopic 3D reconstruction is generated. In an aspect, storage device or memory 102 may include one or more storage devices such as solid-state storage devices such as flash memory chips. Alternatively, or in addition to the one or more solid-state storage devices, memory 102 may include one or more mass storage devices connected to the processor 104 through a mass storage controller (not shown) and a communications bus (not shown).
Although the description of computer-readable media contained herein refers to a solid-state storage, it should be appreciated by those skilled in the art that computer-readable storage media can be any available media that can be accessed by the processor 104. That is, computer readable storage media may include non-transitory, volatile, and non-volatile, removable, and/or non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. For example, computer-readable storage media may include RAM, ROM, EPROM, EEPROM, flash memory or other solid-state memory technology, CD-ROM, DVD, Blu-Ray or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which may be used to store the desired information, and which may be accessed by computing device 125, which may be a personal computer or workstation.
Application 81 may, when executed by processor 104, cause display 106 to present user interface 116. User interface 116 may be configured to present to the user the F3DR, two-dimensional fluoroscopic images, images of the 3D imaging and virtual fluoroscopy image, as shown, for example, in the exemplary screen shots of
Network interface 108 may be configured to connect to a network such as a local area network (LAN) consisting of a wired network and/or a wireless network, a wide area network (WAN), a wireless mobile network, a Bluetooth network, and/or the internet. Network interface 108 may be used to connect between computing device 125 and imaging device 110, e.g., a fluoroscope. Network interface 108 may be also used to receive image data 114. Input device 10 may be any device by means of which a user may interact with computing device 125, such as, for example, a mouse, keyboard, foot pedal, touch screen, and/or voice interface. Output module 112 may include any connectivity port or bus, such as, for example, parallel ports, serial ports, universal serial busses (USB), or any other similar connectivity port known to those skilled in the art.
The user instructions 201 may include a further instruction to click the “Generate Fluoro 3D” button 205 to generate a fluoroscopic 3D reconstruction, which is displayed in the planning user interface 210 of
The planning user interface 210 of
Optionally, the user instructions 211 may include another instruction to mark one or more features that the biopsy needle must avoid with a dotted circle, e.g., dotted circle 214. The one or more features may include anatomical features such as one or more bones, one or more vascular structures, one or more nerve structures, or one or more critical structures, which, if punctured or damaged by the biopsy needle, would cause unnecessary bleeding or harm to the patient. As another option, the user instructions 211 may include a further instruction to mark where to puncture the patient with the needle, i.e., the needle insertion or entry point. The needle insertion point may be marked with crosshairs 216. In other aspects, an entry point may be automatically or manually marked based on determined and/or marked positions of the target and the radiopaque object.
Optionally, the user instructions 211 may include a further instruction to draw a trajectory 217 from the marked needle insertion point to the marked target on the fluoroscopic 3D reconstruction 212. In aspects, the user instructions 211 may include a further instruction to click an optional “Generate Trajectory” button 218 to automatically generate or plan a trajectory from the marked needle insertion point to the marked target and draw or otherwise apply the generated trajectory on the fluoroscopic 3D reconstruction 212.
In aspects, the planning user interface 210 may also allow a clinician to mark the angles of the biopsy tool with respect to the operating table and other rigid structures on the fluoroscopic 3D reconstruction 212 to provide guidance on the trajectory. There may be an intermediate instruction of injecting a substance, e.g., a dye, in the pleura or other body structure. Also, the clinician may be instructed to paint the entry point (e.g., bullseye 25 of
The planning user interface 210 includes the “Back” button 206 to return to the initial fluoroscopic sweep user interface 200 illustrated in
The user instructions 221 include an instruction to navigate the biopsy needle along a portion of the planned trajectory 217 starting at the needle entry point 216. The user instructions 221 include another instruction to perform another fluoroscopic sweep. In some aspects, the other fluoroscopic sweep is a narrower fluoroscopic sweep than the initial fluoroscopic sweep. In other aspects, the another and subsequent fluoroscopic sweeps while navigating the biopsy tool towards the target may be replaced by a 2D fluoroscopic snapshot in order to minimize the patient's exposure to radiation. In aspects, the 2D fluoroscopic snapshots could be taken lateral to the planned trajectory. For example, if the planned trajectory is perpendicular to the bed 20, the fluoroscopic snapshots could be taken at the lateral side. The fluoroscopic snapshots may be taken after each time the biopsy tool is moved a short distance.
The user instructions 221 include a further instruction to click the “Auto Mark Needle Tip” button 225 or to manually mark the biopsy needle tip 226 in the live fluoroscopy window 222. When the “Auto Mark Needle Tip” button 225 is clicked, an image recognition algorithm may process the live fluoroscopic image in the live fluoroscopy window 222 to identify or detect the biopsy needle tip 226 in the live fluoroscopic image.
The user instructions 221 may include a further instruction to click the “Generate Actual Trajectory” button 229 to generate or draw the actual needle trajectory 228 in both the live fluoroscopy window 222 and the navigation window 224. Alternatively, the actual needle trajectory 228 may be applied to or drawn in only the live fluoroscopy window 222 or the navigation window 224. The user instructions 221 include a further instruction to repeat instructions 1-4 until the biopsy needle tip 226 is at a desired position in the target indicated by the target mark 213. As illustrated in the live fluoroscopy window 222 and the navigation window 224, the actual trajectory 229 is not aligned with the planned trajectory 217. In this scenario, the clinician may remove the biopsy needle and retry puncturing the patient's skin with the biopsy needle and perform an initial navigation of the biopsy needle along the planned trajectory 217. Alternatively, the clinician may adjust the navigation direction of the biopsy needle for the next portion of the planned trajectory 217 so that the actual trajectory of the biopsy needle can come into more alignment with the planned trajectory 217.
As illustrated in
Then, an initial fluoroscopic sweep may be performed, the fluoroscopic 3D reconstruction may be generated based on the initial fluoroscopic sweep data, and the fluoroscopic 3D reconstruction may be registered to the preoperative CT images. The registration process enables biopsy needle information (e.g., the biopsy needle's actual trajectory) obtained from fluoroscopy sweeps to be shown on the preoperative CT images. In performing the registration between the fluoroscopic 3D reconstruction and the CT images, the clinician may be prompted to mark a lesion, ribs, or other body structure in both the fluoroscopic 3D reconstruction and in the CT images.
In aspects, the position of the target and the entry point may be determined from a first fluoroscopic 3D reconstruction generated based on the first fluoroscopic images. In some aspects, block 502 may further include determining a pose for each of the captured first fluoroscopic images and generating a first fluoroscopic 3D reconstruction is generated based on the captured first fluoroscopic images and the poses. Then, the first fluoroscopic 3D reconstruction may be displayed in a trajectory planning window. In some aspects, the first fluoroscopic 3D reconstruction may be “cut” into slices to obtain “2D-like” images, which may be displayed and scrolled through by a clinician, e.g., by clicking and dragging the user control 215 of
In some aspects, one or more marks indicating the entry point (e.g., a bullseye), the target, a trajectory to the target, and/or one or more structures to avoid, are applied to the fluoroscopic 3D reconstruction, and the marked first fluoroscopic 3D reconstruction is displayed. The one or more marks may include crosshairs, lines, shapes, or any other suitable marks for indicating the insertion point, the target, the trajectory, and/or the one or more structures to avoid. In some aspects, at least one of the target and the needle shown in the first fluoroscopic 3D reconstruction may be identified and at least one mark indicating at least one of the target and the needle may be applied to the first fluoroscopic 3D reconstruction. Identifying the needle in the first fluoroscopic reconstruction may include receiving a mark indicating a location of the needle on the first fluoroscopic 3D reconstruction, for example, through a user interface or controls operated by a clinician. Identifying the needle in the first fluoroscopic 3D reconstruction may include segmenting the first fluoroscopic 3D reconstruction to determine a location of the needle. In aspects, the needle may be a biopsy needle or an ablation device.
At block 504, a position of a target is determined based on the first fluoroscopic images. Then, at block 506, a position of an entry point is determined based on the position of the target and the first fluoroscopic images.
At block 510, after the needle is inserted into or advanced through the patient's body, an additional fluoroscopic sweep is performed to capture multiple additional fluoroscopic images. In some aspects, block 510 may include determining poses for each of the additional fluoroscopic images, generating an additional fluoroscopic 3D reconstruction based on the additional fluoroscopic images and the poses for each of the additional fluoroscopic images, and displaying the additional fluoroscopic 3D reconstruction. Block 510 may further include registering the current fluoroscopic 3D reconstruction to a previous fluoroscopic 3D reconstruction, and transferring the one or more marks applied to the previous fluoroscopic 3D reconstruction to the current fluoroscopic 3D reconstruction based on the registering. Performing the additional fluoroscopic sweeps may include performing a second fluoroscopic sweep. Performing the first fluoroscopic sweep may include performing the second fluoroscopic sweep to obtain first fluoroscopic images of one or more objects, which may include a metal grid or any suitable radiopaque object. Then, the position of the one or more objects relative to the target may be determined. In turn, the position of the entry point may be determined based on the position of the one or more objects relative to the target.
At block 512, the position and the direction of the inserted needle is determined based on the additional fluoroscopic images. Then, at block 514, a distance between the tip of the needle and the target is determined. In aspects, the distance between the needle and the target may be calculated based on the additional fluoroscopic 3D reconstruction. Then, the distance may be displayed to the clinician and the needle may be advanced the displayed distance by the clinician using length markers on the needle to reach the target with the tip of the needle.
At block 516, the method 500 determines whether the tip of the needle has reached a desired location in the target. If the tip of the needle has not reached the desired location in the target, the method 500 returns to block 510 to perform an additional fluoroscopic sweep to obtain additional fluoroscopic images. Optionally, additional poses for each of the additional fluoroscopic images may be determined and additional fluoroscopic 3D reconstructions may be generated based on the additional fluoroscopic images and the additional poses. Blocks 510-516 are repeated until the tip of the needle has reached the target. If the tip of the needle has reached the target, the method 500 ends at block 518.
At block 604, a fluoroscopic 3D reconstruction is generated based on the captured fluoroscopic images and the poses determined at block 602. Then, at block 606, the fluoroscopic 3D reconstruction is displayed. At block 608, a planned trajectory is applied to the fluoroscopic 3D reconstruction. The planned trajectory may include a target, an insertion point, and a line between the insertion point and the target. In some aspects, a mark indicating a critical structure to avoid may also be applied to the fluoroscopic 3D reconstruction.
At block 610, a live fluoroscopic image, which includes a view of a needle, is displayed and, at block 612, the planned trajectory, which may be indicated by one or more marks are overlaid or displayed on the live fluoroscopic image. The live fluoroscopic image may be displayed in a live fluoroscopy view or window. The live fluoroscopy view may display projections that enable the clinician to guide or navigate a needle so that the needle follows the planned trajectory to the target.
In aspects, a mark indicating a critical structure to avoid may also be applied to the fluoroscopic 3D reconstruction and the mark indicating the critical structure to avoid may be overlaid on the live fluoroscopic image. In aspects, the marks indicating the insertion point, the target, and the critical structure to avoid may be obtained from preoperative CT images. The method 600 may include receiving marked preoperative CT images, registering the fluoroscopic 3D reconstruction to the preoperative CT images, and transferring the markings on the preoperative CT images to the fluoroscopic 3D reconstruction based on the registering. The registering may include determining anatomical features that are in both the fluoroscopic 3D reconstruction and the preoperative CT images, and aligning the fluoroscopic 3D reconstruction and the preoperative CT images based on the anatomical features. The anatomical features may include the target, a lesion, a tumor, or a rib.
At block 614, the advancement of the needle a first distance is displayed in the live fluoroscopic image at two angles. The live fluoroscopic image at two angles provides a view of the direction of the needle. At block 616, the method 600 determines whether the direction of the needle in the live fluoroscopic image is consistent with the planned trajectory overlaid on the live fluoroscopic image. If the direction of the needle is not consistent with the planned trajectory, the direction of the needle is adjusted at block 617 and blocks 614 and 616 are repeated. In some aspects, the needle may be adjusted at block 617 by retracting the needle a small distance before repeating blocks 614 and 616.
If the direction of the needle is consistent with the planned trajectory, advancement of the needle a second distance is displayed in the live fluoroscopic image at one angle, e.g., one of the two angles of block 614 that shows the depth of the needle, at block 618. In aspects, the first distance may be less than the second distance so that one or more small advancement steps are taken to confirm that the needle is advancing in the direction of the planned trajectory before advancing the needle in larger advancement steps towards the target. At block 620, the method 600 determines whether the needle is following the planned trajectory. If the needle is not following the planned trajectory, the trajectory of the needle is adjusted at block 621 and blocks 618 and 620 are repeated until the needle is following the planned trajectory. If the needle is following the planned trajectory, the method 600 ends at block 622.
In aspects, before ending at block 622, the method 600 may include determining whether that the needle is located at the target, and performing one or more additional fluoroscopic sweeps to obtain one or more additional fluoroscopic images in response to determining that the needle is not located at the target. These functions may be repeated until the needle is determined to be located at the target.
The planned trajectory may include an entry point, a target, and a path between the entry point and the target. In aspects, the needle is any percutaneously-inserted device suitable for performing, for example, localization (e.g., using a dye, guide wire, or fiducials), biopsy (e.g., a biopsy needle), or ablation of the target. The actual trajectory of the tip of the needle may be determined by receiving a marking by a clinician or other user of the needle tip on the additional fluoroscopic 3D reconstruction. Alternatively, obtaining the location of the medical device in the additional fluoroscopic images includes automatically segmenting the additional fluoroscopic images to determine the location of the medical device.
In some aspects, a 3D CT reconstruction or 3D volume rendering is generated based on the CT images and a trajectory is applied to the 3D CT reconstruction either by a clinician drawing the trajectory on or otherwise applying the planned trajectory to the 3D CT reconstruction or a software application automatically drawing the planned trajectory on the 3D CT reconstruction. The 3D CT reconstruction may be displayed to a clinician in a suitable user interface. Subsequently, the 3D CT reconstruction may be updated based on the fluoroscopic 3D reconstruction generated at block 710. For example, the actual trajectory of a biopsy device may be updated in the 3D CT reconstruction. The 3D CT reconstruction may provide more detail than the fluoroscopic 3D reconstruction, which, for example, better enables a clinician not to hit an anatomic feature to be avoided with the biopsy device. The anatomic feature may include a bone, a vascular structure, or any other critical structure.
At block 708, a fluoroscopic sweep of at least a portion of the patient's body that includes the target and the percutaneously-inserted device inserted at the insertion point is performed to capture fluoroscopic images. At block 710, a fluoroscopic 3D reconstruction is generated based on the captured fluoroscopic images. At block 712, the fluoroscopic 3D reconstruction is registered to the preoperative CT images. The registration process may involve recognizing the same features in both the fluoroscopic 3D reconstruction and the preoperative CT images, and aligning the fluoroscopic 3D reconstruction and the preoperative CT images with each other based on the recognized features. At block 714, the markings from the preoperative CT images are transferred to the fluoroscopic 3D reconstruction based on the registering.
At block 716, an orientation of the inserted percutaneously-inserted device and a distance between the inserted percutaneously-inserted device and the target are determined based on the fluoroscopic 3D reconstruction. Then, before ending at block 724, the orientation and the distance are displayed to guide advancement of the percutaneously-inserted device toward the target at block 718.
At block 808, a live fluoroscopic image, which includes a view of a needle, is displayed and, at block 810, the planned trajectory, which may be indicated by one or more marks are overlaid or displayed on the live fluoroscopic image. The live fluoroscopic image may be displayed in a live fluoroscopy view or window. In some aspects, the one or more marks indicating one or more critical structures to avoid may be overlaid on the live fluoroscopic image at block 810. At block 812, the live fluoroscopic image is adjusted so that the displayed planned trajectory appears as a point or a bullseye view. Then, at block 814, at the displayed point, the needle is advanced the distance of the planned trajectory using length markers on the needle. Block 814 may include calculating and displaying a distance from the tip of the needle to the target based on the fluoroscopic 3D reconstruction of block 806. In aspects, the calculated distance may be displayed to the clinician overlaid on or adjacent to the live 2D fluoroscopy image. Then, the clinician can use the length marks on the needle to advance the needle the distance displayed to the clinician.
In aspects, one or more additional fluoroscopic sweeps are performed, and the resulting additional fluoroscopic images are processed until the needle, e.g., a biopsy device, is located at the target. For example, at block 816, an additional fluoroscopic sweep is performed to obtain additional fluoroscopic images poses for each of the additional fluoroscopic images after advancement of the needle. At block 818, an additional fluoroscopic 3D reconstruction is generated based on the additional fluoroscopic images and the poses for each of the additional fluoroscopic images, and, at block 820, the additional fluoroscopic 3D reconstruction is displayed.
At block 822, the method 800 determines whether the needle tip is located at the target. For example, the method 700 may include determining whether the clinician has clicked a button or otherwise indicated that the needle tip is or is not located at the target. If the needle tip is not located at the target, the method 800 returns to block 816 to perform an additional fluoroscopic sweep to obtain additional fluoroscopic images. Blocks 816-820 are repeated until the needle tip is located at the target. If the biopsy device is located at the target, the method 800 ends at block 824.
In some aspects, the initial sweep (e.g., the sweep of block 808) may be a full, complete, or wide sweep and the subsequent, secondary sweeps (e.g., the sweep of block 816) may be a partial or narrow sweep to, for example, minimize radiation exposure to the clinicians and/or the patient.
In aspects, location information for the biopsy device may be obtained, e.g., from the location sensor 28 disposed on the biopsy tool 27. Determining the location of the biopsy device may include generating an electromagnetic field, e.g., by the transmitter mat 56, sensing the electromagnetic field by one or more electromagnetic sensors (e.g., location sensor 28) disposed on the biopsy device, and determining the 3D coordinates of the biopsy device based on the sensed electromagnetic field. In some aspects, 3D views showing the 3D position and/or trajectory of the location sensor 28 versus the 3D position of the target may be displayed to a clinician. The clinician may use these 3D views to guide the biopsy device towards the target without a live fluoroscopy view. This may shorten the medical procedure by reducing the number of steps needed. If the target moves, additional sweeps may be performed to confirm or adjust the target position.
In some aspects, the location sensor 28 can serve as a marker, which, when using preoperative CT images, can be registered to either the navigation system or the fluoroscopic 3D reconstruction. In other aspects, when not using pre-operative CT images as described herein, the location sensor 28 can be used for registering the navigation system to the fluoroscopic 3D reconstruction and the biopsy device or other percutaneously-inserted device can be navigated “on” the fluoroscopic 3D reconstruction.
In aspects, information from the marked fluoroscopic 3D reconstruction may be incorporated into a 3D electromagnetic (EM) navigation view of a needle.
At block 906, a trajectory, including an entry point and a target, is marked in the fluoroscopic 3D reconstruction. The method 900 may further include marking a critical structure to avoid in the fluoroscopic 3D reconstruction. Then, at block 908, a location and an orientation of a needle is determined using an EM navigation system including an EM sensor disposed on the needle, after slight insertion of the needle at the entry point. Determining the location and orientation of the needle may include generating an electromagnetic field, sensing the electromagnetic field by one or more electromagnetic sensors disposed on the needle, determining the 3D coordinates and orientation of the needle based on the sensed electromagnetic field, and generating the 3D EM navigation view based on the 3D coordinates and orientation of the needle.
At block 909, the EM navigation system is registered to the fluoroscopic 3D reconstruction based on the determined location and orientation of the needle. Registering the EM navigation system to the fluoroscopic 3D reconstruction may include identifying the EM sensor in the fluoroscopic 3D reconstruction, and registering the fluoroscopic 3D reconstruction to the 3D EM navigation view based on the identified EM sensor. At block 910, a 3D electromagnetic (EM) navigation view of the needle is generated based on the registering. The 3D EM navigation view may be generated based on the location and orientation of a needle determined by EM sensors and detectors of the EM navigation system.
At block 912, the trajectory markings are displayed in the 3D EM navigation view based on the registering. Then, before ending at block 916, advancement of the needle is displayed in the 3D EM navigation view at block 914. In aspects, the method 900 may further include performing a second fluoroscopic sweep to obtain second fluoroscopic images after navigation of the needle to the target and confirming that the needle is at the target based on the second fluoroscopic images.
In aspects, the number of iterations of performing fluoroscopic sweeps and generating the fluoroscopic 3D reconstruction depends on the location of the needle insertion or entry point and the location of the target, e.g., a lesion. For example, one iteration may be needed to get from the needle entry point to the lung, one or two iterations may be needed to verify that the needle got to the pleura, and one further iteration may be needed if the lesion is close to the pleura. If the lesion is further away from the pleura, one iteration may be needed every half inch or every inch until the needle reaches the lesion. For example, up to seven iterations may be needed to reach the lesion. And, in total, five to ten iterations may be needed to navigate the biopsy needle from the entry point to the lesion.
From the foregoing and with reference to the various figure drawings, those skilled in the art will appreciate that certain modifications can also be made to this disclosure without departing from the scope of the same. For example, although an aspect of the systems and methods is described as usable with a biopsy tool, the systems and methods described herein may be utilized with systems that utilize treatment devices, such as ablation devices. Additionally, it is appreciated that the above-described systems and methods may be utilized in other target regions such as the liver. Further, the above-described systems and methods may also be usable for transthoracic needle aspiration procedures.
Detailed embodiments of this disclosure are disclosed herein. However, the disclosed embodiments are merely examples of the disclosure, which may be embodied in various forms and aspects. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for teaching one skilled in the art to variously employ this disclosure in virtually any appropriately detailed structure.
The image-guidance systems of this disclosure may be separate from or integrated with an energy device or a separate tool and may include MRI, CT, fluoroscopy, ultrasound, electrical impedance tomography, optical, and/or device tracking systems. In aspects, the image-guided, percutaneously-inserted device may be used for biopsy, ablation, or localizing tumors, for example, with dye, a guide wire, or fiducials. Methodologies for locating the percutaneously-inserted device according to some aspects of this disclosure include electromagnetic (EM), infra-red (IR), echolocation, optical, or others. Tracking systems may be integrated to an imaging device, where tracking is done in virtual space or fused with preoperative or live images.
It should be understood that various aspects disclosed herein may be combined in different combinations than the combinations specifically presented in the description and accompanying drawings. For example, one or more of the blocks of one of
In one or more examples, the described techniques may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored as one or more instructions or code on a computer-readable medium and executed by a hardware-based processing unit. Computer-readable media may include non-transitory computer-readable media, which corresponds to a tangible medium such as data storage media (e.g., RAM, ROM, EEPROM, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer).
Instructions may be executed by one or more processors, such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor” as used herein may refer to any of the foregoing structure or any other physical structure suitable for implementation of the described techniques. Also, the techniques could be fully implemented in one or more circuits or logic elements.
While several aspects of the disclosure have been shown in the drawings, it is not intended that the disclosure be limited thereto, as it is intended that the disclosure be as broad in scope as the art will allow and that the specification be read likewise. Therefore, the above description should not be construed as limiting, but merely as exemplifications of particular embodiments. Those skilled in the art will envision other modifications within the scope and spirit of the claims appended hereto. For example, although the disclosure refers to a biopsy needle or tool, it is contemplated that the biopsy needle or tool may be replaced by an ablation device, a localization, or any suitable image-guided percutaneously-inserted medical device.
Number | Name | Date | Kind |
---|---|---|---|
5057494 | Sheffield | Oct 1991 | A |
5321113 | Cooper et al. | Jun 1994 | A |
5930329 | Navab | Jul 1999 | A |
5963613 | Navab | Oct 1999 | A |
6003517 | Sheffield et al. | Dec 1999 | A |
6484049 | Seeley et al. | Nov 2002 | B1 |
7233820 | Gilboa | Jun 2007 | B2 |
8165660 | Pfister et al. | Apr 2012 | B2 |
8335359 | Fidrich et al. | Dec 2012 | B2 |
8542900 | Tolkowsky et al. | Sep 2013 | B2 |
8694075 | Groszmann | Apr 2014 | B2 |
8706184 | Mohr et al. | Apr 2014 | B2 |
8827934 | Chopra et al. | Sep 2014 | B2 |
9265468 | Rai | Feb 2016 | B2 |
9375268 | Long | Jun 2016 | B2 |
9918659 | Chopra et al. | Mar 2018 | B2 |
10004558 | Long | Jun 2018 | B2 |
10154239 | Casas | Dec 2018 | B2 |
10194897 | Cedro et al. | Feb 2019 | B2 |
10373719 | Soper et al. | Aug 2019 | B2 |
10376178 | Chopra | Aug 2019 | B2 |
10405753 | Sorger | Sep 2019 | B2 |
10478162 | Barbagli et al. | Nov 2019 | B2 |
10480926 | Froggatt et al. | Nov 2019 | B2 |
10524866 | Srinivasan et al. | Jan 2020 | B2 |
10555788 | Panescu et al. | Feb 2020 | B2 |
10569071 | Harris et al. | Feb 2020 | B2 |
10603106 | Weide et al. | Mar 2020 | B2 |
10610306 | Chopra | Apr 2020 | B2 |
10638953 | Duindam et al. | May 2020 | B2 |
10639114 | Schuh et al. | May 2020 | B2 |
10674970 | Averbuch et al. | Jun 2020 | B2 |
10674982 | Barak et al. | Jun 2020 | B2 |
10682070 | Duindam | Jun 2020 | B2 |
10702137 | Deyanov | Jul 2020 | B2 |
10706543 | Donhowe et al. | Jul 2020 | B2 |
10709506 | Coste-Maniere et al. | Jul 2020 | B2 |
10716525 | Weingarten et al. | Jul 2020 | B2 |
10772485 | Schlesinger et al. | Sep 2020 | B2 |
10796432 | Mintz et al. | Oct 2020 | B2 |
10823627 | Sanborn et al. | Nov 2020 | B2 |
10827913 | Ummalaneni et al. | Nov 2020 | B2 |
10835153 | Rafii-Tari et al. | Nov 2020 | B2 |
10885630 | Li et al. | Jan 2021 | B2 |
20020147462 | Mair et al. | Oct 2002 | A1 |
20030013972 | Makin | Jan 2003 | A1 |
20040120981 | Nathan | Jun 2004 | A1 |
20050222554 | Wallace et al. | Oct 2005 | A1 |
20080045938 | Weide et al. | Feb 2008 | A1 |
20080097187 | Gielen et al. | Apr 2008 | A1 |
20080123921 | Gielen | May 2008 | A1 |
20080154526 | Li et al. | Jun 2008 | A1 |
20080161684 | Li et al. | Jul 2008 | A1 |
20100215213 | Mielekamp et al. | Aug 2010 | A1 |
20130303945 | Blumenkranz et al. | Nov 2013 | A1 |
20140035798 | Kawada et al. | Feb 2014 | A1 |
20150042643 | Shibata et al. | Feb 2015 | A1 |
20150148690 | Chopra et al. | May 2015 | A1 |
20150265368 | Chopra et al. | Sep 2015 | A1 |
20160000517 | Kehat et al. | Jan 2016 | A1 |
20160157939 | Larkin et al. | Jun 2016 | A1 |
20160183841 | Duindam et al. | Jun 2016 | A1 |
20160192860 | Allenby et al. | Jul 2016 | A1 |
20160287344 | Donhowe et al. | Oct 2016 | A1 |
20170035380 | Barak et al. | Feb 2017 | A1 |
20170112571 | Thiel et al. | Apr 2017 | A1 |
20170112576 | Coste-Maniere et al. | Apr 2017 | A1 |
20170209071 | Zhao et al. | Jul 2017 | A1 |
20170265952 | Donhowe et al. | Sep 2017 | A1 |
20170311844 | Zhao et al. | Nov 2017 | A1 |
20170319165 | Averbuch | Nov 2017 | A1 |
20180078318 | Barbagli et al. | Mar 2018 | A1 |
20180144092 | Flitsch et al. | May 2018 | A1 |
20180153621 | Duindam et al. | Jun 2018 | A1 |
20180235709 | Donhowe et al. | Aug 2018 | A1 |
20180240237 | Donhowe et al. | Aug 2018 | A1 |
20180256262 | Duindam et al. | Sep 2018 | A1 |
20180263706 | Averbuch | Sep 2018 | A1 |
20180279852 | Rafii-Tari et al. | Oct 2018 | A1 |
20180325419 | Zhao et al. | Nov 2018 | A1 |
20190000559 | Berman et al. | Jan 2019 | A1 |
20190000560 | Berman et al. | Jan 2019 | A1 |
20190005687 | Weingarten | Jan 2019 | A1 |
20190008413 | Duindam et al. | Jan 2019 | A1 |
20190038365 | Soper et al. | Feb 2019 | A1 |
20190065209 | Mishra et al. | Feb 2019 | A1 |
20190110839 | Rafii-Tari et al. | Apr 2019 | A1 |
20190175062 | Rafii-Tari et al. | Jun 2019 | A1 |
20190175799 | Hsu et al. | Jun 2019 | A1 |
20190183318 | Froggatt et al. | Jun 2019 | A1 |
20190183585 | Rafii-Tari et al. | Jun 2019 | A1 |
20190183587 | Rafii-Tari et al. | Jun 2019 | A1 |
20190192234 | Gadda et al. | Jun 2019 | A1 |
20190209016 | Herzlinger et al. | Jul 2019 | A1 |
20190209043 | Zhao et al. | Jul 2019 | A1 |
20190216548 | Ummalaneni | Jul 2019 | A1 |
20190239723 | Duindam et al. | Aug 2019 | A1 |
20190239831 | Chopra | Aug 2019 | A1 |
20190250050 | Sanborn et al. | Aug 2019 | A1 |
20190254649 | Walters et al. | Aug 2019 | A1 |
20190269470 | Barbagli et al. | Sep 2019 | A1 |
20190269818 | Dhanaraj et al. | Sep 2019 | A1 |
20190269819 | Dhanaraj et al. | Sep 2019 | A1 |
20190272634 | Li et al. | Sep 2019 | A1 |
20190298160 | Ummalaneni et al. | Oct 2019 | A1 |
20190298451 | Wong et al. | Oct 2019 | A1 |
20190320878 | Duindam et al. | Oct 2019 | A1 |
20190320937 | Duindam et al. | Oct 2019 | A1 |
20190336238 | Yu et al. | Nov 2019 | A1 |
20190343424 | Blumenkranz et al. | Nov 2019 | A1 |
20190350659 | Wang et al. | Nov 2019 | A1 |
20190365199 | Zhao et al. | Dec 2019 | A1 |
20190365479 | Rafii-Tari | Dec 2019 | A1 |
20190365486 | Srinivasan et al. | Dec 2019 | A1 |
20190380787 | Ye et al. | Dec 2019 | A1 |
20200000319 | Saadat et al. | Jan 2020 | A1 |
20200000526 | Zhao | Jan 2020 | A1 |
20200008655 | Schlesinger et al. | Jan 2020 | A1 |
20200030044 | Wang et al. | Jan 2020 | A1 |
20200030461 | Sorger | Jan 2020 | A1 |
20200038750 | Kojima | Feb 2020 | A1 |
20200043207 | Lo et al. | Feb 2020 | A1 |
20200046431 | Soper et al. | Feb 2020 | A1 |
20200046436 | Tzeisler et al. | Feb 2020 | A1 |
20200054399 | Duindam et al. | Feb 2020 | A1 |
20200054408 | Schuh et al. | Feb 2020 | A1 |
20200060771 | Lo et al. | Feb 2020 | A1 |
20200069192 | Sanborn et al. | Mar 2020 | A1 |
20200077870 | Dicarlo et al. | Mar 2020 | A1 |
20200078023 | Cedro et al. | Mar 2020 | A1 |
20200078095 | Chopra et al. | Mar 2020 | A1 |
20200078103 | Duindam et al. | Mar 2020 | A1 |
20200085514 | Blumenkranz | Mar 2020 | A1 |
20200109124 | Pomper et al. | Apr 2020 | A1 |
20200129045 | Prisco | Apr 2020 | A1 |
20200129239 | Bianchi et al. | Apr 2020 | A1 |
20200138514 | Blumenkranz et al. | May 2020 | A1 |
20200138515 | Wong | May 2020 | A1 |
20200142013 | Wong | May 2020 | A1 |
20200155116 | Donhowe et al. | May 2020 | A1 |
20200155232 | Wong | May 2020 | A1 |
20200170623 | Averbuch | Jun 2020 | A1 |
20200170720 | Ummalaneni | Jun 2020 | A1 |
20200179058 | Barbagli et al. | Jun 2020 | A1 |
20200188021 | Wong et al. | Jun 2020 | A1 |
20200188038 | Donhowe et al. | Jun 2020 | A1 |
20200205903 | Srinivasan et al. | Jul 2020 | A1 |
20200205904 | Chopra | Jul 2020 | A1 |
20200214664 | Zhao et al. | Jul 2020 | A1 |
20200229679 | Zhao et al. | Jul 2020 | A1 |
20200242767 | Zhao et al. | Jul 2020 | A1 |
20200275860 | Duindam | Sep 2020 | A1 |
20200297442 | Adebar et al. | Sep 2020 | A1 |
20200315554 | Averbuch et al. | Oct 2020 | A1 |
20200330795 | Sawant et al. | Oct 2020 | A1 |
20200352427 | Deyanov | Nov 2020 | A1 |
20200364865 | Donhowe et al. | Nov 2020 | A1 |
20200383750 | Kemp et al. | Dec 2020 | A1 |
20210000524 | Barry et al. | Jan 2021 | A1 |
Number | Date | Country |
---|---|---|
0013237 | Jul 2003 | BR |
0116004 | Jun 2004 | BR |
0307259 | Dec 2004 | BR |
0412298 | Sep 2006 | BR |
112018003862 | Oct 2018 | BR |
1644519 | Dec 2008 | CZ |
186540 | Sep 2016 | CZ |
2709512 | Aug 2017 | CZ |
2884879 | Jan 2020 | CZ |
1774909 | Apr 2007 | EP |
1644519 | Dec 2008 | EP |
3413830 | Sep 2019 | EP |
3478161 | Feb 2020 | EP |
3641686 | Apr 2020 | EP |
3644885 | May 2020 | EP |
3644886 | May 2020 | EP |
3749239 | Dec 2020 | EP |
PA03005028 | Jan 2004 | MX |
PA03000137 | Sep 2004 | MX |
PA03006874 | Sep 2004 | MX |
225663 | Jan 2005 | MX |
226292 | Feb 2005 | MX |
PA03010507 | Jul 2005 | MX |
PA05011725 | May 2006 | MX |
06011286 | Mar 2007 | MX |
246862 | Jun 2007 | MX |
2007006441 | Aug 2007 | MX |
265247 | Mar 2009 | MX |
284569 | Mar 2011 | MX |
Entry |
---|
Extended European search report issued in European Application No. 20198223.8 dated Dec. 9, 2020, 13 pages. |
Extended European Search Report issued in European Application No. 20198236.0 dated Dec. 8, 2020, 15 pages. |
Arianna Ferrari et al., “Digital chest tomosynthesis: the 2017 updated review of an emerging application”, Ann Transl Med 2018; 6(5): 91, pp. 1-8. |
Nelson et al., “Improved targeting accuracy of lung tumor biopsies with scanning-beam digital x-ray tomosynthesis image guidance”, Medical Physics, vol. 43, No. 12, Dec. 2016, 9 pages. |
Number | Date | Country | |
---|---|---|---|
20210085268 A1 | Mar 2021 | US |
Number | Date | Country | |
---|---|---|---|
62905151 | Sep 2019 | US |