Systems and methods for ablation visualization

Information

  • Patent Grant
  • 11707329
  • Patent Number
    11,707,329
  • Date Filed
    Monday, August 12, 2019
    5 years ago
  • Date Issued
    Tuesday, July 25, 2023
    a year ago
Abstract
The visualization method includes displaying three-dimensional image data of at least one anatomical feature of a patient, receiving user input of the target for placing an ablation needle in the at least one anatomical feature of the patient, determining the position and orientation of the ablation needle based on the user input, displaying an image of a virtual ablation needle in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation, receiving user input of parameters of operating the ablation needle, and displaying a three-dimensional representation of the result of operating the ablation needle according to the input parameters.
Description
INTRODUCTION

This disclosure relates to visualizing a treatment procedure and, more particularly, to systems and methods for visualizing the anatomy of a patient based on parameters of operating an ablation probe.


BACKGROUND

Computed tomography (CT) images are commonly used to identify objects, such as physiological structures, in a patient's body. In particular, CT images can be used by physicians to identify malignant tissue or problematic structures in a patient's body and to determine their location within the body. Once the location is determined, a treatment plan can be created to address the problem, such as planning a pathway into the patient's body to remove malignant tissue or planning procedures for accessing and altering the problematic structures. Ablation of tumors is an example of a more targeted approach to tumor treatment. In comparison to traditional body-wide types of cancer treatment, such as chemotherapy, ablation technologies are more targeted and limited, but are just as effective. Thus, such approaches are beneficial in providing targeted treatment that limits unnecessary injury to non-problematic tissue or structures in the patient's body, but they require the assistance of more complex technical tools. Accordingly, there continues to be interest in developing further technical tools to assist with targeted treatment of tissue or structural problems in a patient's body.


SUMMARY

This disclosure relates generally to visualization systems and methods for visualizing potential ablation size in three dimensions overlaid on or incorporated into images of patient anatomy. In one aspect, this disclosure features a method of performing an ablation procedure. The method includes displaying three-dimensional image data of at least one anatomical feature of a patient, receiving user input of the target for placing an ablation needle in the at least one anatomical feature of the patient, determining the position and orientation of the ablation needle based on the user input, displaying an image of a virtual ablation needle in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation, receiving user input of parameters of operating the ablation needle, and displaying a three-dimensional representation of the result of operating the ablation needle according to the input parameters.


In aspects, the method includes receiving further user input of movement of an ablation needle in the at least one anatomical feature of the patient, determining the new position and orientation of the ablation needle based on the further user input, and displaying the virtual ablation needle in the three-dimensional image according to the determined further position and orientation.


In aspects, the method includes receiving user input of other different parameters of operating the ablation needle, and displaying a three-dimensional representation of the result of operating the ablation needle according to the other different parameters.


In aspects, the method includes displaying a default ablation zone relative to the target, receiving user input adjusting the size of the ablation zone, calculating ablation time based on the adjusted size of the ablation zone, and displaying the calculated ablation time.


In aspects, the method includes calculating a tip distance based on the adjusted size of the ablation zone, and displaying the calculated tip distance.


In aspects, the parameters of operating the ablation needle include power level and ablation needle type.


In aspects, the three-dimensional representation of the result of operating the ablation needle is a three-dimensional representation of at least one of a temperature profile, an ablation zone, potential histological zones, a plurality of temperatures, confidence intervals, a heated zone, or probability of cell death with respect to the position of the ablation needle.


In aspects, displaying the three-dimensional image includes displaying a multi-plane view including at least two of a coronal view, a sagittal view, an axial view, or a three-dimensional view of that ablation needle and the ablation zone.


In aspects, the method includes displaying user-selectable icons for selecting a skin view, a muscle view, or a bone view of at least a portion of the patient.


In aspects, the method includes displaying a single-slice view including a coronal view, a sagittal view, or an axial view, and displaying user-selectable icons for selecting between the coronal view, the sagittal view, or the axial view.


In aspects, the method includes simultaneously displaying a different one of the coronal view, the sagittal view, or the axial view, displaying a user-movable cross-section line on the different one of the coronal view, the sagittal view, or the axial view, and displaying a slice corresponding to the position of the cross-section line.


In aspects, the three-dimensional image data is at least one of computed tomography image data, magnetic resonance image data, or ultrasound image data.


In aspects, the method includes displaying a snapshot button, receiving user selection of the snapshot button, and recording an image including an image of the anatomy, a target, an ablation zone, and text indicating the power level and the ablation time.


In another aspect, this disclosure features an ablation visualization system, which includes a display, a processor, and a memory having stored thereon instructions which are executed by the processor. When the instructions are executed by the processor, the processor displays, on the display, three-dimensional image data of at least one anatomical feature of a patient on the display, prompts a user for input of a target for placing an ablation needle in the at least one anatomical feature of the patient, receives user input of the target for placing the ablation needle in the at least one anatomical feature of the patient, determines a position and orientation of the ablation needle based on the user input, displays, on the display, an image of a virtual ablation needle in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation, prompts the user for input of parameters of operating the ablation needle, receives user input of parameters of operating the ablation needle, and displays, on the display, a three-dimensional representation of a result of operating the ablation needle according to the input parameters.


In aspects, the instructions, when executed by the processor, further cause the processor to prompt for further user input of movement of an ablation needle in the at least one anatomical feature of the patient, receive further user input of movement of an ablation needle in the at least one anatomical feature of the patient, determine a new position and orientation of the ablation needle based on the further user input, and display the virtual ablation needle in the three-dimensional image according to the new position and orientation.





BRIEF DESCRIPTION OF THE DRAWINGS

The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.


Various aspects and features of this disclosure are described below with references to the drawings, of which:



FIG. 1 is a block diagram of an exemplary system for ablation visualization in accordance with aspects of this disclosure;



FIG. 2 is a diagram of exemplary regions of a patient to which the disclosed systems and methods may be applied;



FIGS. 3A and 3B are exemplary display interfaces for viewing and selecting patient image data that is used to view or create a surgical plan in accordance with aspects of this disclosure;



FIG. 4 is an exemplary display interface for adding targets to patient image data in accordance with aspects of this disclosure;



FIGS. 5A and 5B are exemplary display interfaces for viewing and modifying an added target in accordance with aspects of this disclosure;



FIGS. 6A and 6B are exemplary display interfaces for viewing and modifying various ablation parameters in accordance with aspects of this disclosure;



FIG. 7 is an exemplary display interface for showing different types of views of the placement and operation of an ablation instrument in accordance with aspects of this disclosure;



FIG. 8 is an exemplary display interface illustrating the bounds of a selectively adjustable ablation zone in an axial view in accordance with aspects of this disclosure;



FIG. 9 is an exemplary display interface for showing a temperature profile within an ablation zone in accordance with aspects of this disclosure;



FIG. 10 is an exemplary display interface for showing a temperature gradient surrounding an ablation zone in accordance with aspects of this disclosure;



FIG. 11 is an exemplary display interface for showing a temperature profile on a three-dimensional view of a liver during a simulated or actual ablation procedure in accordance with aspects of this disclosure;



FIG. 12 is an exemplary display interface for showing a temperature profile on a three-dimensional view of a lung during a simulated or actual ablation procedure in accordance with aspects of this disclosure;



FIG. 13 is an exemplary display interface illustrating the bounds of an ablation zone in multiple plane views in accordance with aspects of this disclosure;



FIG. 14 is an exemplary display interface illustrating the bounds of an ablation zone in multiple plane views in accordance with aspects of this disclosure; and



FIG. 15 is a flow diagram of an exemplary operation of a system in accordance with aspects of this disclosure.





DETAILED DESCRIPTION

Embodiments of this disclosure are now described in detail with reference to the drawings in which like reference numerals designate identical or corresponding elements in each of the several views. As used herein, the term “clinician” refers to a doctor, a nurse, or any other care provider and may include support personnel. Throughout this description, the phrase “in embodiments” and variations on this phrase generally is understood to mean that the particular feature, structure, system, or method being described includes at least one iteration of the disclosed technology. Such phrase should not be read or interpreted to mean that the particular feature, structure, system, or method described is either the best or the only way in which the embodiment can be implemented. Rather, such a phrase should be read to mean an example of a way in which the described technology could be implemented, but need not be the only way to do so.


In order for physicians or clinicians to perform ablations and achieve good clinical outcomes, they need to achieve an ablative margin. Achieving an ablative margin requires understanding where the ablation device will create lethal heating. This is a three-dimensional problem and involves understanding the gradient of temperature created by a particular ablation device at a particular power and time. Some ablation devices provide tables of power, time, and ablation size. These tables are often derived from Ex Vivo models and provide the mean value for each dimension, but leave out statistical information such as sample size and standard deviation. The information in these tables is used during a procedure to select the correct power and time via a manual process where, for example, the physician or clinician visualizes the ablation device on a CT scan and uses a linear measurement tool to estimate the location of the planned ablation on a screen. This process, however, does not take into account the three-dimensional nature of ablation planning, nor does it take into account the inherent variability of the ablation device being used.


This disclosure provides multiple visualization techniques that allow for visualizing potential ablation size in three dimensions overlaid on or incorporated into images of patient anatomy. By overlaying or incorporating visualizations that include different measures of variability in device performance on images of actual patient anatomy, the visualization techniques or methods of this disclosure maximize the information communicated in a single visualization.


The visualization techniques provide an understanding of ablation potential in the context of the actual patient anatomy. The overlay or incorporation of ablation information on CT, MR, or US image data brings context to ablation performance data. The visualization techniques allow for ablation performance to be visualized with respect to actual and/or planned needle position and patient anatomy at the same time in a single view, thereby recreating the entire context of the ablation. This can be done before the needle is placed as a planning step or after the actual needle is in place to aid in final adjustments of needle position or power and time.


The methods of this disclosure also allow for the physician or clinician to visualize temperatures, histological damage, potential histological zones, confidence intervals or probability of cell death, and ablation potential with reference to the needle and the patient anatomy.


Referring now to FIG. 1, there is shown a block diagram of a system 100, which includes a computing device 102 such as, for example, a laptop, desktop, workstation, tablet, or other similar device, a display 104, and an ablation system 106. The computing device 102 includes one or more processors 108, interface devices 110 (such as communications interface and user interface), memory and storage 112, and/or other components generally present in a computing device. The display 104 may be touch sensitive, which enables the display to serve as both an input and output device. In various embodiments, a keyboard (not shown), mouse (not shown), or other data input devices may be employed.


Memory/storage 112 may be any non-transitory, volatile or non-volatile, removable or non-removable media for storage of information such as computer-readable instructions, data structures, program modules or other data. In various embodiments, the memory 112 may include one or more solid-state storage devices such as flash memory chips or mass storage devices. In various embodiments, the memory/storage 112 can be RAM, ROM, EPROM, EEPROM, flash memory or other solid state memory technology, CD-ROM, DVD, Blu-Ray or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computing device 102.


Computing device 102 may also include an interface device 110 connected to a network or the Internet via a wired or wireless connection for the transmission and reception of data. For example, computing device 102 may receive computed tomographic (CT) image data 214 of a patient from a server, for example, a hospital server, Internet server, or other similar servers, for use during surgical ablation planning. Patient CT image data 114 may also be provided to computing device 202 via a removable memory.


In the illustrated embodiment, the memory/storage 112 includes CT image data 114 for one or more patients, information regarding the location and orientation of an ablation probe 116, various user settings 118 (which are described below), and various software that perform the operations described herein 120.


In various embodiments, the system 100 includes an ablation system 106 that includes a generator (not shown) and an ablation probe that includes an ablation antenna. The ablation system 106 will be described in more detail later herein.


In accordance with an aspect of this disclosure, the software 120 of FIG. 1 includes a treatment planning module which guides a clinician in identifying a target for ablation treatment, a target size, a treatment zone, and an access route to the target. As used herein, the term “clinician” refers to any medical professional (e.g., doctor, surgeon, nurse, physician assistant, technician, or the like) for planning, performing, monitoring and/or supervising a medical procedure involving the use of the embodiments described herein. The treatment planning module can generate a user interface screen for presenting information and receiving clinician input. The clinician can select a patient data set corresponding to a patient via the user interface. With reference also to FIG. 2, a patient data set can be selected based on, for example, a region where the ablation target is located, such as a lung region 202, a liver region 204, a kidney region 206, or another region of the patient. The patient data set includes CT image data for the selected region 202-206, which will be described below in connection with FIGS. 3-7.


As persons skilled in the art will understand, CT image data are x-ray scans of “slices” of a patient's anatomy. Although each slice views the anatomy from a particular angle, image data across multiple “slices” can be used to generate views of the anatomy from other angles. Based on the position and orientation of an ablation probe, an image of the anatomy can be generated for a probe-axial view, a probe-sagittal view, and a probe-coronal view. FIGS. 3-6 illustrate examples of these different views for a lung region 202. The lung region 202 is merely exemplary, and as mentioned above, other regions of the body can be viewed as well.



FIGS. 3A and 3B are exemplary display interfaces for viewing and selecting patient image data that is used to view or create a surgical plan in accordance with aspects of this disclosure. To import image data into the system, the user, such as a clinician, selects the “Import Images” icon 312 and selects an image data file in a file system browser window (not shown), which may be displayed in response to a user selecting the “Import Images” icon 312.


After importing image data into the system, information identifying the image data is arranged in the display interface of FIG. 3A. The information for identifying the image data includes patient name 302, patient identification (ID) 304, birth date 306, and date on which the image data for a particular patient was accessed 308. The clinician can use this information to navigate to the image data of a given patient for whom a surgical plan will be created. The display interface includes “Tips for creating a plan” to guide the clinician through the step of creating a surgical plan. This allows the clinician to navigate through the various display interfaces of the system without having to open a separate window, such as a help window. Alternatively, the clinician may quickly locate the image data of a patient by entering a patient name, identification number, or other identifying information in the search field and selecting the search execution button.


To create a new plan or open an existing plan for a patient, the clinician selects the patient name in the display interface of FIG. 3A, which causes the display interface of FIG. 3B to be displayed. The display interface of FIG. 3B shows one or more series of images corresponding to a selected patient. For each series of images, there is a “Create New Plan” icon, which, when selected by a clinician starts the process of creating a surgical plan. If one or more plans have been created for a particular series of images, an “Open Plan” icon is provided, which, when selected by a clinician, allows the clinician to review and/or revise that particular plan. Each series of images also shows compatibility level 314, which may indicate the compatibility of the series of images with the system. For example, a lower compatibility level may indicate that the series of images make it difficult to take advantage of the planning features of the system.


When the clinician selects the “Create New Plan” icon or button shown in FIG. 3B, the “Visualize” interface is displayed, as shown in FIG. 4. The “Visualize” interface allows the clinician to browse different views of the series of images to locate a lesion or other target surgical site and to add a target. The “Visualize” interface of FIG. 4 includes selectable icons 404 to select from among multiple views. The multiple views may include a Maximum Intensity Projection (MW) view, a single slice view (which is shown as selected in FIG. 4), a multi-plane view, and a three-dimensional (3D) view. In the single slice view, the “Visualize” interface of FIG. 4 also includes selectable icons 406 for selecting between an axial view (which is shown as selected in FIG. 4), a coronal view, and a sagittal view. In the single slice view, the clinician can browse through a series of images to locate a lesion. When the clinician locates a lesion or target surgical site, the clinician may click a mouse button at the center of a legion to place a crosshair icon 412 at the center of the legion. The clinician then adds that target location to the series of images by selecting the “Add a target” button 402, which causes the system to open a target window and corresponding tab, as illustrated in FIG. 5A.


Upon opening the target window, a multiplane view may initially be shown. The multiplane view includes a coronal image, a sagittal image, an axial image, and a 4 cm cube centered on the target location. In the target window, a target region centered at the target location may initially be identified by a circle 504 of a predetermined size, a selectably movable point 506 at the target location, and a plurality of selectably movable points 508 for changing the size and shape of the target region in each of the coronal view, the sagittal view, and the axial view. Textual information regarding the initial target region 502a may also be displayed in the target window. The textual information may include the width (x), the height (y), the depth (z), the volume, and the density of the target region. In this example, the width is 1.5 cm, the height is 1.5 cm, the depth is 1.5 cm, the volume is 1.77 cm2, the density (average) is −889 HU, and the density (standard deviation) is 87 HU.


The clinician may change the size and shape of the target region 504 by selecting and moving one or more of the plurality of point 508. For example, may change the size and shape of the target region 504 as illustrated in FIG. 5B, which results in new textual information 502b. In this example, the new width is 1.0 cm, the new depth is 1.1 cm, the new volume is 0.83 cm2, the density (average) is −924 HU, and the density (standard deviation) is 38 HU. When the clinician finishes defining the size and shape of the target region 504, the clinician may select the “Add an Ablation Zone” button 402.


When the clinician selects the “Add an Ablation Zone” button 402, the display interface of FIG. 6A is displayed. The display interface of FIG. 6A shows an ablation antenna or probe 611, an initial ablation zone 612, a selectably movable point 614 at the target location, and a plurality of selectably movable points 616 for changing the size and shape of the ablation zone are displayed in each of the axial view, the coronal view, and the sagittal view. Also, a three-dimensional wireframe representation of the ablation zone is displayed in the 4 cm Cube view 622, and textual information and user-selectable operation settings regarding the ablation antenna and the ablation zone are displayed in the target window.


The user-selectable operation settings include a zone chart setting 602, a power setting 604, and an antenna type setting 608. The textual information includes an ablation time 606, ablation zone information 605, an insertion depth 607, and a tip distance 609. The ablation zone information 605 includes the width (x), the height (y), the volume, the min margin, and the max margin of the ablation zone. In the example of FIG. 6A, the initial time 606 is 8 minutes and 30 seconds, the initial width is 3.5 cm, the initial height is 4.0 cm, the initial volume is 25.75 cm3, the initial min margin is 1.0 cm, the initial max margin is 1.5 cm, the initial insertion depth 607 is “--”, and the initial tip distance 609 is 0.5 cm.


The zone chart setting 602, the power setting 604, and the antenna type setting 608 are each configured to receive user input though a pull-down menu. The pull-down menu for the zone chart setting 602 may include a “Lung (ex vivo)” menu item and a “Lung (in vivo)” menu item that a user can select between. The pull-down menu for the power level setting 604 may include a “45 W” menu item, a “75 W” menu item, and a “100 W” menu item. In some embodiments, the pull-down menu for the power level setting 604 may include other menu items corresponding to other power levels, depending on the type of microwave generator that is being used or other factors. In other embodiments, the power level setting 604 may be set via a text field in which the user can enter a power level value. The pull-down menu for the antenna type setting 608 may include a “15 cm” menu item, a “20 cm” menu item, and a “30 cm” menu item.


As shown in FIG. 6A, the zone chart setting 602 is initially set to “Lung (ex vivo)”, the power setting is initially set to 100 W, and the antenna type setting 608 is initially set to “Emprint 20 cm”. The textual information including the time information 606 is configured to change based on the user selecting one or more of the user-selectable menu items, the user changing the size or position of the ablation zone, or the user changing the position of the antenna. This allows the user to visualize an ablation procedure if one or more parameters of the ablation procedure are changed. In this way, the user can determine the optimal parameters for the ablation procedure to be performed.



FIG. 6B illustrates changes to the user interface of FIG. 6A as a result of the user changing the settings or parameters of the surgical procedure. The user may change the settings or parameters, for example, to avoid heating nearby anatomical structures. As shown in FIG. 6B, the zone chart setting 602 is changed from “Lung (ex vivo)” to “Lung (in vivo)”, the power setting is changed from 100 W to 45 W, and the antenna type setting 608 is changed from “Emprint 20 cm” to “Emprint 15 cm”. Also, the position, size, and shape of the ablation zone 612 is changed as a result of the user manipulating the user-movable points 614 and 616 with a user input device such as a mouse. In particular, the user changed the position of the ablation zone 612 by moving point 614 to a different position and the user changed the size and shape of the ablation zone 612 by moving one or more of the points 616 to different positions in one or more of the views of FIGS. 6A and 6B. Further, the antenna is repositioned by selecting and moving the point 617. As shown in FIG. 6B, the point 617 is moved to indicate the outside surface of the patients' body.


As a result the changes in the settings, the time information 606 changed from 8 minutes and 30 seconds to 2 minutes and 30 seconds, the width of the ablation zone changed from 3.5 cm to 2.7 cm, the height changed from 4.0 cm 3.2 cm, the volume changed from 25.75 cm3 to 12.30 cm3, the min margin changed from 1.0 cm to 0.5 cm, the max margin changed from 1.5 cm to 1.2 cm, the insertion depth 607 changed from “--” to 6.8 cm, and the tip distance 609 changed from 0.5 cm to 0.3 cm. In the example of FIG. 6B, the user was able to change the settings so that the ablation zone would not intersect with nearby anatomical features. In this way, the user interface of FIGS. 6A and 6B allow a user to change various settings or parameters of an ablation procedure while viewing the axial, coronal, sagittal, and 4 cm Cube views of the patient's relevant anatomy. When a user selects the “3D” icon, a cross-sectional, three-dimensional, skin view of the patient's entire body is shown in FIG. 7. This view may help the clinician to plan an ablation procedure in the context of the entire body of the patient.



FIG. 7 shows the ablation antenna inserted into the patient's lung according to the parameters set in FIG. 6B. FIG. 7 also shows the target as a solid three-dimensional sphere and shows the ablation zone as a three-dimensional wireframe. In the three-dimensional view of FIG. 7, the user may switch the user interface from the skin view to a muscle view, a bone view, or a lung view, by selecting the corresponding icon.


In embodiments, images of the various display interfaces of FIGS. 4-7 may be captured by selecting the camera icon. The clinician may capture images of the display interfaces and save them in a patient's electronic medical records so that the clinician may later reference the images in preparation for or during an ablation procedure.



FIG. 8 is an exemplary display interface illustrating the bounds of a selectively adjustable ablation zone in an axial view in accordance with aspects of this disclosure. As described above, user movable points are provided to enable a user to change the size of the ablation zone. However, the user interface places bounds on the size of the ablation zone to reflect the actual capabilities of the ablation antenna. As illustrated in FIG. 8, the ablation zone has a lower bound 802, a middle bound 804, and an upper bound 806. The ablation zone bounds may depend on, among other things, the available power level settings. For example, the upper bound 806 may relate to a 100 W power level setting, the middle bound 804 may relate to a 75 W power level setting, and the lower bound 802, may relate to a 45 W setting.



FIG. 9 is an exemplary display interface for showing a temperature profile within an ablation zone in accordance with aspects of this disclosure. The display interface may show a target with a circle 902 of a first color (e.g., green), an ablation zone with a circle 904 of a second color (e.g., red) different than the first color, and an area 906 within the circle 904 may be filled in with a range of colors (e.g., yellow, green, blue) representing a temperature profile when the ablation antenna is activated.



FIG. 10 is an exemplary display interface showing a temperature profile represented by a temperature gradient 1002 in a multi-plane view of a patient's liver region in accordance with aspects of this disclosure. More specifically, FIG. 10 shows an ablation antenna inserted into a liver 1000 and a circle 1004 representing an ablation zone about a distal portion of the ablation antenna and surrounding an ablation target. The circle 1004 is surrounded by the temperature gradient 1002 to demonstrate projected temperature levels on structures and/or portions of the liver 1000 that neighbor the ablation target. In aspects of this disclosure, the temperature gradient 1002 may be animated and be represented by any one or more colors.



FIG. 11 is an exemplary display interface for showing a temperature profile on a three-dimensional view of a liver 1010 during a simulated or actual ablation procedure in accordance with aspects of this disclosure. A temperature profile may be represented by a range of colors 1020 (e.g., red, orange, yellow, green, blue) on the surface of the liver 1010.



FIG. 12 is an exemplary display interface for showing a temperature profile on a three-dimensional view of a lung 1030 during a simulated or actual ablation procedure in accordance with aspects of this disclosure. A temperature profile may be represented by a range of colors 1040 on the surface of various portions of the lung 1030 to demonstrate projected temperature levels on structures and/or portions of the lung 1030 that neighbor an ablation target.



FIGS. 13 and 14 are exemplary display interfaces illustrating the bounds of an ablation zone in a multi-plane (e.g., axial, coronal, sagittal) view in accordance with aspects of this disclosure. The user interface places bounds on the size of the ablation zone to reflect the actual capabilities of the ablation antenna. As illustrated in FIG. 13, the ablation zone has a lower bound 1050 and an upper bound 1052. As illustrated in FIG. 14, one or more of the multi-plane (e.g., axial, coronal, sagittal) views illustrates an average bound 1060 on the size of the ablation zone, which represents an average of a lower bound 1062 and an upper bound 1064 illustrated in a detailed view of the user interface (e.g., in the upper right window of the user interface). The ablation zone bounds may depend on, among other things, the available power level settings. For example, the upper bounds 1052, 1064 may relate to a 150 W power level setting and the lower bounds 802, 1062 may relate to a 75 W setting.



FIG. 15 is a flow diagram of an exemplary operation of a system in accordance with aspects of this disclosure. The system may be operated as part of a planning phase or as part of an actual ablation procedure. At block 1102, a three-dimensional image of at least one anatomical feature is displayed. For example, a three-dimensional image of a liver may be displayed. At block 1104, user input of a target for placing an ablation needle in the at least one anatomical feature of the patient is received. The user input may include a user manipulating a mouse to select a location on the three-dimensional image. At block 1106, the position and orientation of the ablation needle is determined based on the user input. At block 1108, an image of a virtual ablation needle is displayed in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation. At block 1110, user input of one or more parameters of operating the ablation needle are received. The one or more parameters may include the type of ablation needle or the power level. And, at block 1112, a three-dimensional representation of the result of operating the ablation needle according to the input parameters is displayed.


While several embodiments of the disclosure have been shown in the drawings, it is not intended that the disclosure be limited thereto, as it is intended that the disclosure be as broad in scope as the art will allow and that the specification be read likewise. Any combination of the above embodiments is also envisioned and is within the scope of the appended claims. Therefore, the above description should not be construed as limiting, but merely as exemplifications of particular embodiments. Those skilled in the art will envision other modifications within the scope and spirit of the claims appended hereto. For example, while this disclosure makes reference to some parameters relevant to an ablation procedure, this disclosure contemplates other parameters that may be helpful in planning for or carrying out an ablation procedure including a type of microwave generator, a power-level profile, or a property of the tissue being ablated.

Claims
  • 1. A method of performing an ablation procedure, comprising: displaying three-dimensional image data of at least one anatomical feature of a patient;receiving user input of an ablation target for placing an ablation needle in the at least one anatomical feature of the patient;determining a position and orientation of the ablation needle based on the user input;displaying an image of a virtual ablation needle in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation of the ablation needle;receiving user input of parameters of operating the ablation needle;displaying a three-dimensional representation of a projected ablation zone surrounding the ablation target and surrounding a distal portion of the virtual ablation needle, wherein the representation of the projected ablation zone is generated in response to the received user input of parameters and the determined position and orientation of the ablation needle; anddisplaying a three-dimensional representation of a projected temperature gradient extending radially outward from an outer border of the representation of the projected ablation zone, the projected temperature gradient surrounding the representation of the projected ablation zone, the distal portion of the virtual ablation needle, and the ablation target, wherein the representation of the projected temperature gradient demonstrates a projected temperature level of tissue in proximity to the ablation target that would result from operation of the ablation needle according to the received user input of parameters.
  • 2. The method of claim 1, further comprising: receiving further user input of movement of the ablation needle in the at least one anatomical feature of the patient;determining a new position and orientation of the ablation needle based on the further user input; anddisplaying the image of the virtual ablation needle in the three-dimensional image data according to the determined new position and orientation.
  • 3. The method of claim 1, further comprising: receiving user input of other different parameters of operating the ablation needle; anddisplaying the three-dimensional representation of the projected ablation zone that would result from operation of the ablation needle according to the other different parameters.
  • 4. The method of claim 1, further comprising: receiving user input adjusting a size of the representation of the projected ablation zone;calculating ablation time based on the adjusted size of the representation of the projected ablation zone; anddisplaying the calculated ablation time.
  • 5. The method of claim 1, wherein the parameters of operating the ablation needle include a power level and an ablation needle type.
  • 6. The method of claim 1, further comprising displaying a three-dimensional representation of at least one of a temperature profile, potential histological zones, a plurality of temperatures, confidence intervals, a heated zone, or probability of cell death with respect to the position of the ablation needle.
  • 7. The method of claim 1, wherein displaying the three-dimensional image data includes displaying a multi-plane view including a coronal view, a sagittal view, and an axial view of the three-dimensional image data.
  • 8. The method of claim 1, further comprising displaying user-selectable icons for selecting a skin view, a muscle view, or a bone view of at least a portion of the patient.
  • 9. The method of claim 1, further comprising displaying a single-slice view including a coronal view, a sagittal view, or an axial view; and displaying user-selectable icons for selecting between the coronal view, the sagittal view, or the axial view.
  • 10. The method of claim 1, further comprising: simultaneously displaying a coronal view, a sagittal view, and an axial view of the three-dimensional image data;displaying a cross-section line on at least one of the coronal view, the sagittal view, or the axial view; anddisplaying a slice corresponding to the position of the cross-section line.
  • 11. The method of claim 1, wherein the three-dimensional image data is computed tomography image data, magnetic resonance image data, or ultrasound image data.
  • 12. The method of claim 1, further comprising: displaying a snapshot button;receiving user selection of the snapshot button; andrecording an image including an image of the anatomy, the target, the representation of the projected ablation zone, and text indicating a power level and an ablation time.
  • 13. The method of claim 1, wherein displaying the three-dimensional image data includes displaying a three-dimensional view of the ablation needle and the projected ablation zone.
  • 14. An ablation visualization system, comprising: a display;a processor; anda memory having stored thereon instructions, which, when executed by the processor, cause the processor to: display, on the display, three-dimensional image data of at least one anatomical feature of a patient on the display;prompt a user for input of an ablation target for placing an ablation needle in the at least one anatomical feature of the patient;receive user input of the ablation target for placing the ablation needle in the at least one anatomical feature of the patient;determine a position and orientation of the ablation needle based on the user input;display, on the display, an image of a virtual ablation needle in the three-dimensional image data of the at least one anatomical feature of the patient according to the determined position and orientation of the ablation needle;prompt the user for input of parameters of operating the ablation needle;receive user input of parameters of operating the ablation needle;display, on the display, a three-dimensional representation of a projected ablation zone surrounding the ablation target and surrounding a distal portion of the virtual ablation needle, wherein the representation of the projected ablation zone is generated in response to the received user input of parameters and the determined position and orientation of the ablation needle; anddisplay, on the display, a three-dimensional representation of a projected temperature gradient extending radially outward from an outer border of the representation of the projected ablation zone, the projected temperature gradient surrounding the representation of the projected ablation zone, the distal portion of the virtual ablation needle, and the ablation target, wherein the representation of the projected temperature gradient demonstrates a projected temperature level of tissue in proximity to the ablation target that would result from operation of the ablation needle according to the received user input of parameters.
  • 15. The ablation visualization system of claim 14, wherein the instructions, when executed by the processor, further cause the processor to: prompt for further user input of movement of the ablation needle in the at least one anatomical feature of the patient;receive further user input of movement of the ablation needle in the at least one anatomical feature of the patient;determine a new position and orientation of the ablation needle based on the further user input; anddisplay the image of the virtual ablation needle in the three-dimensional image data according to the new position and orientation.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of the filing date of provisional U.S. Patent Application No. 62/717,041, filed Aug. 10, 2018, the entire contents of each of which are incorporated herein by reference.

US Referenced Citations (217)
Number Name Date Kind
5515160 Schulz et al. May 1996 A
5528699 Obata et al. Jun 1996 A
5776062 Nields Jul 1998 A
5788636 Curley Aug 1998 A
5799099 Wang et al. Aug 1998 A
5810008 Dekel et al. Sep 1998 A
5817022 Vesely Oct 1998 A
5825908 Pieper et al. Oct 1998 A
5836954 Heilbrun et al. Nov 1998 A
5842473 Fenster et al. Dec 1998 A
5873822 Ferre et al. Feb 1999 A
5891030 Johnson et al. Apr 1999 A
5902239 Buurman May 1999 A
5953013 Shimizu Sep 1999 A
5954648 Van Der Brug Sep 1999 A
5957844 Dekel et al. Sep 1999 A
5967980 Ferre et al. Oct 1999 A
6002808 Freeman Dec 1999 A
6006126 Cosman Dec 1999 A
6019724 Gronningsaeter et al. Feb 2000 A
6052477 Wang et al. Apr 2000 A
6081577 Webber Jun 2000 A
6112112 Gilhuijs et al. Aug 2000 A
6112113 Van Der Brug et al. Aug 2000 A
6119033 Spigelman et al. Sep 2000 A
6165181 Heilbrun et al. Dec 2000 A
6167296 Shahidi Dec 2000 A
6195444 Simanovsky et al. Feb 2001 B1
6203497 Dekel et al. Mar 2001 B1
6216029 Paltieli Apr 2001 B1
6226542 Reisfeld May 2001 B1
6259943 Cosman et al. Jul 2001 B1
6285902 Kienzle, III et al. Sep 2001 B1
6298262 Franck et al. Oct 2001 B1
6301495 Gueziec et al. Oct 2001 B1
6332089 Acker et al. Dec 2001 B1
6334847 Fenster et al. Jan 2002 B1
6338716 Hossack et al. Jan 2002 B1
6341231 Ferre et al. Jan 2002 B1
6343936 Kaufman et al. Feb 2002 B1
6379302 Kessman et al. Apr 2002 B1
6381483 Hareyama et al. Apr 2002 B1
6440071 Slayton et al. Aug 2002 B1
6442417 Shahidi et al. Aug 2002 B1
6466815 Saito et al. Oct 2002 B1
6470207 Simon et al. Oct 2002 B1
6477275 Melikian et al. Nov 2002 B1
6487432 Slack Nov 2002 B2
6505065 Yanof et al. Jan 2003 B1
6529758 Shahidi Mar 2003 B2
6539247 Spetz Mar 2003 B2
6540679 Slayton et al. Apr 2003 B2
6546279 Bova et al. Apr 2003 B1
6553152 Miller et al. Apr 2003 B1
6574493 Rasche et al. Jun 2003 B2
6612980 Chen et al. Sep 2003 B2
6669635 Kessman et al. Dec 2003 B2
6675032 Chen et al. Jan 2004 B2
6694163 Vining Feb 2004 B1
6711429 Gilboa et al. Mar 2004 B1
6724930 Kosaka et al. Apr 2004 B1
6731966 Spigelman et al. May 2004 B1
6733458 Steins et al. May 2004 B1
6751361 Wagman Jun 2004 B1
6754374 Miller et al. Jun 2004 B1
6772002 Schmidt et al. Aug 2004 B2
6812933 Silver Nov 2004 B1
6892090 Verard et al. May 2005 B2
6909913 Vining Jun 2005 B2
6920347 Simon et al. Jul 2005 B2
6925319 McKinnon Aug 2005 B2
6947786 Simon et al. Sep 2005 B2
6961405 Scherch Nov 2005 B2
6968224 Kessman et al. Nov 2005 B2
6969352 Chiang et al. Nov 2005 B2
6973202 Mostafavi Dec 2005 B2
7035461 Luo et al. Apr 2006 B2
7043055 Silver May 2006 B1
7043064 Paik et al. May 2006 B2
7050845 Vilsmeier May 2006 B2
7161596 Hoile Jan 2007 B2
7171255 Holupka et al. Jan 2007 B2
7204254 Riaziat et al. Apr 2007 B2
7215990 Feussner et al. May 2007 B2
7251352 Sauer et al. Jul 2007 B2
7259762 Tanacs et al. Aug 2007 B2
7302288 Schellenberg Nov 2007 B1
7333644 Jerebko et al. Feb 2008 B2
7343026 Niwa et al. Mar 2008 B2
7379572 Yoshida et al. May 2008 B2
7383073 Abovitz et al. Jun 2008 B1
7450749 Rouet et al. Nov 2008 B2
7452357 Vlegele et al. Nov 2008 B2
7457443 Persky Nov 2008 B2
7491198 Kockro Feb 2009 B2
7492930 Leitner et al. Feb 2009 B2
7496173 Goldman et al. Feb 2009 B2
7499743 Vass et al. Mar 2009 B2
7519218 Takemoto et al. Apr 2009 B2
7536041 Pekar et al. May 2009 B2
7567697 Mostafavi Jul 2009 B2
7570987 Raabe et al. Aug 2009 B2
7581191 Rice et al. Aug 2009 B2
7593505 Saracen et al. Sep 2009 B2
7623250 Moctezuma de la Barrera et al. Nov 2009 B2
7630753 Simon et al. Dec 2009 B2
7636420 Spies et al. Dec 2009 B2
7639853 Olivera et al. Dec 2009 B2
7643663 Wiemker et al. Jan 2010 B2
7672705 Lachaine et al. Mar 2010 B2
7689019 Boese et al. Mar 2010 B2
7780084 Zhang et al. Aug 2010 B2
7809184 Neubauer et al. Oct 2010 B2
7831082 Holsing et al. Nov 2010 B2
7844087 Ray et al. Nov 2010 B2
7853305 Simon et al. Dec 2010 B2
7856130 Suri et al. Dec 2010 B2
7860331 Lal et al. Dec 2010 B2
7860548 McIntyre et al. Dec 2010 B2
7873400 Moctezuma De La Barrera et al. Jan 2011 B2
7874987 Altmann et al. Jan 2011 B2
7876937 Schildkraut et al. Jan 2011 B2
7876939 Yankelevitz et al. Jan 2011 B2
7876942 Gilboa Jan 2011 B2
7892224 Hartlep et al. Feb 2011 B2
7894663 Berg et al. Feb 2011 B2
7899513 Phillips et al. Mar 2011 B2
7907772 Wang et al. Mar 2011 B2
7912258 Warmath et al. Mar 2011 B2
7916918 Suri et al. Mar 2011 B2
7920911 Hoshino et al. Apr 2011 B2
7953265 Sirohey et al. May 2011 B2
7957572 Von Berg et al. Jun 2011 B2
7970174 Goldbach Jun 2011 B2
8000442 Lachaine et al. Aug 2011 B2
8010180 Quaid et al. Aug 2011 B2
8019133 Knoplioch et al. Sep 2011 B2
8023712 Ikuma et al. Sep 2011 B2
8023734 Jolly et al. Sep 2011 B2
8036435 Partain et al. Oct 2011 B2
8045778 Blaffert et al. Oct 2011 B2
8046052 Verard et al. Oct 2011 B2
8750568 Frank et al. Jun 2014 B2
9439622 Case et al. Sep 2016 B2
9439623 Frank et al. Sep 2016 B2
9439627 Case et al. Sep 2016 B2
9498182 Case et al. Nov 2016 B2
20010027272 Saito et al. Oct 2001 A1
20030135115 Burdette et al. Jul 2003 A1
20030151665 Uchiyama Aug 2003 A1
20030163142 Paltieli et al. Aug 2003 A1
20040015070 Liang et al. Jan 2004 A1
20040034297 Darrow et al. Feb 2004 A1
20040138555 Krag et al. Jul 2004 A1
20050215854 Ozaki et al. Sep 2005 A1
20060229594 Francischelli et al. Oct 2006 A1
20070238961 Vilsmeier et al. Oct 2007 A1
20080063136 Ohyu et al. Mar 2008 A1
20080081982 Simon et al. Apr 2008 A1
20080097186 Biglieri et al. Apr 2008 A1
20080119712 Lloyd May 2008 A1
20080123921 Gielen et al. May 2008 A1
20080123927 Miga et al. May 2008 A1
20080167547 Bova et al. Jul 2008 A1
20080200794 Teichman et al. Aug 2008 A1
20080200926 Verard et al. Aug 2008 A1
20080200927 Hartmann et al. Aug 2008 A1
20080208041 Gilboa Aug 2008 A1
20080214922 Hartmann et al. Sep 2008 A1
20080221650 Turner et al. Sep 2008 A1
20080232656 Voegele Sep 2008 A1
20080242978 Simon et al. Oct 2008 A1
20080262345 Fichtinger et al. Oct 2008 A1
20080285854 Kotake et al. Nov 2008 A1
20090028436 Yoshino et al. Jan 2009 A1
20090124896 Haras May 2009 A1
20090198126 Klingenbeck-Regn Aug 2009 A1
20090221908 Glossop Sep 2009 A1
20090292201 Kruecker Nov 2009 A1
20090312629 Razzaque et al. Dec 2009 A1
20100063392 Nishina et al. Mar 2010 A1
20100063496 Trovato et al. Mar 2010 A1
20100076305 Maier-Hein et al. Mar 2010 A1
20100121189 Ma et al. May 2010 A1
20100121190 Pagoulatos et al. May 2010 A1
20100168763 Zhao et al. Jul 2010 A1
20100179529 Podhajsky et al. Jul 2010 A1
20100208963 Kruecker et al. Aug 2010 A1
20100217117 Glossop et al. Aug 2010 A1
20100249771 Pearson et al. Sep 2010 A1
20100250209 Pearson et al. Sep 2010 A1
20100259474 Hildreth Oct 2010 A1
20100268223 Coe et al. Oct 2010 A1
20100274124 Jascob et al. Oct 2010 A1
20100277655 Sun Nov 2010 A1
20100295931 Schmidt Nov 2010 A1
20100298705 Pelissier et al. Nov 2010 A1
20100312103 Gorek et al. Dec 2010 A1
20100322489 Tizhoosh et al. Dec 2010 A1
20110015628 Dalal et al. Jan 2011 A1
20110118596 Vining et al. May 2011 A1
20110129154 Shimodaira Jun 2011 A1
20110137156 Razzaque et al. Jun 2011 A1
20110137168 Lee et al. Jun 2011 A1
20110160569 Cohen et al. Jun 2011 A1
20110251483 Razzaque et al. Oct 2011 A1
20120050258 Kay et al. Mar 2012 A1
20120136242 Qi et al. May 2012 A1
20120277585 Koenig et al. Nov 2012 A1
20130142410 Dwivedi et al. Jun 2013 A1
20130316318 Frank et al. Nov 2013 A1
20140206988 Ramachandran Jul 2014 A1
20150282786 Anand Oct 2015 A1
20160038247 Bharadwaj Feb 2016 A1
20160038248 Bharadwaj Feb 2016 A1
20180008341 Brannan Jan 2018 A1
20180028267 Onik Feb 2018 A1
Foreign Referenced Citations (15)
Number Date Country
1788693 Jun 2006 CN
101249001 Aug 2008 CN
201422889 Mar 2010 CN
10 2008 013611 Sep 2009 DE
1571581 Sep 2005 EP
1649822 Apr 2006 EP
9515729 Jun 1995 WO
9703609 Feb 1997 WO
0139124 May 2001 WO
2006089426 Aug 2006 WO
2007113815 Oct 2007 WO
2008017051 Feb 2008 WO
2008058520 May 2008 WO
2012025842 Mar 2012 WO
2012066446 May 2012 WO
Non-Patent Literature Citations (3)
Entry
Mathias Markert “Development of an assistance system for open liver surgery usable in clinical interventions” submitted Jan. 31, 2011, <http://mediatum.ub.tum.de?id=1007285> .
International Search Report and Written Opinion of the International Searching Authority issued in corresponding Appl. No. PCT/US2019/046152 dated Oct. 31, 2019 (16 pages).
Kosaka A. et al. “Augmented Reality System for Surgical Navigation Using Robust Target Vision”, Proceedings 2000 IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000. Hilton Head Island, SC, Jun. 13-15, 2000, pp. 187-194. cited by applicant.
Related Publications (1)
Number Date Country
20200046435 A1 Feb 2020 US
Provisional Applications (1)
Number Date Country
62717041 Aug 2018 US