Motion compensation for medical imaging and associated systems and methods

Information

  • Patent Grant
  • 9237860
  • Patent Number
    9,237,860
  • Date Filed
    Friday, June 5, 2009
    16 years ago
  • Date Issued
    Tuesday, January 19, 2016
    9 years ago
Abstract
Medical imaging and localization methods and systems for producing a motion-compensated image of a planning target volume (PTV) of a patient. In one embodiment, an imaging and localization system includes sensors that are positioned to receive an electromagnetic location signal from one or more active markers affixed to or adjacent a PTV. A signal processing component can produce real-time localization data corresponding to the location signal, and a system interface can receive such localization data. The system interface can also receive raw image data from an imaging subsystem and process the raw image data based on the localization data. For example, the imaging subsystem can include a computed tomography (CT) imaging system and image slices or frames can be binned based on the localization data.
Description
TECHNICAL FIELD

The present disclosure is generally related to systems and methods for compensating for patient motion in medical imaging applications.


BACKGROUND

In medical imaging of the thoracic cavity, intrafraction movement can have a significant effect on medical imaging quality. Intrafraction movement occurs when the patient and/or organs within the thoracic cavity (e.g., lungs, heart, etc.) move during an imaging scan. Respiratory movement, in particular, can have a significant affect on image quality and can produce a variety of artifacts (e.g., blur, distortion, etc.) within a medical image. These imaging artifacts can in turn affect the extent to which medical practitioners can rely on a medical image for diagnoses or for treatment planning. For example, in order to compensate for any uncertainties created by imaging artifacts, radiologists may need to over estimate a target treatment volume in a patient.


In general, there exist several techniques to account for respiratory motion in medical imaging. For example, these techniques can including breath-hold, respiratory gating or breathing coaching, which employs visual and/or audio feedback regarding when a patient should inhale and/or exhale. By controlling a patient's breathing in either of these manners, the amount of image artifacts can be reduced to an extent. Unfortunately, however, such breath-hold and coaching techniques are limited. For example, only about 40% of the lung cancer population can hold their breath for an adequate duration of time, and breathing coaching cannot account for irregular motion of the lungs (or irregular motion of tissues adjacent the lungs).





BRIEF DESCRIPTION OF THE DRAWINGS

In the drawings, identical reference numbers identify similar elements or acts. The sizes and relative positions of elements in the drawings are not necessarily drawn to scale. For example, the shapes of various elements and angles are not drawn to scale, and some of these elements are arbitrarily enlarged and positioned to improve drawing legibility. Further, the particular shapes of the elements as drawn are not intended to convey any information regarding the actual shape of the particular elements, and have been solely selected for ease of recognition in the drawings.



FIGS. 1A and 1B are isometric, schematic diagrams of an imaging and localization system configured in accordance with an embodiment of the disclosure.



FIG. 2 is an isometric, schematic diagram of the system of FIG. 1 showing an embodiment of an imaging subsystem in more detail.



FIG. 3 is an isometric, schematic diagram of the system of FIG. 1 showing an embodiment of a localization subsystem in more detail.



FIG. 4 is a block diagram of the system of FIG. 1 showing an embodiment of a system interface in more detail.



FIG. 5 is a flow diagram of a method of operating the system of FIG. 1.



FIG. 6 is a flow diagram of an embodiment of a method for binning or filtering raw image data of an imaging subsystem based on real-time localization data of a localization subsystem.



FIGS. 7-14 are various plots and images showing results from an experimental study that evaluated the performance of motion-compensated computed tomography (CT) imaging.





DETAILED DESCRIPTION

A. System Overview


The following disclosure is directed towards compensating for intrafraction movement in medical imaging systems, for example, compensating for respiratory motion in 4-dimensional computed tomography (CT) imaging systems, including axial CT, helical CT, and ciné CT imaging systems. Well-known characteristics often associated with medical imaging systems and associated signal/data processing have not been shown or described in detail to avoid unnecessarily obscuring the description of the various embodiments. Those of ordinary skill in the relevant art will understand that additional embodiments may be practiced without several of the details described below, and that other embodiments may include aspects in addition to those described below.



FIG. 1A is a schematic diagram illustrating a representative embodiment of an imaging and localization system 100. The system 100 can include an imaging subsystem 110, a localization subsystem 140, and a system interface 170 operably coupled to the imaging and localization subsystems 110 and 140. In operation, the imaging subsystem 110 produces raw image data by scanning a planning target volume (PTV) within a patient. The raw image data can include image slices or image frames (e.g., raw image data computed from multiple image slices). The PTV, for example, can be a portion of an organ and/or tissue within the thoracic cavity of the patient (e.g., a lung, tumor, etc.).


The localization subsystem 140, in operation, employs one or more (implantable) active markers 142 adjacent the PTV that can be tracked or monitored. The active markers, for example, are electrically active, generating detectable electromagnetic pulses when excited by an external source (described further with reference to FIG. 3). When positioned within range of individual markers 142, the localization system 140 can track the real-time location of these markers, and thus can also track the real-time location of the PTV. In many embodiments, and described in more detail below, the localization subsystem 140 tracks the real-time location of the PTV, while the imaging subsystem 110 concurrently scans the PTV.



FIG. 1B is a schematic diagram illustrating the system 100 and, more specifically, the localization system 140 and the patient (including the PTV) positioned within the imaging system 110. When positioned in such a manner, the imaging and localization subsystems 110 and 140 can concurrently produce and transmit both raw image data of the PTV and real-time localization data regarding the PTV to the system interface 170. The system interface 170 (e.g., a computer and/or operator station) can in turn (a) receive the raw image data and the real-time localization data, and (b) filter, bin, or otherwise process the raw image data based on the real-time localization data. In particular, the system interface 170 can yield motion-compensated image data that is synchronized with the real-time location of the PTV. Such synchronization mitigates or eliminates imaging artifacts in the image data, resulting in high quality image data that can be used or evaluated in combination with a variety of diagnostic and/or treatment planning procedures.


In contrast to the system 100, conventional imaging systems are limited by conventional techniques for (intrafraction) motion compensation. For example, as described above, breath-hold, respiratory gating can only be used on a minority of the cancer patient population, and breathing coaching can at best only account for predicable types of respiratory motion. Furthermore, other types of motion compensation techniques can only provide coarse approximations of a PTV location. For example, one conventional technique employs surrogate markers that are externally affixed to a patient's chest or abdomen, at a location that roughly corresponds with the internal location of a PTV. An Infrared radiation source can be used to illuminate the surrogate markers, and a video camera can use reflected radiation to track the movement of the surrogate. Although providing a rough approximation of respiratory motion, surrogate markers are not directly located at the PTV, and thus they are limited in accuracy. For example, during a patient's breathing cycle, the distance between the PTV and the surrogate markers can vary non-linearly and/or unpredictably. Surrogate markers cannot account for these non-linear and unpredictable types of motion. Consequently, in many surrogate marker techniques, the localization data is a predictive measure that determines a suitable gating window. For example, such a gating window can predict the times at which a patient's lungs should be in an inhaled or exhaled state, and raw image data can accordingly be gathered when the lungs are in either of these states. However, if a patient's breathing deviates from this predicted cycle, the raw image data loses accuracy.


Another conventional motion compensation technique includes implanting opaque fiducials at a PTV and correlating fiducial motion with the motion of surrogate markers. For example, a PTV can be imaged for a first period of time, and the location of the fiducials can be separately assessed during a second period of time. This process can be repeated until a motion map of the PTV is acquired, for example, by imaging once every 10 seconds and monitoring the fiducials between the imaging intervals. Generally, this type of motion compensation technique is complicated, has time-intensive set-up procedures, and is only effective for a minority of patients. Furthermore, the fiducial motion is not captured in real-time, nor does the fiducial motion represent accurate organ or tissue motion over relatively short time intervals (e.g., for intervals that are less than 10 seconds).


Embodiments of the imaging and localization system 100, however, overcome these and other limitations associated with conventional motion compensation techniques. For example, the raw image data of the system 100 can be processed based on the real-time PTV position and not exclusively on the predicted or approximated position of the PTV (i.e., derived by surrogate markers or from breath coaching techniques). Accordingly, the system 100 can account for a variety of non-linearities and unpredictable motion of the PTV. In addition, individual markers can transmit a unique electromagnetic signal, allowing the system interface 130 to identify and distinguish between individual markers. Surrogate markers and opaque fiducials, on the other hand, can only be discriminated optically with complicated and manually-intensive imaging procedures. Furthermore, unlike externally positioned surrogate markers, which can be inadvertently repositioned (e.g., by physical contact with the surrogate marker), implantable markers generally remain fixed to soft tissue within a patient and move with motion of soft tissue that they are rigidly fixed to. Accordingly, the implantable markers can reduce or eliminate patient-setup positioning errors.


B. System Components and Operation



FIGS. 2-4 are schematic diagrams illustrating, respectively, the imaging subsystem 110, the localization subsystem 140, and the system interface 170 of the imaging and localization system 100 in more detail. Referring to FIG. 2, the imaging subsystem 110 can include a gantry 212 having a gantry bore 214, a table or couch 216 for carrying the patient and positioning the patient within the gantry bore 214, and one or more imaging device(s) 218 (drawn in phantom) disposed at or within the gantry 212. For example, the imaging device(s) 218 can include an X-ray source (e.g., a linear accelerator) that freely rotates about a central axis 220 of the gantry 220, as well as X-ray detectors that also rotate about the axis 220 or are stationary.


The imaging subsystem 110 also includes a signal processing unit 230 (drawn in phantom) that is operably coupled with the imaging device(s) 218 for collecting raw image data regarding the PTV of the patient. In general, the signal processing unit 230 can carry out a variety of processing algorithms that assemble or otherwise construct image slices or image frames (based on the image slices), for example, to produce a digitally reconstructed radiograph (DRR). In several embodiments, the signal processing unit 230 can also produce raw image data according to a digital imaging and communication in medicine (DICOM) standard or interface with another system (not shown) that formats data into such a standard. In addition, in many examples, the signal processing unit 230 is also operably coupled with the gantry 212 and/or the table 216 for positioning the imaging device(s) 218 with respect to the PTV. For example, the signal processing unit 230 can be configured to control the rate at which the imaging device(s) 218 rotate and/or the rate at which the table 216 moves through the gantry bore 214. In many embodiments, the configuration of the gantry 212, gantry bore 214, table 216, the imaging device(s) 218, and/or the processing unit 230 is suited for CT-based imaging. However, in other embodiments, other imaging subsystems can include additional or alternative components that are suited for other types of imaging (e.g., fluoroscopic, MRI, PET, or ultrasonic imaging). Furthermore, although not shown in FIG. 2, the signal processing unit 230 can also be coupled with various input/output components (e.g., keyboards, displays, touchscreens, etc.) for allowing an operator to program/control the imaging subsystem 110.


Turning now to FIG. 3, the localization subsystem 140 includes an excitation source 352 and a sensor array 354 of individual sensors or coils 356 spaced apart from one another in a known geometry relative to one another. In many embodiments, the excitation source 352 and the sensor array 354 can be carried by a support apparatus 358, which can in turn position the excitation source 352 and sensor array 354 within the gantry bore 214 (FIG. 2) to detect the location of individual markers 142 (and thus the patient PTV). However, in other embodiments the excitation source 352 and sensor array 354 can be mounted to or otherwise attached to the gantry 212 (FIG. 2). Additional embodiments of the localization subsystem 140 can also include, for example, the excitation sources, sensor arrays, and signal processing components shown and described in U.S. Pat. Nos. 6,977,504 and 7,026,927 and U.S. application Ser. No. 10/749,860, all of which are incorporated herein by reference.


The localization subsystem 140 also includes one or more signal processing/control components 360 that can be integrated into a common assembly with the excitation source 352 and sensor array 354 and/or can be incorporated into a stand-alone operator interface (not shown). In general, the signal processing/control components 360 are configured to (a) use the excitation source 352 to wirelessly deliver a pulsed magnetic field to energize individual markers 142, (b) receive a location signal from the individual markers 142 via the sensor array 354, and (c) periodically calculate a location of the individual markers 142 in a reference frame. Because the individual sensors 356 of the sensor array 354 are spaced apart from one other in a known geometry relative to each other, the absolute location of the individual markers can be determined by a known location (e.g., a known location of one of the sensors 356 or another external sensor). The dimensionality of localization, alignment and/or registration of the PTV can be determined based on the known location and the number of active markers deployed at or adjacent the PTV. For example, a single marker defines a single position in localization space; two markers can be used to register the distance between two points. Furthermore, two independent sets of two markers attached to soft tissue can be used to register and align two vectors relative to each other. And, three markers attached to soft tissue relative to each other can be used to define a plane including rotational angles of the plane (i.e. pitch, yaw and roll). A variety of processing algorithms can provide such a determination, examples of which are set forth in U.S. application Ser. Nos. 10/679,801; 10/749,478; 10/750,456; 10/750,164; 10/750,165; 10/749,860; and 10/750,453, all of which are incorporated herein by reference.


In many applications, active markers can be directly affixed to the PTV or to organs or tissues adjacent the PTV. For example, in applications that image portions of the lungs or lung tumors, a bronchoscope along with visual and fluoroscopic guidance can be used to direct a guide wire to bronchi locations. The guide wire can be used to wedge the individual markers as distally as possible to maximize fixation of the markers within the lungs. In many embodiments, such techniques are minimally invasive, and can position the markers as close as possible to the PTV without employing more invasive techniques. In other embodiments, however, other types of procedures may be used to position the markers at the PTV (e.g., laparoscopic implantation).


Embodiments of the individual marker 142 can include a transponder 343 contained in a biocompatible capsule 344 and having a core 345, a coil 347 around the core 345, and a capacitor 349 electrically coupled to the coil 347. The core 345 is typically composed of ferrite, and the coil 347 includes a plurality of windings of a wire around the core 345. The transponder 343 is a resonating circuit that receives the wirelessly transmitted magnetic excitation energy and produces a wirelessly transmitted location signal in response to the excitation signal. The transponder 343 accordingly has a resonant frequency at which the magnetic excitation energy energizes the transponder 343. In many embodiments, individual markers 142 can have transponders having resonant frequencies that are different from one another (e.g., for signal discrimination). Embodiments of suitable markers can also include, for example, the markers shown and described in U.S. Pat. Nos. 7,135,978, 6,918,919 and 6,822,570; U.S. application Ser. Nos. 10/334,700; 10/679,801; 10/382,123; 10/745,097; 10/746,888; and 11/166,801, all of which are incorporated herein by reference.


Referring now to FIG. 4, the system interface 170 can include communication ports 472 and 474, at least one signal processor 476 operably coupled to the communication ports 472 and 474, and a memory 478 storing processing instructions that are executable by the processor 476. The one or more communication ports 472 are operably coupled with the imaging subsystem 110, the one or more communication ports 474 are operably coupled with the localization subsystem 140, and the signal processor 478 can process data received at these ports (i.e., raw image data and real-time localization data) based on the processing instructions stored at the memory 478. In many embodiments, the system interface 170 also includes input/output components 480 (e.g., a keyboard, mouse, display, printer), which, for example, can allow an operator of the system interface 170 to (a) edit or modify image data, (b) control data processing, and/or (c) view graphical or numeric output corresponding to raw image data, real-time localization data, and/or motion-compensated image data.


Embodiments of the system interface 170 can also include a synchronization link 482 and/or a feedback link 484 between the system interface 170 and the imaging subsystem 110. The imaging subsystem 110 can communicate a synchronization signal over the synchronization link 482, and the system interface 170 can communicate a feedback signal over the feedback link 484. In many examples, the synchronization signal can be used by the system interface 170 to align raw image data with real time localization data. For example, to compensate for respiratory motion, a synchronization signal can create a time indicator as to when a patient's lungs are fully inhaled and/or fully exhaled. Accordingly, the real-time localization data can then be correlated with the phase of a patient's breathing cycle based on this time indicator. The feedback signal, on the other hand, can be used by the system interface 170 to communicate with the imaging subsystem 110. For example, if the PTV location shifts out of detection range (or in some other type of manner), the system interface 170 can use the feedback signal to alert the imaging subsystem 110 that such a shift has occurred. The imaging subsystem 110 can accordingly halt operation or adjust image processing (at the processing unit 230) based on the feedback signal.


In other, non-illustrated embodiments, the system interface 170 can be a simple communication link between the imaging subsystem 110 and localization subsystem 140. In several embodiments, the imaging subsystem 110 and the localization subsystem 140 can be configured to communicate signals and transmit data between one another. For example, the processing unit 230 (FIG. 2) of the imaging subsystem 110 could receive real-time localization data from the localization subsystem 110 and use such data in the processing or assembly of image slices or image frames. Additionally or alternatively, the processing unit 230 could also use the real-time localization data to gate one or more of the imaging devices 218 (FIG. 2).



FIG. 5 is a flow diagram showing an embodiment of operating the imaging and localization system 100. The method first includes implanting one or more active markers at a patient PTV (block 500) and positioning the PTV within proximity of the imaging and localization subsystems (block 502). Next, the localization subsystem tracks the location of the PTV (blocks 510-516) and the imaging subsystem scans (and images) the PTV (block 520 and 522). To track the location of the PTV, the localization subsystem uses the excitation source to excite the individual markers (block 510), and detects the location of the markers by “listening” for the location signal (block 512). The localization subsystem can then generate real-time localization data using these location signals (block 514) and communicate the real-time localization data to the system interface (block 516). Concurrent with the location tracking, the imaging subsystem produces raw image data frames regarding the PTV (block 520) and communicates this data to the system interface (block 522). The system interface, in turn, processes the raw image data based on the real-time localization data (block 530).


In one embodiment, the system interface bins raw image data (or image slices or frames) based on the real-time location of the PTV and outputs motion-compensated image data corresponding to a particular subset of the binned image data. For example, the raw image data can be binned (or filtered) based on a range of locations associated with the PTV (described further with reference to FIG. 6). Additionally or alternatively, the binning of the raw image data can be further based on a specific time or window in patient's breathing cycle (e.g., determined by a synchronization signal). In other embodiments, the system interface can employ other types of image processing. For example, in addition to or in lieu of binning the imaging frames, the system interface can adjust or calibrate the processing algorithm of the imaging subsystem based on the real-time localization data of the localization subsystem. In such an example, the system interface 170 could use the feedback link 484 (FIG. 4) to provide an adjustment or calibration.



FIG. 6 is a flow diagram showing an embodiment of binning raw image data in more detail. The method first includes identifying location datums within the real-time localization data of the localization subsystem (block 600). For example, individual location datums can include a time stamp and spatial coordinates (e.g., 2-dimensional or 3-dimensional) regarding an individual marker. Next, one or more sets of these datums are associated with individual sets of raw image data produced by the imaging subsystem (block 602). The individual sets of raw image data are then binned based on the location datums (block 604). For example, individual sets of raw image data can be binned based on a specific window of time stamp values and/or a specific window of spatial coordinates attributed to each of the sets of raw image data. After binning the individual sets of raw image data, a subset of the raw image data can be output as motion-compensated image data (block 606).


C. Imaging Subsystems


Described below are examples of various imaging subsystems that can be employed with the imaging and localization system 100. In general, the imaging subsystems described below are CT based imaging systems. However, it is contemplated that other types of imaging subsystems can be employed within embodiments of the imaging and localization system 110, for example, such as those that use other types of radiographic and non-radiographic imaging techniques (e.g., digital tomosynthesis, fluoroscopic, MRI, and/or ultrasonic imaging).


In general, CT imaging methods include employing an X-ray source and detector to scan a patient PTV at a target depth. In many examples, portions of the anatomy adjacent a PTV can be blurred-out of an imaging frame by implementing a variable depth of field between an X-ray source and detector. An imaging frame of a PTV, accordingly, is sharp and detailed, and well contrasted against background anatomy. As described above, modern methods of CT imaging are generally carried out by rotating an X-ray source about a central axis of a gantry bore and using a detector (either stationary or rotating) on an opposite side of the PTV to acquire multiple image “slices” of a PTV. These image slices can accordingly be assembled based on a variety of processing algorithms to produce an image frame.


Embodiments of the imaging subsystem can employ a variety of CT image acquisition techniques, including axial, helical or spiral, or cine CT techniques. In axial CT, a patient table is stationary while an X-ray source rotates (and optionally a detector) about the PTV to acquire image slices. The table can be subsequently moved, and additional image slices can be acquired after the table returns to a stationary position. Such slices can be assembled or aggregated to create a volumetric representation of a PTV. Helical CT is similar to axial CT, but moves the patient smoothly through the gantry bore while concurrently rotating the X-ray source. Helical CT scans are advantageous in that a large patient volume can be acquired in about 20-60 seconds. In addition, Helical CT scans do not have an interscan delay that normally occurs when repositioning the table in axial CT. Ciné CT, on the other hand, produces a time sequence of axial images, and is generally used when the temporal nature of a PTV is important (e.g., to evaluate blood flow).


Additionally, other embodiments of CT systems can further include multislice and/or dual source systems. Multislice CT scanners can employ axial or helical imaging techniques, but are configured to include more than one detector. In general, increasing the number of detectors allows increased x-ray source rotation speeds. For example, modern multi-slice CT systems can include up to 256 detectors, creating up to 256 corresponding slices (simultaneously). Dual source CT systems can similarly employ axial or helical imaging techniques, but have an increased temporal resolution attributed to a reduction in rotation angle required to acquire a complete image. In addition, dual source CT can employ dual energy imaging, which can be used to differentiate between anatomy that would otherwise be unobtainable differentiable with a single source CT system. For example, dual source CT systems may be used to enhance the differentiation between tumors and adjacent tissues.


Embodiments of CT systems that can be incorporated into the imaging subsystem can also include, for example, the SOMATOM™ CT systems manufactured and sold by Siemens AG.


D. Compatibility of Localization Subsystem with Radiation Delivery Devices


In many embodiments, the localization subsystem 140 is used in proximity to a radiation delivering imaging device (e.g., a linear accelerator) of the imaging subsystem 110. Accordingly, aspects of the imaging and localization system 100, and particularly, the localization subsystem 140 may be adversely interfered with by the operation of such a radiation delivery device (not only the emitted radiation, but the circuitry of the delivery component itself). Therefore, the system 100 can be adapted to mitigate adverse influences attributed to the radiation delivery device.


As one example, components of the localization subsystem 140 (e.g., the excitation source, sensor array 354, and signal processing components 360) can be arranged so that radiation does not travel through the localization subsystem 140. For example, referring to FIG. 2, the excitation source 353, the sensor array 354 and/or the signal processing components 360 can be mounted to the gantry 212 and configured to rotate concurrently with the imaging device(s) 218 such that components of the localization subsystem 140 are never in the line of site of radiation delivering devices.


As another example, the localization subsystem 140 can include a matched filter or other device that can detect the presence of interference due to the operation of radiation delivering devices, or any other interfering device that operates in a pulsed mode. If such interference is detected, the localization subsystem 140 can be operative to discard received location signals that occurred in that timeframe. Additionally or alternatively, the localization subsystem 140 can also include radiation hardened circuitry that is generally impervious to ionizing radiation (below a threshold level). Referring again to FIG. 2, many of the circuit elements of the excitation source 352, sensor array 354, and/or the signal processing components 360 can be manufactured using a variety of radiation hardened techniques and design rules. For example, there are many known techniques in the aerospace industry for mitigating single event transients (SETs), single event upsets (SEUs), or other soft errors that would otherwise occur in the presence of radiation.


E. Experimental Analysis



FIGS. 7-14 are various plots and images showing results of an experimental study that evaluated the performance of motion-compensated CT imaging. More specifically, the experimental study evaluated tumor correlated computed tomography TCCT imaging on a phantom model of respiratory motion. The goals of this study were to 1) simultaneously electromagnetically track and CT scan active markers, 2) read and synchronize data from the subsystems, and 3) select appropriate images for reconstruction. The experimental study was carried out using an active marker-based localization system, embodiments and examples of which have been described above.


1. Experimental Setup


An active marker was embedded in wax, placed adjacent to a 2 cm diameter acrylic ball and the two objects placed on a 4D Phantom stage. The 4D Phantom was placed on the table of a CT scanner (Brilliance CT Big Bore Scanner—Philips, Andover, Mass.) and moved in a trajectory recorded from a canine breathing study that was performed using variable forced ventilation. The sensor array of the localization subsystem was placed over the 4D Phantom stage in the gantry bore. A real-time location signal was generated that was synchronized to the CT acquisition while the CT scans were acquired and the localization sub-system monitored the active marker position The CT scanner was operated in a ciné axial mode with 15 acquisitions per table position. Each image was created with a 0.44 second rotation time followed by a 0.5 second dead time. Each table position contained 16 contiguous 1.5 mm thick for a total of 2.4 cm width. There was a 2-4 s pause between adjacent table position acquisitions.


2. Localization Subsystem Performance Measured by 4D Phantom


The accuracy of the localization sub-system in a CT scanner environment was determined by comparing the real-time localization data output to 4D Phantom position. FIG. 7 is a plot showing the analysis of motion of an active marker location and the 4D phantom location in the vertical direction. FIG. 8 is a plot showing the localization sub-system's RMS localization error was 0.6 mm, slightly greater than when the localization subsystem was run outside of the CT, but still acceptable for clinical applications.


3. Image Data Processing


Active marker positions were successfully synchronized with the CT images. The longitudinal (in/out of table) position and the associated x-ray on signals from the CT table are shown in FIG. 9. In FIG. 9, the active marker positions at the midpoints of the X Ray On signals are represented by x's superimposed on the active marker position curve, and the change in longitudinal baseline with each set of pulses represent the table movement in between different table positions. FIG. 10 shows the 3D active marker position over the study. The active marker positions at the midpoints of the X Ray On signals are represented by crosses.


The CT image frames were binned with respect to the measured active marker position. A sample transverse image and its corresponding active marker trace are shown in FIG. 11. A feature mapping program was written in Matlab to assess the center of the stage in each image for the fourth table position. FIG. 12 is a plot showing the lateral locations of the stage compared against the localization subsystem-measured lateral coordinates for each of the 240 images acquired at one of the table positions. The correlation coefficient is 0.954, consistent with the precision of the localization subsystem in the CT environment. FIGS. 13 and 14 are plots showing the CT image frame lateral positions and the active marker lateral positions versus time and plotted against each other. Note spread of image locations corresponds to limits of image processing algorithm of the CT scanner.


F. Conclusion


The above description of illustrated embodiments, including what is described in the Abstract, is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Although specific embodiments of and examples are described herein for illustrative purposes, various equivalent modifications can be made without departing from the spirit and scope of the invention, as will be recognized by those skilled in the relevant art.


The various embodiments described above can be combined to provide further embodiments. For example, aspects of the system interface 170 can be omitted or incorporated into one or both of the imaging and location subsystem (e.g., the communication ports 472 and 474, the processor 476, and the memory 478). In addition, in other embodiments, the system interface 470 can be used more generally to provide a communication link between the imaging and localization subsystems 110 and 140. Further, all the U.S. patents, U.S. patent application publications, U.S. patent applications, foreign patents, foreign patent applications, and non-patent publications referred to in this specification and/or listed in the Application Data Sheet are incorporated herein by reference, in their entirety. Aspects of this disclosure and referenced disclosures can be modified, if necessary, to employ systems, devices, and concepts of the various patents, applications, and publications to provide yet further embodiments.


These and other changes can be made to various embodiments in light of the above-detailed description. In general, in the following claims, the terms used should not be construed as limiting the specific embodiments disclosed in the specification and the claims, but should be construed to include all imaging and localization subsystems (and aspect thereof) that operate in accordance with the claims to provide systems and methods for imaging a selected target volume within a patient based on the real-time location of one or more active markers. Accordingly, the invention is not limited, except as by the appended claims.

Claims
  • 1. A system for producing image data regarding a planning target volume (PTV) within a patient's body, the system comprising: at least one marker that is implantable at the PTV within the patient;sensors positioned to receive an electromagnetic location signal that is broadcast by the active marker, the sensors being spaced apart from each other in a known geometry relative to each other;a signal processing component configured to produce real-time localization data corresponding to the location signal based on the known geometry, the real-time localization data including a plurality of time stamps and spatial coordinates associated with a corresponding one of the time stamps; anda system interface operably coupled with the signal processing component, the system interface also being operably coupled with an imaging system that scans the PTV to produce raw image data regarding the PTV, wherein the imaging system uses a processing algorithm to construct at least one of image frames and image slices, andwherein the system interface is configured to concurrently receive both the real-time localization data from the signal processing component and raw image data from the imaging system while it simultaneously scans the PTV,bin the raw image data based on a specific window of the real-time location data while the imaging system is still scanning the PTV,output motion compensated image data corresponding to a particular subset of the binned raw image data,detect a change in location of the PTV out of a detection range based on the real-time localization data, andin response to detecting the change in location of the PTV, direct the imaging system to adjust image processing by calibrating the processing algorithm as it is still scanning the PTV.
  • 2. The system of claim 1 wherein the system interface is further configured to provide a feedback signal to the imaging system, the feedback signal configured to adjust the algorithm for constructing the at least one of the image frames and the image slices.
  • 3. The system of claim 1 wherein the PTV includes a portion of a lung or a tissue adjacent the lung.
  • 4. The system of claim 3 wherein the marker is implanted at the bronchi of a lung.
  • 5. The system of claim 1 wherein the system further comprises an excitation source for transmitting excitation energy, and wherein the marker further comprises a wireless transponder configured to wirelessly transmit the location signal in response to the excitation energy.
  • 6. The system of claim 1 wherein the system interface is operably coupled with a computed tomography (CT) imaging system, the CT imaging system producing the raw image data.
  • 7. A system for creating motion-compensated image data regarding an interior and volumetric portion of a patient, comprising: an imaging subsystem that scans the volumetric portion to produce raw image data regarding the volumetric portion and uses a processing algorithm to construct at least one of image frames and image slices;markers fixable at a position located at the volumetric portion and being excitable to produce an identifiable marker signal, the markers being spaced apart from each other in a known geometry relative to each other;sensors and one or more associated signal processing/control components that detect the identifiable marker signal and output real-time localization data based on the identifiable marker signal and the known geometry while the imaging system scans the volumetric portion, the real-time localization data including a plurality of time stamps and spatial coordinates associated with a corresponding one of the time stamps; anda processor that processes the raw image data by binning the raw image data while the imaging subsystem scans the volumetric portion, the binning being based, at least in part, on a specific window of the time stamps and the corresponding spatial coordinates, the processor being operably coupled with the imaging subsystem, the sensors, and the associated signal processing/control components of the sensors,wherein the processor outputs motion compensated image data corresponding to a particular subset of the binned raw image data, andwherein the processor detects a change in location of at least one of the active markers out of a detection range based on the spatial coordinates the processor receives while the imaging system simultaneously scans the volumetric portion, and wherein the processor directs the imaging subsystem to adjust image processing by calibrating the processing algorithm in response to detecting the change in location while the imaging system is still scanning the volumetric portion.
  • 8. The system of claim 7 wherein the markers comprise a first magnetic transponder having a first resonant frequency and a second magnetic transponder having a second resonant frequency different than the first resonant frequency.
  • 9. The system of claim 7, wherein the imaging subsystem further comprises: a source device positioned to deliver X-ray radiation to the volumetric portion;a detector device positioned to receive at least a portion of the X-ray radiation; anda processing unit coupled to the detector device for producing the raw image data based on received X-ray radiation.
  • 10. The system of claim 9 wherein the system further includes a gantry for carrying the source device and the detection device, the gantry being configured to at least partially rotate at least one of the source and the detection device about the volumetric portion of the patient.
  • 11. The system of claim 10 wherein the gantry also carries the sensors, and is configured to rotate the sensors concurrently with the source device so that the X-ray radiation does not pass through the sensors.
  • 12. A method for eliminating motion artifacts in an imaging system, the method comprising: receiving time-varying localization data from a sensor array in communication with a plurality of active markers disposed within a patient's body, the sensor array being located outside the patient, the plurality of active markers being spaced apart from each other in a known geometry relative to each other, and the time-varying localization data being based on the known geometry;concurrent to receiving the localization data, scanning a volumetric portion of the patient's body to produce raw image data regarding the volumetric portion, the volumetric portion including the active markers or the active markers being at the volumetric portion; andwhile scanning the volumetric portion processing the raw image data into image frames regarding the volumetric portion using a processing algorithm to produce image frames,binning the image frames based on a specific window of the real-time localization data, the real-time localization data including a plurality of time stamps and spatial coordinates associated with a corresponding one of the time stamps,outputting motion compensated image data corresponding to a particular subset of the binned raw image data,detecting a shift in location of the volumetric portion out of a detection based on the localization data, andcalibrating the processing algorithm in response to detecting the change in location of the volumetric portion.
  • 13. The method of claim 12 wherein the image frames are produced by a computed tomography scanner.
  • 14. The method of claim 12 wherein the localization data comprises individual datums that include a time stamp and two-dimensional or three-dimensional spatial coordinates.
  • 15. The method of claim 12 wherein the localization data is first localization data, and wherein the method further comprises synchronizing the second localization data with the raw image data prior to processing the image frames based on the first localization data.
  • 16. The method of claim 12 wherein the raw image data is produced by an imaging system, and wherein the method further comprises providing a feedback signal to the imaging system, the feedback signal being based, at least in part, on the localization data.
  • 17. The method of claim 12 wherein volumetric portion includes a portion of the lung, a tumor adjacent the lung, or both.
  • 18. A method, comprising: implanting markers at or adjacent a planning target volume (PTV) within a patient's body, the markers being spaced apart from each other in a known geometry relative to each other;exciting the markers to produce an excitation signal;detecting, in real time, a location of the markers based on the known geometry using an array of sensors positioned to receive the signal;associating a time stamp with each of the detected real-time locations;concurrently receiving the time stamps and associated real-time locations and scanning the PTV using a computed tomography (CT) imaging system, wherein scanning the PTV includes acquiring raw image data regarding the PTV and processing the raw image data into frames using an algorithm; andas the CT imaging system acquires the raw image data providing the image frames to a system interface for binning the image frames in real time, the binned image frames being based on a specific window of the detected real time location of the markers and the time stamp associated with each of the detected real time locations,outputting motion compensated image data corresponding to a particular subset of the binned image frames,receiving a feedback signal at the CT imaging system from the system interface indicating that the PTV has shifted out of a detection range, andmodifying the processing of the raw image data by calibrating the algorithm based on the feedback signal.
  • 19. The method of claim 18 wherein the CT imaging system includes an axial CT, a helical CT, or a ciné CT scanner.
  • 20. The method of claim 18, further comprising using the filtered or binned image data for a medical diagnosis or medical procedure planning.
  • 21. The method of claim 18 wherein individual markers comprise an electromagnetic transponder.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a U.S. National Phase of International Patent Application No. PCT/US2009/046494, filed Jun. 5, 2009, and titled MOTION COMPENSATION FOR MEDICAL IMAGING AND ASSOCIATED SYSTEMS AND METHODS, which claims the benefit of U.S. Provisional Patent Application No. 61/059,119, filed Jun. 5, 2008, and titled MOTION COMPENSATION FOR MEDICAL IMAGING AND ASSOCIATED SYSTEMS AND METHODS; the disclosure of each of these applications is incorporated herein by reference in its entirety.

PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/US2009/046494 6/5/2009 WO 00 6/28/2011
Publishing Document Publishing Date Country Kind
WO2009/149409 12/10/2009 WO A
US Referenced Citations (173)
Number Name Date Kind
3967161 Lichtblau Jun 1976 A
4017858 Kuipers Apr 1977 A
4023167 Wahlstrom May 1977 A
4114601 Abels Sep 1978 A
4123749 Hartmann et al. Oct 1978 A
4127110 Bullara Nov 1978 A
4160971 Jones et al. Jul 1979 A
4222374 Sampson et al. Sep 1980 A
4260990 Lichtblau Apr 1981 A
4343196 Wirth et al. Aug 1982 A
4393872 Reznik et al. Jul 1983 A
4466075 Groch Aug 1984 A
4618822 Hansen Oct 1986 A
4633250 Anderson, III et al. Dec 1986 A
4642786 Hansen Feb 1987 A
4643196 Tanaka et al. Feb 1987 A
4696287 Hortmann et al. Sep 1987 A
4737794 Jones Apr 1988 A
4795995 Eccleston et al. Jan 1989 A
4799495 Hawkins et al. Jan 1989 A
4849692 Blood Jul 1989 A
4909789 Taguchi et al. Mar 1990 A
4936823 Colvin et al. Jun 1990 A
4945305 Blood Jul 1990 A
4945914 Allen Aug 1990 A
4994079 Genese et al. Feb 1991 A
5031634 Simon Jul 1991 A
5050608 Watanabe Sep 1991 A
5062847 Barnes Nov 1991 A
5095224 Renger et al. Mar 1992 A
5099845 Besz et al. Mar 1992 A
5107862 Fabian et al. Apr 1992 A
5117829 Miller et al. Jun 1992 A
5142292 Chang Aug 1992 A
5170055 Carroll et al. Dec 1992 A
5188368 Ryan et al. Feb 1993 A
5198877 Schulz Mar 1993 A
5221269 Miller Jun 1993 A
5222499 Allen et al. Jun 1993 A
5240011 Assa Aug 1993 A
5325873 Hirschi et al. Jul 1994 A
5353804 Kornberg et al. Oct 1994 A
5377678 Dumoulin Jan 1995 A
5397329 Allen Mar 1995 A
5409004 Sloan Apr 1995 A
5411026 Carol May 1995 A
5417210 Funda May 1995 A
5423334 Jordan Jun 1995 A
5425367 Shapiro et al. Jun 1995 A
5425382 Golden et al. Jun 1995 A
5431161 Ryals et al. Jul 1995 A
5446548 Gerig Aug 1995 A
5453686 Anderson Sep 1995 A
5509900 Kirkman Apr 1996 A
5515853 Smith May 1996 A
5526812 Dumoulin et al. Jun 1996 A
5528651 Leksell et al. Jun 1996 A
5546951 Ben-Haim Aug 1996 A
5558091 Acker Sep 1996 A
5568809 Ben-Haim Oct 1996 A
5572999 Funda et al. Nov 1996 A
5617857 Chader Apr 1997 A
5622170 Schulz Apr 1997 A
5622187 Carol Apr 1997 A
5629967 Leksell May 1997 A
5630431 Taylor May 1997 A
5638819 Manwaring et al. Jun 1997 A
5680106 Schrott et al. Oct 1997 A
5681326 Lax Oct 1997 A
5697384 Miyawaki et al. Dec 1997 A
5707362 Yoon Jan 1998 A
5707390 Bonutti Jan 1998 A
5727552 Ryan Mar 1998 A
5735795 Young et al. Apr 1998 A
5745545 Hughes Apr 1998 A
RE35816 Schulz Jun 1998 E
5764052 Renger Jun 1998 A
5769861 Vilsmeier et al. Jun 1998 A
5779638 Vesely Jul 1998 A
5797849 Vesely et al. Aug 1998 A
5805661 Leksell Sep 1998 A
5810851 Yoon Sep 1998 A
5815076 Herring Sep 1998 A
5817022 Vesely Oct 1998 A
5820553 Hughes Oct 1998 A
5823192 Kalend et al. Oct 1998 A
5828770 Leis et al. Oct 1998 A
5830144 Vesely Nov 1998 A
5840148 Campbell et al. Nov 1998 A
5868673 Vesely Feb 1999 A
5868675 Henrion et al. Feb 1999 A
5879297 Haynor et al. Mar 1999 A
5879357 Heaton Mar 1999 A
5899857 Wilk May 1999 A
5902238 Golden May 1999 A
5902310 Foerster May 1999 A
5910144 Hayashi Jun 1999 A
5913820 Bladen Jun 1999 A
5923417 Leis Jul 1999 A
5928137 Green Jul 1999 A
5951481 Evans Sep 1999 A
5987349 Schulz Nov 1999 A
5989265 Bouquet De La Joliniere et al. Nov 1999 A
6015390 Krag Jan 2000 A
6019725 Vesely et al. Feb 2000 A
6026818 Blair et al. Feb 2000 A
6049587 Leksell et al. Apr 2000 A
6052477 Wang et al. Apr 2000 A
6059734 Yoon May 2000 A
6061644 Leis May 2000 A
6064904 Yanof et al. May 2000 A
6067465 Foo et al. May 2000 A
6076008 Bucholz Jun 2000 A
6081238 Alicot Jun 2000 A
6082366 Andra et al. Jul 2000 A
6094007 Faul Jul 2000 A
6097994 Navab Aug 2000 A
6118848 Reiffel Sep 2000 A
6144875 Schweikard Nov 2000 A
6161009 Skurdal et al. Dec 2000 A
6173715 Sinanan et al. Jan 2001 B1
6198963 Haim et al. Mar 2001 B1
6239724 Doron et al. May 2001 B1
6246900 Cosman et al. Jun 2001 B1
6307473 Zampini et al. Oct 2001 B1
6325758 Carol et al. Dec 2001 B1
6363940 Krag Apr 2002 B1
6371379 Dames et al. Apr 2002 B1
6381485 Hunter et al. Apr 2002 B1
6401722 Krag Jun 2002 B1
6405072 Cosman Jun 2002 B1
6416520 Kynast et al. Jul 2002 B1
6474341 Hunter et al. Nov 2002 B1
6611700 Vilsmeier et al. Aug 2003 B1
6675810 Krag Jan 2004 B2
6698433 Krag Mar 2004 B2
6812842 Dimmer Nov 2004 B2
6822570 Dimmer et al. Nov 2004 B2
6838990 Dimmer Jan 2005 B2
6918919 Krag Jul 2005 B2
6963770 Scarantino et al. Nov 2005 B2
7135978 Gisselberg et al. Nov 2006 B2
7657301 Mate et al. Feb 2010 B2
7657302 Mate et al. Feb 2010 B2
7657303 Mate et al. Feb 2010 B2
20020065461 Cosman May 2002 A1
20020193685 Mate et al. Dec 2002 A1
20030088178 Owens et al. May 2003 A1
20030192557 Krag et al. Oct 2003 A1
20030206614 Kendrick et al. Nov 2003 A1
20040092815 Schweikard et al. May 2004 A1
20040123871 Wright et al. Jul 2004 A1
20040125916 Herron et al. Jul 2004 A1
20040133101 Mate Jul 2004 A1
20040138555 Krag Jul 2004 A1
20040158146 Mate Aug 2004 A1
20040176931 Wright Sep 2004 A1
20040236207 Widener et al. Nov 2004 A1
20050059884 Krag Mar 2005 A1
20050140372 Wright et al. Jun 2005 A1
20050154280 Wright et al. Jul 2005 A1
20050154293 Gisselberg Jul 2005 A1
20050195084 Dimmer et al. Sep 2005 A1
20050201510 Mostafavi Sep 2005 A1
20050261570 Mate et al. Nov 2005 A1
20060074301 Meier et al. Apr 2006 A1
20060074302 Meier et al. Apr 2006 A1
20060079764 Wright et al. Apr 2006 A1
20060173294 Ein-Gal Aug 2006 A1
20070153972 Fujishige et al. Jul 2007 A1
20080039746 Hissong et al. Feb 2008 A1
20080226149 Wischmann et al. Sep 2008 A1
20110046481 Mate et al. Feb 2011 A1
Foreign Referenced Citations (40)
Number Date Country
19914455 Oct 2000 DE
0531081 Mar 1993 EP
0 719 420 Nov 1999 EP
1 034 738 Sep 2000 EP
2635259 Feb 1990 FR
2686499 Jul 1993 FR
8-166446 Jun 1996 JP
WO-8808282 Nov 1988 WO
WO-9525475 Sep 1995 WO
WO-9533519 Dec 1995 WO
WO-9608208 Mar 1996 WO
WO-9608999 Mar 1996 WO
WO-9712553 Apr 1997 WO
WO-9736192 Oct 1997 WO
WO-9748438 Dec 1997 WO
WO-9830166 Jul 1998 WO
WO-9838908 Sep 1998 WO
WO-9840026 Sep 1998 WO
WO-9917133 Apr 1999 WO
WO-9927839 Jun 1999 WO
WO-9930182 Jun 1999 WO
WO-9933406 Jul 1999 WO
WO-9935966 Jul 1999 WO
WO-9940869 Aug 1999 WO
WO-9953966 Oct 1999 WO
WO-9958044 Nov 1999 WO
WO-9958055 Nov 1999 WO
WO-9958065 Nov 1999 WO
WO-0024332 May 2000 WO
WO-0038579 Jul 2000 WO
WO-0051514 Sep 2000 WO
WO-0053115 Sep 2000 WO
WO-0065989 Nov 2000 WO
WO-0134049 May 2001 WO
WO-0154765 Aug 2001 WO
WO-0219908 Mar 2002 WO
WO-0239917 May 2002 WO
WO-0239918 May 2002 WO
WO-02100485 Dec 2002 WO
WO-2006113323 Oct 2006 WO
Non-Patent Literature Citations (12)
Entry
Beyer, Thomas, et al. “Dual-modality PET/CT imaging: the effect of respiratory motion on combined image quality in clinical oncology.” European journal of nuclear medicine and molecular imaging 30.4 (2003): 588-596.
Low, Daniel A., et al. “A method for the reconstruction of four-dimensional synchronized CT scans acquired during free breathing.” Medical physics 30.6 (2003): 1254-1263.
Wolthaus, J. W. H., et al. “Fusion of respiration-correlated PET and CT scans: correlated lung tumour motion in anatomical and functional scans.” Physics in medicine and biology 50.7 (2005): 1569.
U.S. Appl. No. 10/416,827, Krag.
Final Office Action, U.S. Appl. No. 09/877,498, Applicant: Calypso Medical Technologies, Inc., Date of Mailing: Feb. 14, 2006, 7 pages.
Decision on Appeal, U.S. Appl. No. 09/877,498, Applicant: Calypso Medical Technologies, Inc., Date of Mailing: May 27, 2009, 16 pages.
Final Office Action; U.S. Appl. No. 09/877,498; Applicant: Calypso Medical Technologies, Inc.; Date of Mailing: Feb. 14, 2006; 7 pages.
Hsiao, K., “Fast Multi-Axis Tracking of Magnetically-Resonant Passive Tags: Methods and Applications,” Feb. 2001, Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, pp. 1-107.
International Preliminary Examination Report; PCT/US02/17876, filed Jun. 5, 2002, in the name of Calypso Medical Technologies, Inc.
International Search Report dated Jan. 24, 2003, PCT Application No. PCT/US/29390.
European Search Report dated Jun. 28, 2011, EP Application No. 10185512.
International Search Report and Written Opinion dated Jul. 28, 2009, PCT Application No. US09/046494, 12 pages.
Related Publications (1)
Number Date Country
20110249880 A1 Oct 2011 US
Provisional Applications (1)
Number Date Country
61059119 Jun 2008 US