The present invention relates to medical imaging. More particularly, the present invention relates to methods involving the use of radial endobronchial ultrasound. More particularly, the present invention relates to methods involving the use of images obtained using radial endobronchial ultrasound imaging to provide localization of targets, such as lesions, in medical images.
Radial endobronchial ultrasound is a medical imaging technique whereby ultrasound waves are emitted radially from a probe positioned within a bronchial passageway of a patient. The ultrasound waves are processed to produce a medical image showing a cross-section (e.g., a “slice”) of the patient's tissue around the bronchial passageway.
In an embodiment, a method includes obtaining at least one preoperative image from an imaging modality; identifying, on the at least one preoperative image, at least one element located within an area of interest; obtaining at least one intraoperative image; highlighting the at least one element on the at least one intraoperative image; navigating a radial endobronchial ultrasound probe to the area of interest using the at least highlighted at least one element; acquiring a plurality of radial endobronchial ultrasound images; extracting a plurality of two-dimensional representations of the element, each of the plurality of two-dimensional representations of the element being extracted from a corresponding one of the plurality of radial endobronchial ultrasound images; reconstructing a three-dimensional representation of the element from the plurality of two-dimensional representations of the element; and projecting a two-dimensional projection of the three-dimensional representation of the element on at least one of the at least one intraoperative image.
In an embodiment, the step of projecting the two-dimensional projection of the three-dimensional representation of the element on the at least one of the at least one intraoperative image is performed in real time.
In an embodiment, a method includes removing the radial endobronchial ultrasound probe from the area of interest; and navigating a further endobronchial tool to the area of interest. In an embodiment, a method includes performing a procedure on the element using the further endobronchial tool. In an embodiment, a method includes removing the further endobronchial tool; navigating the radial endobronchial ultrasound probe to the area of interest; acquiring a plurality of updated radial endobronchial ultrasound images; extracting a plurality of updated two-dimensional representations of the element, each of the plurality of updated two-dimensional representations of the element being extracted from a corresponding one of the plurality of updated radial endobronchial ultrasound images; and reconstructing an updated three-dimensional representation of the element from the plurality of two-dimensional representations of the element.
In an embodiment, a method includes calculating distances between a center of the radial endobronchial ultrasound probe and a plurality of boundary points on a boundary of the target; and estimating a margin size for an ablation based on a maximum one of the distances. In an embodiment, the at least one intraoperative image includes an X-ray.
In an embodiment, the three-dimensional representation of the element is used as a prior for volume reconstruction from at least one of the intraoperative images. In an embodiment, a method also includes registering the three-dimensional representation of the target to a three-dimensional computed tomography volume; and projecting the three-dimensional representation of the element from the three-dimensional computed tomography volume on at least one of the at least one intraoperative image. In an embodiment, the three-dimensional computed tomography volume is a preoperative computed tomography scan volume or a three-dimensional computed tomography volume reconstructed from the at least one intraoperative image.
In an embodiment, a method includes navigating a radial endobronchial ultrasound probe to an area of interest; acquiring a plurality of radial endobronchial ultrasound images and a plurality of intraoperative images, each of the plurality of radial endobronchial ultrasound images corresponding to one of the plurality of intraoperative images and to a different position of the ultrasound probe; extracting a radial endobronchial ultrasound probe tip position from each of the intraoperative images; generating a database of pairs of the intraoperative and endobronchial ultrasound images, each pair corresponding to a specific probe tip position and orientation in the preoperative image coordinate system; removing the radial endobronchial ultrasound probe from the area of interest; navigating a further endobronchial tool to the area of interest; acquiring a further plurality of intraoperative images; extracting a position of the further endobronchial tool from the further plurality of intraoperative images; identifying one of the pairs in the database that corresponds most closely to the position of the further endobronchial tool; and displaying the ultrasound image corresponding to the identified one of the pairs.
In an embodiment, the further endobronchial tool is a biopsy instrument or an ablation catheter. In an embodiment, a method includes obtaining at least one preoperative image from an imaging modality; identifying, on the at least one preoperative image, at least one element located within an area of interest; obtaining at least one intraoperative image; highlighting the at least one element on the at least one intraoperative image, wherein the step of navigating the radial endobronchial ultrasound probe to the area of interest is performed using the highlighted at least one element.
In an embodiment, a method includes navigating a radial endobronchial ultrasound probe to an area of interest; selecting a confirmed position of the radial endobronchial ultrasound probe; acquiring at least one intraoperative image of the area of interest while the radial endobronchial ultrasound probe is positioned in the confirmed position; extracting a position of the radial endobronchial ultrasound probe from at least one of the at least one intraoperative image; and overlaying the confirmed position of the endobronchial ultrasound probe on at least one of the at least one intraoperative image.
In an embodiment, a method includes acquiring at least two further intraoperative images, each of the at least two further intraoperative images having a known geometric relation of the confirmed position of the radial endobronchial ultrasound probe; reconstructing the confirmed position in three-dimensional space based on the at least two further intraoperative images; and overlaying the confirmed position of the radial endobronchial ultrasound probe on at least one of the further intraoperative having a known geometric relation.
In an embodiment, a method includes removing the radial endobronchial ultrasound probe; and navigating a further endobronchial instrument to the confirmed position, whereby accurate positioning of the further endobronchial instrument is ensured. In an embodiment, the further endobronchial instrument is a biopsy instrument or an ablation catheter.
The present invention will be further explained with reference to the attached drawings, wherein like structures are referred to by like numerals throughout the several views. The drawings shown are not necessarily to scale, with emphasis instead generally being placed upon illustrating the principles of the present invention. Further, some features may be exaggerated to show details of particular components.
The figures constitute a part of this specification and include illustrative embodiments of the present invention and illustrate various objects and features thereof. Further, the figures are not necessarily to scale, some features may be exaggerated to show details of particular components. In addition, any measurements, specifications and the like shown in the figures are intended to be illustrative, and not restrictive. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a representative basis for teaching one skilled in the art to variously employ the present invention.
Among those benefits and improvements that have been disclosed, other objects and advantages of this invention will become apparent from the following description taken in conjunction with the accompanying figures. Detailed embodiments of the present invention are disclosed herein; however, it is to be understood that the disclosed embodiments are merely illustrative of the invention that may be embodied in various forms. In addition, each of the examples given in connection with the various embodiments of the invention which are intended to be illustrative, and not restrictive.
Throughout the specification and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrases “in one embodiment” and “in some embodiments” as used herein do not necessarily refer to the same embodiment(s), though they may. Furthermore, the phrases “in another embodiment” and “in some other embodiments” as used herein do not necessarily refer to a different embodiment, although they may. Thus, as described below, various embodiments of the invention may be readily combined, without departing from the scope or spirit of the invention.
The term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of “a,” “an,” and “the” include plural references. The meaning of “in” includes “in” and “on.”
As used herein, an “absolute roll” refers to an orientation of a portion of an image with respect to an absolute (e.g., global) frame of reference. As used herein, a “relative roll” refers to the amount a current roll has changed relative to a reference roll. A reference roll is a roll which cannot be moved and may be preassigned.
In some embodiments, the method of the present invention uses imagery obtained using a radial endobronchial ultrasound (“REBUS”) probe to improve the clinical outcome of endobronchial procedures. In some embodiments, the REBUS provides radial ultrasound images inside a patient's bronchial airways. In some embodiments, the REBUS can be used in addition to the methods described in PCT/US15/56489, PCT/US14/67328, and PCT/US15/10381, which are hereby incorporated by reference in its entireties. PCT/US15/56489 discloses a method to augment an intraoperative imagery (e.g., but not limited to, X-ray, C-arm, etc.) with data from a preoperative imagery (e.g., but not limited to computerized tomography, magnetic resonance imaging) in order to assist a physician during endobronchial procedures. In some embodiments, the method includes detecting dense tissues (e.g., tissues which have a 10%, 20%, 30%, 40%, 50%, etc. increased density compared with surrounding tissues), such as lesions, inside the lungs.
In some embodiments of the methods of the present invention, an intraoperative image (e.g., an image obtained during a procedure using an imaging technique such as, but not limited to, a fluoroscopic image) and a REBUS image are acquired simultaneously. In some embodiments of the method of the present invention, an intraoperative image and a REBUS image are not acquired simultaneously (e.g., a REBUS image is acquired and a fluoroscopic intraoperative image is acquired subsequently). In some embodiments, the 3D position of the REBUS probe tip in the preoperative image is acquired using the methods, e.g., but not limited to, described in PCT/US15/56489. In some embodiments, a plurality of images of the REBUS probe tip are generated (e.g., but not limited to preoperative images). In some embodiments, the methods of the present invention further produce a database of pairs of intraoperative and REBUS images when each pair is corresponding to a specific probe tip position and orientation in the preoperative image coordinate system. In some embodiments, database can be queried or searched using the following non-limiting example: finding a nearest pair that matches the pre-marked position in the preoperative image.
In some embodiments, a method uses a set of REBUS images acquired in proximity to target area (e.g., a defined area which includes a target such as a lesion) and their position and orientation in three-dimensional space to reconstruct the outline and/or topology of the target (e.g., but not limited to, a lesion), which will be referred to herein as a “reconstructed 3D target.” In some embodiments, the reconstructed 3D target can be modified on the intraoperative image, e.g., but not limited to, projecting over or highlighting the reconstructed 3D target.
In some embodiments, a non-limiting example of target reconstruction can be performed in accordance with the method 100 shown in
In step 130, intraoperative images (i.e., images acquired during a procedure) are received. In some embodiments, the intraoperative images are two-dimensional images. In some embodiments, the intraoperative images are three-dimensional images. In some embodiments, the intraoperative images are any known suitable type of medical images (e.g., fluoroscope images such as X-ray images).
In step 140, a region of interest is highlighted in the intraoperative images.
In some embodiments, the steps 100-140 are performed in accordance with the exemplary methods described in International Patent Application No. PCT/IB2015/002148, the contents of which are incorporated herein by reference in their entirety. In some embodiments, the steps 100-140 are performed in accordance with the process shown in
Referring now to
Continuing to refer to
In some embodiments, the coarse registration process of step 240 is performed based on an intensity-based automatic registration method using multiple intraoperative (e.g., X-ray) images and the preoperative CT volume. In some embodiments, the method is iterative. In some embodiments, for each optimization step high quality digitally reconstructed radiographs (“DRR”) are generated and then compared against acquired intraoperative (e.g., X-ray) images. In some embodiments, the method 200 uses the registration techniques disclosed in, Khamene et al., “Automatic registration of portal images and volumetric CT for patient positioning in radiation therapy,” Medical Image Analysis 10 (2006) 96-112, which is hereby incorporated by reference in its entirety. In some embodiments, such registration can be implemented, as a non-limiting example, as intensity-based and/or as feature based, depending on the specific medical application. In some embodiments, intensity-based and feature based registration are as described by David et al., “Intensity-based Registration versus Feature-based Registration for Neurointerventions,” Medical Vision Laboratory, Dep't of Engineering Science, University of Oxford, England, which is hereby incorporated by reference in its entirety. In some embodiments, point-based registration is implemented using known anatomical landmarks on a patient's chest. In some embodiments, at least one known landmark can be marked on a CT image and/or fluoroscopic image. In some embodiments, special markers are attached to the patient's chest during procedure to improve/increase detectability on a fluoroscopic image.
Continuing to refer to
In some embodiments, the fine registration process of step 260 includes the steps of: (1) Feature Identification: identifying a set of relevant features in the two images, such as edges, intersections of lines, region contours, regions, etc; (2) Feature Matching: establishing correspondence between the features (i.e., each feature in the sensed image is be matched to its corresponding feature in the reference image); each feature is identified with a pixel location in the image, and these corresponding points are usually referred to as control points; (3) Spatial Transformation: determining the mapping functions that can match the rest of the points in the image using information about the control points obtained in the previous step; and (4) Interpolation: resampling the sensed image using the above mapping functions to bring it into alignment with the reference image. Some embodiments use an area-based approach, which is also referred to as correlation-like methods or fine registration (i.e., template matching), such as described in Fonseca et al., “Registration techniques for multisensor remotely sensed imagery,” PE & RS-Photogrammetric Engineering & Remote Sensing 62 (9), 1049-1056 (1996), which describes the combination of feature detection and feature matching. In some embodiments, the method 200 is suited for templates which have no strong features corresponding to an image, since the templates operate directly on the bulk of values. In some embodiments, matches are estimated based on the intensity values of both image and template. In some embodiments, techniques that are used include squared differences in fixed intensities, correction-based methods, optimization methods, mutual information, or any combination thereof. In some embodiments, fine registration is performed automatically. In some embodiments, fine registration includes aligning a 2D projection of an anatomical structure from a CT scan obtained through coarse registration with correspondent anatomical structure extracted from fluoroscopic image.
Continuing to refer to
Referring back to
In step 180, the three-dimensional shape of the target is reconstructed based on the extracted target contours from the REBUS images and the known position and orientation of the probe tip from the intraoperative image. In some embodiments, the extracted target contours and the known position and orientation of the probe tip define a REBUS-based real space. In some embodiments, the process of reconstruction includes the steps of: (1) mapping each two-dimensional target contour point to a voxel in the REBUS-based real space; (2) traversing the REBUS-based real space and generating a three-dimensional target model by marking every voxel surrounded by or belonging to contour points as an target voxel; and (3) applying a surface-extraction algorithm to generate a three-dimensional target surface from the reconstructed target model. In some embodiments, the 3D volume is reconstructed from the set of REBUS slices in accordance with the techniques described in Zang et al., “3D Segmentation and Reconstruction of Endobronchial Ultrasound”, Medical Imaging 2013: Ultrasonic Imaging, Tomography and Therapy (Vol. 8675).
In some embodiments, the position and orientation of the REBUS probe may be identified using the techniques described in International Pat. App. No. PCT/IB17/01376, the contents of which are incorporated herein by reference in their entirety. In some embodiments, a radiopaque pattern is provided on the REBUS probe in order to facilitate such identification. In some embodiments, the position and orientation of the REBUS probe may be performed in accordance with the process shown in
In step 190, the reconstructed target shape is projected. In some embodiments, the complete or partial 3D target is be segmented from this volume and projected over the intraoperative image combining the values of the projected and source volume depending on the application need. In some embodiments, the 3D target is highlighted over the inter-operative image. In some embodiments, the full or partial target is used for the registration between the preoperative image and postoperative image.
In some embodiments, a 3D target reconstructed from REBUS images is registered to a 3D volume sourced from the preoperative image. In an embodiment, the registration between the two volumes is based on matching voxel intensity values. In an embodiment, the registration between the 3D target and the volume is performed by generating a binary volume from the 3D target shape and then registering two volumes based on matching voxel intensity values. In an embodiment, the registration between the 3D target and the volume is performed by extracting geometric shapes of anatomical structures from the volume and then matching these geometric shapes with the shape of the 3D target. In an embodiment, the registration between the 3D target and the volume is based on co-aligning the centers of mass or moments of inertia. In some embodiments, the pose of an intraoperative image relative to a preoperative image can be further calculated from the known position (e.g., location and orientation) of the REBUS probe on the intraoperative image.
In some embodiments, a 3D target reconstructed from REBUS images is registered to the 3D volume reconstructed from plurality of intraoperative images, wherein the reconstruction methods, not limited to, described in PCT/US15/56489.
In some embodiments, a 3D target reconstructed from REBUS images is used as a prior for volume reconstruction from plurality of intraoperative images. In some embodiments, the area of interest can be segmented on reconstructed 3D target. In some embodiments, only a partial shape or area of interest is reconstructed from REBUS images, registered to the volume sourcing from the preoperative image and enhanced or accomplished using the information from preoperative image to obtain additional information in the volume reconstructed from intraoperative images.
In some embodiments, the 3D target is reconstructed from different interoperative images, such as a set of fluoroscopic images or CT images, with known relative pose. In some embodiments, such reconstruction can be performed using a back projection algorithm or any other reconstruction algorithm utilized in computational tomography. In some embodiments, a 3D target or area of interest reconstructed from REBUS images can be co-registered with compatible 3D volume or area of interest sourcing additional interoperative image, preoperative image or combination of thereof.
In some embodiments, the non-limiting examples above rely on determination of the REBUS probe tip position and orientation with respect to the intraoperative image. In some embodiments, the methods disclosed in International Pat. App. No. PCT/IB17/01376 can be used in the methods of the present invention. In some embodiments, a radio opaque pattern attached to the catheter can be used to determine its pose with respect to the intraoperative imaging device. In some embodiments, a helical spring pattern, which is asymmetric relatively to the axes of catheter rotation, can be used to determine the relative or absolute roll for each REBUS slice. In some embodiments, an angular measurement device attached to the probe tip of the catheter can be used to determine the relative or absolute roll for each REBUS slice. In some embodiments, adjacent frames can be stabilized by minimizing rotation difference between these frames.
In some embodiments, a REBUS image with an associated position and orientation is used to locate bronchial tree bifurcations on the intraoperative image. In some embodiments, such bifurcations may be used as additional fiducial points in the registration process described in International Pat. App. No. PCT/US15/56489. In some embodiments, an airway bifurcation may be visible on a REBUS image and can be detected by means of image processing (e.g., but not limited to, highlighting). In some embodiments, since (1) at least one image of REBUS is of a probe tip acquired by an intraoperative device, and at the same time (2) at least one image of bifurcation is acquired by REBUS itself, the position and orientation of the bifurcation may be marked in the intraoperative image. In some embodiments, the bifurcation position in the intraoperative image and its corresponding position in the preoperative image as additional fiducial points to improve to accuracy for the registration process can obtained using the methods described in, e.g., but not limited to, PCT/US15/56489.
In some embodiments, a location of the REBUS probe at the selected position of a REBUS image can be marked or confirmed on a corresponding intraoperative image; this may be referred to as “REBUS confirmation”. In some embodiments, the location of the tip of the REBUS probe as seen in the intraoperative image can be marked, stored and shown at any time, even after the REBUS probe has been retracted and another device (e.g., other endobronchial instruments or endotherapeutic accessories) has been introduced into the same area inside the patient. In some embodiments, a compensation for respiratory motion and tissue deformation caused by endobronchial instruments may be provided, whereby the displayed position of selected and stored location of the REBUS are adjusted. In an embodiment, the movement compensation is performed by tracking anatomical structures and/or instruments. In some embodiments, the position of the REBUS probe corresponding to the “REBUS confirmation” can be extracted from multiple intraoperative images having different poses with known geometric relation, thereby providing the 3D reconstruction of the REBUS probe and, particularly, the calculated position of the tip of the REBUS probe. In some embodiments, the 3D location of the “REBUS confirmation” location can be projected onto any future intraoperative image(s).
The clinical applications of the exemplary embodiments can be illustrated through the following prophetic examples:
1) The exemplary embodiments may provide the physician with REBUS images during the time when the REBUS probe is retracted and replaced by other endobronchial tool, such as biopsy forceps or ablation probe.
2) Where the target is a lesion, by reconstructing the shape of the lesion in three dimensions with respect to the position of the REBUS probe, the maximal distance from the center of the probe to the boundary of the lesion can be determined. This information may provide an appropriate margin size for an ablation procedure when the REBUS probe has been removed and replaced by an ablation catheter. For example, the margin can be calculated by multiplying a constant by the maximal distance from the center of the REBUS probe to the boundary of the lesion.
3) The desired area of interest can be marked on the acquired REBUS images and an endobronchial tool can be guided to the desired preplanned location (e.g., marking on the acquired REBUS images).
4) A method to obtain an accurate localization of the target during, e.g., but not limited to, an endobronchial biopsy or ablation. The method can be performed using the following steps:
1. Endobronchial navigation to the target region. The navigation may be performed using, e.g., but not limited to, the navigation method described in PCT/US15/56489.
2. Acquire REBUS images in the target area.
3. Extract the target from the REBUS images and project it on the intraoperative image.
4. Position the endobronchial tool in the target center and perform an ablation or biopsy.
5. If ablation was performed and the anatomy changed, steps 2-4 may be performed again.
While a number of embodiments of the present invention have been described, it is understood that these embodiments are illustrative only, and not restrictive, and that many modifications may become apparent to those of ordinary skill in the art. Further still, the various steps may be carried out in any desired order (and any desired steps may be added and/or any desired steps may be eliminated).
This application is an international (PCT) application relating to and claiming the benefit of commonly-owned, copending U.S. Provisional Patent Application No. 62/510,729, entitled “METHODS FOR USING RADIAL ENDOBRONCHIAL ULTRASOUND PROBES FOR THREE-DIMENSIONAL RECONSTRUCTION OF IMAGES AND IMPROVED TARGET LOCALIZATION,” filed May 24, 2017, the contents of which are incorporated by reference herein in their entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/IB2018/000624 | 5/24/2018 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62510729 | May 2017 | US |