The present disclosure generally relates to image processing and analysis. More specifically, the present disclosure relates to a computer-implemented methods and devices for performing three-dimensional blood vessel reconstruction using a single-view angiographic image and refining the three-dimensional blood vessel reconstruction.
Rotational two-dimensional (2D) X-ray angiographic images provide valuable geometric information on vascular structures for diagnoses of various vascular diseases, such as coronary artery diseases and cerebral diseases. After a contrast agent (usually an x-ray opaque material, such as iodine) is injected into the vessel, the image contrast of the vessel regions is generally enhanced. Three-dimensional (3D) vascular tree reconstruction using the 2D projection images is often beneficial to reveal the true 3D measurements, including diameters, curvatures and lengths, of various vessel segments of interests, for further functional assessments of the targeted vascular regions.
Extant 3D reconstruction methods typically rely on 2D vessel structures segmented from multiple X-ray images from different imaging projection angles (such as a primary angle and a secondary angle). Usually, 2D vessel centerlines are first extracted from the segmented vessel regions, and 3D centerlines are then computed by establishing the proper projection imaging system geometry. One technical challenge presented by extant methods is the foreshortening issue. The vessel lengths are slightly different when viewed from different angles due to the nature of the projection imaging, causing foreshortening. Generally, foreshortening may be reduced by avoiding using images containing pronounced foreshortening vessel segments (represented with darker intensity) for 3D reconstruction. However, at least some level of foreshortening frequently occurs due to the curved geometrical nature of vessels and due to physiological motion of the patient during the imaging process (e.g., due to respiratory motion and cardiac motion).
Embodiments of the disclosure address the above problems by systems and methods for improved three-dimensional blood vessel reconstructions.
Embodiments of the present disclosure include computer-implemented methods and devices for performing three-dimensional blood vessel reconstruction using a single-view projection image and then refining the three-dimensional reconstruction based on optical path lengths of the blood vessel obtained through different approaches.
In one aspect, the disclosure is directed to a computer-implemented method for performing three-dimensional blood vessel reconstruction. The computer-implemented method includes receiving a first two-dimensional image of a blood vessel of a patient, where the first two-dimensional image is a projection image acquired in a first projection direction. The method further includes reconstructing, by a processor, a three-dimensional model of the blood vessel based on at least the first two-dimensional image. The method additionally includes adjusting the three-dimensional model of the blood vessel, based on a comparison of a first optical path length determined from a second two-dimensional image of the blood vessel of the patient and a second optical path length determined from the three-dimensional model.
In another aspect, the disclosure is further directed to a device for performing three-dimensional blood vessel reconstruction. The device includes an interface, which is configured to receive a first two-dimensional image of a blood vessel of a patient, where the first two-dimensional image is a projection image acquired in a first projection direction. The device further includes a processor, which is configured to reconstruct a three-dimensional model of the blood vessel based on at least the first two-dimensional image, and adjust the three-dimensional model of the blood vessel, based on a comparison of a first optical path length determined from a second two-dimensional image of the blood vessel of the patient and a second optical path length determined from the three-dimensional model.
In yet another embodiment, the disclosure is directed to a non-transitory computer-readable medium, having instructions stored thereon. The instructions, when executed by a processor, perform a method for performing three-dimensional blood vessel reconstruction. The method includes receiving a first two-dimensional image of a blood vessel of a patient, where the first two-dimensional image is a projection image acquired in a first projection direction. The method further includes reconstructing a three-dimensional model of the blood vessel based on at least the first two-dimensional image. The method additionally includes adjusting the three-dimensional model of the blood vessel, based on a comparison of a first optical path length determined from a second two-dimensional image of the blood vessel of the patient and a second optical path length determined from the three-dimensional model
Capable of using only one projection view to perform the initial reconstruction of a 3D vessel model, the disclosed method and device can reduce the amount radiation exposure for doctor and patients. They also relax requirement for obtaining 3D vessel reconstruction, as it removes the stringent requirements for traditional multi-view reconstruction algorithm, which requires at least two projection views from sufficiently different angles that both show the target vessel clearly without overlapping with other nearby vessels. Reconstructing from a single-view projection image is also faster compared to multi-view reconstruction, which requires finding correspondence points among different views.
By using the same or another projection image to further refine the 3D model, the disclosed method and device may further make full use of the intensity (e.g., grayscale) distribution pattern of the two-dimensional vessel (when filled with contrast agent) which is normally neglected and three-dimensional projection path information implied therein, effectively reducing the foreshortening phenomenon in the three-dimensional reconstruction, thereby improving the reconstruction accuracy of three-dimensional vascular tree. The scheme of the present disclosure assists the reconstruction of the three-dimensional image by considering the image pixel intensity information, and improves the reconstruction accuracy.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
In the drawings, which are not necessarily drawn to scale, like numerals may describe similar components in different views. Like numerals having different letter suffixes may represent different instances of similar components. The drawings illustrate generally, by way of example, but not by way of limitation, various embodiments, and together with the description and claims, serve to explain the disclosed embodiments. When appropriate, the same reference numbers are used throughout the drawings to refer to the same or like parts. Such embodiments are demonstrative and not intended to be exhaustive or exclusive embodiments of the present method, device, or non-transitory computer readable medium having instructions thereon for implementing the method.
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.
This description may use the phrases “in one embodiment,” “in another embodiment,” “in yet another embodiment,” or “in other embodiments,” all referring to one or more of the same or different embodiments in the present disclosure. Moreover, an element which appears in a singular form in the specific embodiments do not exclude that it may appear in a plurality (multiple) form. An “optical path” may be a geometric path of rays propagating within a subject (not a vacuum). Accordingly, an “optical path length” may be the length of a geometric path along which the rays propagate in the subject. Consistent with the disclosure, terms such as “first” and “second” are used, which can refer to the same or different components or items. For example, a “first two-dimensional image” and a “second two-dimensional image” can be the same or different images, and “first projection direction” and a “second projection direction” can be the same or different images.
In some embodiments, the 3D information inference step 210 could be implemented by a depth-based reconstruction or a model-based reconstruction, or a hybrid of thereof. For example,
In some embodiments, the 3D information inference step 210 can be performed by an inference learning model. The inference learning model may be a machine learning model or a deep learning model trained to infer the 3D information from a 2D projection image. The inference learning model can be trained using sample single-view images and their corresponding 3D model projection annotations. The 3D model projection annotation can be obtained in various ways. In some embodiments, another modality from which the 3D model can be readily obtained. For example, a 3D CT angiographic image can be obtained from which the 3D model can be constructed. The projection parameters can be derived from geometric parameters recorded by the imaging acquisition device (e.g., an imaging scanner). These parameters can also be refined by optimizing the alignment of projected 3D model and angiographic images. In some embodiments, the 3D model projection annotation can be obtained using multi-view 3D model reconstruction algorithm. In some embodiments, the 3D model projection annotation can also be synthetic data obtained by first rendering a 3D model and then projecting the 3D model to produce a synthetic single-view projection image using an image generator/renderer. The synthetic data could be realistic given a powerful image generator/renderer. In yet some embodiments, human annotator can finetune annotations of 3D model and projection parameter.
Returning to
Accordingly, for depth-based inference (e.g.,
For deformable model-based inference (e.g.,
The process 100 may proceed to step 102: acquiring a second two-dimensional image in a second projection direction of a blood vessel and the reconstructed three-dimensional model of the blood vessel. In some embodiments, process 100 may skip step 101 and proceed directly to acquiring step 102 to acquire an already reconstructed three-dimensional model of the blood vessel from a stereoscopic imaging device.
In some embodiments, the first two-dimensional image (i.e., the single-view 2D image) based on which the blood vessel three-dimensional model is reconstructed may be also used as the second two-dimensional image. That is, the single-view 2D image may be reused as the second two-dimensional image referred to in the present disclosure. In that case, the first and second two-dimensional images are actually the same image, and the first and second projection directions are actually the same projection direction. By using the single-view projection image as both the image for initial reconstruction of the 3D model (step 101) and the image for refining the reconstructed 3D model (steps 103 and 104), the entire reconstruction process requires only one projection image acquired from a single projection direction. Accordingly, radiation exposure can be reduced, image acquisition is simplified, and reconstruction can be faster.
In other embodiments, the second two-dimensional image is a two-dimensional image captured by the imaging device, which is different from the first two-dimensional image based on which the three-dimensional model of the blood vessel is reconstructed. For example, when the imaging device captures two two-dimensional images in the first projection direction, and the second projection direction, respectively. One of the images (e.g., the one acquired in the first projection direction) may be used to reconstruct the three-dimensional model of a blood vessel as described in connection with step 101, and the other image (e.g., the second image) obtained in the other direction (e.g., the second projection direction) serves as the second two-dimensional image mentioned in present disclosure. In some alternative embodiments, both the first and the second image may be used to reconstruct the three-dimensional model of a blood vessel as described in connection with step 101 (in which case, a multi-view reconstruction is performed), and one of the images may be used as the second two-dimensional image in step 102. In yet some alternative embodiments, when the imaging device captures two two-dimensional images in the first projection direction and captures one two-dimensional image in the second projection direction, one of the two two-dimensional images in the first projection direction and the one two-dimensional image captured in the second projection direction may be used to reconstruct the three-dimensional model of the blood vessel in step 101, and the other two-dimensional image in the first projection direction may be used as the second two-dimensional image in step 102. In addition, for example, when the imaging device continuously captures a two-dimensional image in at least one or more projection directions, one image out of the obtained image sequences may serve as the second two-dimensional image.
After the acquisition step 102 is completed, a simulated optical path length determining step 103 is performed. At step 103, the simulated optical path length within the blood vessel at a position (i.e., at least one position) in the second projection direction may be determined based on the three-dimensional model of the blood vessel.
In one embodiment, the size at the position of the blood vessel in the second projection direction (a first X-ray transmission direction) in the three-dimensional model of the blood vessel may be determined as the simulated optical path length within the blood vessel at the corresponding position.
For example, as shown in
In another embodiment, the simulated optical path length xC may be obtained by radius estimation. As shown in
x
C
=D/sin θ Equation (1)
Then it proceeds to a three-dimensional reconstruction adjustment step 104. At the step 104, reconstruction parameters of the three-dimensional model of the blood vessel may be adjusted based on the simulated optical path length (xC1, xC2, . . . , and xCn) within the blood vessel at the position(s) in the second projection direction, intensity value at the corresponding position(s) of the blood vessel in the second two-dimensional image, and a relationship between intensity value at each position of a blood vessel in a second-dimensional image and an optical path length at the corresponding position. The adjusted reconstruction parameters may be utilized for three-dimension vessel reconstruction, so that a foreshortening of three-dimension vessel reconstruction may be rectified.
As shown in
It is observed that there is an inherent relationship between the intensity value at each position of the blood vessel in the two-dimensional image and the optical path length at the corresponding position, under the same contrast agent injection condition for the same patient. When optical path length at each position of a blood vessel in a two-dimensional image are denoted by xC, exp[xC] has an inherent relationship with the intensity value, such as gray values gC, at the corresponding position of the blood vessel in a two-dimensional image, such as an approximately linear relationship. The value obtained by removing background from the intensity value (such as the gray-scale value gC) and logarithmically processing the intensity value at each position of the blood vessel, has a linear relationship with the optical path length xC at the corresponding position, as shown in
The above-mentioned inherent relationship (for example, a linear relationship) can be explained through the following approximate derivation.
Specifically, the relationship between x-ray attenuation and optical path in the contrast agent may be defined by following Equation (2).
where II is the incident beam intensity, IT is the transmitted beam intensity, □μ/ρ is the mass attenuation coefficient and x is the optical path. In addition, the subscripts c and o represent contrast agent and organ (e.g., the vessel), respectively. In the absence of contrast agent, the x-ray beam absorption, due to the organ alone, is described by Equation (3).
where IB is the transmitted beam intensity with only background.
By incorporating Equation (3) into Equation (2), the relationship between the intensity of the light transmitted through the blood vessel at each position and the optical path length xC at the corresponding positions can be obtained, see Equation (4).
The light transmitted the blood vessel at each position thereof may be incident onto a flat panel detector, so as to obtain a gray-scale two-dimensional image. Thereby, the intensity of the light transmitted through the blood vessel at each position is converted to intensity value (for example, gray value) at the corresponding position of the blood vessel in the two-dimensional image. It is verified that the conversion to grayscale does not destroy the described inherent relationship, so that the inherent relationship between the intensity of the light transmitted through the blood vessel at each position and the optical path length xC at the corresponding position is maintained between the intensity value at each position of the blood vessel in the two-dimensional image and the optical path length xC at the corresponding position. Therefore, the reconstruction of the three-dimensional model can be guided by using the intensity values at the position(s) of the blood vessel in the two-dimensional images. Compared with the existing three-dimensional reconstruction technology that ignores the intensity value of two-dimensional images, the present disclosure considers the above relationship so that the reconstruction accuracy of the three-dimensional model can be improved. Hereinafter, in order to facilitate description, the transition between the intensity of the light transmitted through the blood vessel at the position(s) of and the intensity value at corresponding position(s) of the corresponding blood vessel in the two-dimensional image is ignored, and IT is used to denote the intensity value at position(s) of the blood vessel in the two-dimensional image, and IB is used to denote the background intensity at the corresponding position(s) of the blood vessel in the two-dimensional image.
In some embodiments, the reconstruction of the three-dimensional image may be assisted using the linear relationship between the processed intensity value at each position in the two-dimensional image and the optical path length xC at the corresponding position. By applying the logarithm to both sides of formula (4), the following Equation (5) can be obtained.
It can be seen that the optical path length xC through the contrast agent is proportional to the processed image (i.e., that obtained by removing background from and logarithmically processing the image). That is, the value resulted by removing background and logarithmically processing the intensity value at each position of the blood vessel has a linear relationship with the optical path length xC at the corresponding position, as shown in
In some embodiments, the value resulted by removing background from and logarithmically processing the intensity value at each position of the blood vessel in a two-dimensional image, i.e., ln(IT)−ln(IB), may be obtained by the following steps: calculating the logarithm of the intensity value at each position of the blood vessel to obtain a first processed value ln(IT); calculating the logarithm of background intensity value at each position of the blood vessel to obtain a second processed value ln(IB); and subtracting the second processed value from the first processing value, so as to obtain the value resulted by removing background from and logarithmically processing the intensity value, ln(IT)−ln(IB).
In some embodiments, the relationship between the intensity value at each position of the blood vessel in a two-dimensional image and the optical path length at the corresponding position may be established in advance, in previous angiography and three-dimensional reconstruction of the same patient under the same contrast injection conditions, or the relationship is established in advance for part of the blood vessel in the same angiography and three-dimensional reconstruction. In some embodiments, in the same angiography and three-dimensional reconstruction, the foreshortening phenomenon may be avoided or decreased for the part of blood vessel, which is easy to achieve. And thus an accurate optical path length may be obtained by a reconstructed three-dimensional model based on the part of blood vessel, so that an accurate relationship between the intensity value at each position of the blood vessel in the two-dimensional image and the optical path lengths at the corresponding position may be established in advance. After the relationship is established in advance, the relationship can be recalled directly in a subsequent application scenario that satisfies the same contrast agent injection conditions.
When a three-dimensional model is reconstructed for a specific patient, difference in physiological characteristics (such as blood viscosity, respiratory motion, cardiac motion, etc.) and/or contrast agent parameters (such as injection time and injection volume) between previous and later angiographies may be small for him/her. Therefore, the relationship established in advance in the previous angiography and three-dimensional reconstruction or the relationship established in advance for the part of the blood vessel in the same angiography and three-dimensional reconstruction can be continuously adapted to the same patient. Compared to adopting the relationship obtained by means of the reconstruction of the three-dimensional model for other patients, it facilitates improving the accuracy of the reconstructed three-dimensional model of the blood vessel for the specific patient.
In the following embodiment, as shown in
At step 5041, a first processed image resulted by removing background from and logarithmically processing the second two-dimensional image may be calculated. For example, in some embodiments, the step of calculating the first processed image may include (not illustrated in the drawings): calculating the logarithm of an intensity value of each pixel of the second two-dimensional image to obtain a third logarithmically processed image; then, inpainting intensity values of the blood vessel portion in the second two-dimensional image based on intensity values of the background pixels of the periphery of the blood vessel portion; calculating the logarithm of an intensity value of each pixel of the imprinted second two-dimensional image, so as to obtain a fourth logarithmically processed image; and then subtracting the fourth logarithmically processed image from the third logarithmically processed image, so as to obtain the first processed image, which is exemplified by
At step 5042, the optical path length within the blood vessel at the position(s) may be estimated using the aforesaid linear relationship, which may be established in advance, based on the first processed image.
At step 5043, the optical path length within the blood vessel at the position(s) may be compared with the determined simulated optical path length within the vessel blood at the corresponding position(s), and the size of the blood vessel at the corresponding position(s) in the second projection direction in the three-dimensional model thereof may be elongated based on the comparison.
In one embodiment, the step 5043 may include: determining difference between the optical path length within the blood vessel at the position(s) and the simulated optical path length at the corresponding position(s) of the blood vessel; providing a warning if the difference is greater than a first predetermined threshold, otherwise, elongating the size of the blood vessel at the corresponding position(s) in the X-ray transmitting direction in the three-dimensional model of the blood vessel based on the difference, so as to eliminate the difference.
The first predetermined threshold may be a value preset empirically by a person skilled in the art, and this value is used to reflect the allowable degree of deviation of the reconstructed three-dimensional model. If the difference is greater than the first predetermined threshold, it means the deviation of the reconstructed three-dimensional model is relatively great, so a warning may be provided to draw the attention of a user (such as a surgeon or the like). Besides, if the difference is less than or equal to the first predetermined threshold, the size of the blood vessel at the corresponding position in the X-ray transmission direction in the three-dimensional model may directly modified (e.g., elongated) to eliminate the difference, thereby generating a calibrated blood vessel three-dimensional model.
In the following embodiment, as shown in
It begins with an acquisition step 702. At step 702, a reconstructed three-dimensional model of a blood vessel and a second two-dimensional image in the second projection direction corresponding thereto may be acquired.
Then, a simulated light path length determining step 703 is performed. At step 703, simulated optical path length within the blood vessel at position(s) in the second projection direction may be determined based on the three-dimensional model of the blood vessel.
After that, a three-dimensional reconstruction adjustment step is performed. With reference to
At step 7042, the second processed image resulted by removing the background from and logarithmically processing a simulated two-dimensional projection image may be estimated using the linear relationship based on the determined simulated optical path length, wherein the simulated two-dimensional projection image is obtained by projecting the three-dimensional model of the blood vessel in the second projection direction. That is, according to the simulated optical path length and the linear relationship, processed (including background removing and logarithmically processing) intensity values may be obtained for corresponding position of the blood vessel (more specifically, each pixel point of the blood vessel region).
At step 7043, the first processed image may be compared with the second processed image, and the reconstruction parameters of the three-dimensional model of the blood vessel may be adjusted based on the comparison. The adjusted reconstruction parameters may be used to perform three-dimensional blood vessel reconstruction using the X-ray angiographic images.
Specifically, the pixel value (i.e., the processed intensity value) at each pixel position of the first processed image may be compared with the processed intensity value at the corresponding pixel position of the second processed image, and the reconstruction parameters of the three-dimensional model of the blood vessel may be adjusted based on the comparison. And three-dimensional vessel model reconstruction may be performed using the adjusted reconstruction parameters.
In some embodiments, the cost function is set as the difference obtained by the above comparison, and the reconstruction parameters of the three-dimensional model of the blood vessel may be adjusted by minimizing the cost function.
In other embodiments, with the cost function defined as the above difference, steps of the three-dimensional vessel reconstruction and reconstruction parameters may be performed iteratively, the cost function may be calculated and fed into an optimizer to update the reconstruction parameters and the corresponding three-dimensional blood vessel tree geometry (i.e., generating a calibrated three-dimensional model of a blood vessel). For example, the reconstruction parameters can be gradually updated using a Newton iteration method or the like until optimized reconstruction parameters are obtained. The optimized reconstruction parameters may be used to reconstruct an accurate three-dimensional model of the blood vessel.
In still other embodiments, step 7043 may include: determining a difference between the first processed image and the second processed image; providing a warning if the difference is greater than a second predetermined threshold, otherwise, adjusting the reconstruction parameters of the three-dimensional model of the blood vessel based on the comparison, so as to eliminate the difference.
The second predetermined threshold may be a value preset empirically by a person skilled in the art, which reflects the allowable degree of deviation of the reconstructed three-dimensional model. If the difference is greater than the second predetermined threshold, the deviation of the reconstructed three-dimensional model is considered as relatively great, so a warning is provided to draw the attention of the user (such as a surgeon or the like), and if the difference is less than or equal to the second predetermined threshold, the reconstruction parameters of the three-dimensional model of the blood vessel may be directly adjusted, thus adjusting the corresponding three-dimensional geometry of the vessel tree.
In some embodiments, the acquisition unit 902 may acquire a vascular medical image from the medical image database 935. The acquired vascular medical image may include a three-dimensional model of the blood vessel and/or a second two-dimensional image in the second projection direction corresponding to the three-dimensional model of the blood vessel. In some other embodiments, the acquiring unit 902 can directly acquire the three-dimensional model of a blood vessel and/or a second two-dimensional image corresponding to the three-dimensional model of the blood vessel in the second projection direction from an external device such as a medical image acquisition device (not shown). In still other embodiments, the acquisition unit 902 may acquire the above model and/or image from an image data storage device (not shown). In a modified embodiment, the acquisition unit 902 can acquire the required models and images from at least two of the above sources.
In one embodiment, device 900 may further include a three-dimensional model reconstruction unit 901. The three-dimensional model reconstruction unit 901 is configured to generate a reconstructed blood vessel three-dimensional model based on a single-view projection image (for example, the second two-dimensional image in the second projection direction may be used). The three-dimensional model reconstruction unit 901 may be connected to any one of the medical image database 935, the image acquisition device, and the image data storage device, so as to acquire the two-dimensional image(s) based on which the reconstruction is performed. The acquisition unit 902 may acquire the reconstructed three-dimensional model of a blood vessel from the three-dimensional model reconstructing unit 901. In one embodiment, the acquisition unit 902 may also acquire, from the three-dimensional model reconstructing unit 901, the first two-dimensional images, based on which the blood vessel three-dimensional model is reconstructed, and the device 900 may use the first two-dimensional image as the second two-dimensional image.
The acquisition unit 902 transmits the acquired three-dimensional model of the blood vessel and the corresponding second two-dimensional image in the second projection direction to the simulated optical path length determining unit 903. The simulated optical path length determination unit 903 transmits the determined simulation optical path length to the three-dimensional reconstruction adjustment unit 904, so that it may adjust reconstruction parameters of the three-dimensional model of the blood vessel, based on the simulated optical path length within the blood vessel at position(s) thereof in the second projection direction, intensity value at the corresponding position(s) of the blood vessel on the second two-dimensional image, and a relationship between intensity value at each position of a blood vessel in a second-dimensional image and optical path length at the corresponding positions. Thus the adjusted reconstruction parameters may be adopted to reconstruct the three-dimensional blood vessel model using X-ray angiographic images. In some embodiments, the three-dimensional reconstruction adjustment unit 904 may output a calibrated three-dimensional model of a blood vessel.
For the specific implementation steps and methods of each unit of the device 900, reference may be made to corresponding steps and methods detailed in the foregoing method embodiments, and the description of which are omitted.
It is contemplated that the devices and methods disclosed in the embodiments may be implemented using a computer device. In some embodiments, the medical image processing device 1000 may be a dedicated smart device or a general-purpose smart device. For example, the medical image processing device 1000 may be a computer customized for image data acquisition and image data processing tasks, or a server placed in the cloud. For example, the device 1000 may be integrated into an image acquisition device. Optionally, the device may include or cooperate with a three-dimensional model reconstruction unit for generating a reconstructed three-dimensional model based on the two-dimensional images acquired by the image acquisition device.
The medical image processing device 1000 may include an image processor 1002 and a memory 1003, and may additionally include at least one of an input/output 1004 and an image display 1005.
The image processor 1002 may be a processing device including one or more general-purpose processing devices such as a microprocessor, a central processing unit (CPU), a graphics processing unit (GPU), and the like. More specifically, the image processor 1002 may be a complex instruction set computing (CISC) microprocessor, a reduced instruction set computing (RISC) microprocessor, a very long instruction word (VLIW) microprocessor, a processor running other instruction sets, or a processor that runs a combination of instruction sets. The image processor 1002 may also be one or more dedicated processing devices such as application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs), digital signal processors (DSPs), system-on-chip (SoCs), and the like. As would be appreciated by those skilled in the art, in some embodiments, the image processor 1002 may be a special-purpose processor, rather than a general-purpose processor. The image processor 1002 may include one or more known processing devices, such as a microprocessor from the Pentium™, Core™ Xeon™, or Itanium® family manufactured by Intel™, the Turion™, Athlon™, Sempron™ Opteron™, FX™, Phenom™ family manufactured by AIVID™, or any of various processors manufactured by Sun Microsystems. The image processor 1002 may also include graphical processing units such as a GPU from the GeForce®, Quadro®, Tesla® family manufactured by Nvidia™, GMA, Iris™ family manufactured by Intel™, or the Radeon™ family manufactured by AMD™. The image processor 1002 may also include accelerated processing units such as the Desktop A-4 (6, 8) Series manufactured by AIVID™, the Xeon Phi™ family manufactured by Intel™.
The disclosed embodiments are not limited to any type of processor(s) or processor circuits otherwise configured to meet the computing demands of identifying, analyzing, maintaining, generating, and/or providing large amounts of imaging data or manipulating such imaging data to calibrate a three-dimensional vessel model or to manipulate any other type of data consistent with the disclosed embodiments. In addition, the term “processor” or “image processor” may include more than one processor, for example, a multi-core design or a plurality of processors each having a multi-core design. The image processor 1002 can execute sequences of computer program instructions, stored in memory 1003, to perform various operations, processes, methods disclosed herein.
The image processor 1002 may be communicatively coupled to the memory 1003 and configured to execute computer-executable instructions stored therein. The memory 1003 may include a read only memory (ROM), a flash memory, random access memory (RAM), a dynamic random-access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM, a static memory (e.g., flash memory, static random-access memory), etc., on which computer executable instructions are stored in any format. In some embodiments, the memory 1003 may store computer-executable instructions of one or more image processing program(s) 923 and the data generated when the image processing program(s) are performed. The computer program instructions can be accessed by the image processor 1002, read from the ROM, or any other suitable memory location, and loaded in the RAM for execution by the image processor 1002, so as to implement each step of above methods. The image processor 1002 may also send/receive medical image data to/from storage 1003. For example, memory 1003 may store one or more software applications. Software applications stored in the memory 1003 may include, for example, an operating system (not shown) for common computer systems as well as for soft-controlled devices. Further, memory 1003 may store an entire software application or only a part of a software application (e.g., the image processing program (s) 923) to be executable by the image processor 1002. In some embodiments, the image processing program 923 may include the simulated optical path length determining unit 903 and the three-dimensional reconstruction adjustment unit 904 shown in
The image processor 1002 may execute an image processing program 923 to implement a method for three-dimensional vessel reconstruction using X-ray angiographic images. In some embodiments, when the image processing program 923 is executed, the image processor 1002 may associate the acquired reconstructed blood vessel three-dimensional model with the adjusted reconstruction parameters and the generated calibrated blood vessel three-dimensional model and store them in memory 1003. Alternatively, the image processor 1002 may associate the acquired reconstructed blood vessel three-dimensional model with the adjusted reconstruction parameters and the generated calibrated blood vessel three-dimensional model and send them to the medical image database 2000 via the network interface 1001.
It is contemplated that the device may include one or more processors and one or more memory devices. The processor(s) and storage device(s) may be configured in a centralized or distributed manner.
The device 1000 may include one or more digital and/or analog communication device (input/output 1004). For example, the input/output 1004 may include a keyboard and a mouse that allow the user to provide an input.
Device 1000 may be connected to the network through network interface 1001. The network interface 1001 may include a network adapter, a cable connector, a serial connector, a USB connector, a parallel connector, a high-speed data transmission adapter such as optical fiber, USB 3.0, lightning, a wireless network adapter such as a WiFi adapter, a telecommunication (3G, 4G/LTE, etc.) adapters. The network may provide the functionality of local area network (LAN), a wireless network, a cloud computing environment (e.g., software as a service, platform as a service, infrastructure as a service, etc.), a client-server, a wide area network (WAN), and the like.
The device 1000 may further include an image display 1005. In some embodiments, the image display 1005 may be any display device suitable for displaying a vascular angiographic image(s) and the three-dimensional reconstruction results. For example, the image display 1005 may be an LCD, CRT, or LED display.
Various operations or functions are described herein that may be implemented as software code or instructions or as software code or instructions. Such content may be directly executable source code or difference code (“incremental” or “block” code) (“object” or “executable” form). The software codes or instructions may be stored in a computer-readable storage medium and, when executed, may cause the machine to perform the described functions or operations and include any mechanism for storing information in a form accessible by the machine (e.g., computing devices, electronic systems, etc.), such as recordable or non-recordable media (e.g., read-only memory (ROM), random access memory (RAM), disk storage media, optical storage media, flash memory devices, etc.).
Although described using X-ray images, imaging modalities in the disclosed systems and methods may be alternatively or additionally applied to other imaging modalities where the pixel intensity varies with the distance traveled by imaging particles, such as CT, cone beam computed tomography (CBCT), Spiral CT, positron emission tomography (PET), single-photon emission computed tomography (SPECT), etc.
Following long-standing patent law convention, the terms “a”, “an”, and “the” refer to “one or more (at least one)” when used in this application, including the claims. Thus, for example, reference to “a unit” includes a plurality of such units, and so forth.
As used herein, the term “and/or” when used in the context of a listing of entities, refers to the entities being present singly or in combination. Thus, for example, the phrase “A, B, C, and/or D” includes A, B, C, and D individually, but also includes any and all combinations and combinations of A, B, C, and D.
Another aspect of the disclosure is directed to a non-transitory computer-readable medium storing instructions which, when executed, cause one or more processors to perform the methods, as discussed above. The computer-readable medium may include volatile or non-volatile, magnetic, semiconductor, tape, optical, removable, non-removable, or other types of computer-readable medium or computer-readable storage devices. For example, the computer-readable medium may be the storage device or the memory module having the computer instructions stored thereon, as disclosed. In some embodiments, the computer-readable medium may be a disc or a flash drive having the computer instructions stored thereon.
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed system and related methods. Other embodiments will be apparent to those skilled in the art from consideration of the specification and practice of the disclosed system and related methods.
It is intended that the specification and examples be considered as exemplary only, with a true scope being indicated by the following claims and their equivalents.
The present application is a continuation-in-part of U.S. application Ser. No. 16/895,573, filed Jun. 8, 2020, which is a continuation of U.S. application Ser. No. 16/106,077, filed Aug. 21, 2018, which claims the benefits of priority to U.S. Provisional Application No. 62/592,595, filed Nov. 30, 2017, now U.S. Pat. No. 10,709,399. The present application further claims the benefits of priority to U.S. Provisional Application No. 63/248,999, filed Sep. 27, 2021. The contents of all these applications are incorporated herein by reference in their entireties. The application further incorporates by reference the content of U.S. Provisional Application No. 62/591,437, filed Nov. 28, 2017.
Number | Date | Country | |
---|---|---|---|
62592595 | Nov 2017 | US | |
63248999 | Sep 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16106077 | Aug 2018 | US |
Child | 16895573 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16895573 | Jun 2020 | US |
Child | 17497980 | US |