The present application relates to the field of motion estimation and, in particularly, relates to a system and method for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue.
Target motion estimation is an important research direction in the field of medical image processing. In a medical image, a target in a soft tissue easily moves autonomously due to respiration or changes in position due to the influence of organ displacement and deformation. In clinical applications such as needle biopsy and real-time tracking radiotherapy, risks are brought to treatment accuracy; and in the imaging field, edge blur or artifacts of a target area may be caused. Therefore, the estimation on a motion signal of a target has a significant application value. In a clinical application environment, the uncertainty of the treatment can be reduced; and in the field of fast imaging, the method is conducive to removing the artifact or blurring phenomenon from the target area.
A traditional method for performing motion estimation on a target in a soft tissue is a technical solution based on registration, and the core thereof is to obtain an optimal deformation vector field to describe the displacement of each voxel in the soft tissue, so as to estimate the displacement of the target. The method can specifically include a deformation registration method based on gray scale and a deformation registration method combined with a biomechanical model.
The deformation registration based method is based on a three-dimension (three-dimension, 3D) image (denoted as V0). During the motion estimation process, a two-dimension (two-dimension, 2D) projection image (denoted as Ip) of the target tissue is acquired quickly by using the X-ray imaging technology. With Ip as a reference, a new 3D image (denoted as V1) is generated by deforming V0, until the projection image I1 of the V1 is optimally matched with Ip, and the corresponding V1 contains the motion displacement information of the target. The “optimal matching” in the method is usually based on the measurement criterion of image gray scale, which is easily affected by gray scale changes and noise. The time taken by the optimization iteration process required for registration also typically limits the rapidity of motion estimation. The registration combined with the biomechanical model is to integrate the knowledge of morphology, material analysis, tissue elasticity of an anatomical structure and the like into the registration process, the boundary condition limitation of registration is added, the registration accuracy is improved, and it is difficult to accurately describe the biomechanical characteristics of the anatomical structure during this process.
The technical problem to be solved by the present application is to overcome the shortcomings of the prior art and provide a system for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue. The system includes: an acquisition unit, a reference input unit, two surface extraction units, a target position extraction unit, a feature calculation unit, and a target motion estimation unit.
The acquisition unit is used for obtaining an actually captured image of the soft tissue; the reference input unit is used for inputting a reference image of the soft tissue; the surface extraction units are used for extracting soft tissue surfaces from the actually captured image and the reference image of the soft tissue; the target position extraction unit is used for extracting a reference position of the target from the reference image of the soft tissue; the feature calculation unit is used for calculating a deformation feature of the surface of the soft tissue in the actually captured image relative to the surface in the reference image; and the target motion estimation unit is used for calculating and outputting motion displacement estimation of the target based on the deformation feature and the reference position of the target.
Another object of the present application is to provide a method for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue, which is achieved through the following steps:
S1: a reference input unit uses medical imaging equipment to capture a soft tissue image as a reference image Iref.
In at least one embodiment of the present application, the “medical imaging equipment” includes: computed tomography (computed tomography, CT), cone-beam computed tomography (cone-beam computed tomography, CBCT), and ultrasonography.
In at least one embodiment of the present application, before the image is captured, one or more markers are implanted into a target area, and then the soft tissue image is captured by using the medical imaging equipment.
S2: a target position extraction unit identifies, calculates and outputs a reference position (denoted as tref) of the target from Iref.
S3: a surface extraction unit extracts a soft tissue surface (denoted fref) from Iref.
In at least one embodiment of the present application, the method of “extracting the soft tissue surface” includes: an automatic edge recognition algorithm based on adaptive threshold segmentation or a fully convolutional neural network model.
S4: an acquisition unit uses medical imaging equipment to capture an actually captured image (denoted as Ii) of the soft tissue.
In at least one embodiment of the present application, the “medical imaging equipment” includes: CT, CBCT, and ultrasonography.
S5: the surface extraction unit extracts a soft tissue surface (denoted as fi) from Ii.
S6: a feature calculation unit calculates and outputs a deformation feature (denoted as Ψi) of fi relative to fref.
In at least one embodiment of the present application, the method of “calculating and outputting the deformation feature of fi relative to fref” is: inputting fi and fref into Ñ neural network models M={Mj|j=1, . . . , Ñ}, and the output results of the Ñ models together constitute the deformation feature Ψi={Ψi,j|j=1, . . . , Ñ}.
S7: Ψi and trefare input into a target motion estimation unit, and a target motion estimation model (m) in the unit calculates and outputs motion estimation (denoted as {circumflex over (t)}i) of the target based on Ψi and tref.
The design method of the “Ñ neural network models M={Mj|j=1, . . . , Ñ}” in the “feature calculation unit” in the step S6 is as follows.
(1) Modeling: establishing a fully convolutional neural network model (denoted as FCN), wherein an input layer of FCN is 2 pieces of surface data, the hidden layers include {l1, . . . lN−i}, and an output layer is lN . A deformation vector field (denoted as ϕk) of the 2 pieces of surface data is output by lN.
(2) Collecting training data: using medical imaging equipment to collect multiple groups of soft tissue images Ik, extracting changing surfaces fk (k=1, 2, . . . , n) of the soft tissue from Ik, taking any one of the surfaces {fk|k=1, . . . , n} as a reference surface (denoted as fref), and taking the rest surfaces as changing surfaces to form training sample pairs {(fk, fref)|k=1, . . . ,n and k≈ref} together.
In at least one embodiment of the present application, the “medical imaging equipment” includes four-dimension (four-dimension, 4D) CT, 4D CBCT, and three-dimension ultrasonography. The soft tissue surface is directly delineated or identified by using automatic threshold segmentation or the neural network from the collected images.
(3) Training and optimizing FCN: inputting {(fk, fref)|k=1, . . . , , n and k≈ref} into FCN, performing iterative optimization on the model by using unsupervised learning, setting a loss function as the difference between fk and a generated surface. The generated surface (ϕk·fref) is achieved by applying ϕk on fref, and when the loss function is optimal, terminating the optimization.
In at least one embodiment of the present application, the index for measuring the difference between ϕk·fref and fk is the sum of minimum distances from all points in ϕk19 fref to fk.
(4) Constructing M={Mj|j=1, . . . ,Ñ}: in the layer structure {l1, . . . lN} of the trained FCN, and taking Ñ layers ({lk
If kj=1, then Mj is only composed of an input layer and an output layer (l1).
In at least one embodiment of the present application, the “{lk
The design method of the “target motion estimation model (m)” in the step S7 is as follows.
(1) Data collection: using medical imaging equipment to capture multiple groups of soft tissue images Ip, and identifying and calculating a soft tissue surface fp and a target position tp from Ip, wherein p=1, 2, . . . , n′.
In at least one embodiment of the present application, the “medical imaging equipment” includes 4D CT and three-dimension ultrasonography. The soft tissue images acquired by 4D CT are 10 groups of 3D CT scans which are uniformly sampled within one breathing cycle. Each 3D CT image contains a 3D image of the soft tissue surface and the target. Or, by implanting one or more markers into the target area of the soft tissue, surface changes of the soft tissue and the displacement of the marker are captured by using the 3D ultrasonography. The soft tissue surface and the target position are directly delineated or identified by using automatic threshold segmentation or the neural network from the collected images.
(2) Calculating a deformation feature Ψp to form training data: randomly taking a group from the collected {(fp, tp)|p=1,2, . . . , n′} as reference samples, marking the same as (fref, tref), taking the rest as change samples {(fp, tp)|p=1,2, . . . n′ and p≠ref}, and inputting {(fp, fref)|p=1, 2, . . . , n′ and p≠ref} into the feature calculation unit to generate the deformation feature Ψp. The deformation feature constitutes training data {(Ψp, tp, tref)|p=1,2, . . . , n′ and p≠ref} together with tp and tref.
(3) Fitting the target motion estimation model (m): the input data of the model is tref and Ψp, the output is the displacement estimation (denoted as {circumflex over (t)}p) of the target. Through iterative optimization, the difference between {tilde over (t)}p output by the model and a true value tp thereof is minimized.
In at least one embodiment of the present application, the “difference between {tilde over (t)}p output by the model and the true value tp thereof” is measured by the Euclidean distance between {tilde over (t)}p and tp.
In at least one embodiment of the present application, the “target motion estimation model (m)” includes: a neural network model, a fully convolutional neural network model, a linear model, and a support vector machine model.
The present application provides a system and method for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue. The beneficial effects of the system and the method are: the displacement of the target is estimated depending on the deformation registration method in the prior art, the registration process is a process of obtaining an optimal solution through repeated calculations, and this process increases the time delay for estimating the displacement of the target; and on the other hand, the optimal solution for registration is often based on the gray scale similarity between a deformed image and a reference image, which determines that the accuracy of target displacement estimation in the prior art is affected by the gray difference, low contrast, noise and the like of the images. Different from the prior art, the solution provided by the present application is to fit the deformation feature with the displacement of the target, and does not include the process of repeated calculations, thus ensuring the timeliness of the estimation; and on the other hand, the deformation feature mentioned in the present application is only about the deformation of the soft tissue surface, unlike the prior art where the entire image is used as a registration object, the present application only focuses on the deformation of the surface profile, thus reducing the impact of the image quality on the accuracy. Therefore, the present application can guarantee the accuracy and the timeliness of target displacement estimation at the same time, and has a significant advantage compared with the prior art.
The present application will be further explained in conjunction with the drawings and examples.
A system for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue, as shown in
A method for estimating the motion of a target inside a tissue based on surface deformation of the soft tissue, as shown in
S1: a reference input unit uses medical imaging equipment to capture a soft tissue image as a reference image Iref.
In at least one embodiment of the present application, the “medical imaging equipment” includes: CT, CBCT and ultrasonography.
In at least one embodiment of the present application, before the image is captured, one or more markers are implanted into a target area, and then the soft tissue image is captured by using the medical imaging equipment.
S2: a target position extraction unit identifies, calculates and outputs a reference position (denoted as tref) of the target from Iref.
S3: a surface extraction unit extracts a soft tissue surface (denoted as fref) from Iref.
In at least one embodiment of the present application, the method of “extracting the soft tissue surface” includes: an automatic edge recognition algorithm based on adaptive threshold segmentation or a fully convolutional neural network model.
S4: an acquisition unit uses medical imaging equipment to capture an actually captured image (denoted as Ii) of the soft tissue.
In at least one embodiment of the present application, the “medical imaging equipment” includes: CT, CBCT, and ultrasonography.
S5: the surface extraction unit extracts a soft tissue surface (denoted as fi) from Ii.
S6: a feature calculation unit calculates and outputs a deformation feature (denoted as Ψi) of fi relative to fref.
In at least one embodiment of the present application, the method of “calculating and outputting the deformation feature of fi relative to fref” is: inputting fi and fref into Ñ neural network models M={Mj|j=1, . . . , Ñ}, and the output results of the Ñ models together constitute the deformation feature Ψi={Ψi,j|j=1, . . . , Ñ}.
S7: Ψi and tref are input into a target motion estimation unit, and a target motion estimation model (m) in the unit calculates and outputs motion estimation (denoted as {circumflex over (t)}i) of the target based on Ψi and tref.
The design method of the “Ñ neural network models M={Mj|j=1, . . . , Ñ}” in the “feature calculation unit” in the step S6 is:
(1) Modeling: establishing a fully convolutional neural network model (denoted as FCN), wherein an input layer of FCN is 2 pieces of surface data, the hidden layers include {l1, . . . lN−1} and an output layer is lN. A deformation vector field ϕk of the 2 pieces of surface data is output by IN.
(2) Collecting training data: using medical imaging equipment to collect multiple groups of soft tissue images Ik, extracting changing surfaces fk(k=1, 2, . . . , n) of the soft tissue from Ik, taking any one of the surfaces { fk|k=1, . . . ,n} as a reference surface (denoted as fref), and taking the rest surfaces as changing surfaces to form training sample pairs {(fk, fref)|k=1, . . . ,n and k≠ref} together.
In at least one embodiment of the present application, the “medical imaging equipment” includes 4D CT, 4D CBCT, and three-dimensional ultrasonography. The soft tissue surface is directly delineated or identified by using automatic threshold segmentation or the neural network from the collected images.
(3) Training and optimizing FCN: in the embodiment shown in
In at least one embodiment of the present application, the index for measuring the difference between ϕk·fref and fk is the sum of minimum distances from all points in ϕk·fref to fk.
(4) Constructing M={Mj|j=1, . . . , Ñ}: in the embodiment shown in
If j=1, then M1 is only composed of an input layer and an output layer (li).
The design method of the “target motion estimation model (m)” in the step S7 is:
(1) data collection: using medical imaging equipment to capture multiple groups of soft tissue images Ip, and identifying and calculating a soft tissue surface fp and a target position tp from Ip, wherein p=1,2, . . . , n′.
In at least one embodiment of the present application, the “medical imaging equipment” includes 4D CT and three-dimensional ultrasonography. The soft tissue images acquired by 4D CT are 10 groups of 3D CT scans which are uniformly sampled within one breating cycle. Each 3D CT image contains a 3D image of the soft tissue surface and the target. Or, by implanting one or more markers into the target area of the soft tissue, surface changes of the soft tissue and the displacement of the markers are captured by using the 3D ultrasonography. The soft tissue surface and the target position are directly delineated or identified by using automatic threshold segmentation or the neural network from the collected images.
(2) Calculating a deformation feature Ψp to form training data: randomly taking a group from the collected {(fp, tp)|p=1,2, . . . n′} as reference samples, marking the same as (fref, tref), taking the rest as change samples {(fp,tp)|p=1,2, . . . ,n′ and p≠ref}, and inputting {(fp,fref)|p=1,2, . . . , n′ and p≠ref} into the feature calculation unit to generate the deformation feature Ψp. The deformation feature constitutes training data {(Ψp, tp, tref)|p=1,2, . . . , n′ and p≠ref} together with tp and tref.
(3) Fitting the target motion estimation model (m): in the embodiment shown in
In at least one embodiment of the present application, the “difference between {circumflex over (t)}p output by the model and the true value tp thereof” is measured by the Euclidean distance between {circumflex over (t)}p and tp.
In at least one embodiment of the present application, the “target motion estimation model (m)” includes: a neural network model, a fully convolutional neural network model, a linear model, and a support vector machine model.
The core of the technical solution of the present application for realizing target displacement estimation is fitting, that is, correlating the target displacement with the deformation feature of the soft tissue surface using a mathematical expression. Since the speed at which the mathematical expression gives an estimated value is only related to the own calculation speed of a computer, the present application can ensure the timeliness of estimation. The source of the motion of the target in the soft tissue is the deformation of the soft tissue, and the most obvious deformation is a change on its surface, therefore, in the solution provided by the present application, the feature describing the surface deformation of the soft tissue is used as an association object, and the deformation feature is extracted by reconstructing the output of the hidden layers of the neural network, the function of the neural network is to realize the matching of the actually measured surface and the reference surface, and in order to avoid the influence of the gray scale of the image, the matched object adopts a surface contour extracted from the image.
Number | Date | Country | Kind |
---|---|---|---|
202011118867.9 | Oct 2020 | CN | national |
The present application is a continuation of International Application No. PCT/CN2021/080052, filed on Mar. 10, 2021, which claims priority to Chinese Application No.202011118867.9, filed on Oct. 19, 2020, the contents of both of which are incorporated herein by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
20170109871 | Nakano | Apr 2017 | A1 |
20180005391 | Miga | Jan 2018 | A1 |
20190057505 | Pheiffer | Feb 2019 | A1 |
20190057517 | Koizumi | Feb 2019 | A1 |
20190325572 | Mansi | Oct 2019 | A1 |
20200146635 | Wang | May 2020 | A1 |
20200219272 | Pizer et al. | Jul 2020 | A1 |
Number | Date | Country |
---|---|---|
102949240 | Mar 2013 | CN |
103295234 | Sep 2013 | CN |
103530884 | Jan 2014 | CN |
104318553 | Jan 2015 | CN |
112330603 | Feb 2021 | CN |
Entry |
---|
International Search Report (PCT/CN2021/080052); Date of Mailing: Jul. 19, 2021. |
Number | Date | Country | |
---|---|---|---|
20220130048 A1 | Apr 2022 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2021/080052 | Mar 2021 | WO |
Child | 17571562 | US |