The present disclosure relates to the technical field of Computed Tomography (CT), and in particular, to an imaging and reconstruction method.
Computed tomography (CT) is an important research direction in the fields of medical imaging and computer vision graphics. This technology can reconstruct the internal structure of the object by measuring the amount of light absorbed by the scene in different directions. This technology has a wide range of application scenarios, such as medical imaging, industrial monitoring, aviation security inspection and cultural relics protection.
In particular, it is of great scientific and application value to apply tomography technology to dynamic scenes. For example, mechanical inspection and medical diagnosis both require three-dimensional reconstruction of high-speed dynamic scenes. However, extending traditional CT to dynamic scenes faces a critical challenge: since high-quality reconstruction results are often based on intensive sampling in different directions, when the scene changes rapidly, the intensive sampling process must be completed in a short time to avoid the afterimage problem in reconstruction and ensure the reconstruction quality. This makes it necessary for dynamic scene-oriented CT to far exceed the high sampling ability of traditional methods.
Over the past decades, various studies have proposed different algorithms for CT acquisition and reconstruction of dynamic scenes. For a specific dynamic phenomenon, the properties of the observed scene can be used for solution. For example, Chen et al. proposed a scanning reconstruction algorithm (Chen Guang-Hong, Theriault-Lauzier Pascal, Tang Jie, Nett Brian, Leng Shuai, Zambelli Joseph, Qi Zhihua, Bevins Nicholas, Raval Amish, Reeder Scott. 2011. Time-resolved interventional cardiac C-arm cone-beam CT: An application of the PICCS algorithm. IEEE transactions on medical imaging. 31, 4, 907-923). However, such methods are limited to specific scene characteristics and lack universality. Some studies aim to improve sampling speed by reducing the number of point light source samples and making strong prior assumptions for reconstruction. However, the dependence on these assumptions limits the applicability of the methods to certain scenarios (Zang Guangming, Idoughi Ramzi, Wang Congli, Bennett Anthony, Du Jianguo, Skeen Scott, Roberts William L., Wonka Peter, Heidrich Wolfgang. 2020. TomoFluid: reconstructing dynamic fluid from sparse view videos. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1870-1879). Therefore, it is urgent to propose a computed tomography acquisition and reconstruction algorithm for general dynamic scenes.
The present disclosure aims at solving the problems that the existing computed tomography scanning method has insufficient acquisition ability in dynamic scenes and a strong prior assumption is required for reconstruction, and provides a method which greatly improves the acquisition ability and is oriented to general scenes.
In order to achieve the above object, the present disclosure provides a method for computed tomography imaging and reconstruction based on learning, which measures the scene density distribution in an illumination multiplexing manner, including the following steps:
where Wraw represents a parameter to be trained; Wl corresponds to an illumination intensity matrix during imaging, with a size of k×ns, ns represents a number of light sources of the scanning device, and k represents a number of samples; and fw represents a mapping, and is configured for transforming Wraw, so that the generated illumination intensity matrix Wl corresponds to a possible illumination intensity of the light sources;
Further, in the step (1), the method of generating the CT images further includes: randomly placing several objects with different densities in an effective area of the scene, and generating the CT images based on a selected ray model according to positions of the light sources and sensors obtained by calibration.
Further, the ray model is a linear absorption model, with an equation as follows:
where I represents a matrix composed of the CT images of different light sources, with a size of ns×nd, an element Iij in I represents a reading of a jth sensor when a ith light source emits light at a maximum intensity in a given density field, x represents a vector after the density field is discretized into voxels, with a length being a number of the voxels nv, K represents a Radon transform represented by a third-order tensor, x3 represents a mode-3 product of the tensor and the vector, ⊙ represents an element-by-element multiplication between matrices, and Ĩ represents I when the density field is 0 everywhere.
Further, in the step (2), a relationship between the linear fully connected layer and the input is as follows:
where I represents a matrix composed of the CT images of different light sources, with a size of ns×nd.
Further, in the step (2), the neural network for reconstruction is expressed as follows:
where M is mapped into Sinogram Dnn by fnn, and a density field reconstruction result xnn is obtained by using a computed tomography reconstruction method frecon.
Further, the computed tomography reconstruction method is realized by using a filtered back projection method.
Further, in the step (2), a loss function used for training is expressed as follows:
where is used to evaluate a density field reconstruction quality,
is used to allow the illumination intensity to have a specific property, gw represents a function adopted for evaluating the property of the illumination intensity, and λr and λp are used to balance the weights between different loss functions.
Further, the loss function for evaluating the density field reconstruction quality can be expressed as:
Further, the function gw for actually evaluating the property of the illumination intensity can be designed according to different scenes, and the following two design methods are given, but not limited thereto:
I, for a dynamic scene requiring high-speed scanning, gw(Wl)=−Σ|Wl|, so that a value of Wl tends to be binary.
II, for a scene requiring low-dose scanning, gw(Wl)=Σ|Wl|, so that the value of Wl tends to be minimized.
The present disclosure has the beneficial effects that the computed tomography imaging and reconstruction method of illumination multiplexing is obtained through learning. Compared with traditional computed tomography methods for dynamic scenes, which are mainly designed for setups using point light source, the method according to embodiments of the present disclosure can sample in all ray spaces, thereby greatly improving the acquisition efficiency. Thanks to the processor to implement the method, the density field can be reconstructed without strong prior assumptions. Therefore, the method is suitable for general scenes. Compared with the traditional illumination multiplexing imaging method, in the method according to embodiments of the present disclosure, the processor can use a neural network to acquire the used illumination intensity and reconstruction algorithm, thereby greatly reducing the required number of inputs, and achieving the high quality reconstruction. The proposed computed tomography method can collect and reconstruct dynamic scenes, which cannot be implemented by the traditional methods under the same hardware conditions. In addition, this method is not limited to a specific scanning device, and thus has great value of general applicability.
In order to make the purpose, technical solution and advantages of the present disclosure more clear, the present disclosure will be further described in detail with the attached drawings and specific embodiments.
The present disclosure provides a method for computed tomography imaging and reconstruction based on learning, which measures the scene density distribution in an illumination multiplexing manner. The steps are shown in
In an embodiment, the scanning device includes at least two light sources and at least one sensor, each sensor can receive light from a plurality of light sources at the same time, and the intensities of all light sources are adjustable; the target scene is in part or all of the light path formed by the light source and the sensor.
In one embodiment, the scene is detected by visible light, and
Different ray models can be selected to generate CT images. In this embodiment, a linear absorption model is selected, but it is not limited to this. The formula of the linear absorption model is as follows:
where I is a matrix composed of the CT images of different light sources, with a size of ns×nd, an element Iij in I is a reading of a jth sensor when a ith light source emits light at a maximum intensity in a given density field, x is a vector after the density field is discretized into voxels, with a length being a number of the voxels nv, K is a Radon transform represented by a third-order tensor, x3 is a mode-3 product of the tensor and the vector, ⊙ is an element-by-element multiplication between matrices, and Ĩ is I when the density field is 0 everywhere; the element Ĩij in Ĩ is a reading of the jth sensor when the ith light source emits light at a maximum intensity in the empty scene.
where Wraw is a parameter to be trained; Wl corresponds to an illumination intensity matrix during imaging, with a size of k×ns, ns is a number of light sources of the scanning device, and k is a number of samples; fw is a mapping, which is used for transforming Wraw, so that the generated illumination intensity matrix Wl can correspond to a possible illumination intensity of the light sources; in this embodiment, a Sigmoid function is selected for fw to transform Wraw to [0,1], so that the elements of Wl have practical physical significance; and the relationship between the linear fully connected layer and the input is as follows:
where M represents a measured value matrix of the sensor.
In an embodiment, the neural network used for reconstruction is expressed as follows:
where M is mapped into Sinogram Dnn by fnn, and then using a Computed Tomography reconstruction method frecon to obtain a density field reconstruction result xnn is obtained by using a computed tomography reconstruction method frecon.
In this embodiment, the measured value matrix is divided into 24 sub-matrices according to the acquisition modules, and the sub-matrix with the acquisition module serial number of Gi is shifted to the left by Gi×8 pixels in rows, so that all sub-matrices share the same parameterized form to take advantage of the rotation invariance of the scanning geometry in this embodiment. The moved measurement matrix is reconstructed by 24 decoding networks sharing parameters to obtain the sinogram of each acquisition module, which is assembled into the reconstructed sinogram after the serial number of the acquisition module is moved back to the original position. Finally, the density field is reconstructed from the differentiable 3D-FBPNet. It should be noted that the network structure shown here is the technical solution of one embodiment of the present disclosure, but not the limitation. The loss function for training can be expressed as follows:
where is used to evaluate a density field reconstruction quality,
is used to allow the illumination intensity to have a specific property, gw represents a function adopted for evaluating the property of the illumination intensity, and λr and λp are used to balance the weights between different loss functions. λr and λp are preferably 1.0 and 1e−5.
In order to improve the scanning speed, the illumination intensity should be encouraged to increase, therefore gw is preferably:
Corresponding to the embodiment of the method for computed tomography imaging and reconstruction based on learning, the present disclosure further provides an embodiment of a learning-based Computed Tomography imaging and reconstruction apparatus.
The learning-based Computed Tomography imaging and reconstruction apparatus provided by the embodiment of the present disclosure comprises a memory and one or more processors. The memory stores executable codes, and when the executable codes are executed by the processors, the method for computed tomography imaging and reconstruction based on learning in the embodiment is implemented.
The embodiment of the learning-based Computed Tomography imaging and reconstruction apparatus of the present disclosure can be applied to any device with data processing capability, which can be a device or device such as a computer. The embodiment of the apparatus can be realized by software, or by hardware or a combination of hardware and software. Taking software implementation as an example, as a logical device, it is formed by reading the corresponding computer program instructions in the non-volatile memory into the memory through the processor of any equipment with data processing capability. In the hardware level, in addition to the processor, memory, network interface, and nonvolatile memory, any device with data processing capability where the apparatus in the embodiment is located may further include other hardware according to the actual function of the device with data processing capability, which will not be described here again.
The embodiment of the present disclosure further provides a computer-readable storage medium, on which a program is stored, which, when executed by a processor, implements the method for computed tomography imaging and reconstruction based on learning in the above embodiment.
The computer-readable storage medium can be an internal storage unit of any device with data processing capability as described in any of the previous embodiments, such as a hard disk or a memory. The computer-readable storage medium can further be an external storage device of any device with data processing capability, such as a plug-in hard disk, Smart Media Card (SMC), SD card, Flash Card and the like provided on the device. Further, the computer-readable storage medium can further include both internal storage units and external storage devices of any device with data processing capability. The computer-readable storage medium is used for storing the computer program and other programs and data required by any device with data processing capability, and can further be used for temporarily storing data that has been output or will be output. The above is only the preferred embodiment of one or more embodiments of this specification, and it is not intended to limit one or more embodiments of this specification. Any modification, equivalent substitution, improvement and the like made within the spirit and principle of one or more embodiments of this specification shall be included in the protection scope of one or more embodiments of this specification.
The present application is a continuation of International Application No. PCT/CN2023/102613, filed on Jun. 27, 2023, the contents of which is incorporated herein by reference in its entirety.
| Number | Date | Country | |
|---|---|---|---|
| Parent | PCT/CN2023/102613 | Jun 2023 | WO |
| Child | 18777615 | US |