The present invention relates generally to diagnostic imaging. More specifically, it relates to methods for image reconstruction in computed tomography and magnetic resonance imaging.
Image reconstruction is an inverse problem that solves for a computational image based on sampled sensor measurement. Sparsely sampled image reconstruction poses addition challenges due to limited measurements.
In this work, we propose an implicit Neural Representation learning methodology with Prior embedding (NeRP) to reconstruct a computational image from sparsely sampled measurements. The method differs fundamentally from previous deep learning-based image reconstruction approaches in that NeRP exploits the internal information in an image prior, and the physics of the sparsely sampled measurements to produce a representation of the unknown subject. No large-scale data is required to train the NeRP except for a prior image and sparsely sampled measurements. In addition, we demonstrate that NeRP is a general methodology that generalizes to different imaging modalities such as CT and MRI. We also show that NeRP can robustly capture the subtle yet significant image changes required for assessing tumor progression.
Conventionally, a neural network is trained to learn the mapping from the sampled measurement data to reconstruct images based on a large-scale training database. The proposed NeRP model learns the network, i.e. multi-layer perceptron (MLP), to map the image spatial coordinates to the corresponding intensity values. The neural network learns the continuous implicit neural representation of the entire image by encoding the full image spatial field into the weights of MLP model. Image reconstruction is then reduced to simply querying the model embedded in the network. The image reconstruction problem is transformed into a network optimization problem. Instead of searching for the best match in the image space, we reconstruct the image by searching for it in the space of networks weights. The MLP is trained by matching the subsampled measurements in raw data space (e.g. projection space sampling for CT or frequency space sampling for MRI), then integrating the forward model of the corresponding imaging system. For sparse sampling, the measurements may not provide sufficient information to precisely reconstruct images of the unknown subject due to the ill-posed nature of the inverse problem. The proposed NeRP framework exploits prior knowledge from a previous image for the same subject. This is particularly applicable to clinical protocols where patients are scanned serially over time, such as monitoring tumor response to therapy. The implicit neural representation first embeds the internal information of the prior image into the weights of MLP. This serves as the initialization for the search for the representation of the target image. Starting from this prior-embedded initialization, the network can be optimized to reach an optimal point in the function space given only sparsely sampled measurements. Finally, the learned MLP can generate the image reconstruction by traversing all the spatial coordinates in the image space. Note that NeRP requires no training data form external subjects except for the sparsely sampled measurements and a prior image of the subject.
The main contributions of this work are:
1) We present a novel deep learning methodology for sparsely sampled medical image reconstruction by learning the implicit neural representation of image with prior embedding (NeRP). Our method requires no training data from external subjects and can be easily generalized across different imaging modalities and different anatomical sites.
2) We propose a prior embedding method in implicit neural representation learning by encoding internal information of the prior image into network parameters as the initialization of network optimization, which enables sparsely sampled image reconstruction.
3) We present extensive experiments for both 2D and 3D image reconstruction with various imaging modalities, including CT and MRI, and demonstrate the effectiveness and generalizability of the proposed NeRP method. In particular, we show that our method is robust to capture subtle yet significant structural changes such as those due to tumor progression.
Commercial Applications:
(1) Tomographic imaging with sparsely sampling to significantly speed up data acquisition and/or reduce the potential radiation side effects.
(2) Magnetic resonance imaging from under-sampled k-space for accelerated acquisition and reconstruction.
(3) Simplified hardware design with substantially reduced efforts in data acquisition.
(4) Other imaging modalities and devices/systems that can be formulated as an inverse problem.
Although many existing methods have shown the advantages of deep learning models for medical image reconstruction, there are increasing concerns regarding the limitations of the current deep learning approaches. First, training deep neural networks is data-intensive and requires large-scale datasets. This may prevent many practical applications due to the difficulty of data collection. The limited availability of specific image modalities or images of rare diseases may make it difficult to acquire sufficient training data for deep learning modeling. In addition, there is a common concern about the robustness and reliability of deep reconstructed images. For example, recent study finds that the small but significant structural changes in tumors or lesions may not be accurately captured in the deep reconstructed images. Finally, the generalization capability of deep networks is unclear. The trained deep networks may produce poor performance when generalized to data out of the training distribution, such as across different imaging modalities or across different anatomical sites. The model generalizability is related to not only the training data distribution but also the network structure. For example, it is a non-trivial task to directly transfer a deep network developed for MRI reconstruction to the task of CT reconstruction due to different sensor measurements fields. Because of these issues and limitations, new insights are needed in developing deep learning-based image reconstruction algorithms.
The proposed NeRP method provides a new perspective to the problems of image reconstruction, which promises to overcome the shortcomings of previous deep reconstruction approaches. First, NeRP requires no training data from external subjects, but the sparsely sampled sensor measurements data for the target image and a prior image from a previous scan of the same subject. In addition, from our experiments, the reconstructed images from NeRP are more robust and reliable, and can capture the small structural changes such as tumor or lesion progression. The implicit image priors captured by network structure and the prior embedding can effectively incorporate the prior knowledge in the reconstruction process, which makes it possible for the network to capture and reconstruct the fine structural details in the resultant images. As a result of the previous two points, NeRP can be more easily generalized to different image modalities, different anatomical sites, and different dimensionalities in the image reconstruction task. The relaxed requirements for training data make the method more transferrable and applicable across various applications. The proposed NeRP is a general methodology for medical image reconstruction with promising advantages over mapping-learning-based deep reconstruction methods.
Beyond image reconstruction, the proposed method can be extended to solve other inverse problems with measurements (or sparse measurements) acquired at different time points or under different environments/conditions. Broadly, the proposed method is able to tackle problems with spatial coordinate-dependence, such as dose (or other physical quantities such as wave, heat, magnetic field, sound waves, and so forth) distribution prediction, semantic segmentation, and cross-modality image translation. In these cases, the output of MLP network would be corresponding values with semantics in different contexts. Additionally, the proposed method has potential to deal with the longitudinal problems, where the time dimension will be added as another dimension in the input of MLP network. In this way, both the temporal and spatial information and correlation are embedded into network parameters through neural representation learning, which can then be used for solving the longitudinal problems in clinical protocol to reveal quantities such as tumor progression and image content changes.
Image reconstruction is conventionally formulated as an inverse problem, with the goal of obtaining the computational image of an unknown subject from measured sensor data. For example, projection data are measured for computed tomography imaging (CT) while frequency domain (k-space) data are sampled for magnetic resonance imaging (MRI). To reconstruct artifact-free images, dense sampling in measurement space is required to satisfy the Shannon-Nyquist theorem. However, in many practical applications it would be desirable to reconstruct images from sparsely sampled data. One important application is reducing radiation dose in CT imaging. Another application is accelerating MRI. The ill-posed nature of the sparse sampling image reconstruction problem poses a major challenge for algorithm development. Many approaches have been studied to solve this problem. One widely used approach is to exploit prior knowledge of the sparsity of the image in a transform domain, such as in compressed sensing, where total-variation, low-rankness, and dictionary learning have been applied [1]-[7].
With the unprecedented advances in deep learning, deep neural networks driven by learning from large-scale data have achieved impressive progress in many fields including computational image reconstruction. Many research works have introduced deep learning models for medical imaging modalities such as CT and MRI [8], [9]. The key of these deep learning approaches is training convolutional neural networks (CNNs) to learn the mapping from raw measurement data to the reconstructed image by exploiting the large-scale training data as shown in
Although these works show the advantage of deep learning for medical image reconstruction, they have also exposed some limitations. For example, the acquisition of large-scale training data sets can be a bottleneck, the reconstructions may not be robust when deployed to unseen subjects, the reconstructions can be unstable with subtle yet significant structural changes such as tumor growth, and there can also be difficulties generalizing to different image modalities or anatomical sites [16]. To address these limitations, we propose a new insight for deep learning methodology for image reconstruction. We propose to learn the implicit Neural Representation of an image with Prior embedding (NeRP), instead of learning the reconstruction mapping itself. This is an essentially different perspective from previous deep learning-based reconstruction methods.
The NeRP is illustrated in
The main contributions of this work are:
1) We present a novel deep learning methodology for sparsely sampled medical image reconstruction by learning the implicit neural representation of image with prior embedding (NeRP). Our method requires no training data from external subjects and can be easily generalized across different imaging modalities and different anatomical sites.
2) We propose a prior embedding method in implicit neural representation learning by encoding internal information of the prior image into network parameters as the initialization of network optimization, which enables sparsely sampled image reconstruction.
3) We present extensive experiments for both 2D and 3D image reconstruction with various imaging modalities, including CT and MRI, and demonstrate the effectiveness and generalizability of the proposed NeRP method. In particular, we show that our method is robust to capture subtle yet significant structural changes such as those due to tumor progression.
Related Work
Deep Learning for Medical Image Reconstruction
Most previous work in deep learning-based medical image reconstruction builds the deep neural networks to learn the mapping from sampled measurements to reconstructed images [8]-[11]. Zhu et al. [8] introduced a CNN model for manifold learning and mapping between the sensor and image domain, which accelerates the acquisition and reconstruction for MRI imaging. Meanwhile, Shen et al. [9] proposed a deep network with a transformation module in high-dimensional feature space to learn the mapping from 2D projection radiographs to 3D anatomy for reconstructing volumetric CT images from sparse views. Later on, adversarial learning was introduced to train the CNN model for obtaining better reconstruction image quality [10], [11]. Mardani et al. [10] used generative adversarial networks (GAN) to model the low-dimensional manifold of MR images. Ying et al. [11] adopted the adversarial loss in the network training objective for 2D-3D CT reconstruction. Although GAN-based methods can reconstruct more realistic images with sharper image quality, the generated fine details and structures may not be reliable due to the synthetic nature of GAN models [11].
Recently, human prior knowledge has been considered to guide the construction of deep learning models [12]-[15]. For example, geometry-integrated deep learning methods were proposed for CT image reconstruction [12] and X-ray projection synthesis [15] by incorporating the physics and geometry priors of imaging model to build a unified deep learning framework. Lin et al. [13] presented a dual-domain network for metal artifact reduction in CT imaging with a differentiable Radon inversion layer connecting sinogram and image domains. Würfl et al. [14] mapped filtered back-projection to neural networks and introduced a cone-beam back-projection layer. Despite the integration of physics and geometry priors, these methods still follow the same methodology of training the network to learn the mapping from measurements to image. In contrast, our method uses the network to learn continuous implicit neural representation of entire image by mapping spatial coordinates to corresponding intensity values. This is fundamentally different from previous deep learning-based reconstruction approaches.
Implicit Neural Representation Learning
In recent years, the introduction of neural representations has completely changed representation and generation of images in deep learning [17]-[22]. First, Eslami et al. [17] introduced a generative query network to learn the scene representation from input images taken from different viewpoints, and used this representation to predict the image from unobserved viewpoints. Then, Sitzmann et al. [18] proposed a scene representation network to represent scenes as continuous functions that map world coordinates to local feature representation and formulate the image formation as a differentiable ray-marching process. Mildenhall et al. [19] used the fully-connected network to represent scenes and optimized neural radiance fields to synthesize novel views of scenes with complicated geometry and appearance. Later on, more researches this year further improve the algorithms of implicit neural representation and extend to broader applications. [20] showed that Fourier feature mapping enabled MLP to learn high-frequency functions when using MLP to represent complex objects or scenes, while [21] demonstrated that periodic activation functions helped to represent complex signals and their derivatives with fine details. Moreover, Martin-Brualla et al. [22] extended neural radiance fields [19] to address real-world problems and enabled synthesis of novel views of complex scenes from unstructured internet photo collections with variable illumination.
Although much attention has been attracted for neural representation in natural images or photographic images, few studies have been done in the medical domain. In [19] some early discussions of CT and MRI reconstruction were included in one of the validation experiments, but the reconstructed image qualities are far from satisfactory. A work by Sun et al. on arXiv [23] is relevant. However, it only focused on using MLP to represent the measurement field instead of image, and then relying on other existing reconstruction methods to obtain reconstructed images. Our method aims to learn and optimize the neural representation for the entire image, and can directly reconstruct the target image by incorporating the forward model of imaging system.
Image Prior Embedding
Deep neural networks have shown great performance in image generation and reconstruction, because of its capability to learn the image priors from a large scale of example images during training [9]. In addition, Ulyanov et al. [24] showed that the network structure itself can also capture low-level image statistics priors, where a randomly-initialized neural network given noise inputs can be used as a handcrafted image prior for inverse problems. Moreover, Gandelsman et al. [25] used multiple deep image prior networks for image decomposition in various applications. In our work, beyond leveraging such image priors through optimization in the function space of network's parameters, we also take advantage of another image prior unique in medical domain. In medical imaging, it is common that one patient may have multiple imaging scans over time for the purpose of treatment assessment, or for image-guided interventions. Although the images are taken at different subject states, earlier scanned images can still provide useful prior knowledge of the patient anatomy. Our neural representation method proposes a simple yet effective way to embed this prior information and facilitate the reconstruction of target image.
Method
Problem Formulation
First, we mathematically formulate the inverse problem for computational image reconstruction. The forward process of imaging system can be modeled as:
y=Ax+e (1)
where x is the image of the unknown subject while y is the sampled sensor measurements. Matrix A represents the forward model of the imaging system, and e is the acquisition noise.
Image reconstruction aims to recover the computational image x of the unknown subject, given the measurements y from sensors. In the problem of sparsely sampled image reconstruction, the measurements y are undersampled in sensor space due to either acceleration acquisition, as in MRI, or reduction of radiation, as in CT. The inverse problem for sparse sampling is ill-posed, and is typically formulated as an optimization problem with regularization:
where ε(Ax, y) is the data term, which measures the errors between Ax and y and guarantees the data consistency to sensor measurements. Function ε can be different distance metrics such as L1 or L2 norm. ρ(x) is the regularizer term characterizing the generic image prior. The regularizer ρ(x) can be determined in many different ways to capture the various image characteristics, such as total variation of the image to enforce smoothness, or sparsity in a transform domain as in compressed sensing.
Neural Representation for Image
In implicit neural representation learning, the image is represented by a neural network as a continuous function. The network θ with parameters θ can be defined as:
θ
:c→v with c∈[0,1)n,v∈ (3)
where the input c is the normalized coordinate index in the image spatial field, and the output v is the corresponding intensity value in the image. The network function θ maps coordinates to the image intensities, which actually encodes the internal information of entire image into the network parameters. Thus, network structure
θ with the parameters θ is also regarded as the neural representation for the image. Note that, theoretically, a random image in any modality or in any dimension x∈
n can be parameterized by the network using this method. Below we introduce the specific network structure used in our method.
Fourier Feature Embedding
Since Fourier features are shown to be effective for networks to learn high-frequency functions [20], we use a Fourier feature mapping γ to encode the input coordinates c before applying them to the coordinate-based network. Thus, the encoded coordinates are:
γ(c)=[cos(2πBc), sin(2πBc)]T (4)
where matrix B represents the coefficients for Fourier feature transformation. Following [20], entries of matrix B are sampled from Gaussian distribution (0, σ2), where σ is a hyperparameter characterizing the standard deviation of the prior distribution. After the Fourier feature embedding, the input to the network
θ is the encoded coordinates γ(c).
Multi-Layer Perceptron Network
The network θ is implemented by a deep fully-connected network or multi-layer perceptron (MLP). The coordinate-based MLP parameterizes the continuous function to represent the entire image. This function is defined by the network structure as well as the network parameters. In the next section, we will describe in detail how to obtain the network parameters through optimization. For the network structure, the model depth and width of MLP are hyper-parameters, characterizing the representative capability of the MLP model. Moreover, we use the periodic activation functions in our MLP model after each fully-connected layer, which are demonstrated to effectively represent fine details in signals [21].
NeRP for Sparsely Sampled Image Reconstruction
Next, we introduce how the proposed implicit neural representation learning with prior embedding (NeRP) is used to solve image reconstruction problem. The goal is to recover the image x of the target subject, given corresponding sparsely sampled measurements y and a prior image xpr. Note that xpr and x are different scans for the same subject, but at different time points. These capture the changing state of the subject such as the tumor progression for monitoring patient response to therapy.
Prior Embedding
In the first step, we embed the prior image xpr into the network. We use the coordinate-based MLP ϕ introduced in Sec.III.B to map the spatial coordinates to corresponding intensity values in prior image xpr. That is,
ϕ: ci→xipr, where i denotes the coordinate index in image spatial field. Given all the coordinate-intensity pairs in prior image {ci, xipr}i=1N with a total of N pixels in the image, the randomly-initialized MLP is optimized based on the objective:
After optimization, the internal information of prior image xpr is encoded into the MLP network ϕ* with the corresponding network parameters ϕ*. For clarity, we use
pr to denote the prior-embedded MLP network, i.e. xpr=
ϕ*=
pr.
Network Training
Given the prior-embedded MLP pr and measurements y, we train the network to learn the neural representation of the target image. Based on the formulation in Eq. (2), the unknown target image x is parametrized by a coordinate-based MLP
θ with parameters θ. Thus, the data term is defined as
where the optimization in image space is transformed to the optimization in the space of MLP's parameters. Furthermore, the regularizer ρ(x) is replaced by the implicit image priors from network parametrization, including the internal information from prior image embedded in pr as well as the low-level image statistics prior captured by network structure itself
θ [24]. Thus, the optimization subjection in Eq. (2) can be formulated as follows:
The network θ is trained by minimizing the L2-norm loss, which is initialized by the prior-embedded network
pr.
Note that forward model A is adapted to the corresponding imaging system, such as Radon transform for CT imaging and Fourier transform for MRI imaging. The operation A is differentiable, which enables training the network θ in an end-to-end fashion.
Image Inference
Finally, after the network is well trained, the reconstruction image can be generated by inferring the trained network across all the spatial coordinates in the image field. That is: x*: {ci, θ*(ci)}i=1N, where i denotes the coordinate index in image spatial field. This is denoted in short as x*=
θ* in Eqs. (6) and (7). Filling the intensity values at all the coordinates in image grid constitutes the final reconstruction image x*.
Technical Details of NeRP
In our implementation, we construct an 8-layer MLP network with a width of 256 neural nodes for CT reconstruction, where each fully-connected layer is followed by the periodic activation function [21] except for the last layer. For MRI reconstruction, we increase the MLP width to 512 layers. We will discuss and analyze the influence of different network structures in next Section. The Fourier feature embedding [20] size is 256, where the hyper-parameter for the standard deviation of the coefficient's Gaussian distribution is set as 3 for MRI reconstruction and 4 for CT reconstruction. For prior embedding, the training objective in Eq. (5) is optimized by the Adam optimizer with a learning rate of 0.0001. The total training iterations are 1000 for 2D images and 2000 for 3D images. Next, given the prior-embedded MLP as the initialization, the reconstruction network is trained by optimizing the objective in Eq. (7) using the Adam optimizer with a learning rate of 0.00001. Usually we train 1000 iterations for 2D images and 2000 iterations for 3D images. We implemented our networks using PyTorch [26]. For the differentiable forward model A, the Radon transform or forward projection operation for CT imaging is realized by using Operator Discretization Library (ODL) [27]. The non-uniform Fast Fourier Transform (NUFFT) for MRI imaging is implemented based on the torchkbnufft package [28].
To evaluate the proposed NeRP method, we conducted experiments for 2D/3D CT and MRI image reconstruction with sparsely sampling. For CT image reconstruction we assume 20 projections equally distributed across a semi-circle. We compute parallel-beam projections for 2D CT and cone-beam projections for 3D CT. For MRI image reconstruction, 40 radial spokes are sampled in k-space with golden angle as the angular interval, as shown in
Datasets
Pancreas 4D CT Data
For CT image reconstruction, we collected a pancreas 4D CT data from a clinical patient. The 4D CT data is a temporal scan with a sequential series of 3D CT images over a respiratory cycle. Due to respiratory motion there is continuous motion in the CT images at different time points. In the first row of
Head and Neck CT and Lung CT Data
To further validate the generalization of the proposed NeRP algorithm, we collected two clinical patient cases including a head and neck CT case and a lung CT case. For each case there are two longitudinal 3D CT images scanned for the same patient at time points during treatment with radiation therapy. The goal is to follow tumor volume to assess response to therapy. In the data preprocessing, we firstly conduct rigid image registration to align the two CT images at the same position. Then, we use NeRP to reconstruct the latter 3D CT image while using the earlier 3D CT image as the prior image.
Brain Tumor Progression MRI Data
For MRI image reconstruction we conducted experiments on a public dataset for brain tumor progression [29] [30]. This dataset includes MRI images from 20 subjects with primary newly diagnosed glioblastoma. The patients were treated with surgery and standard concomitant chemo-radiation therapy (CRT) followed by adjuvant chemotherapy [30]. For each patient, there are two MRI exams included, which were within 90 days following CRT completion and at tumor progression.
Thus, the tumor changes can be clearly observed by comparing the two MRI exams of the same patient at different time points. In addition, each MRI exam contains multi-modality MRI images including T1-weighted, and contrast-enhanced T1-weighted (T1c), T2-weighted, FLAIR. In our experimental setting, we set the first MRI exam as the prior image and aim to reconstruct the MRI image in the second exam. This is tested for different MR image modalities respectively.
Experiments on 2D CT Image Reconstruction
In
Experiments on 3D CT Image Reconstruction
To evaluate the effectiveness of NeRP in a higher-dimensional reconstruction task, we conducted experiments for 3D CT image reconstruction. In the first experiment on the pancreas 4D CT data, we use the entire phase-1 3D CT as the prior image and aim at reconstructing the 3D CT image at phase 6 with image size of 128×128×40 after image cropping and resizing. Note that phase 1 and phase 6 are exactly the inhale and exhale phases during this 4D CT scan, which have the largest structural difference.
For comparison, we also conducted experiments and show the results of other reconstruction methods. First, we show the reconstruction results for “NeRP w/o prior” as an ablative study by removing the prior embedding. Comparing the image quality, we can see that the prior embedding effectively contributes to reconstructing high-quality image with sparse sampling. Moreover, we compare with the analytic reconstruction algorithm with filtered back projection (FBP). The back-projected operation adjoint to the cone-beam projection in the forward model can reconstruct the 3D image from the given 2D projections with filter correction. As shown in
Going beyond 4D CT data, we also evaluated the clinical radiation therapy patient data with both head and neck CT and lung CT. The quantitative results for 3D CT reconstruction evaluated by PSNR and SSIM metrics are reported in Table I on different anatomic sites including pancreas CT, head and neck CT and lung CT with all comparison methods. To evaluate the effectiveness under different settings, we also compare the reconstruction results with different number of projections. Our proposed NeRP method achieves the best performance in either metric for all the 3D CT image cases with 10/20/30 projections respectively, outperforming all the other methods without using prior image embedding. Reconstructed images for the head-and-neck CT case are shown in
Experiments on 2D MRI Image Reconstruction
We conducted experiments to evaluate the proposed method for MRI image reconstruction. We aimed to reconstruct 2D MRI images with sparsely sampled frequency space (k-space) data by using a radial sampling pattern for data acquisition, which is widely used in clinical MRI. The 2D NUFFT for radial sampling is used as the forward model to compute and sample k-space data as shown in
Experiments on 3D MRI Image Reconstruction
Using the same dataset with brain tumor regression, we further evaluated the 3D MRI image reconstruction. In this case, the entire 3D MRI volume in the first exam is used as the prior image in order to reconstruct the 3D MRI image in the second exam for the same patient. The forward model is the 3D NUFFT to compute and sample 3D k-space data. The whole learning framework of NeRP is similar to that of 2D MRI reconstruction except for using a 3D coordinate index. In pre-processing, all the 3D MRI images are cropped and resized to 128×128×24.
Quantitative results of 3D MRI image reconstruction evaluated by PSNR and SSIM metrics are reported in Table II for different image modalities including T1, T1c, T2 and FLAIR. We compare the results of different reconstruction methods with 30/40/50 sampled radial spokes, respectively. From the Table II, our NeRP method achieves better performance than other methods without using prior image for all the image modalities. Reconstructed 3D MRI images for FLAIR modality are demonstrated in
Analysis of Network Structure
For the proposed NeRP algorithm, one important issue is to set a proper network structure for the MLP backbone. The MLP network parameters serve as the variables to expand the function space for network optimization and seeking the optimal reconstructed images. The number of network parameters is related to the depth and width of the MLP, i.e., the number of layers and the number of neurons in each layer. To analyze the influence of network structures, we conduct ablation study to obtain the reconstruction results with changed MLP depth and width as shown in Table III. Here, the 3D pancreas CT image is reconstructed from 20 projections while the 3D T1 MRI image is reconstructed from 40 radial spokes. From Table III, we see that the reconstruction results are not very sensitive to the change of network depth or width, which indicates the proposed NeRP algorithm is robust to the specific choice of network structure. In experiments, we also observe that training the MLP model could be more difficult with more layers, where the insufficient optimization may cause worse reconstruction results.
Analysis of Sparse Sampling Ratio
To better analyze the influence of sparse sampling ratio, we use NeRP to reconstruct 3D CT and MRI images with different number of sampled projections or radial spokes.
Although many existing methods have shown the advantages of deep learning models for medical image reconstruction [31], there are increasing concerns regarding the limitations of the current deep learning approaches. First, training deep neural networks is data-intensive and requires large-scale datasets. This may prevent many practical applications due to the difficulty of data collection. The limited availability of specific image modalities or images of rare diseases may make it difficult to acquire sufficient training data for deep learning modeling. In addition, there is a common concern about the robustness and reliability of deep reconstructed images. For example, Antun et al. [16] find that the small but significant structural changes in tumors or lesions may not be accurately captured in the deep reconstructed images. Finally, the generalization capability of deep networks is unclear. The trained deep networks may produce poor performance when generalized to data out of the training distribution, such as across different imaging modalities or across different anatomical sites. The model generalizability is related to not only the training data distribution but also the network structure. For example, it is a non-trivial task to directly transfer a deep network developed for MRI reconstruction to the task of CT reconstruction due to different sensor measurements fields. Because of these issues and limitations, new insights are needed in developing deep learning-based image reconstruction algorithms.
The proposed NeRP method provides a new perspective to the problems of image reconstruction, which promises to overcome the shortcomings of previous deep reconstruction approaches. First, NeRP requires no training data from external subjects, but the sparsely sampled sensor measurements data for the target image and a prior image from a previous scan of the same subject. In addition, from our experiments, the reconstructed images from NeRP are more robust and reliable, and can capture the small structural changes such as tumor or lesion progression. The implicit image priors captured by network structure and the prior embedding can effectively incorporate the prior knowledge in the reconstruction process, which makes it possible for the network to capture and reconstruct the fine structural details in the resultant images. As a result of the previous two points, NeRP can be more easily generalized to different image modalities, different anatomical sites, and different dimensionalities in the image reconstruction task. The relaxed requirements for training data make the method more transferrable and applicable across various applications. The proposed NeRP is a general methodology for medical image reconstruction with promising advantages over mapping-learning-based deep reconstruction methods.
In this work, we propose a new deep learning-based medical image reconstruction methodology by learning implicit neural representations with prior embedding (NeRP), which efficiently incorporates the prior knowledge and learns to reconstruct the target image through implicit neural representations. Through the experiments for 2D/3D MRI and CT image reconstruction, we show that the proposed NeRP algorithm is able to provide high-quality reconstruction images even with sparsely sampled measurements data. The NeRP approach possesses a number of unique advantages: (1) requires no training data from external subjects for developing networks; (2) accurate reconstruction of small and detailed changes such as anatomic structure or tumor progression; (3) broad applicability to different body sites, different imaging modalities and different patients. For medical images, it is common that a patient is scanned multiple times for clinical diagnosis or treatment follow-up, for the purpose of treatment planning or monitoring of the changes in tumor volume before and after therapy. In a longitudinal image series, previous scans can provide useful prior knowledge for NeRP image reconstruction. The effectiveness of NeRP and advantages of prior embedding have been demonstrated in the extensive experiments.
This application claims priority from U.S. Provisional Patent Application 63/210,433 filed Jun. 14, 2021, which is incorporated herein by reference.
This invention was made with Government support under contract CA256890 awarded by the National Institutes of Health, and under contract CA227713 awarded by the National Institutes of Health. The Government has certain rights in the invention.
Number | Date | Country | |
---|---|---|---|
63210433 | Jun 2021 | US |