The present invention generally relates to techniques for image recovery. More specifically, the present invention relates to an image restoration processor applying a framework to generation of sparsity regularizers for image recovery.
Principal component analysis (PCA) refers to seeking an underlying low-dimensional linear subspace from high-dimensional data. It is widely used for dimensionality reduction, and abounds in many applications such as computer vision, bioinformatics, as well as signal and image processing. Although PCA works well in the presence of zero-mean Gaussian noise, it is vulnerable to outliers. To overcome this drawback, robust PCA (RPCA) is proposed to decompose the outlier-contaminated data into a sum of low-rank and sparse matrices. RPCA has been applied in hyperspectral image restoration, shadow removal and video separation, to mention a few. The low-rank plus sparse decomposition model can be considered as a weighted linear combination of rank and l0-norm minimization problem, which is NP-hard. To address the issue, a common method is to replace the rank function and the l0-norm with the nuclear norm and the l1-norm, respectively, leading to a convex optimization formulation, which is known as principal component pursuit (PCP). Besides, it is proved that minimizing the weighted linear combination of the nuclear norm and l1-norm minimization attains exact low-rank and sparse component recovery with high probability, if the incoherence conditions are satisfied. To solve PCP, many efficient algorithms based on singular value thresholding (SVT), accelerated proximal gradient and augmented Lagrange multipliers, are developed.
Convex relaxations are relatively easy in optimization, but their estimates are known to be biased. For example, when the nuclear norm is employed to find the low-rank component, it underestimates all nonzero singular values and shrinks them with the same constant. It has been shown that shrinking less the larger singular values will attain better restoration performance in image denoising and inpainting, as well as background subtraction. There are two main strategies to alleviate the bias induced by the nuclear norm. The first is to weigh the singular values differently via updating the weights per iteration, which is similar to the reweighted l1 minimization. One of related works extends this technique to low-rank matrix approximation, and adopts a weighted nuclear norm minimization (WNNM) as a surrogate to rank minimization and l1-norm to resist outliers. Besides, the weighted Schatten-p-norm minimization as a generalization of WNNM is suggested. Moreover, to facilitate weight adaption, a weighted minimax-concave penalty (WMCP) is developed, which uses a clever trick to unfold the minimax-concave penalty (MCP) and utilizes the alternating direction method of multipliers (ADMM) to find the solution.
The second strategy employs nonconvex sparsity-inducing regularizers to reduce the estimation bias. Many studies have shown the superiority of nonconvex regularizers over the convex relaxation approaches. One of the studies develops a fast proximal algorithm with nonconvex regularizers such as the smoothly clipped absolute deviation (SCAD) and MCP, for low-rank matrix learning. One of the studies suggests robust PCA via adopting the p-norm (0<p≤1) as the regularizers to impose constraints on the low-rank and sparse terms, but the lp-norm does not have a closed-form expression for its proximal operator, except for three special cases of
A modified lp-norm is then devised to impose low-rank and sparse constraints to achieve the low-rank and sparse decomposition, resulting in good performance in background separation. In addition, other nonconvex regularizers, including the exponential-type penalty (ETP) and the Laplace function, are applied for low-rank matrix recovery via iteratively reweighted nuclear norm. Nevertheless, neither of them has the explicit expressions for their corresponding proximal operators.
Although numerous nonconvex regularizers have been suggested, they still have the aforementioned limitations. The field of low-rank matrix recovery faces challenges in handling large-scale datasets, complex noise models, and incomplete or corrupted data. Traditional methods may not be sufficient in addressing these challenges, necessitating the development of novel approaches that can effectively handle these issues and provide accurate recovery results.
It is an objective of the present invention to provide an apparatus and a method to address the aforementioned issues in the prior arts.
In accordance with one aspect of the present invention, an image restoration processor is provided. The image restoration processor includes an image receiver, a matrix converter, a framework memory, a modifier, a matrix decomposer, and an image reconstructor. The image receiver is configured to receive one or more degraded images. The matrix converter is configured to reshape image data of the one or more degraded images into one or more target matrices. The framework memory is configured to store a framework for generating one or more sparsity-inducing regularizers, which enables generation of the one or more sparsity-inducing regularizers and derivation of their theoretical properties and closed-form proximity operators. The modifier is configured to read the framework memory and apply the framework to modify an M-estimator, accordingly output a hybrid M-estimator and the corresponding sparse regularizer. The matrix decomposer is configured to receive the one or more target matrices and the sparse regularizer and to apply the sparse regularizer to a robust principal component analysis (RPCA) approach via decomposing one of the one or more target matrices into a sum of low-rank and sparse matrices. The image reconstructor integrates outcomes of decomposition from the matrix decomposer to form one or more recovered images from the one or more degraded images, utilizing the low-rank and sparse matrices.
In accordance with one aspect of the present invention, an image restoration processor is provided. The image restoration processor includes an image receiver, a matrix converter, a framework memory, a modifier, a matrix decomposer, an image reconstructor, and an image selector. The image receiver is configured to receive one degraded image. The matrix converter is configured to reshape image data of the degraded image into a target matrix. The framework memory is configured to store a framework for generating one or more sparsity-inducing regularizers, which enables generation of the one or more sparsity-inducing regularizers and derivation of their theoretical properties and closed-form proximity operators. The modifier is configured to read the framework memory and apply the framework to modify more than one M-estimator and accordingly output more than one hybrid M-estimator and more than one sparse regularizer corresponding to the more than one hybrid M-estimator. The matrix decomposer is configured to receive the target matrix and the different sparse regularizers and to apply the different sparse regularizers to a robust principal component analysis (RPCA) approach via decomposing the matrix into sums of low-rank and sparse matrices. The image reconstructor integrates outcomes of decomposition from the matrix decomposer to form more than one recovered image from the degraded image using different sparse regularizers, utilizing the low-rank and sparse matrices. The image selector is configured to receive the recovered images from the reconstructor and determine which one of the recovered images is to be outputted according to image evaluation metrics for all of the recovered images.
In one embodiment, a framework is devised to generate different nonconvex sparsity-inducing regularizers under some relatively mild conditions. Although the resultant regularizers may be nonconvex, it can be proven that their Moreau envelopes are convex, and their analytical solutions are given. In addition, the framework is applied to three popular M-estimators, namely, Welsch, Cauchy and German-McClure, to generate the corresponding sparse regularizers, because they have achieved considerable success in image restoration, compressed sensing and subspace clustering. Although these M-estimators provide implicit regularizers (IRs) via half-quadratic optimization, sparsity cannot be achieved (See Appendix A). While the sparsity promoting regularizers associated with the three M-estimators are generated in this work. Moreover, an effective strategy is derived for the hyperparameter selection and their physical meanings are explained.
The sparse regularizers generated in the framework of the present invention can be considered as nonconvex surrogates to enforce low-rank-plus-sparse decomposition. Different from PCP and WNNM based RPCA (WNNM-RPCA), which both employ the 1-norm to combat sparse outliers, the devised nonconvex regularizers are adopted in the present invention. The resultant optimization problem is tackled via ADMM with convergence guarantees. Briefly, there are contributions provided via the above configuration:
Embodiments of the invention are described in more details hereinafter with reference to the drawings, in which:
In the following description, image restoration processors using framework for generating sparsity-inducing regularizers and the likes are set forth as preferred examples. It will be apparent to those skilled in the art that modifications, including additions and/or substitutions may be made without departing from the scope and spirit of the invention. Specific details may be omitted so as not to obscure the invention; however, the disclosure is written to enable one skilled in the art to practice the teachings herein without undue experimentation.
For clearer illustrations of the embodiments of the present invention, notations and basic operations are stated before the descriptions of the embodiments.
where X, M and E are the corrupted, low-rank and sparse matrices, respectively, and the nuclear norm ∥M∥* is the sum of singular values of M. The 1-norm of E, i.e., ∥E∥1 is the regularizer to handle sparse outliers, and A is the regularization parameter, which controls the relative weight between the nuclear norm and regularizer. However, the nuclear norm underestimates the nonzero singular values, especially large singular values that contain major information of a matrix. To alleviate this effect, the weighted nuclear norm is suggested, which is extended from the weighted
1-norm for vectors, resulting in:
where φ1(·) and φ2(·) are nonconvex penalty functions.
whose solution is solved by the operator:
which is called the proximity operation of φ(·). In fact, the proximity operator has been extensively used in signal denoising, image restoration and sparse representation. When φ(·) is convex, (4) is strictly convex and has the optimal solution. While when φ(·) is nonconvex, (4) is generally a nonconvex problem. Recent studies show that the solution to (4) can be attained even when φ(·) is nonconvex, but it requires iterations, that is, the explicit proximity operator cannot be found. In addition, one related work generates a series of nonconvex IRs to achieve robustness using a half-quadratic optimization, but their resultant IRs are not sparsity-inducing regularizers, which limits their application. The common sparsity-promoting regularizer is the 1-norm, and its proximity operator can be found in (10) which will be discussed below. However, as afore-mentioned, it suffers a bias problem.
It is one of the purposes of the present invention to address the problem of restoring a matrix with low-rank constraint in the presence of sparse outliers. Nonconvex regularizers may not have the closed-form proximity operators, thus iterations are needed to find their expressions, which increase a computational load. To address this issue, a framework in accordance with various embodiments is devised to generate sparsity-inducing regularizers with a closed-form proximity operator. Although the resultant regularizers may be nonconvex, their Moreau envelopes are convex.
The framework is then applied to three popular M-estimators, namely, Welsch, Cauchy and German-McClure functions, and their associated sparsity-inducing regularizers are first generated, which enrich the variety of regularizers. Next, a parameter selection strategy for the resultant regularizers is also proposed. Moreover, the regularizers are exploited for low-rank and sparse decomposition applications, leading to three algorithms based on the alternating direction method of multipliers with convergence guarantees. The property that any limit point generated is a critical point is proven as well. Finally, extensive numerical experiments based on synthetic and real-world data are conducted to demonstrate the validity of the developed approaches.
Firstly, the framework for sparsity inducing regularizer generation is discussed.
The curves of lh and lt are shown in
where φ(y)=φh(y)=|y| for the Huber function, while for the truncated-quadratic function:
The solutions to (8) for φh(y) and φt(y), known as proximity operator, are:
respectively, implying that φh(y) and φt(y) are sparsity-inducing regularizers. It is worth noting that the proximity operator for the 0-norm is the same as (11).
In the present invention, a framework is provided to generate different sparsity-inducing regularizers via generalizing (6) and (7) for |x|>λ as illustrated in
where g(x) is a continuous function and g′(x)≥0 for x>0, while a and b are constants to make l(x) continuously differentiable at x=λ. Thus, a=λ/g′(λ)>0(g′(λ)≠0), and b=λ2/2−ag(λ). It is easy to see that (6) and (7) are a special case of (12).
If h′(x)>0 and h″(x)>0 for x>λ, namely, h(x) is strictly convex when x>λ, then (12) can be used to generate sparsity-inducing regularizers.
and if ƒ(x) is convex, the conjugate of ƒ*(x) is ƒ(x), namely,
Thus,
Combining (15) and (18), it is easy to obtain:
Similar to the regularizers in the related works, the expression of φ(y) is generally unknown, which is referred to as implicit regularizer (IR). Before solving (19), the following lemma is introduced.
Since the solution to (19) is the same as that to (18), according to Lemma 1, the solution to y ·in (19) is:
which means that the IR generated by lg,λ(x) can make the solution sparse, and Pφg,λ(x) is nondecreasing since h(x) is strictly convex for |x|>λ.
Next, the reason of sparsity-promoting regularizer generation is illustrated via (12), compared with other convex functions in
Second, generalization via g(x) is discussed. It is worth mentioning that g(x) can represent many convex or nonconvex functions. In various embodiments, a nonconvex g(x), and exploit g(x) for three M-estimators, namely, Welsch, Cauchy, and German-McClure are considered, resulting in three different sparsity-inducing regularizers. Compared with the case g(x) as convex or concave functions for x>0, it is difficult to analyze those M-estimators because they are nonconvex and nonconcave.
Regarding modification to the Welsch Function, the Welsch function is a M-estimator, whose expression is:
It is because minimizing (12) has been shown to be equivalent to maximizing the correntropy criterion when the Gaussian kernel is adopted as a correntropy function with σ being the kernel parameter. Thus, the provided theory of the present invention is applied to the Welsch function. When g(x)=ρwelsch(x), according to (12), the hybrid Welsch function is:
which can be rewritten via (19) as:
The solution is:
Regarding modification to the Cauchy Function, unlike the Welsch M-estimator, which is bounded from above, the Cauchy function is unbounded from above, whose expression is:
where γ is the scale parameter of the Cauchy distribution. Thus, when g(x)=ρcauchy(x), the hybrid Cauchy function can be obtained as:
where
In addition, there is:
and based on (21), its solution is:
Regarding modification to the German-McClure Function, the expression of German-McClure (GMC) M-estimator is:
where τ>0 is a scale parameter. Thus, when g(x)=ρgmc(x), there is:
which amounts to:
where φτ,λ(y) is the sparse regularizer related to lτ,λ(x). Employing (21), the solution to (32) is:
In various embodiments, a parameter selection strategy can be applied to the above modified functions.
There are four parameters for lσ,λ(x), φσ,λ(y), lγ,λ(x), φγ,λ(y), lσ,λ(x) and φσ,λ(y), i.e., λ, σ, γ and τ. Apparently, knowing the physical meaning of a parameter can help choose parameters for their value. First, the Huber function lh(x) can be rewritten via (8) as:
where λ is a positive weighting parameter that controls sparsity, namely, y=0 for |x|≤λ. Similarly, λ for φσ,λ(γ), φγ,λ(y) and φτ,λ(y) in (24), (28) and (32), respectively, also controls sparsity. That is, y=0 for |x|≤1. However, different from the Huber function that has the explicit λ|y|, the expressions of φσ,λ(y), φγ,λ(y) and φσ,λ(y) are generally unknown. Nevertheless, they all have a smaller bias than λ|y|, which is illustrated in
Besides, the proximal operator for λ|y| in (10) treats all values of x equally and shrinks x with the same bias λ as indicated in (10). In fact, for many real-world applications, the bias for different x should not be the same. For example, when addressing low-rank matrix recovery, larger singular values of an observed matrix correspond to the dominant information, thus it is better to shrink them less. Mathematically, the bias can be described via the difference between y=x and the proximal operator for different regularizers when x>1:
and Δd is required to be non-increasing as x increases, implying that g″(x)≤0, namely, g(x) is concave when x>λ. For the hybrid Welsch function, when x>λ,
thus the obtained is σ≤√{square root over (2)}x. Since x>λ, there is:
Similarly, for lγ,λ(x), when x>λ.
and it is obtained that γ≤x. Thus for lγ,λ(x), there is:
Finally, for the hybrid GMC function,
is obtained to yield:
The above is summarized in the following proposition.
Next, applications to low-rank matrix recovery are discussed, including mathematical preliminaries, algorithms for low-rank-plus-sparse matrix decomposition, and complexity.
Regarding mathematical preliminaries, the theory derived in the above is applied to the scalar case, and when it is extended to vectors, matrices, and the singular values of matrices, it is necessary to give the corresponding definitions. Besides, the proofs of the following propositions can be found in Appendices D, E and F, respectively.
where φ·,λ(·) can be φσ,λ(·), φγ,λ(·) and φτ,λ(·), and σ, γ and τ are positive constants.
where φ·,λ(·) can be φσ,λ(·), φγ,λ(·) and φτ,λ(·), while σ, γ and τ are positive constants.
whose solution is:
where Pφ·,λ(·) is an element-wise operator, and Pφ·,λ(·) can be Pφσ,λ(·), Pφγ,λ(·) as well as Pφτ,λ(·). Besides, when σ≤√{square root over (2)}λ, γ≤λ and τ≤√{square root over (3)}λ/2, the obtained is:
which is the 1-norm of s.
However, employing the nuclear norm to achieve low-rank recovery will lead to a biased solution because the nuclear norm is based on the 1-norm of singular values. To address this issue, the non-convex regularizers φ(·) of the present invention are used to replace the
1-norm. Compared with other nonconvex regularizers, ours have closed-form proximity operators.
If the proximity operator Pφ·,λ is monotone, then the solution for (50) is:
where s* satisfies s1*≥ . . . ≥si*≥ . . . ≥sr* with i=1, 2, . . . , r, which is determined as:
Regarding algorithms for low-rank-plus-sparse matrix decomposition, the regularizers of the present invention are applied to RPCA via decomposing the target matrix into a sum of low-rank and sparse matrices, which is formulated as:
where φ·, 1/ρ(·) and φ·, λ/ρ(·) are the regularizers included in TABLE II.
The problem (52) can be solved by ADMM, and its augmented Lagrangian is:
which amounts to:
where Λ contains the Lagrange multipliers, the last term is the augmented term and ρ>0 is the augmented Lagrangian parameter. The ADMM updates the primal and dual variables at (k+1)-th iteration via:
The exact expressions at the (k+1)th iteration, i.e., {Mk+1, Ek+1, Λk+1}, are derived as follows.
Update of E: E is updated via:
where the constant term is ignored since it does not affect the solution to Ek+1. Invoking (21), there is:
Update of M: Given Ek+1 and Λk, the low-rank matrix M is updated via:
According to Lemma 2, the solution is:
Theorem 2. The {Mk, Ek} generated by the proposed algorithm satisfy the following properties:
Let {Mk
Regarding complexity, similar to PCP and WNNM-RPCA, when finding the low-rank matrix, the proposed algorithms involve the computation of an SVD per iteration, whose complexity is (min(m,n)mn), where m and n are the row and column lengths of the degraded matrix, respectively. In addition, the complexity of calculating the sparse matrix is
(mn). Therefore, solving the low-rank component dominates the computational time at each iteration.
In order to verify an image restoration processor of the present invention, extensive experiments on both synthetic and real-life data are conducted. In experimental results, SIR-HW, SIR-HC and SIR-HG, together with SIR-HU and SIR-HT, are assessed. Parameter settings for σ=√{square root over (2)}λ, γ=1 and τ=√{square root over (3)}λ/2 are suggested. In addition, three benchmark techniques, namely, PCP, WNNM-RPCA, and DPRPCA are employed. The experimental results include two parts, (1) synthetic data and (2) real-world image restoration.
Regarding synthetic data, the data matrix X∈ is generated by a sum of low-rank matrix M=UVT, where U∈
and V∈
with r being the rank, and sparse matrix E. n=m is set for convenience. The entries of U and V satisfy the standard Gaussian distribution, and the locations of the sparse outliers E, which has S non-zero entries uniformly distributed in [−500,500], are drawn independently from a Bernoulli distribution. Similar to some related works, r=pr×m and S=ps×m2. In the experiments, m=400 is set, and pr and ps vary from 0.01 and 0.05 with a step size of 0.02. Besides, two evaluation metrics are employed, namely, relative reconstruction error of the low-rank matrix REE=∥M−{circumflex over (M)}∥F2/∥M∥F2, where {circumflex over (M)} is the estimated low-rank matrix, and the estimated rank {circumflex over (r)}. All the competing methods, including PCP and WNNM-RPCA, are set with the same parameters and stopping conditions. That is, λ=1/√{square root over (max(m,n))} and μ=1.05, while ∥X−Mk−Ek∥F/∥X∥F<10−7 and the maximum iteration number of 1000 as termination conditions for all algorithms. Moreover, the performance of all approaches is evaluated using the average results of 100 independent runs.
1-norm and PCP also employs the
1-norm to achieve low-rank and sparse decomposition. Besides, compared with PCP. WNNM-RPCA and DPRPCA (p=0.6), SIR-HW, SIR-HC and SIR-HG can recover more cases, and among all the techniques, SIR-HG has the biggest success area (If REE<10−4, it is denoted as a success recovery). TABLE III tabulates the average estimated rank by different algorithms. It is observed that SIR-HW and SIR-HG give more accurate rank estimates than the remaining methods for different rank parameters.
= 0.
)
indicates data missing or illegible when filed
In addition,
The sections (e)-(g) of
Regarding hyperspectral image restoration,
Hyperspectral imaging has numerous applications such as environmental monitoring, mineral exploration and urban planning. However, they may be contaminated by noise during the acquisition process. To restore hyperspectral images (HSIs), low-rank modeling has been found to be very useful because of the strong correlation along the spectral direction. The features as above are applied to HSI restoration. Three quantitative metrics, i.e., peak signal-to-noise ratio (PSNR), structure similarity index measure (SSIM) and root mean square error (RMSE), are employed. The PSNR, SSIM and RMSE of HSI data are calculated by:
where M∈ and {circumflex over (M)}∈
denote the ground truth and estimated matrix of the HSI, and Mj and {circumflex over (M)}j are the j-th band images of M and {circumflex over (M)}, respectively. Besides, the higher the PSNR and SSIM, and the smaller the RMSE, the better the restoration quality.
The sub-image of the Washington DC Mall contains 191 spectral bands with 256*256 pixels per band. The gray values of each band are normalized onto [0, 1], and then each band is vectorized as a vector, which is stacked to construct the pure matrix M∈. Impulsive noise is generated by the built-in command of ‘imnoise(I, ‘salt & pepper’, ρ)’ in MATLAB, where I is the original matrix, and ρ is the normalized noise intensity. The relationship between p and the signal-to-noise ratio (SNR) is ρ=1/SNR. In addition, to fairly compare different algorithms, the best weighting parameter λ for each method is selected based on the lowest RRE. Impulsive noise with different ρ is added to M, and the average denoising results by 10 independent runs are tabulated in Table IV. It is observed that SIR-HC, SIR-HW and SIR-HG, achieve good restoration results, and among them, SIR-HG is in total the best method since it attains better recovery performance for all cases. Although DPRPCA (p=0.6) has smaller SSIM for one case, its choice of ρ is crucial because it employs
p-norm to achieve sparseness. In order to provide a visual comparison, three bands of HSIs are chosen to form a pseudo-color image shown in
Furthermore, all methods are assessed using the real HYDICE urban dataset. Similar to related works, the bands 104-108, 139-151 and 207-210 are removed since they are seriously contaminated by noise, thus the sub-data with dimensions 256×256×188 are used.
= 0.1
= 0.2
= 0.3
= 0.4
indicates data missing or illegible when filed
Regarding multispectral image restoration,
Multispectral images (MSIs) are different from HSIs because of their higher spatial resolution. In this section, two datasets, i.e., CAVE and Harvard, are used. The CAVE dataset consists of 32 different objects, and each object has 31 spectral bands with dimensions 512×512. The Harvard database includes 50 MSIs of real-world indoor and outdoor scenes. For each set of MSIs, it has 31 spectral bands with spatial resolution 1392×1040.
It can be seen that SIR-HC, SIR-HW and SIR-HG have higher PSNR values for most of the bands in both datasets. Note that all methods perform bad in the first few bands, because there exists a blur in those bands. To provide visual comparison,
620
indicates data missing or illegible when filed
With the above derivation and verification, in various embodiments, an image restoration processor is provided to perform modification to function and recovery to degraded images.
The image receiver 102 is configured to receive at least one degraded image to be recovered. The matrix converter 104 is electrically coupled with the image receiver 102 and is configured to reshape image data of the degraded image into at least one target matrix. The degraded image or the target matrix can serve as input data to be processed by the image restoration processor 100. The matrix converter 104 is electrically coupled with the matrix decomposer 108 and is further configured to send the target matrix to the matrix decomposer 108.
The modifier 106 is electrically coupled with the framework memory 110. The framework memory 110 can be configured to store a framework as discussed above (e.g., the Proposition 1 as discussed above) for generating one or more sparsity-inducing regularizers, which enables generation of the one or more sparsity-inducing regularizers and derivation of their theoretical properties and closed-form proximity operators. The modifier 106 can receive a M-estimator from an external source. For example, users can optionally input a M-estimator to the modifier 106 for further modification. In one embodiment, the image restoration processor 100 may further include a user interface configured to receive an M-estimator from an external source and transmit an information signal to inform the modifier 106 of the received M-estimator. The modifier 106 can be configured to read the framework memory 110 and apply the framework to modify the input M-estimator. The input M-estimator can be modified by the modifier 106, as discussed above, and then the modifier 106 can accordingly output a hybrid M-estimator and generate the corresponding sparse regularizer. In an embodiment, the sparse regularizer is generated by the hybrid M-estimator in the modifier 106 and then output by the modifier 106. In some embodiment, the input M-estimator is selected from Welsch Function, Cauchy Function, and German-McClure Function. In various embodiments, the generated sparse regularizer is nonconvex and has closed-form proximity operators.
The parameter selector 120 is electrically coupled with the modifier 106 and is configured to perform the parameter selection strategy as discussed above. Specifically, the parameter selector 120 can determine a relationship between a scale parameter of the sparse regularizer and a positive weighting parameter controlling sparsity.
The matrix decomposer 108 is configured to receive the target matrix and the sparse regularizer. In various embodiments, the matrix decomposer 108 can receive the target matrix and the sparse regularizer via accessing the first cache 130 and the second cache 132. For example, the first cache 130 is configured to receive the target matrix from the matrix converter 104 and store the target matrix, the second cache 132 is configured to receive the sparse regularizer from the modifier 106 and store the sparse regularizer, when the matrix decomposer 108 is triggered to perform decomposition, the matrix decomposer 108 can obtain the target matrix and the sparse regularizer via accessing the first cache 130 and the second cache 132, respectively. After the obtaining, the matrix decomposer 108 can be configured to apply the sparse regularizer to a RPCA approach, as discussed above, via decomposing the target matrix into a sum of low-rank and sparse matrices. The computation module 122 can work together to decompose the target matrix and solve the problem of the equation (52) as afore-mentioned with the matrix decomposer 108. The computation module 122 can be configured to perform an ADMM approach following with the RPCA approach performed by the matrix decomposer 108.
The image reconstructor 140 is electrically coupled with the matrix decomposer 108 and is configured to integrate outcomes of decomposition from the matrix decomposer 108 to form at least one recovered image from the original degraded image, utilizing the low-rank and sparse matrices. In an embodiment, the image restoration processor 100 can further include a display in electrical communication with the image reconstructor 140 and configured to display the recovered image and may further display a difference of the image before and after restoration.
The image restoration processor 100 can be applied as a visual rehabilitation pipeline or a digital image remastering framework, which combines specific steps or processes to restore images to their original visual state. In this regard, the image restoration processor 100 can be used for a scenario that dramatically reduces image capture resources by deliberately undersampling an image to be captured. This permits transfer of the undersampled image in less time and consuming fewer hardware resources; for example, an image transferred wirelessly will require less spectral bandwidth; an image transferred electronically will use less RAM/cache as the image is transferred over a network. As such, the image restoration processor 100 of the present invention improves network operation, particularly for networks that transmit considerable image, video, and/or radar data.
For example, hyperspectral imaging is a popular style of image acquisition which captures a multitude of two-dimensional images at different frequencies to yield more information from a scene. If the images are spectrally undersampled, image recovery according to the present invention may be performed to reconstruct the complete images. Similarly, other types of images, such as radar and MIMO images, which are related to antenna technology used in wireless communications, may also suffer from corruption during transmission. In such cases, the image restoration processor 100 can be effectively applied to restore these images, mitigating any potential misunderstandings or misinterpretations caused by the corruption. This can help ensure accurate and reliable image processing for radar and MIMO applications, enhancing the overall performance and reliability of these systems. Accordingly, in various embodiments, the image data is from one or more degraded hyperspectral images or from one or more degraded multispectral images.
In the image restoration processor 200 illustrated in
The image selector 260 is configured to receive the recovered images from the image reconstructor 240 and determine which one of the recovered images is to be outputted according to the image evaluation metrics for all of the recovered images. In one embodiment, the image evaluation metrics include SNR (Signal-to-Noise Ratio), SSIM (Structural Similarity Index), and RMSE (Root Mean Square Error). The recovered image with the best image evaluation metrics can be transmit to the display 250 from the image selector 260 for displaying. In various embodiments, the display 250 electrically communicating with the image selector 260 is configured to show which function is introduced as the sparse regularizer for the output recovered image, so as to inform the user of which sparse regularizer is applied to output eventually. In various embodiments, the image restoration processor 200 can further include an image comparator 262 configured to compare the recovered images and quantify recovery performance thereof, as the tables above (e.g., TABLE I to TABLE V), thereby informing users of the quantified comparation of the recovered images.
As described above, in accordance with the various embodiments of the present invention, a framework is provided to generate sparsity promoting regularizers along with a theoretical analysis. The framework of the present invention can be applied to three M-estimators, namely, Welsch, Cauchy and GMC, and produce the IRs of those functions. It is the first time that sparsity regularizers associated with these three functions are generated. Moreover, the obtained regularizers are applied to low-rank recovery, and we propose three corresponding algorithms with convergence guarantees. Numerical examples based on both simulations and real-world data demonstrate that the image recover method of the present invention can consistently achieve outstanding restoration performance. As such, the image restoration processor of the present invention can be provided to perform the image recovery.
The Appendices as afore mentioned are provided as follows.
where φλ(y) is the associated implicit regularizer (IR). The IRs of Welsch, Cauchy and German-McClure M-estimators are found in Table VI, but they cannot achieve sparseness. When φλ(y)=λ|y|, the solution to (61), denoted as Pφ(x), is zero for |x|≤λ. That is to say, the 1-norm can attain sparseness and set y as zero when the magnitude of x is less than a threshold. However, when the penalty φλ(y) is considered as the IRs of these M-estimators, sparseness cannot be attained since the solution is zero if and only if x=0. Besides, to clearly illustrate the conclusion that for regularizers tabulated in Table VI, only the
1-norm can attain sparseness, the curves of Pφ(x) are shown in
1
Proof of Proposition 1: Since ƒ*(y) is convex in (16), it can be known that
in (18) with respect to (w.r.t.) y are concave, thus m(y) is convex w.r.t. y. That is, (19) is a convex problem.
where the penultimate equation is obtained because lg,λ(x) is an even function. Thus, φg,λ(y) is symmetric.
Besides, it is obtained via (21):
Then, the solution x* in (63) is discussed in terms of two cases, namely, y>0 and γ=0. Moreover, q (x)=x·y−ƒ(x) is defined, which is a concave function since ƒ(x) is convex, thus the solution x* for (63) satisfies ∇q(x*), and γ=ƒ′(x*) is obtained.
is unique, and it satisfies:
implying that y<x* due to ·g′(|x|)sign(x)>0 since g(x) is a monotonically increasing function, and by (63)(x*=y+λφ′g,λ(y), note that ∂φg,λ(y) is replaced with φ′g,λ(y) because φg,λ(y) is differentiable for |y|>0), the obtained is φ′g,λ(y)>0. Therefore, φg,λ(y) increases with y for y>0 and φg,λ(y) is nonnegative because φg,λ(0)=0.
If y1·y2>0 and it is first assumed that y1>0 and y2>0, then for any y>0, there is (65) and by (16), it is obtained that y increases with x* increasing. Combining x*=y+λφ′g,λ(y) and (65), it is known φ′g,λ(γ)=α/λ·g′(|x*|) sign (x*). Besides, it is assumed that g(x) is a concave function for x>λ and then g′(x) decreases with x increasing. Therefore, φ′g,λ(y) decreases as y increases. That is, φg,λ(y) is a concave function. Thus,
Then, there is φg,λ(y1+y2)≤φg,λ(y1)+φg,λ(y2). On the other hand, if y1·y2>0, y1<0 and y2<0, it is easy to obtain via (62):
Finally, if y1·y2<0 and it is supposed that y1>0 and y2<0, the conclusion is drawn immediately from φg,λ(y1+y2)≤φg,λ(y1+|y2|)≤φg,λ(y1)+φg,λ(|y2|)=φg,λ(y1)+φg,λ(y2).
Proof of Proposition 2: When, σ≤√{square root over (2)}λ, γ≤λ and τ≤√{square root over (3)}λ/2, it is known that Δd for φσ,λ(y), φγ,λ(y) and φτ,λ(y) is non-increasing. When |x|=λ, Δd obtains the maximum value, i.e., λ, thus Δd≤λ for the above three regularizers. While when φ(y)=λ|y|, Δd=λ, thus the proof is complete.
Proof of Proposition 3: According to Definition 3, φ·,λ(y) is separable, thus there is:
with solution being:
thus,
where Pφ·,λ(·) is a point-wise operator.
Besides, when σ≤√{square root over (2)}λ, γ≤1 and τ≤√{square root over (3)}λ/2, it can be known via Proposition 2:
thus, there is ∥x−Pφ·,λ(x)∥2≤√{square root over (n)}λ
Proof of Proposition 4: The proof follows exactly as in the case of Proposition 3 because Definition 4 shows that φ·,λ(Y) is separable.
In addition, when σ≤√{square root over (2)}λ, γ≤λ and τ≤√{square root over (3)}λ/2, via Proposition 2, there is:
thus, there is ∥X−Pφ·,λ(X)∥F≤√{square root over (mn)}λ
Proof of Proposition 5: Let X=U Diag(s) VT be the SVD of a rank-r matrix X∈, where s=[s1, s2, . . . , sγ]T is the vector of singular values, and
there is:
where the last inequality is due to Proposition 3.
Therefore,
is obtained.
Before the proof, the definition of a critical point and the following proposition are first stated as follows.
then, {M, E, Λ} is a critical point of ƒ.
According to (17), there is
By (21) or the section (b) of
is checked. Thus, h1(σ, λ) increases with λ and σ. Similarly, it can be verified that h2(γ,λ) and h3(τ,λ) increases with λ and their corresponding parameters γ and τ, respectively.
Let Uk Diag(sk)(Vk)T be the SVD of the matrix
where sk=[s1k, s2k, . . . , srk]T is the vector of singular values, r is the rank of {tilde over (M)}k and r«min(m,n)). Thus, {Λk} is bounded because:
where a is owing to Proposition 3.
According to the update equations of Ek+1 and Λk, it is obtained:
Thus, there is:
where the last inequality is owing to Proposition 4.
Similarly, according to the updates of Mk+1 and Λk, the obtained is:
Thus, there is:
It is easy to obtain:
However, the preceding Frobenius norm conditions cannot guarantee the boundedness of Mk and Ek. Next, their boundedness will be established via the boundedness of ρ
After updating Λk+1 and ρk+1, there is:
where b and c are due to Proposition 6 and (66), respectively, and Q is a constant w.r.t. u. Combining (67) and (68), there is:
Thus, there is:
Given M0, E0 and Λ0 are bounded, since
it is known that ρ
ρ
ρ
ρ
ρ
By Bolzano-Weierstrass theorem, the boundness of {Mk, Ek, Λk} suggests that there exists at least one accumulation point {M*, E*, A*} for {Mk, Ek, Λk}. That is, there exists a subsequence {Mk
Since the proximal operator is the closed-form solution to the Moreau envelope of the sparsity-inducing regularizers, that is, Ek+1 and Mk+1 are the minimizers of (56) and (58), respectively, although the regularizers are nonconvex, there is:
In addition,
Thus, the following can be obtained via combining (69) and (70):
Therefore, any accumulation point {M*, E″, Λ*} is a critical point.
Motivated by PCP, the parameter λ is chosen as c/√{square root over (max(m,n))}, i.e., λ=c/√{square root over (max(m,n))}, where c is a constant. The value of c is investigated in three cases, that is, different matrix rank, ratio of outliers and matrix dimensions.
It can be seen that the range of proper values of c for the proposed algorithms is larger than that of PCP, and compared with PCP, the matrix rank and dimensions have little impact on the choice of c for the proposed methods. Besides, the range of c for all techniques decreases with the increase of ratio of outliers, thus it is set λ=1/√{square root over (max(m,n))} in the synthetic data for convenience, because all the methods including PCP have a comparable recovery performance although λ is not the optimal value for the current settings.
The functional units and modules of the image restoration processor and methods in accordance with the embodiments disclosed herein may be embodied in hardware or software. That is, the claimed image restoration processor may be implemented entirely as machine instructions or as a combination of machine instructions and hardware elements. Hardware elements include, but are not limited to, computing devices, computer processors, or electronic circuitries including but not limited to application specific integrated circuits (ASIC), field programmable gate arrays (FPGA), microcontrollers, and other programmable logic devices configured or programmed according to the teachings of the present disclosure.
Computer instructions or software codes running in the computing devices, computer processors, or programmable logic devices can readily be prepared by practitioners skilled in the software or electronic art based on the teachings of the present disclosure.
The image restoration processor may include computer storage media, transient and non-transient memory devices having computer instructions or software codes stored therein, which can be used to program or configure the computing devices, computer processors, or electronic circuitries to perform any of the processes of the present invention. The storage media, transient and non-transient memory devices can include, but are not limited to, floppy disks, optical discs, Blu-ray Disc, DVD, CD-ROMs, and magneto-optical disks, ROMs, RAMS, flash memory devices, or any type of media or devices suitable for storing instructions, codes, and/or data.
The image restoration processor may also be configured as distributed computing environments and/or Cloud computing environments, wherein the whole or portions of machine instructions are executed in distributed fashion by one or more processing devices interconnected by a communication network, such as an intranet, Wide Area Network (WAN), Local Area Network (LAN), the Internet, and other forms of data transmission medium.
The foregoing description of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art.
The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated.