The present invention relates generally to methods for video and image processing. More specifically, it relates to image restoration and reconstruction techniques.
Atmospheric imaging turbulence caused by variation of refractive index along the optical transmission path can strongly affect the performance of a long-distance imaging system. It produces geometric distortion, space-and-time-variant defocus blur, and motion blur (if the exposure time is not sufficiently short).
Existing restoration algorithms for this problem can generally be divided into two main categories. The first is based on a multi-frame reconstruction framework and the other is known as “lucky exposure”. The main problem with multi-frame reconstruction algorithms is that in general they cannot accurately estimate the actual point spread function (PSF), which is spatially and temporally changing, hence limiting their performance. The “lucky exposure” approach employs image selection and fusion methods to reduce the blurring effects caused by turbulence. One shortcoming of this method is that even though turbulence caused blur is strongly alleviated through the fusion process, the output still suffers from blur caused by diffraction-limited PSF.
Accordingly, an embodiment of the present invention provides a method for restoring a single image from an image sequence acquired in anisoplanatic (i.e., air turbulence) scenarios. The method is designed to reduce the spatial variation of PSFs over the whole image space through a non-parametric kernel regression algorithm, so that the blur can be approximately treated as spatially invariant, and the latent image content is then estimated globally. The method is capable of alleviating geometric deformation and space-and-time-varying blur caused by turbulence, recovering details of the scene and significantly improving the visual quality.
An overview of an embodiment of the method is illustrated in
The techniques of the present invention may be implemented, for example, as an image restoration and reconstruction computer system for processing image data input to the system or stored by the system, where the image data have geometric distortion and space and time-variant blur due to atmospheric turbulence. The present invention may also be realized as a digital storage medium tangibly embodying machine-readable instructions executable by a computer, where the instructions implement the techniques of the invention described herein to perform image restoration and reconstruction.
In one aspect, the invention provides a method for restoring a single image from an image sequence acquired in aniso-planatic scenarios. The 3-D physical scene is assumed to be static, while the air between the scene and sensor is affected by atmospheric turbulence. The approach is designed to reduce the space and time-variant deblurring problem to a shift invariant problem. It focuses on the observed regions convolved by near-diffraction-limited PSFs, which can be viewed as space and time-invariant, and thus can be estimated along with the latent sharp image through a blind deconvolution algorithm. This framework is capable of alleviating both geometric deformation and blur, and significantly improving the visual quality.
The imaging process can be modeled as
g
k
=H
k
F
k
f+n
k (1)
where gk represents the k-th observed image frame, f denotes the (unknown) ideal image, Fk represents the geometric deformation matrix, Hk represents the blurring matrix, and nk represents additive noise (which is defined as zero mean and i.i.d). The image reconstruction problem is to estimate the ideal image f based on the observed image data gk.
In some prior approaches to this reconstruction problem, specific models (e.g., Gaussian) are used to describe the shape of local PSFs caused by air turbulence, but these do not fit practical cases very well. In the real world, while turbulence-caused PSFs can look quite arbitrary, they still have some common characteristics. For example, usually they are restricted within a small region, and contain a “core,” which is the peak intensity of a PSF. In the above model, since geometric distortion is separated from the blurring matrix, the core of each local PSF (also the entry with highest value in each row) is typically located on the diagonal of Hk.
To estimate the latent image f, a new restoration framework is provided, which contains three main steps, as outlined in
1. Non-rigid image registration (step 102).
2. Fusion-based near-diffraction-limited image restoration (step 106).
3. Single image blind deconvolution (step 110).
The first step 102 registers each frame onto a fixed reference grid, removing geometric distortion from the observed data. The second step 106 fuses the registered sequence to produce a single image convolved by near-diffraction-limited PSFs, which have short support and can be approximately viewed as spatially invariant. The fused image is finally deconvolved in the third step 110 through a blind deblurring algorithm based on a natural image prior. Details about each step are given in the following subsections.
Non-Rigid Image Registration
Assuming that the local turbulent motion has a zero-mean distribution, the geometric distortion can be removed by simply averaging the observed frames. Such averaged image, though even more blurred than the observed data, can serve as a reference frame to register each observed frame. A B-spline based non-rigid registration method may be directly implemented in the approach. For example, in prior work by the present inventors, such a B-spline based non-rigid registration method is disclosed in the article “Image reconstruction from videos distorted by atmospheric turbulence,” In SPIE Electronic Imaging, Conference 7543 on Visual Information Processing and Communication, San Jose, Calif., January 2010. This method incorporates a symmetry constraint that can effectively improve the estimation accuracy.
The techniques of the present invention are related to the question of how the registration operator physically changes local PSFs. Assume that the warping matrix Fk is non-singular. The registration operator can then be denoted as Fk−1. Left-multiplying both sides of eq. 1 by Fk−1 we have
q
k
={tilde over (H)}
k
f+ñ
k (2)
where
{tilde over (H)}
k
=F
k
−1
H
k
F
k (3)
ñ
k
=F
k
−1
n
k
and where
q
k
=F
k
−1
g
k
is the registered image.
The blurring matrices {tilde over (H)}k and Hk are related by a similarity transformation. Since a similarity transformation preserves the spectrum (eigenvalues) of Hk, the transformation should not significantly change the physical shape of PSFs. In particular, if the motion is pure translational and applies only integer movement (pixel to pixel), then Fk is a block-circulant matrix with each row and column containing precisely a single 1 and with 0 s everywhere else. That is, we can approximate Fk by a permutation matrix, which means that FkT=Fk−1. For such matrices, the corresponding similarity transformation circulates the rows and columns of Hk. In other words, the registration operator only moves the locations of local PSFs without changing their physical shape. Of course the overall motion caused by turbulence is not translational. However, the local motion inside an iso-planatic region can be well-approximated as translational. If the support of any local PSF is also small enough, then inside an isoplanatic region the transformation Fk−1HkFk preserves the physical shape of local PSF. Meanwhile, it can be shown that the diagonal entries of Hk still remain on the diagonal of Hk with their locations permutated. Furthermore, since Fk can be approximated as a permutation matrix (pixel to pixel movement) we can also write:
cov(ñk)=Fk−1cov(nk)Fk≈σn2I (4)
Near-Diffraction Limited Image Restoration
By decomposing the overall blur {tilde over (H)}k=H+Δ{tilde over (H)}k, equation (2) can be rewritten as
where H is diffraction-limited blur and z=Hf denotes the diffraction-limited image. The matrix Δ{tilde over (H)}k represents turbulence-caused blur, and ek=Δ{tilde over (H)}kf is the corresponding blurring artifact. The diffraction-limited blur is not a priori known, but is assumed to be space and time invariant, and there are numerous blind deconvolution algorithms available for removing it. Estimating Δ{tilde over (H)}k, on the other hand, is not trivial. However, if we treat ek as an additive noise, then it is still possible to estimate the diffraction-limited image z. This requires an analysis of the statistical properties of ek, which we present next.
Both {tilde over (H)}k and H can be viewed as row-stochastic (i.e., each row consists of non-negative real numbers that sum up to 1), which means {tilde over (h)}iT1=hiT1=1, and thus
Δ{tilde over (h)}iT1=({tilde over (h)}i−hi)T1=0 (6)
Here Δ{tilde over (h)}iT, {tilde over (h)}iT, and hiT represent the i-th rows of matrices Δ{tilde over (H)}k, {tilde over (H)}k, and H, respectively. Because the support of the PSFs {tilde over (h)}i and hi are restricted to a small isoplanatic region around the i-th pixel denoted as wi, all the non-zero values of Δ{tilde over (h)}i should also be located inside wi:
Δhij=0, ∀j∉i (7)
where Δ{tilde over (h)}ij represents the j-th element of vector Δ{hacek over (h)}i. By defining a diagonal matrix Mi denoting a mask having 1 s in the locations inside wi and 0 s everywhere else, we have MiΔ{tilde over (h)}i=Δ{tilde over (h)}i.
We model the latent sharp image f as a random field whose mean vector is
mf=[mf1, mf2, . . . ]9,
and its covariance matrix is assumed to be diagonal
Cf=diag[σ2f1, σ2f2, . . . ].
The statistics of f are considered piece-wise stationary:
mfj≈mfi, σ2fj≈σ2fi∀j∈wi (8)
and thus
Mimf≈mfiMi1 (9)
So considering eq. 6-9, we have:
which indicates that the mean of ek is zero. Now we have zero-mean additive noise ek, whose covariance matrix is Ce
σ2e
Again, by applying eq. 7 and 8 the above entries can be written as:
σ2e
Combining the additive noise ñk and ek together we define the total additive noise
εk=ñk+ek (13)
Since ñk and ek are independent, the covariance matrix of εk becomes Ck=Ce
{circumflex over (z)}=arg minzΣk(qk−z)TCk−1(qk−z) (14)
For simplicity, we only take the diagonal part of Ck denoted as Uk=diag[uk1, uk2, . . . ], whose i-th diagonal entry can be written as:
u
ki=σ2n+σ2fi∥Δ{tilde over (h)}i∥2 (15)
Now the estimation problem in eq. 14 can be simplified as:
Or in pixel-wise form:
{circumflex over (z)}
i=(Σkuki−1qki)(Σkuki−1)−1 (17)
which may be interpreted as an image fusion process, where the weight value uki−1 is determined by σ2n and σ2fi∥Δhi∥2. If σ2n is temporarily constant, then a larger value of σ2fi∥Δ{tilde over (h)}i∥2 (i.e., more blurry) corresponds ponds to a smaller value of uki−1 (i.e., lower weight).
In practice, noise variance σ2n is viewed as spatially invariant and can be estimated using, for example, median absolute deviation (MAD) method.
Once sufficient observations are collected, diffraction-limited image patches (also the sharpest isoplanatic patches) that occasionally appear due to the turbulence variation can be detected through known sharpness metrics such as Strehl ratio, image gradient, or local intensity variance. These diffraction-limited patches suffer from noise and cannot be used directly as outputs since noise will be magnified in the following deconvolution step. However, they can be utilized as references for calculating the local variances of ek and the weights (15) for the fusion. Let us consider a given isoplanatic region wi centered at the i-th pixel across all the registered frames. Assume that the sharpest patch of wi, which can be approximated as a diffraction-limited patch, is detected in the k′-th frame:
M
i
q
k′
=M
i
Hf+M
i
n
k′. (18)
Then, given the k-th frame we can write the patch difference:
M
i(qk−qk′)=MiΔ{tilde over (H)}kf−Mink′+Mink (19)
Because of the isoplanaticism of local PSF, and of the piecewise constancy of image variance, it can be deduced that:
where |wi| is the cardinality of the set wi. Thus the weight-related value uki in eq. 15 can be approximated by
u
ki=var[Mi(qk−qk′)]−σ2n (21)
It is worth noting that due to the covariance matrix simplification in eq. 16 and the limited registration accuracy, the PSF of the fused image is, in practice, not the pure diffraction-limited PSF. Namely, it includes diffraction-limited blur, residual turbulence blur, and blur caused by registration error. So we call such PSF near-diffraction limited. Because the fusion process strongly reduces the PSF variation, such PSF can be approximately viewed as spatially invariant.
A concise description of the algorithm for the near-diffraction-limited image restoration step is as follows:
Procedure for Restoring A Diffraction-Limited Image from Registered Frames
1. Given a registered frame sequence {qk}, divide each frame into N×N overlapping patches centered at each pixel, and calculate the variance of each patch as a local sharpness measure.
2. For patches centered at the i-th position across all the frames, find the sharpest patch, say in frame k′, as the diffraction-limited reference.
3. Set uk′i=σ2n, and for the remaining patches in frame k≠k′ use eq. 21 to calculate uki.
4. Restore the i-th pixel according to the regression form eq. 17.
5. Go to the (i+1)-th pixel and return to step 2.
Single Image Blind Deconvolution Finally, the unknown image f and the near-diffraction-limited PSF h (which is the vector form of H) can be estimated using a Bayesian image deconvolution algorithm described as:
{circumflex over (f)},ĥ
=arg minf,h{z−HF}+λ1Rf(f)+λ2Rh(h)
where Rf and Rh are the regularization terms based on prior knowledge about the latent sharp image f and the PSF h. Recent research on natural image statistics has shown that image gradients obey heavy-tailed distributions that have most of the mass on small values but give significantly more probability to large values than Gaussian distributions. Based on these studies, several sparsity-based regularization methods have been introduced and have achieved great success in solving the blind deconvolution problem. One example is disclosed in Q. Shan, J. Jia, and A. Agarwala, “High-quality motion deblurring from a single image,” ACM Transactions on Graphics (SIGGRAPH), 2008. Shan's method may be directly applied in the proposed framework as the final step, using their default parameter settings, except that the noise level parameter ‘noiseStr’ is chosen in the range [0.01, 0.05] according to the actual noise level observed in the given data.
The methods have application, for example, to telescopic optical imaging where at least part of the optical path passes through the atmosphere.
This application is a continuation-in-part of U.S. patent application Ser. No. 13/856977 filed Apr. 4, 2013, which claims priority from U.S. Provisional Patent Application 61/620185 filed Apr. 4, 2012, both of which are incorporated herein by reference.
This invention was made with Government support under contract CCF-1016018 awarded by National Science Foundation, and under contract FA9550-07-1-0365 awarded by Air Force Office of Scientific Research. The Government has certain rights in this invention.
Number | Date | Country | |
---|---|---|---|
61620185 | Apr 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13856977 | Apr 2013 | US |
Child | 14108083 | US |