Embodiments of the invention generally relate to optical image acquisition and processing apparatus and methods.
Objects imaged by conventional imaging subsystems are sharply in focus over a limited distance known as the depth of field, which is inversely proportional to the square of the imaging system's numerical aperture for diffraction-limited imaging. Present-day cameras have mechanical focusing means, including automatic systems, to provide high quality images of particular scenes at various object distances. Even with these means, it is difficult to photograph objects clearly that span a large range of such distances. Cameras with a larger depth of focus will clearly provide superior photographs.
Digital processing of image data on a pixel-by-pixel basis has afforded more opportunity for improving and correcting optically imaged scenes. Some of these improvements have related to increasing the depth of field. For example, digital processing has been used to combine images of the same scene taken at different depths of focus to produce a composite image having an extended depth of field. The multiple images take time to collect, are difficult to process, and are generally unsatisfactory for scenes subject to change.
Amplitude attenuation filters have also been used to extend the depth of field. Typically, the attenuation filters are located in the aperture of the imaging systems, leaving inner radii clear but attenuating the outer annulus. However, the filter introduces large amount of light loss, which limits its applications.
More promising attempts have been made that deliberately blur an intermediate image in a systematic way so that at least some information about the imaged object is retained through a range of focus positions and a non-ideal impulse response function remains substantially invariant over the defocus range. Digital processing, which effectively deconvolutes the point spread function, restores the image to a more recognizable likeness of the object through an extended depth of field.
One such example locates a cubic phase mask within the aperture of the imaging system to generate a distance invariant transfer function, thereafter. Digital processing removes the blur. Although significant improvement in the depth of field is achieved, the cubic phase mask is not rotationally symmetric and has proven to be expensive and difficult to fabricate.
Another such example similarly locates a circularly symmetric, logarithmic asphere lens to extend the depth-of-field, which is more economical to manufacture. However, for the log-asphere lens, the impulse response is not perfectly uniform over the full range of operation, and as a result, some degradation is experienced in the image quality of the recovered image.
Reconstruction algorithms for removing the blur of such intermediate images are subject to problems relating to the quality and efficiency of their results. Nonlinear processing algorithms can suffer from slow convergence or stagnation and produce images with reduced contrast at high spatial frequencies.
Our studies of circularly symmetric multifocal lenses have revealed that a controlled amount of spherical aberration provides a desirable distance-invariant blur that leads to superior depth-of-field imaging. According to an embodiment of the invention, a multifocal lens for extending the depth of field can be based on any standard imaging arrangement modified to incorporate a third-order spherical aberration as well as higher-order spherical aberrations. Non-limiting examples of such standard imaging arrangements include Petzval lenses, Cooke lenses, and double Gauss lenses.
In addition, our studies have found that a further improvement in depth of field imaging, particularly at diffraction-limited resolution throughout the extended depth of field, can be realized, according to an embodiment of the invention, by obscuring a center portion of the aperture of the multifocal imaging subsystem to narrow the impulse response for close-in distances (β-design) or for far distances (γ-design). This increases the range in distance over which the impulse response is invariant. The center portion of the multifocal imaging subsystem can be a major contributor to the variation in the impulse response with distance. The combination of a central obscuration with a properly designed multifocal imaging can be used to further extend the depth of field, or to support higher resolution imaging through the extended depth of field.
Referring to
The multifocal imaging subsystem 12 includes a single or multiple element lens 22 and a phase plate 24. The lens 22 can be a conventional lens having at least one spherical surface arranged for ideal imaging and the phase plate 24 may be arranged to contribute a predetermined amount of spherical aberration. A central obscuration 26 can also be located within an aperture 28 of the multifocal imaging subsystem 12 for further improving performance. According to various embodiments, the phase plate 24 can be fabricated separately and aligned with the lens 22 as shown, or the optical contribution of the phase plate can be incorporated into a surface of lens 22, such as in the form of a logarithmic lens. Although both the lens 22 and the phase plate 24 are typically transmissive, either or both of the lens 22 and the phase plate 24 can be alternatively fashioned as reflective surfaces such as in telescopic photography applications. The central obscuration 26 can also be realized in different ways in various embodiments, such as by adding a central stop within the aperture 28 or by arranging for an annular pattern of illumination that has center darkness. A pre-existing central stop can also be used, e.g., a secondary mirror of a telescope.
Other imaging systems contemplated by embodiments of the invention include multiple lens elements such as for dealing with chromatic aberrations or other imaging requirements. The embodiments of the invention provide increased flexibility within such multiple lens element designs to distribute the desired amount of spherical aberration among a plurality of the lens elements. For example, at least two of the lens elements could be formed as logarithmic aspheres, each incorporating a portion of the desired spherical aberration.
The image detection device 14, which collects an intermediate image 30 of object 20 that is generally blurred, can be fashioned as a pixilated CCD (charge coupled device) or CMOS (complementary metal oxide semiconductor) detector or other light sensitive device. The detector pixels can be arranged as a two-dimensional array, as a one-dimensional array, or even as a single detector pixel. Any pixel combinations short of two dimensions are preferably subject to scanning for collecting enough information to complete a two dimensional intermediate image 30. However, one-dimensional imaging can be used for particular applications.
The digital processing subsystem 16 can include a computer-processing device having a combination of hardware and software for the purpose of image processing. The digital processing subsystem 16 can be incorporated into a camera system that also includes the multifocal imaging subsystem 12, or, alternatively, the digital processing subsystem 16 can be arranged as a standalone image-processing computer. The primary purpose of the digital processing subsystem 16 is to sharpen the intermediate image 30. An inverse filter or its modifications, e.g., Wiener filter, can be used for this purpose. According to an embodiment of the invention, a nonlinear algorithm such as an iterative maximum entropy algorithm is used to sharpen the intermediate image 30. If the maximum entropy algorithm is used, an optional acceleration factor, referred to herein as a metric parameter, can be chosen to optimize the speed and convergence of the algorithm.
The digitally processed image, which is referred to herein as a recovered image 32, is outputted to the display device 18, which can be a CRT (cathode ray tube), LCD (liquid crystal display), or other display device appropriate for viewing purposes. Alternatively, the display device 18 can be omitted and the recovered image 32 can be inputted to other functional hardware/software. For example, the recovered image 32 can be input to a pattern recognition system or a machine vision system. If the recovered image 32 is used for these latter purposes, then the digital processing subsystem can be incorporated into the pattern recognition or machine vision system. The digital processing device can become optional according to an embodiment depending on the amount of blur in the intermediate image.
The illustrative integrated computational imaging system 10 is applicable to binary or gray scale or color imaging. It is also applicable to a range of different wavelengths including infrared imaging.
An optical diagram of a modified multifocal imaging subsystem 12 for producing the intermediate image 30 according to an embodiment of the invention is illustrated in
In optical system design, the ideal imaging components and the aberrations can be described as follows:
φ/k=W+φideal/k (1)
wherein the phase delay φ is measured in radians, k equals 2π/λ0 where λ0 is the average wavelength of illumination, and W is the optical path difference (O.P.D.) in micrometers.
For the ideal imaging system, it is well known that
φideal=k(√{square root over (t2+r2)}−t+√{square root over (s02+r2)}−s0) (2)
in which the phase delay φideal is measured in radians for a perfect diffraction limited lens, r is the radial coordinate in plane I, and s0 is the focused object position, and k=2π/λ0.
For the ideal lens, as an example with s0=1500 mm, t=62.5 mm, R=8 mm, and λ0=0.5 μm, from a power series expansion of φideal, one can readily find:
which Equation (3) is valid in the non-paraxial regime.
For the OPD, W, of the two types of logarithmic aspheres denoted by subscript β and subscript γ, one can express them as follows:
where s0 is the center of the depth of field range, λ0 is the wavelength of illumination in vacuum, and the expression of φP for the two types of logarithmic aspheres, φPβ and φPγ, can be written as follows.
From a power series expansion of Equations (5) or (7), it can be appreciated for purposes of the embodied invention that additional spherical aberration is the dominant feature of the purposeful blur that is being introduced. This will be made more evident within a description of some specific embodiments.
For completing a design based on Equations (4)-(8), the desired range for the depth of field s1, s2 can be selected along with representative values for t, R, δR, s0, and λ0. Thereafter, the variables aβ, Aβ, and φPβ (or aγ, Aγ, and φPγ) can be computed. From these, Equation (4) can be used to compute the aberration term W.
The logarithmic aspheres described above are examples of a multifocal lens that can be constructed in accordance with an embodiment of the invention. From a more general point of view, a multifocal lens useful for extended depth-of-field imaging can be composed of any standard imaging arrangement that is designed to incorporate a predetermined amount of spherical aberration, including third-order spherical aberration as well as higher-order spherical aberrations. For example, non-limiting examples of such standard imaging and projection arrangements such as Petzval lenses, Cooke lenses, and double Gauss lenses can be used for these purposes.
For describing a multifocal lens in terms of a range of aberrations, it is useful to expand the aberration function φP in a series in terms of (r/R). For example, if the design parameters: s0=1500 mm, s1=1400 mm, s2=1615 mm, t=62.5 mm, R=8 mm, λ0=0.5 μm, and δR=0, from Equations (4), (5), and (6) are used, a power series expansion of the phase delays of a β-type logarithmic asphere or a logarithmic phase plate is found as follows.
In Table I, the first row of data is the whole phase delay function of the multifocal lens from Equation (1), i.e., φβ(r)=6721.22(r/R)2−45.63(r/R)4+0.32(r/R)6+ . . . . The second row of data is the radian phase delay function for an ideal lens arrangement, e.g., Petzval lens, Cooke lens, double Gauss lens, or Cassegrain system. The third row of data is the aberration terms of the phase delay function, which is the difference between the phase delays of the multifocal lens and an ideal lens. The dominant aberration term in the multifocal lens is the third-order spherical aberration (i.e., the fourth-order term of r/R). For diffraction-limited resolution, the largest allowable OPD is generally 0.25λ. To achieve a ten-fold increase in the depth of field, the spherical aberration is around 3 wavelengths (i.e., 19.28λ/2π=3λ), which is a little over ten times the allowable defocus for diffraction-limited imaging. Good performance of an embodiment of our multifocal lens includes spherical aberration in the amount of 1.8 to 6 wavelengths, while the amount of higher-order spherical aberration is largely insignificant. The multifocal lens designed this way will have an extended depth of field from 6 to 10 times that of a conventional ideal lens.
Another example of our multifocal lens has the same parameter values as above, but with δR/R=0.5 to illustrate the effectiveness of a central obscuration 26. The phase delay for the different terms is shown in the following Table II.
Although the third-order spherical aberration (r/R)4 looks larger than without central obscuration, the effective third-order aberration, i.e., the phase delay difference contributed from spherical aberration between the edge of the lens and the edge of the central obscured block is: 25.76−{25.76×(δR/R)2}=19.32 radians. Thus, the effective third-order aberration amounts are similar for both the full aperture multifocal lens described by Table I and the centrally obscured multifocal lens described by Table II. Accordingly, the good performance centrally obscured multifocal lens has an effective third-order aberration that is still within a range from 1.8 to 6 wavelengths.
From the above description, it is apparent that the multifocal lens having an effective third order spherical aberration in the range of 1.8 to 6 wavelengths, can increase the depth of field six to ten times over that of a conventional lens. This conclusion pertains to any reasonable amount of central obscuration and also is independent of the wavelength of illumination, focal length and best focus object distance.
The second order term, i.e. (r/R)2 of the series expansion, is not relevant to the increase of depth of field, but has the function of changing the position of center of the focus range. For the second order term, we generally pick a value in a way that the aberration W at the inner edge of aperture 28 and the aberration W at the outer edge of central obscuration 26 have similar values to facilitate the phase plate or lens fabrication. In the case that no central obscuration is used, i.e., δR=0, a coefficient of the second order term is selected so that the aberration W at the edge of aperture 28 is zero.
There are different ways that these controlled aberrations can be incorporated into the well-known imaging lens arrangements, e.g., Petzval lens or Cooke lens. For an existing lens arrangement, a simple way is to fabricate the aberration part of the multifocal lens as the phase plate 24, which can be attached to the aperture 28 of the lens arrangement. This method is most effective if the aperture 28 of the lens arrangement is outside the last lens element at the image plane (II) side.
Another method of multifocal lens realization is to incorporate the aberration into the lens design of the logarithmic asphere 34. By modifying a surface parameter of the logarithmic asphere 34, the overall phase delay function can still include the ideal lens part and the aberration part. This method has the advantage that no actual lens element is needed. e.g., the flipover of the well-known lens arrangement introduces large amount of spherical aberration, which could be used as the starting design point. Two important features of this embodiment are that it contains good angular resolution as well as good color correction. The desired amount of spherical aberration can also be distributed among multiple lens elements of the design to proved more design flexibility.
A substantially distance-invariant impulse response is important to the recovery of images having extended depths of focus. A predetermined amount of spherical aberration can be used to produce a more distance-invariant impulse response for effective performance both with and without central obscuration. In the lens with δR=0.5, an optimum amount of spherical aberration has been found to be about 3 waves. However, fairly good image recovery is obtained for a spherical aberration in the range from 1.8 to 6 waves.
The above discussions apply also to the case of the γ-type logarithmic asphere. For the γ-type logarithmic asphere, the coefficients of the power series for W change signs but are otherwise similar, which is shown below by way of example.
For the γ-type logarithmic asphere, the same design parameters can be used including: s0=1500 mm, s1=1400 mm, s2=1615 mm, t=62.5 mm, R=8 mm, λ0=0.5 μm, and δR=0. From Equations (4), (7), and (8), a power series expansion of the phase delays of a γ-type logarithmic asphere or a logarithmic phase plate is found as shown in Table III.
As another example of the multifocal lens, Table IV is based on the same parameter values as above, but with δR/R=0.5 to illustrate the effectiveness of a central obscuration 26. This result can be compared to that in Table II.
A difference between the β-type and γ-type phase plates is the sign change for the second and fourth-order terms. The fourth-order term, which corresponds to third-order spherical aberration, is positive for the γ-type lens and negative for the β-type lens. However, the absolute values of the corresponding third-order spherical aberration terms are similar for the same design range.
To demonstrate the performance of the γ-type lens,
Based on the similarity of the impulse responses over the range of object distances, e.g., s1 to s2, digital processing of the intermediate image 30 can be used to sharpen the images of object points throughout the depth of field. One method of image recovery involves using an inverse filter, such as the Weiner-Helstrom inverse filter. Alternatively, a maximum entropy algorithm can be programmed into the digital processing subsystem. This approach for image recovery according to an embodiment of the invention is set forth below.
For the known, measured, noisy image d, point spread function h, and standard deviation σi,j of noise θ in i,jth pixel of d, and unknown object f, one can write the following relation:
d=h**f+θ (9)
in which the double asterisk (**) is a spatial convolution. For an estimate of the object f, we start with an assumed object f(0) and iterate according to
where:
and Caim is the total number of pixels in the image and <f> is the average of image.
The maximum entropy algorithm is an iterative approach to determining an estimate of object 20. A diagram of the algorithm is shown in
The single point spread function used in the convolution can be calculated as an average of the point spread functions observed for the different focal depths. However, individual focal distances can be weighted differently to adjust the single point spread function for favoring certain object distances over others for compensating for other effects. The single point spread function could also be varied experimentally to achieve desired results for particular applications or scenes.
For each iteration, the new estimation of object is calculated from the earlier estimation by adding three (or four) direction-images with appropriate coefficients. i.e.,
where f(n+1) is the (n+1)th estimation of object, f(n) is the earlier nth estimation, and ei is the ith direction-image.
Thus, two key steps of the algorithm are:
i) What direction-images ei should be used.
ii) How to calculate the corresponding coefficients xi of the direction images.
A new metric parameter γ is introduced as a first step to determining the direction-images ei. The parameter γ adjusts the pixel values of direction-images derived from a steep ascent method. The parameter γ ranges from 0 to 1, although γ>1 is still possible. When this parameter is larger, more emphasis is given to the larger pixel values in the image, and also there exists more deviation of the direction images ei from direction images derived steepest ascent method.
In the second step, Taylor expansions of S and C relative to variables δf are calculated up to second order terms. Hence, the quadratic approximation models St and Ct are established. The quadratic models greatly facilitate the constrained maximization process because these quadratic equations are much easier to solve than the original nonlinear equations in Equations (10) and (11). The diagram of how to find the next estimation of the object is shown in
In order to study the optimum value of metric parameter γ, an extended study has been made of the effect of varying the parameter γ. Three different pictures of varying histograms are used including: binary scene, zebra, and tiger. Each of these pictures has 256×256 pixels with the maximum pixel value scaled to 255. Each picture is blurred using 15 normalized impulse responses with the maximum blur consisting of a 5×5 matrix with 15 non-zero values and 10 zeros in the outer regions. Gaussian noise is added with a standard deviation σ ranging from 0.2 to 1.8 in 9 steps. The metric parameter γ is given 21 values ranging from 0.0 to 1.0. Hence, in these computer simulations there are about 8,000 cases. It is convenient to use an effectiveness parameter for the number of iterations, which is defined by Lσ/D, where L is the number of loops for the maximum entropy calculation to converge, σ is the noise standard deviation, and D is the number of non-zero pixel in the blurring function. In
For the computer simulation, the number of loops L for the maximum entropy recovery is linearly proportional to the area of the point spread function, D, or qualitatively proportional to the severity of the blur. The loop number is also approximately inversely proportional to the standard deviation of the noise, σ.
For a wide variety of pictorial content, it is apparent from
A more rigorous mathematical description of the metric parameter—maximum entropy algorithm follows within which control over the metric parameter .gamma. enables the algorithm converge much faster for a wide range of scenes.
Two operators used in this section are defined in the following for convenience:
To find the solution of {fk} according to the Lagrange multiplier method, a new function is defined as:
Q=S−λC (13)
where is λ is a Lagrange multiplier constant. Now the problem becomes to maximize Q under the constraint C=Caim. Since Q is a function of n variables, where n is the number of pixels in the image, which is usually very large, an iterative numerical method can be used to find the solution. In each iteration, the standard way is first to determine the search directions in which the solution is estimated to lie and then to find the step length along these directions.
The choice of directional images is important in determining the convergence and speed of the algorithm. In the steepest ascent method, the search direction for maximizing Q is ∇Q. But in order to adjust the weight of different pixel values fi, the direction can be modified to be:
eA=f.γ.×∇Q (14)
In the above equation, the new metric parameter γ improves the speed and reliability of the metric parameter-maximum entropy algorithm. For image deblurring in photography, the larger pixel values will have larger weight, so γ>0 is chosen to let the algorithm approach the desired larger pixel value faster. Generally, γ is chosen from 0 to 1. When γ=0, eA becomes the search direction for the steepest ascent method. When γ=1, eA becomes the search direction used by Burch et al. in a paper entitled “Image restoration by a powerful maximum entropy method,” Comput. Visions Graph. Image Process. 23, 113-128 (1983), which is hereby incorporated by reference. Neither the steepest ascent method nor the method of Burch et al. incorporate the metric parameter γ, which provides a new mathematical construction that can be manipulated to increase the speed of convergence and avoid stagnation.
At the maximum point Q, we have:
∇Q=0 (15)
This implies ∇Q·∇Q needs to be minimized, too. Accordingly, the next search direction should be ½ ∇(∇Q·∇Q), or ∇∇Q·∇Q. Here ∇∇Q is the dyadic gradient whose component is defined as follows:
Again, in order to emphasize the bigger pixel values, the direction is modified to be:
eB=f.γ.×[∇∇Q·(f.γ.×∇Q)] (16)
Substitution of Equation (13) into Equations (14) and (16) yields the following:
Observing the above expression, we know that the two directions actually are linear combinations of many directions, which can be treated as separate search directions, viz.,
From Equations (10) and (11), we have:
Substitution of Equations (19) and (20) into (18) yields the components of each search direction as follows:
In the algorithm, e5 and e6 can be disregarded as search directions because when γ=0.5 they decompose to e1 and e2, respectively; and also when γ<0.5, for small pixel values, they both involve dividing by small numbers, which can cause numerical accuracy problems. Accordingly, e1, e2, e3, and e4 are chosen as four search directions.
To simplify the algorithm, three search directions are enough for this problem, we can pick e1, e2, and e3 or we can pick e1, e2, and e4, −e3, as another choice. The algorithm converges at about the same speed, although the latter choice is a little better. Here, λ is a constant chosen by Equations (13) and (15), i.e., λ is given by:
In either case, three directions can be written as e1, e2, and e3 for simplicity. In the calculation of search directions, it is apparent that that the directions e2 and e3 are basically the convolutions related to the real point spread function, h, and object, f, before being shaped to one dimension. Care needs to be taken to make certain that there is not pixel shift or image position shift after the convolution operation.
After three search directions are calculated for the current iteration (n), the next task is to find an estimation of the object for the next iteration f(n+1), which is defined as:
f(n+1)=f(n)+δf (23)
where δf is the change of image for the current iteration. It is defined as linear combinations of search directions, with their coefficients to be determined, i.e.,
δf=x1e1+x2e2+x3e3 (24)
Since S and C are functions of f that vary in a complicated way, a Taylor expansion can be used to calculate their values as a function of the search directions. Retaining up to the quadratic terms, St and Ct can be written as follows:
Substitution of Equation (24) into Equations (25) and (26) yield the following expressions written in matrix form,
In Equations (27) and (28), the notation is defined as follows:
where [ . . . ]T denotes the transpose of matrix. Matrices A, B, M, and N can be calculated from Equations (20) and (21).
Equations (27) and (28) can be simplified by introducing new variables to diagonalize B and N. First, the rotation matrix R is found to diagonalize the matrix B, i.e.,
RBRT=diag(λ1, λ2, λ3) (30)
where diag( . . . ) denotes the diagonal matrix.
A new variable Y is defined as follows
Y=RX (31)
Substitution of Equations (31) and (30) into Equations (27) and (28) yield the following expressions:
Some eigenvalues of B may be very small, and this is discussed in the following two cases.
Case i) Assume that none of λ1, λ2, and λ3 is small.
We introduce Z, such that:
Z=diag(√{square root over (λ1)}, √{square root over (λ2)}, √{square root over (λ3)})Y (34)
Substituting Equation (34) into equations (32) and (33), yields:
A second rotational matrix V is introduced to diagonalize P, i.e.,
VPVT=diag(μ1, μ2, μ3) (37)
and also define U as:
U=VZ (38)
Then, substitution of Equations (37) and (38) into Equations (35) and (36) yield the following expressions:
Combining Equations (31), (34), and (38), yields the identity:
Case ii) Assume λ3 is small relative to λ1 and λ2.
In this case, λ3≈0, y3=0 in Equation (31), and also:
Then, Equations (32) and (33) become:
A second rotation matrix V is introduced, such that
A new variable U is defined as:
U=VZ (46)
then
Combining Equations (31), (42), and (46), yields the identity:
The other case, when two values of λ1, λ2, and λ3 are small, can be treated in a similar way. In general, the following expressions of St and Ct can be written for all cases with the quadratic matrices diagonalized as follows:
The relation between ui and xi (i=1, 2, 3 or i=1, 2, or i=1) can be found in the identities Equations (41) or (49).
Now the maximum entropy problem becomes one to maximize St in Equation (50) under the constraint of Ct=Caim in Equation (51). Ct in Equation (51) has minimum value of:
Clearly, Cmin could be larger than Caim. If this happens, then the maximization of S under the constraint of Ct=Caim will not have any solution. Accordingly, a new constraint is defined that can always be reached as follows:
The Lagrange multiplier method can be used to solve the maximization in Equation (50) by introducing a new variable Qt as follows:
Qt=αSt−Ct (α>0) (54)
where Qt corresponds to the Q wave variable at the left side of Equation (54), (a>0) guarantees that the solution found will maximize the entropy instead of minimizing it.
Substituting Equations (50) and (51) into (54), yields the values of ui that maximize Qt as follows:
where α is determined by solving the following equation, which is derived by the substitution of Equation (55) into Equation (51) and by the use of the constraint in Equation (53):
After α is known, coefficients of x1, x2, and x3 are found by Equations (55) and (41) or (49), and the next assumed object for the next iteration can be calculated by Equations (23) and (24). At each iteration, the negative values in the assumed object are reset to zero.
At each iteration if the constraint of C=Caim is satisfied, whether the entropy is maximized is checked by determining if ∇Q is zero, or whether ∇S and ∇C are parallel by calculating the following value:
The algorithm stops if |test|<0.1.
There is a special case when coefficients of x1, x2, and x3 are too large such that the expressions in Equations (25) and (26) are not accurate. The Burch et al. paper deals with this by introducing a distance penalty parameter. However, if the starting estimation of the object is a uniformly gray picture or the blurred picture, then this complexity can generally be avoided. Only when the starting image is random should the extra parameter be introduced in the algorithm but only through the first several loops. A further description of the metric parameter-maximum entropy algorithm is found in a paper authored by the co-inventors entitled “Computational imaging with the logarithmic asphere: theory” J. Opt. Soc. Am. A, Vol. 20, No. 12, December 2003, which is hereby incorporated by reference.
In addition to increasing the speed of convergence and avoiding stagnation improved deblurring and image recovery are possible. The metric parameter-maximum entropy algorithm or MPME algorithm improves image quality by increasing the contrast of the recovered image. Adjustments to the metric parameter .gamma., particularly to within the range of 0.2 to 0.6 result in a modulation transfer function having a more rectangular form, which preserves contrast of higher spatial frequency components. The effect of the metric parameter γ is also evident on the point-spread function as a reduction in side lobe oscillations apparent in the intermediate image. The final point images are closer to true points with little or no ringing. Disappearance of the oscillating rings also increases contrast.
The MPME algorithm provides an iterative digital deconvolution method capable of starting with any image. An estimate of the next new image can contain a linear combination of directional images. The metric parameter γ modifies the directional images from those provided by conventional maximum entropy algorithms, while reconciling the directional images of the conventional algorithms as integer instances of the metric parameter γ. Preferably, a quadratic Taylor expansion is used to calculate the values of the entropy S and the statistical noise constraint C as functions of the search directions. The modified statistical noise constraint assures an iterative solution of the new image estimate.
The metric parameter-maximum entropy algorithm (MPME) has an important range of applications due to the “box-like” form of the resulting overall modulation transfer function, as shown in
The optional central obscuration 26 of the aperture 28 as apparent in
To illustrate the improved performance introduced by the centrally obscured logarithmic asphere,
A similar simulation using the centrally obscured logarithmic asphere 26 is shown in
The performance improvements, particularly for close-in distances (β-design), is believed due to both the narrower central peak of the average point spread function and the similar oscillating ring structure of the point spread function over the designed object range. These two factors lead to a point spread function that varies less with object distance s, so that the average point spread function used in the digital processing can provide a significantly improved output. Thus, in comparing
To further demonstrate the improved performance of centrally obscured logarithmic asphere, we show in
The improved performance made possible by the invention has particular benefits for photography, which can be observed by comparison from the pictures of
In intensity imaging systems, it is common to characterize their performance by an optical transfer function. Extending this notion to a computational imaging system, in principle, the overall frequency response can be found by dividing the spectrum of the recovered image by that of the input object. Thus, to find the overall frequency response, the images of a point source can be calculated at various object distances, and the maximum entropy algorithm can be applied to these intermediate images to recover the point object. The recoveries can be considered as the combined impulse response of the integrated computational imaging system. A Fourier transform of the recoveries is plotted in
However, the concept of overall transfer function is only an approximate index of the system performance because of the nonlinear digital processing involved. In other words, different overall transfer functions can be expected for various objects. Nonetheless, the transfer function shown in
Thus, in an embodiment of the invention, an imaging system for producing images having an extended depth of field includes a multifocal imaging subsystem that purposefully blurs an intermediate image of the object by introducing a controlled amount of spherical aberration. The spherical aberration provides a more uniform impulse response over a range of focal depths of the multifocal imaging subsystem. In a particular embodiment, third-order spherical aberration is the dominant feature of the purposeful blur. In various embodiments, the spherical aberration can be contributed by a single component/lens element or it can be divided among more than one lens element to increase design flexibility.
According to an embodiment, the multifocal imaging subsystem has a centrally obscured aperture that, among other things, narrows the impulse response. The controlled spherical aberration-induced blur together with the centrally obscured aperture provide a sufficiently narrow and invariant impulse response over an extended depth of focus that achieves diffraction-limited performance over the extended range which exceeds classical limits. The circularly symmetric structure of the imaging subsystem simplifies manufacture and reduces overall costs.
According to an embodiment, the aperture of the multifocal imaging subsystem has a minimum radius, δR, defining an outer limit (radius) of the central obscuration and an inner limit of an annular segment of the aperture, and a maximum radius, R, defining an outer limit of the annular segment and thus the aperture. According to an exemplary embodiment described herein above, a ratio of δR/R greater than or equal to 0.3 provided improved imaging results. As noted herein above, 0≦δR<R, with δR=0 described as a special case. An important consideration is that as δR approaches zero, the light loss due to practical implementation considerations not be excessive; i.e., degrade the image. For small diameter lenses (e.g., R˜1.5 mm), δR/R≧˜⅙ is practical. According to an exemplary embodiment, satisfactory imaging resulted when the phase delays produced within the aperture at δR and R were approximately equal for the center of the designated object range.
Regarding the effects of the central obscuration, each of the intermediate image point spread functions over the extended object range provided by the embodiments of the invention has a central peak and an oscillating ring structure. As mentioned above, the central obscuration provides for narrowing the average point spread function either for the close-in points or the distant object points, depending on the system design. In particular, the central obscuration renders both the widths of the central peaks and the oscillating ring structures more uniform over the range of object distances. Thus the central obscuration tends to remove variant components of the point spread functions produced by the purposeful blur. The observed increase in performance associated with the appropriate central obscuration are believed to be due to the similarities of the point spread functions over the design range of object distances rather than from any direct increase of the depth of field that might otherwise accompany the use of a central obscuration in an ideal imaging system. In particular, the associated improvements in the depth of field, particularly for close-in distances, are believed mainly due to both the narrower central peak of the average point spread function and the similar oscillating ring structures of the point spread functions over the designed object range. These two factors lead to point spread functions that vary less with object distance, so that the average point spread function used in the digital processing can provide a significantly improved output.
According to an exemplary embodiment, at least one lens of the multifocal imaging subsystem can be designed substantially free of spherical aberration, and a phase plate can be designed to produce the spherical aberration that forms the dominant feature of the purposeful blur. The phase plate may be attached to an aperture at a pupil location in the multifocal imaging subsystem.
According to another exemplary embodiment described herein above, an extended depth of field imaging system includes a multifocal imaging subsystem designed as a combination of an ideal imaging component and a spherical aberration component that balances in and out of focus effects through a range of object distances. An intermediate image-detecting device detects the blurred, intermediate images. A computer processing device calculates a recovered image having an extended depth of field based on a correction of the balanced in and out of focus effects through the range of object distances.
As further described above, an embodiment of the invention is directed to a method of designing a multifocal imaging subsystem as a part of an integrated computational imaging subsystem. In an exemplary embodiment, a first component of the multifocal imaging subsystem is designed as an ideal imaging component for imaging an object at a given object distance. A second component of the multifocal imaging subsystem is designed as a spherical aberrator for balancing in and out of focus effects through a range of object distances. Combining the first and second components of the multifocal imaging subsystem produces an intermediate image that is purposefully blurred. The second component of the multifocal imaging subsystem contributes a controlled amount of spherical aberration as the dominant feature of the purposeful blur. Information concerning the intermediate image and the purposeful blur is supplied to a digital processing system for producing a recovered image having an extended depth of field.
Another embodiment of the invention described herein above is directed to a method of recovering an image based on an intermediate image, which includes accessing an intermediate image of a scene and performing an iterative digital deconvolution of the intermediate image using a maximum entropy algorithm. A new image is estimated containing a combination of directional images. These directional images are uniquely altered using a metric parameter to speed convergence toward a recovered image while avoiding points of stagnation. The metric parameter reconciles conventional maximum entropy algorithms at metric parameter values of zero and one. Values of the metric parameter between zero and one advantageously adjust the weight of different pixel values. In an exemplary embodiment, the metric parameter has a value between 0.2 and 0.6. The appropriate choice of the metric parameter contributes to a modulation transfer function having a shape that increases contrast at high spatial frequencies approaching a Nyquist limit. The intermediate image can be produced using a multifocal imaging system, such as an aspheric lens. Typical point spread functions of such lenses have oscillating bases, which reduce image contrast. The metric parameter can be adjusted to significantly reduce side lobe oscillation in the blurred image. The “metric parameter-maximum entropy” algorithm (MPME algorithm) is expected to have wide applicability to digital image processing. The attributes of rapid convergence and the avoidance of stagnation will be advantageous for, among other things, image reconstruction, restoration, filtering, and picture processing.
From the foregoing description, it will be apparent that an improved system, method and apparatus for imaging are provided using a multifocal imaging system in which spherical aberration is the predominate form of blur for intermediate imaging and in which a central obscuration can be used for making more intermediate images more susceptible to correction by digital processing for increasing both resolution and depth of field. Variations and modifications in the herein described system, method, and apparatus will undoubtedly become apparent to those skilled in the art within the overall spirit and scope of the invention.
This application is a continuation application of parent application U.S. Ser. No. 10/908,287 filed on May 5, 2005 now U.S. Pat. No. 7,336,430 and claims priority thereto. This application also claims priority to U.S. Provisional Application No. 60/522,990 filed Nov. 30, 2004 and to U.S. Provisional Application No. 60/607,076 filed Sep. 3, 2004. The disclosures of both provisional applications are hereby incorporated by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
7336430 | George et al. | Feb 2008 | B2 |
Number | Date | Country |
---|---|---|
WO 9314384 | Jul 1993 | WO |
Number | Date | Country | |
---|---|---|---|
20080151388 A1 | Jun 2008 | US |
Number | Date | Country | |
---|---|---|---|
60522990 | Nov 2004 | US | |
60607076 | Sep 2004 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10908287 | May 2005 | US |
Child | 11963033 | US |