Metrics for designing a plenoptic imaging system

Information

  • Patent Grant
  • 9471728
  • Patent Number
    9,471,728
  • Date Filed
    Monday, September 9, 2013
    11 years ago
  • Date Issued
    Tuesday, October 18, 2016
    8 years ago
Abstract
Metrics for characterizing the focusing of a plenoptic imaging system. In one aspect, the metric is based on the high frequency content and/or the blurring of the plenoptic image.
Description
BACKGROUND OF THE INVENTION

1. Field of the Invention


This invention relates generally to plenoptic imaging systems, and more particularly to designing plenoptic imaging systems.


2. Description of the Related Art


The plenoptic imaging system has recently received increased attention. It can be used to recalculate a different focus point or point of view of an object, based on digital processing of the captured plenoptic image. The plenoptic system also finds application in multi-modal imaging, using a multi-modal filter array in the plane of the pupil aperture. Each filter is imaged at the sensor, effectively producing a multiplexed image of the object for each imaging modality at the filter plane. Other applications for plenoptic imaging systems include varying depth of field imaging and high dynamic range imaging.


However, there is limited literature on the optical design of a plenoptic imaging system that goes beyond geometric optics considerations. Thus far, most of the literature about these systems is concentrated in the areas of computer vision, graphics and digital processing algorithms—how to further process the “light field” properties of the data captured at the sensor of a plenoptic system. These algorithms focus on what to do with the plenoptic images after they are captured. In addition, many of those algorithms are not specific to plenoptic imaging systems; they can be directly transferred to completely different configurations, such as multi-lens imaging systems (multiple cameras mounted adjacent to one another). This body of work assumes that the plenoptic imaging system has been well designed, manufactured and assembled. However, there is not much literature explaining how to actually design, manufacture or assemble plenoptic imaging systems.


As one example, consider the “focusing” a plenoptic imaging system. Much is known about how to focus a conventional imaging system, in order to produce a sharp image of the object at the sensor. However, a plenoptic imaging system that is “in focus” does not produce a sharp image of the object at the sensor array. In a plenoptic imaging system, the sensor array is not placed at a plane conjugate to the image plane. Therefore, conventional focusing criteria are not appropriate for focusing a plenoptic system. In order to design a plenoptic system, including diffraction effects, it is desirable to obtain an appropriate evaluation metric that can be used to gauge the quality of the system.


Thus, there is need for such criteria.


SUMMARY OF THE INVENTION

The present invention overcomes the limitations of the prior art by providing approaches to adjust the axial spacing (i.e., “defocus”) of components in a plenoptic imaging system based on the plenoptic image captured by the sensor array.


A plenoptic imaging system includes a primary imaging subsystem, a secondary imaging array, and a sensor array. A plenoptic imaging system that is “in focus” will not produce a sharp image of the object at the sensor array. Thus, focusing metrics different from those used in conventional imaging systems are required. In one approach, the axial position of the secondary imaging array relative to the primary imaging subsystem is adjusted based on increasing a high frequency content in the plenoptic image captured by the sensor array. In different variations, this is achieved by using metrics based on the frequency of ringing in the plenoptic image, on the number of ripples caused by ringing in the plenoptic image, or on the flatness or contrast of ringing in the plenoptic image.


Furthermore, the axial position of the sensor array relative to the secondary imaging array can be adjusted based on reducing a blur in a pupil image function of the plenoptic imaging system. In addition, the distance to the object can be estimated once the in-focus positions of the secondary imaging array and/or sensor array have been determined.


These principles can also be used in other ways. For example, they can be used as the basis for a plenoptic imaging system with autofocus capability. Alternately, they can be used to retrofit a plenoptic sensor module (i.e., secondary imaging array plus sensor array) onto an existing imaging system. As a final example, merit functions based on the plenoptic image can be used in the design process for the plenoptic imaging system.


Other aspects of the invention include methods, devices, systems and applications related to the principles described above.





BRIEF DESCRIPTION OF THE DRAWINGS

The invention has other advantages and features which will be more readily apparent from the following detailed description of the invention and the appended claims, when taken in conjunction with the accompanying drawings, in which:



FIG. 1 is a simplified diagram of a plenoptic imaging system.



FIGS. 2a-2d are simulated images showing the response of an in-focus system.



FIGS. 3a-3d are simulated images showing the response of a system with defocus in


Z2.



FIGS. 4a-4d are simulated images showing the response of a system with defocus in z3.



FIG. 5 is a flow diagram for aligning a plenoptic imaging system.



FIG. 6a is a plot of flatness as a function of z2.



FIG. 6b is a plot of the number of low frequency ripples as a function of z2.



FIG. 7a is a plot of edge sharpness as a function of z3.



FIG. 7b is a plot of edge width as a function of z3.



FIG. 8 is a diagram of a plenoptic imaging system with autofocus.



FIG. 9a (prior art) and FIG. 9b are diagrams of a plenoptic sensor module retrofitted to an existing imaging system.



FIG. 10 is a flow diagram for designing a plenoptic imaging system.





The figures depict embodiments of the present invention for purposes of illustration only. One skilled in the art will readily recognize from the following discussion that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles of the invention described herein.


DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

The figures and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.


Sample Configuration.



FIG. 1 is a simplified diagram of a plenoptic imaging system. The system includes a primary imaging subsystem 110 (represented by a single lens in FIG. 1), a secondary imaging array 120 (represented by a lenslet array) and a sensor array 130. These form two overlapping imaging subsystems, referred to as subsystem 1 and subsystem 2 in FIG. 1.


The spatial coordinates (ξ, η) will be used at the object plane, (x, y) at the pupil plane for imaging subsystem 1, (u, v) at the plane of the lenslet array, and (t, w) at the sensor plane. The primary lens is a distance z1 from the object, the lenslet array is a distance z2 from the primary lens, and the sensor is a distance z3 from the lenslet array. In FIG. 1, the different components are each located at a single plane so the distances between components is easy to define. However, in other systems, the different components may be more complex (e.g., the primary “lens” may be a set of lenses), but the concept of the distances z1, z2, z3 can be generalized using known principles. The focal lengths of the primary lens and the lenslet array are represented by f1 and f2 and their diameters are D1 and D2, respectively. In addition, the different components do not have to be designed exactly as shown in FIG. 1. For example, the “primary lens” could be various combinations of elements, including lenses, mirrors and combinations of the two. Similarly, the secondary imaging array could be a pinhole array, or a reflective array.


In imaging subsystem 1, the object 150 is imaged by the primary lens 110 to produce an image that will be referred to as the “primary image.” This primary lens 110 may be a camera imaging lens, microscope objective lens or any other such imaging system. The lenslet array 120 is placed approximately at the location of the primary image. Each lenslet then images the pupil of the primary lens to the sensor plane. This is imaging subsystem 2, which partially overlaps with imaging subsystem 1. The image created at the sensor array 130 will be referred to as the “plenoptic image” in order to avoid confusion with the “primary image.” The plenoptic image can be divided into an array of subimages, corresponding to each of the lenslets. Note, however, that the subimages are images of the pupil of imaging subsystem 1, and not of the object 150.


Wave Propagation Analysis.


Now analyze this system based on wave propagation. Consider the first imaging subsystem, where the primary lens has a generalized pupil function P1. Its corresponding impulse response is given by












h
1



(

u
,

v
;
ξ

,
η

)


=





j






kz
1





λ
2



z
1



z
2





exp


[



j





k


2


z
2





(


u
2

+

v
2


)


]




exp


[



j





k


2






z
1





(


ξ
2

+

η
2


)


]










x






yP
1



(

x
,
y

)





exp


[



j





k

2



(


1

z
1


+

1

z
2


-

1

f
1



)



(


x
2

+

y
2


)


]



exp


{

-



j





k


z
2




[



(

u
-

M





ξ


)


x

+


(

v
-

M





η


)


y


]



}






,




(
1
)








where λ is the wavelength of imaging, k=2π/λ and the magnification from object to image plane is given by M=−z2/z1. Substituting, x′=x/λz2 and y′=y/λz2 in the integral in Eq. (1) yields











λ
2



z
1



z
2










x







y






P
1



(



x



λ






z
2


,


y



λ






z
2



)



exp


{



j





k

2




(


1

z
1


+

1

z
2


-

1

f
1



)



[



(


x



λ






z
2


)

2

+


(


y



λ






z
2


)

2


]



}


exp


{


-
j






2


π


[



(

u
-

M





ξ


)



x



+


(

v
-

M





η


)



y




]



}





=




λ
2



z
1



z
2



FT


(



P
1



(



x



λ






z
2


,


y



λ






z
2



)



exp


{



j





k

2




(


1

z
1


+

1

z
2


-

1

f
1



)



[



(


x



λ






z
2


)

2

+


(


y



λ






z
2


)

2


]



}


)





|


fx
=

(

u
-

M





ξ


)



fy
=

(

v
-

M





η


)





=


λ
2



z
1



z
2




h
1




(


u
-

M





ξ


,

v
-

M





η



)








(
2
)








where we have defined the term h1′ for convenience as











h
1




(

u
,
v

)


=

FT


(



P
1



(


x





λ






z
2


,

y





λ






z
2



)



exp


{



j





k

2




(


1

z
1


+

1

z
2


-

1

f
1



)



[



(

x





λ






z
2


)

2

+


(

y





λ






z
2


)

2


]



}


)






(
3
)








This reduces Eq. (1) to











h
1



(

u
,

v
;
ξ

,
η

)


=




j






kz
1





exp


[



j





k


2






z
2





(


u
2

+

v
2


)


]




exp


[



j





k


2






z
1





(


ξ

2
+




η
2


)


]






h
1




(


u
-

M





ξ


,

v
-

M





η



)


.






(
4
)







An object having a complex field Uo then produces an image at the plane of the lenslet array, given by,

Ui(u, v)=∫∫dξd ηh1(u,v;ξ,η)Uo(ξ,η)  (5)

Using Eqs. (4) and (5) and substituting ξ′=Mξ and η′=Mη yields











U
i



(

u
,
v

)


=





j





k






z
1




M
2




exp


[



j





k


2






z
2





(


u
2

+

v
2


)


]











ξ







η






U
o



(



ξ


M

,


η


M


)



exp


{



j





k


2






z
1





[



(


ξ


M

)

2

+


(


η


M

)

2


]


}





h
1




(


u
-

ξ



,

v
-

η




)


.









(
6
)







The above equation can be written as a convolution (indicated by the symbol *):











U
i



(

u
,
v

)


=





j






k
1




M
2




exp


[



j





k


2






z
2





(


u
2

+

v
2


)


]





{


[



U
o



(


u
M

,

v
M


)



exp


{



j





k


2






z
1





[



(

u
M

)

2

+


(

v
M

)

2


]


}


]

*


h
1




(

u
,
v

)



}

.






(
7
)








This yields the field of the first image formed by the primary lens at the plane of the lenslet array.


We assume each lenslet has a diameter D2, focal length f2, pupil function given by P2 and there are M′×N′ such lenslets in the array. Then the amplitude distribution of the field Ui after the lenslet array may be written as











U
i




(

u
,
v

)


=



U
i



(

u
,
v

)







m


M









n


N








P
2



(


u
-

m






D
2



,

v
-

n






D
2




)



exp



{




-
j






k


2






f
2





[



(

u
-

m






D
2



)

2

+


(

v
-

n






D
2



)

2


]


}

.









(
8
)







Using the Fresnel transform to propagate this field to the sensor, which is located a distance z3 from the lenslet array, yields











U
f



(

t
,
w

)


=





j






k
3




j





λ






z
3





exp


[



j





k


2






z
3





(


t
2

+

w
2


)


]







m


M









n


N












u




v








U
i



(

u
,
v

)









P
2



(


u
-

m






D
2



,

v
-

n






D
2




)



exp


{




-
j






k


2






f
2





[



(

u
-

m






D
2



)

2

+


(

v
-

n






D
2



)

2


]


}



exp


[



j





k


2






z
3





(


u
2

+

v
2


)


]





exp


[




-
j






k


z
3




(

ut
+
vw

)


]


.











(
9
)








Further simplification gives











U
f



(

t
,
w

)


=






j






k
3




j





λ






z
3





exp


[



j





k


2






z
3





(


t
2

+

w
2


)


]







m


M









n


N






exp


{




-
j






k


2






f
2





[



(

m






D
2


)

2

+


(

n






D
2


)

2


]


}


FT


{



U
i



(

u
,
v

)





P
2



(


u
-

m






D
2



,

v
-

n






D
2




)




exp


[



j





k

2



(


1

z
3


-

1

f
2



)



(


u
2

+

v
2


)


]



}







|



f
x

=


t

λ






z
3






m






D
2



λ






f
2







f
y

=


w

λ






z
3






n






D
2



λ






f
2












(
10
)








Using Eq. (10), the plenoptic image captured at the sensor is given by,











I
f



(

t
,
w

)


=









j






k
3




j





λ






z
3





exp


[



j





k


2






z
3





(


t
2

+

w
2


)


]







m


M









n


N






exp


{




-
j






k


2






f
2





[



(

m






D
2


)

2

+


(

n






D
2


)

2


]


}


FT


{



U
i



(

u
,
v

)





P
2



(


u
-

m






D
2



,

v
-

n






D
2




)




exp


[



j





k

2



(


1

z
3


-

1

f
2



)



(


u
2

+

v
2


)


]



}







|



f
x

=


t

λ






z
3






m






D
2



λ






f
2







f
y

=


w

λ






z
3






n






D
2



λ






f
2












2
.






(
11
)







Simulation Results.


We simulated this analysis in Matlab for a general non-fluorescent, point object and tested the effect of varying the axial distances in the system. The focal lengths of the primary lens and the lenslet array were f1=51 mm and f2=5 mm. The primary lens aperture is a pentagon inscribed in a circle of diameter D1=10 mm. The pentagon shape was selected so that pupil effects and object effects would be more evident. The object is at an object distance z1=500 mm from the primary lens. Each lenslet in the multi-lenslet array is circular with a diameter D2=0.5 mm. The ideal distance of the lenslet array from the plane of the primary lens was calculated to be 56.8 mm, based on z2=1/(1/f1−1/z1). The ideal distance for the sensor from the lenslet array is 5.5 mm, given by z3=1/(1/ f2−1/z2).


For the sake of brevity, we display simulation results for only the on-axis lenslet and on-axis object point. However, the analysis and simulations remain valid for off-axis object points and lenslets. For example, an on-axis object point will produce a primary image that, if in focus, will produce a focused spot on the on-axis lenslet. A similar situation can be produced for all (or a selected subset) of the lenslets in the array by using an object that is an array of point sources, each point source imaged to a corresponding lenslet.


In focus system. FIGS. 2a-2d are simulated images showing the response of this system when it is in focus. Under these conditions, the first imaging subsystem is in perfect focus. Hence the on-axis point object is imaged to a tightly focused spot at the plane of the lenslet array, which corresponds to the impulse response of the first imaging subsystem. It is then truncated by the finite extent of the on-axis lenslet. The truncated impulse response is shown in FIG. 2a. The impulse is displayed as |Ui|0.1 in FIG. 2a in order to emphasize its dim details. The five streaks radiating from the center of the impulse response correspond to the Fourier transform of the five-sided (pentagon) pupil of the primary lens in our simulations. The pentagon shape was selected purely for illustrative purposes. The circular boundary corresponds to the edge of the circular lenslet.


This field Ui′ is then propagated to and captured at the sensor as intensity If, given in Eq. (11). FIGS. 2b-2d show one subimage of the plenoptic image. Since the object is a point source, the subimage is an image of the pentagon shaped pupil of the primary lens. FIG. 2b displays If0.1 and FIG. 2c shows a magnified portion of this. FIG. 2d shows the same magnified portion as FIG. 2c, but If2.5 is displayed in order to show the fine rings at the edges due to the truncation. For convenience, we will refer to this image of the pupil, produced by each lenslet for a point object, as the pupil image function (PIF).


In the ideal case where z1, z2 and z3 are in focus, there are fine fringes in the interior of the pupil image function, and the edges have a small sharp overshoot as seen in FIG. 2d. However, overall, the edges of the PIF are sharp and the interior is relatively flat.


Defocus of z2. Now consider the effect of a defocus error in z2. Assuming accurate z1 and z3, we now test the effect of inaccuracy in the value of z2. We retained z1=500 mm and z3=5.5 mm. But instead of 56.8 mm, we use z2=54 mm. FIGS. 3a-3d are simulated images showing the response of the system with defocus in z2.



FIG. 3a shows the impulse response of the first imaging subsystem, as truncated by the lenslet. As shown in FIG. 3a, defocus in the value of z2 implies that the lenslet array is placed at an axial position where the impulse response of the first imaging subsystem contains some defocus. Therefore, the lenslet array captures and truncates this defocused impulse response of the primary lens. This is further propagated to the sensor and captured as the pupil image function shown in FIGS. 3b-3d. FIGS. 3b and 3c display If2.5 and If0.1, respectively. FIG. 3d is a magnified view of FIG. 3c. The PIF shows ringing in the interior (compare FIGS. 3b and 2b), but FIGS. 3c-3d show that the edges of the pentagon shaped image of the pupil of the primary lens remain relatively sharp. This is probably because, despite the defocus in the first imaging subsystem, the second imaging subsystem which images the pupil of the primary lens at the sensor has greater tolerance to variation in z2 and remains reasonably focused.


Defocus of z3. Now consider the effect of defocus error in z3. Assuming accurate z1 and z2, we now test the effect of inaccuracy in the value of z3. We retained z1=500 mm and z2=56.8 mm. But instead of 5.5 mm, we use z3=5 mm. FIGS. 4a-4d are simulated images showing the response of the system with defocus in z3.



FIG. 4a shows the impulse response of the first imaging subsystem. As expected, inaccuracy in z3 has no influence on the first imaging subsystem. Therefore, the point object is imaged accurately on the lenslet array. However, defocus in z3 creates defocus in the second imaging subsystem. FIGS. 4b and 4c display If2.5 and If0.1, respectively. FIG. 4d is a magnified view of FIG. 4c. FIGS. 4c-4d show that the edges of the pentagon shaped pupil image are now blurred, even though the interior of the PIF is relatively flat as shown in FIG. 4b.


Some Applications.


As shown above, blur in the edges of the PIF is an indicator of defocus in z3 and ringing in the interior of the PIF (more generally, the amount of high frequency content in the PIF) is an indicator of defocus in z2. These indicators can be used in a number of different ways. For example, they can be used to help determine the correct axial spacing of the components within a plenoptic imaging system. Depending on which characteristic is observed, either the distance between the primary imaging subsystem and the secondary imaging array (z2) and/or the distance between the secondary imaging array and the sensor array (z3) may be adjusted. This alignment may occur during assembly of the plenoptic imaging system, or it may be implemented as an autofocus or manual focus adjustment of the plenoptic imaging system in the field. Alignment may also occur when an existing imaging system is converted to a plenoptic imaging system by replacing the existing sensor array with a plenoptic sensor module (i.e., secondary imaging array plus sensor array). In that case, the plenoptic sensor module must be aligned to the primary imaging optics in the existing system.


Another use is during the system design process. It is difficult to design a plenoptic imaging system using conventional lens design software because there are not well-defined metrics that capture the quality of a plenoptic imaging system. However, metrics based on the characteristics above can now be used to optimize the plenoptic imaging system with respect to defocus. As a final example application, the characteristics above can be used in an inverse fashion to estimate the object distance z1. If the optimal distances z2 and z3 are known (e.g., by optimizing the blur and high frequency characteristics), then this can be used to estimate object distance z1.


Possible Metrics.


Flatness / high frequency content metrics. Defocus in z2 (i.e., separation of primary imaging subsystem and secondary imaging array) is characterized by ringing in the interior of the PIF. More generally, it is characterized by the amount of high frequency content in the PIF. As z2 comes into focus, the ringing becomes higher frequency and less modulated (i.e., there are more ridges but with a lower contrast, so the interior becomes flatter). This is because the high frequency content in the PIF is increasing. Different metrics can be used to measure these characteristics. For example, metrics can be based on slope analysis for the detection of maxima or minima, counting the number of peaks and valleys of the ringing (i.e., local maxima and minima), spatial frequency analysis (e.g., percent energy contained above a certain frequency, or distribution of energy across frequencies) or flatness of the ringing (e.g., modulation divided by DC level).


Two examples of automated detection of this criterion use image processing functions in Matlab. One example metric counts the peaks in the interior of the PIF with the function “imregionalmax” in Matlab. This function detects the peaks in a local neighborhood. We sum the total number of peaks detected inside the PIF. When the interior of the PIF becomes flatter, the local peaks (ripples) in it become finer and increase in number. Therefore decreasing defocus in z2 should be characterized by an increasing number of local peaks.


Another example metric is based on the Matlab function “edge” with the option of using the “canny” edge detector. This function is less sensitive to fine local variations and detects longer, continuous edges inside the PIF (unlike imregionalmax where the local peaks may be single pixels). As the PIF becomes flatter, the number of these low frequency peak-to-valley ripples decreases since the PIF is dominated more by the fine local variations. Therefore the canny edge detector finds fewer edges as the defocus of z2 decreases.


Blur metric: The blur around the edges of the PIF can be detected, for example, by using sharpness or contrast detection algorithms. One example is based on taking a derivative of the PIF along either the x or y dimension (rows or columns). The PIF whose derivative shows the maximum slope contains the sharpest edges and should indicate the optimal position of z3. In simulations, we used the Matlab function “diff(PIF,1)” to obtain the derivative. The peak-to-valley difference in diff(PIF,1) is approximated to be the greatest slope for that PIF. This slope should increase as the defocus in z3 decreases.


Another blur metric is based on thresholding the edge of the PIF and counting all the pixels that have a value above the selected threshold. As the blur reduces, the number of pixels in the edges should decrease.


Note that this phenomenon has been described using the word “blur,” but it should be understood that metrics for sharpness, contrast, etc. may also be measures of blur.


Example Automated Alignment



FIG. 5 is a flow diagram for aligning a plenoptic imaging system, based on the principles described above. In this example, it is assumed that the object is at an unknown object distance zi, so that it is not possible to accurately calculate z2 and z3. Instead, z2 and z3 are adjusted as follows. As an initial setting, z3 is set 510 to f2. Then z2 is adjusted 520 based on increasing the high frequency content in the plenoptic image captured by the sensor array. Then z3 is adjusted 530 based on reducing blur in the PIF. Steps 520 and 530 may be iterated, or interchanged. Alternately, z2 and z3 may be adjusted simultaneously based on a metric that accounts for both the high frequency and blur characteristics. More details are given below, using the example plenoptic imaging system introduced previously.


In step 510, the sensor array is moved to the back focal plane of the lenslet array, i.e. z3=f2=5 mm. This can be done experimentally using collimated light and using the lenslet array to focus it to a tight spot at the sensor.


In step 520, we move 522 the lenslet array and sensor array in concert, so that z2 changes while z3 remains constant. The sensor array captures 525 the plenoptic image at different values of z2. A flatness metric is calculated 526 at these different values. In this example, z1=500 mm and z3=f2=5 mm. FIG. 6a plots the number of pixels in the local peaks detected by imregionalmax for z2˜54 mm to 58 mm. This plot shows a maximum at z2=56.8 mm. FIG. 6b plots the number of pixels in the edges detected by the canny edge detector for z2˜54 mm to 58 mm. It shows a minimum, also at z2=56.8 mm. These curves for two different metrics both show 527 that the axial position z2=56.8 mm for the lenslet array gives the flattest interior for the PIF. Thus, the lenslet array is set 529 to this location. So z1=500 mm and z2=56.8 mm.


In step 530, we move 532 the sensor so that z3 changes. A similar procedure to step 520 is used. The sensor array captures 535 the plenoptic image at different values of z3. The blur metric is calculated 536 at these different values. FIG. 7a plots the max-to-min difference in the derivative of the PIF (along the rows), obtained using the function diff(PIF,1) in Matlab. The plot shows a maximum for z3=5.5 mm. FIG. 7b plots the number of pixels in the thresholded edge blur for z3=5.1 mm to 5.8 mm. It also shows a minimum at z3=5.5 mm. These curves for two different blur metrics both show 537 that the axial position z3=5.5 mm for the sensor array reduces the blur the most. Thus, the sensor array is set 539 to this location.


Note that FIGS. 6-7 each show plots for three different lenslets. Recall that the lenslet array contains multiple lenslets, each of which produces a corresponding subimage within the plenoptic image. This array nature can be addressed in different ways. For example, the various metrics (blur, flatness, high frequency content, etc.) can be calculated for each subimage and these metrics combined into a single metric for optimization. Two common methods for combining are based on averaging (including weighted averaging), or on taking the worst case. Also, the metric can be calculated for all subimages, or for just a subset or sampling of the subimages. Different wavelengths in the plenoptic image can be handled in a similar manner. If the plenoptic imaging system has a fixed position for a range of object distances, then the variation in object distance can also be handled in a similar manner (e.g., by taking some combination of the performances at different object distances). Plenoptic imaging systems are commonly used to effect different filtering for different subimages by placing corresponding filters in the pupil. For example, different wavelength filters may be used at different locations in the pupil. The performance over different subimages can also be handled in a similar manner.


More Applications


If the plenoptic imaging system is adjustable, then the principles described above can be implemented in an autofocus mechanism. FIG. 8 is a block diagram of a plenoptic imaging system with autofocus. The autofocus mechanism 850 receives the plenoptic image captured by the sensor array 130 and, based on the captured image, adjusts the distances z2 and/or z3. In one approach, the distances are adjusted based on an algorithm such as described in FIG. 5.


In other applications, z1 may be varied in addition to or in place of z2 and/or z3. For example, if the distance from secondary imaging array to sensor array is fixed (i.e., fixed z3), it may be possible to focus the plenoptic imaging system by adjusting the object distance z1.


In another approach, a lookup table lists z3 values that correspond to different values of z2. The distance z2 is determined based on the captured plenoptic image, and distance z3 is determined based on the lookup table.


As another variation, once z2 and/or z3 are determined, they can be used to estimate the object distance z1. The look-up table may be stored in software in hardware/firmare on a chip.


In yet another approach, the plenoptic imaging system is focused, either manually or automatically, by use of a single adjustment. For example, instead of providing one adjustment for z2 and a second adjustment for z3, z2 and z3 may be coupled and varied together through one adjustment. This coupling may be based on the principles described above. A similar approach can be used for cases where other distances from among z1, z2 and z3 are adjustable.


Similar approaches may be used when an existing imaging system is converted to a plenoptic imaging system, as shown in FIGS. 9a-9b. FIG. 9a shows the existing imaging system, which includes a primary lens 910 and a sensor array 932 that captures the conventional primary image. As shown in FIG. 9b, the conventional sensor array 932 is replaced by a plenoptic sensor module 920, which includes a secondary imaging array 120 and sensor array 130. Retrofit of the plenoptic sensor module includes alignment of the module with the primary imaging subsystem 910, as described above.


In another aspect, the principles described can also be used during the design phase, as shown in the flow diagram of FIG. 10. The pupil image function for a candidate design 1010 is calculated 1020. The PIF is used to calculate a high frequency content metric 1030 and/or a blur metric 1040. If the metrics are sufficiently good, then the lens design is completed 1050. Otherwise, the design is iterated 1060. Other approaches can also be used. For example, different metrics or combinations of metrics could be used. Only the high frequency content metric or only the blur metric could be used, or a metric that combines the two or combines these metrics with other metrics (such as rms wavefront error or spot size) could be used. As another alternative, the metrics could be evaluated without directly calculating the pupil image function.


Although the detailed description contains many specifics, these should not be construed as limiting the scope of the invention but merely as illustrating different examples and aspects of the invention. It should be appreciated that the scope of the invention includes other embodiments not discussed in detail above. Various other modifications, changes and variations which will be apparent to those skilled in the art may be made in the arrangement, operation and details of the method and apparatus of the present invention disclosed herein without departing from the spirit and scope of the invention as defined in the appended claims. Therefore, the scope of the invention should be determined by the appended claims and their legal equivalents.

Claims
  • 1. A computer-implemented method for automatically designing a plenoptic imaging system, comprising: accessing a candidate design for the plenoptic imaging system, the plenoptic imaging system comprising a primary imaging subsystem, a secondary imaging array, and a location for a sensor array;simulating a plenoptic image formed by the candidate design;calculating a metric that is a function of the high frequency content in the plenoptic image; andadjusting the candidate design, based on improving the metric in a manner that increases the high frequency content in the plenoptic image; wherein adjusting the candidate design comprises adjusting a distance z2 to position the secondary imaging array at the primary image location of the primary imaging subsystem, wherein z2 is a distance between the primary imaging subsystem and the secondary imaging array, said z2 adjustment based on improving the metric in a manner that increases the high frequency content in the plenoptic image.
  • 2. The computer-implemented method of claim 1 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a frequency of ringing in the plenoptic image.
  • 3. The computer-implemented method of claim 1 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a number of ripples caused by ringing in the plenoptic image.
  • 4. The computer-implemented method of claim 1 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a flatness of ringing in the plenoptic image.
  • 5. The computer-implemented method of claim 1 wherein the metric accounts for high frequency content in a plurality of subimages within the plenoptic image.
  • 6. The computer-implemented method of claim 1 wherein the metric accounts for high frequency content in a plurality of wavelengths within the plenoptic image.
  • 7. The computer-implemented method of claim 1 wherein the plenoptic imaging system includes different wavelength filters for different subimages within the plenoptic image.
  • 8. The computer-implemented method of claim 1 wherein simulating the plenoptic image comprises simulating a plenoptic image formed by an array of point objects.
  • 9. A computer-implemented method for automatically designing a plenoptic imaging system, comprising: accessing a candidate design for the plenoptic imaging system, the plenoptic imaging system comprising a primary imaging subsystem, a secondary imaging array, and a location for a sensor array;simulating a plenoptic image formed by the candidate design;calculating a metric that is a function of the high frequency content in the plenoptic image; andadjusting the candidate design, based on improving the metric in a manner that increases the high frequency content in the plenoptic image; wherein the metric is also a function of a blur in a pupil image function of the plenoptic imaging system; and adjusting the candidate design further comprises adjusting a distance z3 to position the location of the sensor array at the pupil image location of the secondary imaging array, wherein z3 is a distance between the secondary imaging array and the sensor array, said z3 adjustment based on improving the metric in a manner that reduces the blur in the pupil image function of the plenoptic imaging system.
  • 10. The computer-implemented method of claim 9 wherein the plenoptic imaging system includes different wavelength filters for different subimages within the plenoptic image.
  • 11. The computer-implemented method of claim 9 wherein simulating the plenoptic image comprises simulating a plenoptic image formed by an array of point objects.
  • 12. A non-transitory computer readable medium containing computer readable instructions that, when executed by a computer, cause the computer to perform a method for automatically designing a plenoptic imaging system, the method comprising: accessing a candidate design for the plenoptic imaging system, the plenoptic imaging system comprising a primary imaging subsystem, a secondary imaging array, and a location for a sensor array;simulating a plenoptic image formed by the candidate design;calculating a metric that is a function of the high frequency content in the plenoptic image; andadjusting the candidate design, based on improving the metric in a manner that increases the high frequency content in the plenoptic image; wherein adjusting the candidate design comprises adjusting a distance z2 to position the secondary imaging array at the primary image location of the primary imaging subsystem, wherein z2 is a distance between the primary imaging subsystem and the secondary imaging array, said z2 adjustment based on improving the metric in a manner that increases the high frequency content in the plenoptic image.
  • 13. The computer readable medium of claim 12 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a frequency of ringing in the plenoptic image.
  • 14. The computer readable medium of claim 12 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a number of ripples caused by ringing in the plenoptic image.
  • 15. The computer readable medium of claim 12 wherein adjusting the distance z2 is based on improving the metric in a manner that increases a flatness of ringing in the plenoptic image.
  • 16. The computer readable medium of claim 12 wherein the metric accounts for high frequency content in a plurality of subimages within the plenoptic image.
  • 17. The computer readable medium of claim 12 wherein the metric accounts for high frequency content in a plurality of wavelengths within the plenoptic image.
  • 18. The computer readable medium of claim 12 wherein the plenoptic imaging system includes different wavelength filters for different subimages within the plenoptic image.
  • 19. The computer readable medium of claim 12 wherein: the metric is also a function of a blur in a pupil image function of the plenoptic imaging system; andadjusting the candidate design further comprises adjusting a distance z3 to position the location of the sensor array at the pupil image location of the secondary imaging array, wherein z3 is a distance between the secondary imaging array and the location of the sensor array, said z3 adjustment based on improving the metric in a manner that reduces the blur in the pupil image function of the plenoptic imaging system.
  • 20. The computer readable medium of claim 12 wherein simulating a plenoptic image comprises simulating a plenoptic image formed by an array of point objects.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This application is a division of U.S. patent application Ser. No. 13/113,952, “Focusing and Focus Metrics for a Plenoptic Imaging System,” filed May 23, 2011 and issued on Sep. 10, 2013 and U.S. Pat. No. 8,531,581. The subject matter of the foregoing is incorporated herein by reference in its entirety.

US Referenced Citations (46)
Number Name Date Kind
5629765 Schmutz May 1997 A
5659420 Wakai et al. Aug 1997 A
5729011 Sekiguchi Mar 1998 A
6838650 Toh Jan 2005 B1
7620309 Georgiev Nov 2009 B2
7723662 Levoy et al. May 2010 B2
7732744 Utagawa Jun 2010 B2
7949252 Georgiev May 2011 B1
7956924 Georgiev Jun 2011 B2
7962033 Georgiev et al. Jun 2011 B2
7965936 Raskar et al. Jun 2011 B2
8189065 Georgiev et al. May 2012 B2
8189089 Georgiev et al. May 2012 B1
8228417 Georgiev et al. Jul 2012 B1
8288704 Perlman et al. Oct 2012 B2
8290358 Georgiev Oct 2012 B1
20030231255 Szajewski et al. Dec 2003 A1
20070091197 Okayama et al. Apr 2007 A1
20070252074 Ng et al. Nov 2007 A1
20080131019 Ng Jun 2008 A1
20080152215 Horie et al. Jun 2008 A1
20080165270 Watanabe et al. Jul 2008 A1
20080187305 Raskar et al. Aug 2008 A1
20080193026 Horie et al. Aug 2008 A1
20090041381 Georgiev et al. Feb 2009 A1
20090041448 Georgiev Feb 2009 A1
20090140131 Utagawa Jun 2009 A1
20090185801 Georgiev et al. Jul 2009 A1
20090295829 Georgiev et al. Dec 2009 A1
20100026852 Ng et al. Feb 2010 A1
20100085468 Park et al. Apr 2010 A1
20100141802 Knight et al. Jun 2010 A1
20100194926 Kang et al. Aug 2010 A1
20100265386 Raskar et al. Oct 2010 A1
20100277629 Tanaka Nov 2010 A1
20110129165 Lim et al. Jun 2011 A1
20110149125 Morimoto Jun 2011 A1
20110199458 Hayasaka et al. Aug 2011 A1
20120050562 Perwass et al. Mar 2012 A1
20120050589 Ueno et al. Mar 2012 A1
20120127351 Vlutters et al. May 2012 A1
20120177356 Georgiev et al. Jul 2012 A1
20120229679 Georgiev et al. Sep 2012 A1
20120249819 Imai Oct 2012 A1
20130057749 Hiasa et al. Mar 2013 A1
20130128069 Georgiev May 2013 A1
Foreign Referenced Citations (3)
Number Date Country
2009-229125 Oct 2009 JP
2010-178333 Aug 2010 JP
2011-027901 Feb 2011 JP
Non-Patent Literature Citations (3)
Entry
European Extended Search Report, European Application No. 12164663.2, Sep. 23, 2013, 7 pages.
U.S. Appl. No. 13/113,952, Oct. 26, 2012, 25 pages.
Japanese Office Action, Japanese Application No. 2012-116519, Mar. 29, 2016, 9 pages (with concise explanation of relevance).
Related Publications (1)
Number Date Country
20140019098 A1 Jan 2014 US
Divisions (1)
Number Date Country
Parent 13113952 May 2011 US
Child 14022074 US