Embodiments disclosed herein relate in general to spectral imaging (SI) apparatus, systems and methods and in particular to snapshot (“single shot”) hyperspectral imaging (HSI) using digital cameras and compressed sensing (CS)-based spatial-spectral cube reconstruction (SCR) algorithms.
“Spectral imaging” is commonly understood as referring to imaging with a limited number of wavelengths (e.g. up to 10) in a given wavelength range, e.g. the visible (“VIS”) range or the near-infrared (“NIR”) range. “Hyperspectral imaging” is commonly understood as referring to imaging with a larger number of wavelengths in a given wavelength range, for example between 10 and hundreds or even thousands of wavelengths. “Snapshot” SI and HSI imagers perform simultaneous (instantaneous) acquisition of spatial and spectral data in a single snapshot. The data acquired forms a “spatial-spectral cube” (also referred to herein simply as “spectral cube” or “data cube”) of a source object (also referred to simply as “object” or “scene”). “Spatial-spectral cube”, “spectral cube” and “data cube” are hereinafter used interchangeably. A data cube includes light intensity data in two spatial dimensions and one spectral dimension and is expressed as a three-dimensional (3D) matrix.
Commonly authored and assigned U.S. patent application Ser. No. 13/752,560 titled “Snapshot spectral imaging based on digital cameras” (published as US Pat. Pub. 20130194481), which is incorporated herein by reference in its entirety, teaches compressed sensing (CS)-based snapshot spectral imaging (CS-SSI) in apparatus including an imaging lens, a dispersed image sensor and a restricted isometry property (RIP) diffuser inserted in the optical path between a source image and a pixelated (as. e.g. in a digital camera) image sensor. The RIP diffuser may be one dimensional (1D). It provides a dispersed and diffused image (“DD image”) at the dispersed image sensor. Due to the 1D RIP diffuser optical properties, each pixel in the DD image includes a linear mixture of spectral and spatial information from all pixels of a corresponding column in the DD image. In US 20130194481, full reconstruction of the data cube is performed using a CS-based optimization process to compensate for the underdetermined nature of the problem. The operator performing the linear projections can be described as a “sensing matrix” that has fewer rows than columns and that operates on the data cube to form a DD image. The reconstruction process guarantees full reconstruction of the source object if the sensing matrix satisfies a RIP condition. The RIP diffuser is designed such that the transfer-function (which is identical with the sensing matrix) of an optical imaging system including the diffuser satisfies the RIP condition at each single wavelength (or at a band chosen around a single wavelength).
The solution provided in US 20130194481 performs 1D CS-SCR using block Toeplitz matrices to perform a single 1D transform applied sequentially to columns of an array that comprises all the wavebands images concatenated in a vertical direction. It has been shown that the RIP condition for block Toeplitz matrices is harder to uphold than for random ones, in terms of the sparsity required from the signal one wishes to reconstruct.
In various embodiments there are disclosed methods and apparatus for HSI based on CS principles. In the following description “snapshot spectral imager”, “SSI apparatus” and “SSI camera” are used interchangeably. The reference throughout is to “HSI”, although apparatus and methods disclosed herein can also be used for spectral imaging with a smaller number of wavelengths. Consequently, “spectral imaging” may be used in the following description as a general term for imaging using from three wavelengths (i.e. R(ed), G(reen) and B(lue)) up to hundreds and even thousands of wavelengths.
An apparatus disclosed herein is a snapshot hyperspectral imager. Thus, “apparatus for HSI” “HSI apparatus” and “snapshot hyperspectral imager” may be used interchangeably. Although focused on spectral imaging of a source object, methods and apparatus disclosed herein can also be applied to gray or colored images with various spectra at different spatial positions on the object.
We discovered that the SCR described in commonly owned US 20130194481 can be further improved by application of a two-dimensional (2D) framelet transform separately to the arrays representing different wavebands, instead of the sequential column-wise application of a 1D transform used therein. The application of the 2D framelet transform separately to arrays representing different wavebands of spectral cube data, referred to hereinafter as “2D CS-SCR”, includes application of direct and inverse 2D framelet transforms to the arrays. The direct and inverse framelet transforms are included exemplarily in a split Bregman iteration. The framelet transform inside the Bregman iteration uses a split-to-wavelength bands instead of the split-to-one spatial coordinate end. The 2D CS-SCR disclosed herein provides faster and better (in terms of peak signal-to-noise ratio or “PSNR”) SCR than the 1D CS-SCR described in US 20130194481
We also discovered that for certain objects, the addition of a randomization operation during image acquisition, performed with either an added hardware (HW) optical element (“HW randomizer”) or algorithmically in software (“SW randomizer”), may further improve the 2D CS-SCR results. In some apparatus and method embodiments disclosed hereinbelow, a randomizer is added to a SSI apparatus described in US 20130194481. The randomizer aids in the reconstruction of spectral images of a non-sparse (“regular”) object. A HW randomizer may be implemented as a thin optical element at the image sensor plane. The randomizer causes the DD image data reaching the image sensor (for the HW randomizer) or the DD image data obtained with the image sensor (for the SW randomizer) to become completely random and to result in a “randomized image”. As used herein, “randomized image” refers to the image data obtained after the action of the randomizer on the DD image. In some embodiments, a single random matrix R of the randomizer is taken from a statistical ensemble.
In some embodiments there are provided snapshot spectral imaging apparatus and methods based on digital cameras with minimal hardware changes. The SSI process includes performing 2D CS-SCR from a DD image of a source object, with or without added randomization.
In an embodiment, there is provided apparatus for obtaining a plurality of spectral images of a source object in a snapshot, the apparatus comprising an imaging section of a digital camera that includes a lens and a pixelated image sensor, the imaging section configured to obtain a DD snapshot image Y, and a digital processor configured to perform 2D CS-SCR from snapshot image Y, thereby providing images of the source object in a plurality of spectral bands.
In an embodiment, there is provided a method for obtaining a plurality of spectral images of a source object in a snapshot comprising the steps of obtaining a DD snapshot image Y and performing 2D CS-SCR from snapshot image Y, thereby providing images of the source object in a plurality of spectral bands. In some embodiments, snapshot image Y is obtained by imaging the source object with an imaging section of a digital camera that includes a lens and a pixelated image sensor positioned at an image sensor plane, wherein the DD image is formed through a RIP diffuser that satisfies a RIP condition related to a sensing matrix A. In such embodiments, the 2D CS-SCR includes transposing sensing matrix A into a transposed matrix AT, applying AT to Y to obtain AT Y, applying a 2D direct sparsifying transform D to AT Y to obtain a sparse version d of a reconstructed data cube X, using an inverse transform Ψ to obtain X from d, and processing X by to obtain the images of the source object in the plurality of spectral bands.
In some embodiments, spectral images are reconstructed from a randomized image using spline-based frames. In some embodiments, spline-based frames are applied to reconstruct spectral images from superposition of several randomized monochromatic images.
Aspects, embodiments and features disclosed herein will become apparent from the following detailed description when considered in conjunction with the accompanying drawings. Like elements may be numbered with like numerals in different figures, wherein:
Apparatus 100 is in principle similar to apparatus disclosed in US 20130194481 (e.g. apparatus 200 therein) except that processor 110 is configured to perform 2D CS-SCR instead of the 1D CS-SCR disclosed in US 20130194481. A detailed and enabling example of the 2D CS-SCR process is provided below. Optionally, apparatus 100 may include an added external (to the camera) digital processor 105 configured to perform some or all of the 2D CS-SCR disclosed herein.
The following model is described with reference to a single lens SSI apparatus as in
Suppose that an ideal original image of a source object obtained (without use of diffuser or randomizer) has an intensity distribution I0(x, y; λ), which is a cross section of a data cube at wavelength λ. The RIP diffuser has a complex transmission function P:
P(υ′;λ)=exp[iφ(υ′;λ)] (1)
where φ(υ′; λ) is a phase function of the diffuser at wavelength λ. When installed into the optical system, the RIP diffuser converts the original image to a DD image, since the imaging system ceases to be ideal. The shape and characteristics of the DD image can be calculated as function of P and of the original image. The coherent point-spread function of the system can be calculated as Fourier transform of P:
and describes the system's impulse response for the delta function at input, in complex amplitude of the electromagnetic field, where R is a distance from the exit pupil to the image sensor of the imaging system. If the light is incoherent, one can measure only the intensity of light received by the image sensor. Accordingly, the system's impulse response in intensity is described by the incoherent PSF hI(y′; λ) given by:
h
I(y′;λ)=|h(y′;λ)|2λ2 (3)
A spatially shift invariant model imaging system provides the DD image intensity as a 1D convolution I′=hI{circle around (x)}I of the ideal (“non-dispersed”) image I with the incoherent PSF hI:
I′(x,y′;λ)=∫hI(y′−y;λ)I(x,y;λ)dy (4)
where I(x, y; λl) is the intensity of an ideal image of a spatially incoherent source object obtained by the imaging system without use of diffuser or randomizer at wavelength λl, and x, y are Cartesian coordinates at the image sensor. Note that a 1D convolution is calculated separately for each coordinate x of the ideal image.
Since the DD image is taken with a pixelated image sensor, it is in effect sampled and can be represented as a matrix of the intensity in each pixel. The incoherent PSF can also be represented by a Toeplitz matrix that represents convolution Eq. (4). The image sensor has naturally a discrete pixelated structure characterized by a 2D spatial pitch δx×δy, a number Nx,Ny of pixels and a number Nb of bits per pixel. In an embodiment, an imaging zoom is chosen such that an image blur caused by the RIP diffuser causes the dispersed-diffused image to occupy all Ny pixels in each column and all Nx pixels in each row at the image sensor. Accordingly, an “undiffused-undispersed image” obtained without the RIP diffuser at the same zoom occupies only a smaller number N<Ny of pixels located in a central part of each column and all Nx pixels in each row at the image sensor. The data cube is defined as a 3D array with size Nx×N×L, where Nx,N are spatial dimensions and L is a spectral dimension, i.e. the number of spectral bands or wavelengths in the spectral image. Even though the number Nx×Ny of sensed pixels (i.e. dimensions of experimental data) may be substantially smaller than number of voxels in a targeted 3D data cube with dimensions Nx×N×L, we suggest a solution for the 3D data cube by resorting to a CS approach and making use of implicit redundancy in the image data. The suggested solution provides the data compression rate N×L/Ny.
Following discrete notations customary in CS, we define the following index ranges: a range i=
The voxels of data cube share spatial pitches of the sensor but have a different index range, so their indices are shifted by:
and their Cartesian coordinates are xj and yi+i
N
d
=D
υ′/Δυ′ (6)
vertical straight line strips extending parallel to the u′ axis, with widths Δυ′ and centers
Therefore the RIP diffuser can be described by a complex piece-wise constant pupil function that depends only on the coordinate υ′:
where φk,l is a phase constant within a width Δυ′ of the kth strip on the RIP diffuser, k=
Equations (2) and (3) for the incoherent PSF provide a discrete convolution kernel as a Toeplitz convolution matrix for each wavelength:
where the convolution kernel is:
and Pkl is defined by Eq. (9). Note that array KΔi′,l for fixed l=
X
i,l
(j)
=I(xj,yi+i
A discrete version of the ideal image intensity in each spectral band l=
X=(Xi,l(j),i=
In other words, X is a matrix that represents a spatial-spectral data cube.
Assuming that the optical system allows only 1D dispersion such that the two spatial dimensions x, y of the image are not mixed, each column of a DD image can be considered separately. Each column includes the image data (for the image matrix) and the corresponding transfer function (PSF matrix). Moreover, because the dispersion is only 1D, the columns of the PSF are identical, which allows to drop the column index j for the PSF. Therefore, at each wavelength, Eq. (4) of the continuous 1D convolution can be rewritten as a discrete 1D convolution applied separately for each of Nx image columns. The contribution of light with single wavelength λl to discrete pixels of the DD image can be expressed in discrete form as
where j is the number of a column in the DD image as well as in the data cube and Ki′−i,l are elements of a Toeplitz “convolution matrix”. Equation (15) shows that in a single spectral band, light intensity formed by the imaging lens on the image sensor is described as the discrete convolution of the data cube and elements of a Toeplitz matrix, defined by Eqs. (10) and (11).
SSI Apparatus with Randomizer
The RIP diffuser in the embodiments of
The randomizer randomizes the Toeplitz structure of the initial measurement matrix and allows reconstruction for reasonably sparse real images. The randomizer can be represented as a 2D matrix with random elements Ri′,j, i′=
R
i′,j
I′(xj,yi′;λl),i=
where only a single column (Ri′,j, i′=
R
i′,j≡1,i=
In one embodiment, the randomizer may be implemented as an algorithm and software code for the digital processor of a photo or video camera or an external to camera laptop or desktop computer. In another embodiment, the randomizer may be implemented as hardware, in particular as an optical element placed between the imaging lens and an image sensor of photo or video camera, preferably in close vicinity to, or mounted on the image sensor.
The contribution of light with an entire set of wavelengths to discrete pixels of the DD image in CS for spectral imaging is denoted as Yi′(j) and can be expressed as a sum of the intensities of DD images over all the wavelengths at each image sensor pixel to obtain the sensed intensity:
The non-negative numbers κl (in our computer simulations below κl=1) characterize a relative spectral sensitivity of the image sensor at wavelength λl, and coefficients Ai′,i,l(j) describe the combined effect of the RIP diffuser and the randomizer Ri′,j,
A
i′,i,l
(j)
=R
i′,j
B
i′,i,l
,i=
B
i′,i,l=κlKi′−i,l,i=
Therefore, the randomizer breaks the Toeplitz structure of the sensing matrix at each wavelength, creating an even more random structure for the signal. It changes randomly the amplitude received by each image sensor pixel, thus improving the ability to fulfill the RIP condition.
In an embodiment with a single version of a randomizer (“single randomization” action), multiplication at the images sensor pixels of the acquired DD image pixels with the gray-level pixels of the randomizer may render some of the image sensor pixels actually unused. For example zero or near-zero values of randomizer pixels will cause respective image sensor pixels to appear as zero or near-zero, i.e. actually missing. Use of such a randomizer may lead to some loss of light flux and to a reduction in the throughput and sensitivity of the SI camera. The latter are however very important for reduction of noise in a SI camera with a RIP diffuser and monochromatic image sensor. High throughput and sensitivity of a SI camera as disclosed herein may be achieved by using multiple versions of the SW randomizer (also referred to as “multiple randomization”), described below.
It appears convenient for mathematical considerations to concatenate spectral and vertical spatial dimensions in a data cube, i.e. to substitute two indices i,l by a single index in arrays Xi,l(j) and Ai′,i,l(j). Accordingly, we resort to 1D vectors X(j) with enlarged length NL:
The entire set of Nx column vectors X(j), j=
X=[X
(j)
,j=
with size NL×Nx, which contains all the spectral cube's data. The matrix X can be alternatively split into L spectral dimensions
such that each spectral dimension is described by a sub-matrix Xl of size N×Nx. A Ny×NL dimensional sensing matrix
A=(Ai′,i,l,i′=
can be treated as a block-wise rectangular matrix A=[A1, A2, . . . AL] composed of L sub-matrices Al of size N×N each. Each sub-matrix Al corresponds to a single wavelength. Matrix A provides a high level of randomization by integrating effects of the RIP diffuser and (when present) of the randomizer. We also define a vector of the DD image:
Note that vector X(j) is the data we wish to reconstruct from the one-column sensed vector Y(j). The entire set of Nx column vectors Y(j), j=
Y=[Y
(j)
,j=
with size Ny×Nx. Matrix X is the spectral data we wish to reconstruct from sensed intensity matrix (or snapshot image) Y. Eq. (17) can now be expressed in a matrix form as a multiplication of a vector with length NL over a matrix with dimensions Ny×NL. The multiplication results in a vector with a smaller length Ny:
Y
(j)
=AX
(j). (26)
Accordingly, for the 2D data processing, merged vectors and matrices can be expressed in matrix form as a multiplication of a matrix of size Ny×NL with a matrix with dimensions NL×Nx, resulting in a matrix of smaller size Ny×Nx:
Y=AX. (27)
Equation (26) provides the CS model for each column j, and Eq. (27) provides the CS model for the entire two-dimensional DD image at the image sensor. The CS problem is in reconstruction of a matrix X such that to satisfy Eq. (27) with a given matrix Y.
Due to the compressibility property of the source object, this object can be sparsely represented in a space in which it is sparse. The sparse representation of the source object can be reconstructed from the dispersed image by performing minimization of a functional that comprises the l2 norm of difference between a reconstructed vector of the source object multiplied by the sensing matrix and the dispersed image. Therefore, the l1 norm of the coordinates in the space in which the object is also sparse. The minimization process (with the constraint) can be exemplarily achieved via a split Bregman iteration process [Z. Cai et al., SIAM J. of Multiscale Modeling and Simulation, Vol. 8(2), pp. 337-369, 2009, hereinafter “Cai”]. This process has been known to be an efficient tool for CS reconstruction. The split Bregman iteration is an iterative algorithm involving a closed loop, with the reconstruction constrained l1 error serving as the feedback to the loop and with a shrinking operation that ensures a sparse reconstruction.
Since A is a Ny×NL matrix and L>1, the number of unknown variables NL×Nx is larger than the number Ny×Nx of equations. Accordingly, the problem seems to be ill-posed and cannot be solved in general case. The CS sensing theory addresses, however, a specific case when matrix X is compressible and can thus be represented as a linear transform of a K-sparse matrix d in some (possibly redundant) basis, where d=DX is a matrix having only K non-zero elements with known locations and D is a sparsifying matrix. A sparsifying matrix is a matrix that converts a vector or array to a sparse matrix, i.e. to a matrix having only a small number of non-zero elements. The redundant basis may be implemented exemplarily by resorting to 2D framelet transforms, described in more detail in the “Spline-based frames for spectral image reconstruction” section below. We are hereby applying, for the first time, 2D semi-tight wavelet frames (or framelets) originating from quadratic quasi-interpolating polynomial splines to spectral image reconstruction in CS-based spectral imaging. A detailed description of spline-based frames, and, in particular the development of a variety of low-pass filters h0 from interpolating and quasi-interpolating polynomial splines, may be found in A. Averbuch and V. Zheludev “Interpolatory frames in signal space”, IEEE Trans. Sign. Proc., 54(6), pp. 2126-2139, 2006. A description of framelets may be found in A. Averbuch, P. Neittaanmaki and V. Zheludev, “Splines and spline wavelet methods with application to signal and image processing. Volume I: Periodic splines”, Springer, 2014 (hereinafter “APZFrame”). In particular, we apply a 2D direct linear transform with a sparsifying matrix D to obtain a sparse version d of data cube X. In an embodiment, sparsifying matrix D may be the matrix of a direct 2D framelet transform, which is applied separately to each sub-matrix Xl, l=1, . . . , L, of matrix X:
We apply a 2D inverse linear (exemplarily a frame) transform with matrix Ψ to obtain data cube X from its sparse version described by the K sparse matrix d:
which is a matrix of the inverse 2D framelet transform applied separately to each sub-matrix dl, l=1, . . . , L, of matrix d. Y can be now expressed in the form
Y=AX=AΨd=Θd, (30)
where Θ=AΨ. As well known, the RIP condition of order K in CS (see E. J. Candes and T. Tao, “Decoding by Linear Programming,” IEEE Trans. Information Theory 51(12): 4203-4215 (2005)) demands that any sub-matrix of Θ formed by zeroing all its columns, except for less than K ones, must satisfy the inequality:
(1−δK)∥d∥l
for any K-sparse vector d, where δK>0 is some small number, and ∥d∥l
One of the best known examples for a sensing matrix satisfying the RIP condition is a random matrix or random Toeplitz matrix formed by Gaussian random variables with zero mean and 1/NL variance. In this case, the columns are approximately orthogonal and the RIP condition is satisfied with high probability if:
where 0<C≦1 is a constant. We reconstruct the spectral cube X from a given matrix Y by solving the following constrained minimization problem:
where DX is the block-wise 2D framelet transform of the matrix X as in Eq. (25), the l1 norm of a vector a is |a|l
An approach to solve the minimization problem Eq. (33) was presented in Cai. The process works by introducing several additional variables, which are treated separately. In more detail, following the analysis performed there, the minimization for a linear operator A is performed by an iterative process
where k is a number of iteration, dk, and ck are the intermediate vectors, used to execute iterations, AT denotes a transposed matrix A and
shrink(x,γ)=sgn(x)max(|x|−γ,0). (35)
is a function applied to each component of a matrix. The parameters of the process μ, χ (where χ−1 is a shrinkage threshold) enable to give different significance or weight to the terms in the problem: ∥AXk+1−Y∥l
After completion of the iterations, we have the compressible block-wise matrix for the reconstructed data cube:
X=Ψd (36)
comprising sub-matrices Xl=dl, l=1, . . . , L, where the sparse block-wise matrix d=DX. Reconstructed data cube components Xl=Ψdl containing a reconstructed image in each spectral band are then represented by a matrix with size N×Nx
where l=
at each spectral band.
RIP Diffuser Design with Permutations of Saw-Tooth Diffraction Grating
An exemplary 1D RIP diffuser implementation was described in US 20130194481. In mathematical terms, it is however convenient to scale it to the size of the exit pupil by a coefficient of pupil magnification. The grooves of the RIP diffuser in US 20130194481 were designed for a specific wavelength λdes with phase levels as shown in
The RIP diffuser includes Nd vertical straight line strips extending parallel to the u′ axis, with widths Δυ′ and centers υk′ defined in Eq. (7). The groove depths hk are constant within the width Δυ′ of a kth strip. Each groove depth h causes at a wavelength λ a corresponding phase shift φk,des given by the following paraxial case equation:
where n(λ) is the refractive index at wavelength λ. Since the phase is wavelength-dependent, each groove depth adds a different phase to light with a different wavelength. The phase additions for two different wavelengths are related by:
The approximation in previous equation can be applied because the refractive index n slowly varies with the wavelength. Therefore, if the mask grooves are designed for a specific wavelength λdes, the mask's impact on light with wavelength λ is:
In view of Eq. (36), the phase provided by the RIP diffuser can be described as
where φk,des is phase at the design wavelength λdes at a straight line strip number k on the RIP diffuser, λl is a central wavelength of a of spectral band number l, l=
The coherent point spread function h(y′; λ) associated with the RIP diffuser is also 1D, depending only on coordinate y′ at the image plane, and can be calculated as inverse Fourier transform of the piecewise constant pupil function. Resorting to a known result of the Fourier transform of a rect function as a sin c function:
and resorting to shift properties yields:
where υk′ and Δυ′ are the location and the width of the k-th straight line strip in a 1D RIP diffuser respectively, Pkl is constant within a width Δυ′ of the kth strip on the RIP diffuser and was defined in Eq. (9) through a phase shift φk,l, and k=
In an embodiment disclosed herein, the RIP diffuser design is developed further as follows. The RIP diffuser may be installed at, or in a vicinity of, the entrance pupil of the imaging system. However, in mathematical equations it is convenient to scale it to the size of the exit pupil by a coefficient of pupil magnification. In an embodiment, the RIP diffuser is a 1D thin phase optical element providing changes in phase of an optical light field in a single operating direction and including line grooves extending perpendicular to the operating direction. The RIP diffuser is fabricated of transparent material with the refractive index n(λ) and consists of Nd vertical straight line strips extending parallel to u′. The depths and phases are constant within the width Δυ′ of kth strip and quantized to NQ discrete phase levels equidistant from each other with a phase difference of 2π/NQ. In an embodiment presented here, the design for the RIP diffuser started with a blazed diffraction grating with a saw-tooth profile, as shown in
π mod2π(2πυ′/Λ), (43)
where mod2π(•) function denotes a minimum positive residue of the argument, after subtracting multiples of 2π. The phase function was quantized to NQ discrete phase levels such that the strip widths are Δυ′=Λ/NQ. The total number of strips was chosen to be
N
d
=D
y
/Δυ′=N
Q
D
y/Λ (44)
In an embodiment, a quantized saw-tooth array was created with a number of points NQ in every cycle corresponding to the number of groove depth levels, and with a total number of pixels Nd. Consequently, each point k in the saw-tooth array represents the phase value for one strip:
π mod2π(2πυk′/Λ). (45)
In this embodiment, the number of groove depth levels NQ and the blazing period Λ are limited by practical considerations, i.e. fabrication rules for feature-size and the number of groove depth levels. However, other embodiments may use different numbers of groove depth levels and blazing periods.
A randomization of a 1D blazed saw-tooth array was executed by a spatial permutation of the indices k=
A HW randomizer has a pixelated structure with pixel size and arrangement matching exactly those of the image sensor. Thus, each pixel on the image sensor receives its portion of the intensity of a randomized image created by the diffuser multiplied by a random coefficient. A HW randomizer may be made of any material transparent in the wavelength range of interest, for example glass or plastic. Each HW randomizer pixel has a random transparency value between fully transparent and fully opaque. The HW randomizer is positioned in the optical path between RIP diffuser and image sensor, preferably adjacent to (or alternatively exactly at) the image sensor plane. The randomizer breaks the block Toeplitz structure of the measurement matrix, thus creating a random structure for the signal. It changes randomly the amplitude received by each pixel of the image sensor, thus improving the ability to hold the RIP condition.
In some embodiments, the randomizer design uses pseudo-random numbers from a function to create a matrix of the same size as the image sensor pixel matrix. The values for elements of the randomizer matrix are random, given preferably by independent Gaussian random variables with a standard normal distribution, whose probability density is of the form:
Note that other probability densities may be used for this purpose. In other embodiments, values for elements of the randomizer matrix can be either uncorrelated random variables or pixels of a 1D or 2D random process or field, described by recurrent equations, which are well known to those of ordinary skill in the art. In still other embodiments, values for elements of the randomizer matrix can be deterministic and be defined by a closed form equation, for example an array of lenslets with equal or variable phase shift between the individual lenslets.
The values for elements of the SW randomizer matrix may be same as described above for the HW randomizer. In addition, apparatus and method embodiments with a SW randomizer may use either single randomization or multiple randomization. The latter uses multiple random versions of a SW randomizer, in which zero pixels of one version are followed by non-zero pixels of another version at the same positions on the detector. This results in statistical averaging of the randomization action and enables the entire light flux acquired by the detector to be used efficiently. In an embodiment, the data cube reconstruction algorithm is then run separately with each version of the SW randomizer to provide several versions of a reconstructed data cube. These may be then merged by simple averaging per pixel or, alternatively, by image fusion algorithms. In another embodiment, the data cube reconstruction may be performed using an algorithm that employs a class of measurement matrices that differ only in the SW randomizer version, while relying on the same RIP diffuser. In this embodiment, a single iterative process with multiple random versions of the randomizer will provide directly the reconstructed data cube, based on all the detector pixels.
The “multiple randomization” process may also be described as follows: more randomizing may be provided by resorting to several independent versions R(1), R(2), . . . , R(N
Here we describe in more detail the spline-based frames (or framelets) for spectral image reconstruction in CS-based spectral imaging. Spline-based frames, and in particular the development of a variety of low-pass filters h0 from interpolating and quasi-interpolating polynomials splines was reported previously, see e.g. A. Averbuch and V. Zheludev, “Construction of bi-orthogonal discrete wavelet transforms using interpolatory splines”, Applied and Computational Harmonic Analysis, 12, 25-56, 2002, A. Averbuch, A. B. Pevnyi, and V. A. Zheludev, “Bi-orthogonal Butterworth wavelets derived from discrete interpolatory splines”, IEEE Trans. Signal Processing, 49(11), 2682-2692, 2001, and APZFrame. The spline-based framelet transforms are applied to successive approximations Xk that are derived from the randomized input in the process of Bregman iterations.
A system {tilde over (Ψ)}={{tilde over (φ)}l}l=0L-1, L>N of signals from Π[N], which is a space of N−periodic signals, forms a frame of the space Π[N] if there exist positive constants A and B such that for any signal x={x[k]}εΠ[N] the following inequalities
hold. If the frame bounds A=B, the frame is said to be tight. If {tilde over (Φ)} is a frame, then there exists another frame Φ={φl}l=0L-1 (synthesis) such that any signal x={x[k]}εΠ[N] is represented by
If {tilde over (Φ)} (also called “analysis” frame) is tight, then the synthesis frame can be Φ={tilde over (Φ)}.
The analysis four-channel filter bank {tilde over (H)}={{tilde over (h)}s}s=03 and the synthesis filter bank H={hs}s=03, with down-sampling factor of 2, form a perfect reconstruction (PR) filter bank if any signal x={x[k]}εΠ[N] can be expanded as:
Equation (47) provides a frame expansion of the signal x, where the signals {{tilde over (h)}s[•−2k]}, s=0, . . . , 3, k=0, . . . , N/2−1, constitute an analysis frame, while the signals {hs[•−2k]}, s=0, . . . , 3, k=0, . . . ,N/2−1, form a synthesis frame.
Denote by x0={x[2k]}εΠ[N/2] and by x1={x[2k+1]} the even and the odd polyphase components of a signal xεΠ[N], respectively. Denote ω=e2πi/N.
are the discrete Fourier transform (DFT) of signal x and its polyphase components. hps and {tilde over (h)}ps, p=0, 1, s=0, . . . , 3, are the polyphase components of filters hs and {tilde over (h)}s. ĥps[n] and {tilde over (ĥ)}ps[n], p=0, 1 are their DFT. Denote:
{tilde over (P)}[n] and P[n] are respectively the analysis and synthesis polyphase matrices of the filter banks {tilde over (H)} and. The symbol ( . . . )T means matrix transposition. The direct framelet transform of a signal x of length N, which produces four sets of the coefficients d={ds}, s=0, 1, 2, 3, each of which contains N/2 members, can be represented as:
The inverse framelet transform, which restores the signal from coefficients ds, s=0, 1, 2, 3, is:
Thus, the length—N signal x becomes represented by 2N coefficients from the sets ds, s=0, 1, 2, 3. In that sense, this representation is doubly redundant. The relation P[n]{tilde over (P)}[−n]=I2 (PR), where I2 is the 2×2 identity matrix, is the condition for the pair {{tilde over (H)},H} of filter banks to form a PR filter bank. Filters {tilde over (h)}0 and h0 from the PR filter banks {{tilde over (H)},H} are low-pass.
To extend the framelet transform to the lower resolution scale and to increase the representation redundancy, the transform is applied to the low-frequency coefficients array d0 using analysis polyphase matrix {tilde over (P)}[2n]. The coefficients' array d0 is restored using synthesis polyphase matrix {tilde over (P)}[2n] and P[2n]. Similarly, the transform is extended to further resolution scales using matrices {tilde over (P)}[2mn] and {tilde over (P)}[2mn], m=2, 3, . . . . . The 2D framelet transform of a 2D array thus includes application of a 1D transform to columns of the array, followed by application of a 1D transform to rows of the array.
In an exemplary embodiment, we designed a family of 4-channel PR filter banks with diverse coefficients (see APZFrame). Their polyphase matrices have a specific structure, which is determined by a low-pass filter whose frequency response is ĥ0[n]=ĥ00[n]+ω−nĥ10[n]:
where T2 [n]{tilde over (T)}2[−n]=T3 [n]{tilde over (T)}3 [−n]=1−|ĥ00[n]|2+|ĥ10[n]|2. A filter bank generates the tight frame if T2[n]{tilde over (T)}2[−n]=T3[n]={tilde over (T)}3[−n]. A filter bank generates the semi-tight frame if T2 [n]≠{tilde over (T)}2[−n],T3[n]≠{tilde over (T)}3[−n]. Unlike the tight-frames filter banks, filter banks generating semi-tight frames have linear phase.
In an embodiment, we use the filter bank derived from a quasi-interpolating quadratic spline (see APZFrame). This filter bank generates a semi-tight frame. The frequency responses of the analysis and synthesis filters are:
where the sequences T[n], {tilde over (T)}[n] and G[n] are:
In the process of Bregman iterations, Eqs. (33) and (34), the direct and the inverse 2D framelet transforms are repeated. Each 2D framelet transform is implemented by the application of a 1D framelet transform to columns of the matrices using fast Fourier transforms, followed by a 1D transform of the rows, Eqs. (49) and (50). Polyphase matrices {tilde over (P)}[n] and P[n] defined in Eq. (51), are used for one-level transforms, while polyphase matrices {tilde over (P)}[2mn] and P[2mn], m=2, 3, . . . are used for multi-level transforms.
Various computer simulations of data cube reconstruction for test multispectral source objects sensed with a digital camera equipped with a 1D RIP diffuser and without or with randomizer were run. The simulations were done with Matlab software code.
Simulations of 2D CC-SCR Multispectral Images in Apparatus that Includes a Digital Camera and a 1D RIP Diffuser without Randomizer Simulation using Matlab was performed on the base of the 2D CS-SCR description above. The spectral data source was a fragment “houses” scene number 7 in D. H. Foster, “Hyperspectral images of natural scenes 2004”, http://personalpages.manchester.ac.uk/staff/davidloster/Hyperspectralimages_of_natural_scenes_04.html, 2004 (hereinafter “houses in Porto”). A DD image was obtained by computer simulation of an optical imaging system that includes a digital camera having an imaging lens and a pixelated image sensor, with 1D RIP diffuser inserted in the pupil at the image sensor plane. Each column of the DD image was a linear combination of all the spectral and spatial data in the corresponding source object image column with a sensing matrix. The Bregman iteration process was applied to reconstruct spectral cube information corresponding to M voxels in each of N columns and L=33 spectral bands of the spectral cube. The result is a set of vectors, each vector including all spectral information for each pixel in the corresponding image column. All reconstructed image columns were then placed next to each other, thereby providing the entire spectral information that represents the full spectral cube. Finally, the spectral cube was processed to obtain L separate spectral images of the object by taking consecutive sets of M rows corresponding to required spectral bands. The quality of the 2D CS-SCR results was evaluated by comparing the PSNR of our results with the PSNR achieved in the reported studies. Options without or with randomizer were executed in simulation.
Table 1 summarizes the parameters of the optical system and of the designed RIP diffuser used in the simulations. The parameters fit a 10 Mp camera.
Table 2 provides the minimum number of rows M on the image of the sensor required to satisfy the RIP condition Eq. (31) following Eq. (32), for an image with column size Mimage, L spectral bands and 20% sparsity (the portion of the non-zero values in a “sparse” image).
Each publication mentioned in this application is hereby incorporated by reference in its entirety for all purposes set forth herein. It is emphasized that citation or identification of any reference in this application shall not be construed as an admission that such a reference is available or admitted as prior art. While this disclosure describes a limited number of embodiments, it will be appreciated that many variations, modifications and other applications of such embodiments may be made. For example, while the description refers specifically to a DD image obtained through a particular RIP diffuser, the 2D CS-SCR methods described herein may be applied to any other dispersed-diffused image for which a sensing matrix A is defined. For example, while the description refers specifically to framelet transforms and Bregman iterations, other types of transforms or algorithms may be used for the SCR described herein. Further, while Toeplitz matrices and convolution are described in detail, more general matrices and linear transformations, corresponding to non-paraxial and spatially-variant optical systems and/or optical systems having aberrations may also be used. For example, while 1D diffusion/dispersion is described in detail, 2D diffusion/dispersion may also be used. Thus, the disclosure is to be understood as not limited to framelet transforms, split or other Bregman iterations and 1D diffusion/dispersion. In general, the disclosure is to be understood as not limited by the specific embodiments described herein, but only by the scope of the appended claims.
This application is a 371 of international patent application PCT/IB2014/062270 and is related to and claims priority from U.S. Provisional Patent Application No. 61/836,218 having the same title and filed Jun. 18, 2013, which is incorporated herein by reference in its entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/IB2014/062270 | 6/16/2014 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
61836218 | Jun 2013 | US |