The present invention relates to imaging. More particularly, the present invention relates to enhanced underwater imaging.
Underwater vision is plagued by poor visibility conditions. Direct employment of most computer vision methods (e.g., those based on stereo triangulation or on structure from motion) underwater is difficult. This is due to the particularly challenging environmental conditions, which complicate image matching and analysis. It is important to alleviate these visibility problems, since underwater imaging is widely used in scientific research and technology. Computer vision methods are being used in this mode of imaging for various applications (see, for example, A. Ortiz, M. Simo, and G. Oliver, “A vision system for an underwater cable tracker,” in Machine Vision and Applications, vol. 13, pp. 129-140, 2002) such as mine detection, inspection of underwater power and telecommunication cables, pipelines, nuclear reactors, and columns of offshore platforms. Underwater computer vision is also used commercially to help swimming pool life-guards. As in conventional computer vision, algorithms are sought for navigation and control of submerged robots. In addition, underwater imaging is used for research in marine biology, archaeology and mapping. Moreover, underwater photography is becoming more accessible to the wider public.
What makes underwater imaging so problematic? Underwater, visibility degradation-effects vary as distances to the objects increase. Since objects in the field of view (FOV) are at different distances from the camera, the causes for image degradation are spatially varying. Contrary to this fact, traditional image enhancement tools, e.g., high pass filtering and histogram equalization are typically spatially invariant. Since they do not model the spatially varying distance dependencies, traditional methods are of limited utility in countering visibility problems.
A common approach to improve underwater visibility and color is based on artificial illumination. The most popular realization of this approach (see, for example, B. Skerry and H. Hall, Successful Underwater Photography. New York: Amphoto books, 2002, pp. 25-41) uses an off axis wide-field strobe attached to the camera. A significant problem associated with this is the falloff of scene irradiance with the distance from the strobe. Moreover, “sea snow” may be created by the defocus blur of the strong backscatter from suspended particles at close distances.
To bypass the backscatter problem, advanced research underwater imaging systems use specialized active radiation hardware (see, for example, G. D. Gilbert and J. C. Pernicka, “Improvement of underwater visibility by reduction of backscatter with a circular polarization technique,” App. Opt., vol. 6, pp. 741-746, 1967). Yet, the range of such systems is limited, for the reason mentioned with respect to a wide field torch: at some distance the source's falloff leads to too low levels of scene irradiance. Such systems tend to be highly power consuming, complex and expensive. These problems are avoided by passive computer vision which exploits natural illumination. When available, natural illumination exists all over the scene, alleviating the need to project energy towards objects.
It was demonstrated decades ago (J. N. Lythgoe and C. C. Hemmings, “Polarized light and underwater vision,” Nature, vol. 213, pp. 893-894, 1967) that polarization filtering can enhance contrast in passive underwater vision. Yet, using the raw result of simple optical filtering may have a limited effect, indicating that some post processing is needed, based on acquisition of both components of polarized light. One approach is based on a simple subtraction of the differently polarization filtered images (see, for example, L. J. Denes, M. Gottlieb, B. Kaminsky, and P. Metes, “AOTF polarization difference imaging,” in Proc. SPIE Advances in Computer-Assisted Recognition, vol. 3584, 1999, pp. 106-115) or displays the degree of polarization (DOP) (for example, M. P. Rowe, E. N. Pugh, Jr., J. S. Tyo, and N. Engheta, “Polarization-difference imaging: a biologically inspired technique for observation through scattering media,” Optics Letters, vol. 20, pp. 608-610, 1995). That approach has fundamental disadvantages. It assumes that polarization is associated with the object radiation, rather than the causes for the degradation of this signal. However, due to depolarization, that assumption becomes invalid as distances increase.
Our approach is based on physics-based model that accounts for degradation effects that depends of the object distance (from the camera and the light source), and on optical phenomena that are associated with these effects. In particular, we exploit the fact polarization can be associated with the prime visibility disturbance, which we wish to remove (veiling light). The association of polarization to veiling light has recently been utilized in an attempt to enhance visibility in a limited setting where illumination was perpendicular to the line of sight (LOS) (see P. C. Y. Chang, J. C. Flitton, K. I. Hopcraft, E. Jakeman, D. L. Jordan, and J. G. Walker, “Improving visibility depth in passive underwater imaging by use of polarization,” App. Opt, vol. 42, pp. 2794-2802, 2003). That method used ad-hoc processing. Such enhancement methods are far from inverting the image formation process and recovering the objects. In contrast, our approach inverts the general physical model, thus the recovered image is similar to clear visibility appearance.
It is a purpose of the present invention to provide a novel method and system for enhancing underwater imaging.
There is thus provided, in accordance with some preferred embodiments of the present invention, a method for enhancing underwater imaging affected by image degradation effects, the method comprising:
Furthermore, in accordance with some preferred embodiments of the present invention, the image characteristics comprise at least one of the characteristics group consisting of: contrast, color, sharpness, brightness.
Furthermore, in accordance with some preferred embodiments of the present invention, compensating distance-dependent underwater degradation effects relating to the distance of illumination sources from the scene comprises compensating effects attributed to the underwater depth of the scene.
Furthermore, in accordance with some preferred embodiments of the present invention, compensating effects attributed to the underwater depth of the scene comprises white-balancing. Furthermore, in accordance with some preferred embodiments of the present invention, the physics-based mathematical model comprises an inversion of an image-formation model including backscatter.
Furthermore, in accordance with some preferred embodiments of the present invention, the invention includes regularization.
Furthermore, in accordance with some preferred embodiments of the present invention, the image-formation model that is inverted is approximated such that the approximation error is not discernible.
Furthermore, in accordance with some preferred embodiments of the present invention, acquiring of at least one image of the underwater scene comprises acquiring at least two images in different imaging conditions.
Furthermore, in accordance with some preferred embodiments of the present invention, said at least two images are acquired in different resolution.
Furthermore, in accordance with some preferred embodiments of the present invention, acquiring at least two images in different imaging conditions comprises acquiring at least two images of the scene in different polarizing states on the imaging device.
Furthermore, in accordance with some preferred embodiments of the present invention, acquiring said at least two images comprises acquiring said at least two images simultaneously.
Furthermore, in accordance with some preferred embodiments of the present invention, the reconstructed image comprises three-dimensional rendering of the scene.
Furthermore, in accordance with some preferred embodiments of the present invention, the information regarding distances of parts of the scene relative to the imaging device is used to reconstruct a distance map of the scene.
Furthermore, in accordance with some preferred embodiments of the present invention, the imaging device comprises a camera.
Furthermore, in accordance with some preferred embodiments of the present invention, the imaging device comprises at least two cameras.
Furthermore, in accordance with some preferred embodiments of the present invention, determining of information regarding distances of parts of the scene relative to the imaging device comprises extracting the information from said at least one image.
Furthermore, in accordance with some preferred embodiments of the present invention, there is provided a system for enhancing underwater imaging affected by image degradation effects, the system comprising:
a and
a illustrates an undesired effect, photoelasticity of the window of the imaging device that affects polarization.
b illustrates placement of a polarizer externally to the imaging device to minimize the effect shown in
a illustrates an undesired effect: the transmittance of a flat window is polarization-dependent at oblique incidence.
b illustrates a spherical dome concentric with the center of projection that eliminates the effects shown in
In this work we develop a physics-based approach for recovery of visibility when imaging underwater scenes, preferably in natural illumination. Since it is based on the models of image formation, the approach automatically accounts for dependencies on object distance, and facilitates retrieving a distance map of the scene as a by-product. The method presented is fast to perform, and according to a preferred embodiment of the present invention relies on raw images taken through different states of a polarizing filter. These raw images have slight photometric differences with respect to one another. The differences serve as initial cues for our algorithm factoring out turbidity effects. It is interesting to note that marine animals too use polarization for improved vision.
To demonstrate the approach, we have built an underwater polarization imaging system composed of both custom and off-the-shelf components (the considerations for selecting the components are described hereinafter). We used the method for experimenting at sea. Significant improvements of contrast, color and visibility range are obtained by the approach of the present invention.
The intensity and color of each scene point essentially depend two unknown pointwise characteristics: the albedo of the object point and the distance to that point. The multiple acquisitions of the scene, say through the polarizing filter enables our algorithm to solve for the object intrinsic appearance as well as for the object distance. Note that in principle the method of the present invention may be applied to a single image taken, if the distances of different parts of the scene from the imaging device and illumination sources are known, are measured or are estimated.
As depicted in
The signal is composed of two components, termed direct transmission and forward scattering. As a light ray progresses from the object towards the camera, part of its energy is lost due to scattering and absorption. The fraction which does reach the camera is the direct transmission (3), given by
D(x, y)=Lobject(X, Y)e−CZ (1)
where z is the distance to the object, which depends on the pixel coordinates x and y, while c is the attenuation coefficient. Here Lobject is the object radiance we would have sensed, had there been no scattering and absorption along the line of sight (LOS).
The attenuation coefficient is given by c=a+b, where a is the absorption coefficient and b is the total scattering coefficient of the water. The scattering coefficient b expresses the ability of an infinitesimal water volume to scatter flux in all directions. Integrating over all solid angles {right arrow over (Θ)}
where θ is the scattering angle relative to the propagation direction. The angular scattering coefficient b(θ) is sometimes referred to as the phase function. Note that the variables a, b(θ), c and Lobject are all functions of the wavelength λ.
The forward scattering component is similar to the direct transmission. However, it represents light scattered forward at small angles relative to the LOS. This creates image blur given by the convolution
F(x, y)=D(x, y)*gz(x, y) (3)
where D is given by Eq. (1) and gz is a point spread function (PSF). The PSF is parameterized by the distance z, since the further the object, the wider the support of the blur kernel.
There are several models in the literature for the form of the underwater PSF (for example, B. L. McGlamery, “A computer model for underwater camera system,” in Proc. SPIE, vol. 208, 1979, pp. 221-231). Since the PSF depends on the hydrosols suspended in the water, the models are typically parameterized by various empirical constants. For example, the model described by McGlamery (and others) is of the form
gz=(e−γz−e−cz)F−1{Gz} where Gz=e−Kzw (4)
where K>0 and γ are empirical constants, F−1 is the inverse Fourier transform, and ω is the spatial frequency in the image plane. Note that Gz is a low pass filter. The effective frequency “width” of Gz is inversely proportional to z. This expresses the increase of spatial blur spread for distant objects. The constant γ is limited to |γ|≦c.
Accounting for both the direct transmission (1) and the forward scattering (3), we define the signal as
S=D+F (5)
We define an effective object radiance Lobjecteffective as
Lobjecteffective−Lobject+Lobject*gz (6)
It is a somewhat blurred version of Lobject. From Eqs. (1,3,5), the signal is
S=e−czLobjecteffective (7)
We claim that in practice the prime cause for underwater image degradation is not blur, but rather veiling light.
Veiling light does not originate from the object on the LOS. Rather, light coming from illumination sources is scattered towards the camera (
The source illuminates the particles on the LOS from a direction {right arrow over (r)}=(θ, ρ) relative to the LOS, with intensity Isource. The contribution of this source to the veiling light is (see B. L. McGlamery, “A computer model for underwater camera system,” in Proc. SPIE, vol. 208, pp. 221-231, 1979)
where f is the focal length of the camera and I0 is the distance between the lens and the underwater housing port. This integral accounts for scattering into the LOS at some distance I, followed by attenuation until reaching the camera. It also accounts for geometric projection of the irradiance on the detector, via the ratio f/(I+I0).
It is now shown that we can considerably simplify the expression for veiling light, relative to Eq. (8). We can do this because typically f/(c−1+/l0)<<1. Consider typical ranges of values as c−1 ε [3 m,10 m], f ε [20 m,50 m],l0≈80 mm, and object distance in the order of meters. We assessed the integral numerically. It is shown in
where the correction coefficient is given by
For instance, as shown in
B({right arrow over (r)})approx=B∞({right arrow over (r)})(1−e−cz) (11 )
This close-form expression is much simpler than Eq. (8). It is easily seen that the farther that object, the larger the veiling light is. In Eq. (11), the variable
B4({right arrow over (r)})≡kIsource({right arrow over (r)})b(θ) (12)
expresses the veiling light in a LOS which extends to infinity in the water. Summing up the contribution from light sources at all directions, the total veiling light is
is a scalar termed the water background, which depends on λ. The veiling light B implicitly depends on x and y due to its explicit dependence on the distance z.
We now discuss the contribution of each of the above mentioned components. We performed a simulation of underwater imaging. The effects of water are simulated using a model for oceanic water (see C. D. Mobley, Light and Water: Radiative Transfer in Natural Waters. San-Diego Academic Press, 1994, ch. 3,5), with a moderate concentration of hydrosols. Starting from an object radiance of Lobject, we calculated Lobjecteffective, which accounts for blur by forward scattering (without attenuation and veiling light). We simulated the effects of varying distances by setting a spatially varying distance map to the scene. Incorporating veiling and attenuation effects as well as blur, the total image irradiance is
Itotal=S+B=e−czLobjecteffective+B (15)
The visibility strongly deteriorated at image Itotal: even objects at moderate distances were swamped in a veiling light and become obscured. Veiling light affected the color and contrast of all objects.
We made a quantitative analysis of the simulated images. As a criterion for image quality, we look at contrast. The reason is that measurement of contrast underlies the determination of the modulation transfer function (MTF) of media. Contrast between two points v=1,2 is usually defined as |I1-I2|/|I1+I2|, where Iv is the intensity at point v. We generalize this definition, such that in a region having N pixels, we use
where STD {Iv} is the standard deviation of the N intensity values. To calculate the contrast in a region of a color image having N pixels, we use
where χ is an index of the chromatic channel, and
The contrast of the effective object C(Lobjecteffective) and the sensed image C(Itotal) are calculated. Relative to the contrast of the original scene, the degradation of contrast Is
Before inspecting these values, we define an image resulting from all degradation effects excluding blur.
Ĩ(no
with a respective contrast degradation measure
The contrast degradation in the scene simulated scene was calculated as a function of distance. The resulting Δobjecteffective indicate that the effect of blur alone is not negligible, especially at long distances. Nevertheless, Δ(no
Underwater scattering involves polarization effects. We exploit these effects to compensate for underwater visibility degradation. Let us describe the models for these effects. Consider a narrow source which illuminates the scattering particles residing on the LOS. The narrow source and the LOS from the camera to the object define a plane of incidence. We may divide the veiling light into two polarization components, that are parallel and perpendicular to this plane, B∥({right arrow over (r)}) and B⊥({right arrow over (r)}), respectively. Typically, B⊥({right arrow over (r)})>B∥({right arrow over (r)}), i.e., the veiling light is partially polarized perpendicular to the plane. Moreover, if the illumination is done by an artificial source, the illumination can be fully polarized, for example it can be fully linearly polarized or circularly polarized. The intentional polarization of the source enables an increased polarization of the veiling light.
In general, irradiance of the LOS is not due to a narrow source. However, the LOS is generally illuminated in a highly anisotropic manner. In the particular case of natural underwater light, illumination originates from the sun and sky above the water. This light is restricted to cone around the zenith called the optical manhole or Snell's window. Scattering in the water leads to illumination of the LOS from other directions as well (e.g., from below and from the sides), but with lower intensity. Changes caused by this scattering to the angular irradiance distribution gradually depend of the underwater depth. As we go deeper under the water surface, the natural illumination reaches an asymptotic distribution, which is strongly peaked around the zenith.
Due to the anisotropy of the LOS irradiance, typical underwater natural veiling is significantly partially polarized. As depth increases, this polarization reaches an asymptotic value, which is maintained to infinite depth. The direction of polarization depends on the sun's bearing (unless it is occluded by clouds) and the viewing direction. However, this orientation typically tends to be approximately horizontal, for horizontal lines of sight.
In order to sense the different polarization components, we image the scene through a polarizing filter (
B=Bmax+Bmin (22)
where B is given by Eq. (13). The DOP of the veiling light is defined by:
p≡(Bmax+Bmin)/B (23)
As for the signal S, we assume that it has insignificant influence on the polarization of the measured light, relative to the influence of the veiling light. Four reasons are listed for this assumption:
Nevertheless, note that this assumption may not hold at very close distances in a relatively good visibility, if the object strongly polarizes light. When the image areas corresponding to such objects are small, we detect and treat them by relying on the image content around them, using morphological and linear filtering.
We now describe the first part of our visibility restoration method: image acquisition. In particular, we describe our apparatus. When a polarizing filter is mounted, the sensed intensity at each image pixel changes as a function of the state of the filter (for linear polarizers it is their orientation angle). Similarly to veiling light, there are two orthogonal polarizer states corresponding to extremum values of the intensity, Imax and Imin, where
Itotal=Imax+Imin (24)
while Itotal is given by Eq. (15). Since we assume that the signal polarization is insignificant, the polarizer modulates only the veiling light. Therefore, the raw images corresponding to the extrema of the intensity measure are
Imax=S/2+Bmax and Imax=S/2+Bmin (25)
Note that Imax is the image taken at the “worst state” of the polarizer, where the disturbing veiling is maximal. On the other hand, Imin is the image taken at the “best state” of the polarizer, where the disturbing veiling is minimal. For this reason, some past methods were based on simple acquisition of Imin. However, in most cases, Imin is still a highly degraded version of Lobject, where the degradation is spatially varying and is affected by the distance of the object from the camera and from the illumination source.
It is not necessary to use the minimum and maximum intensity values. Almost any two images (i.e., polarizer states) that modulate the veiling light may be used. Nevertheless, the extrema images provide the strongest modulation, yielding the most reliable outcome. The exception is when the two images have identical values, as occurs when a linear polarizer mounted on the camera is oriented at ±45° relative to the veiling light polarization.
In order to acquire the raw images, we built a custom system for underwater polarimetric imaging, which we term the Aqua-Polaricam. Several specifications determined its design, as detailed in the following. A camera was placed in a housing manufactured by Sealux, which is commercially available. We close the housing with a 120 mm dome port made of glass, while an inverted-circular polarizer was attached externally to it. The surrounding water flows to the space between the external polarizer and the dome, through several openings in the housing's interface to the polarizer mount. We use the Nikon DI00 digital SLR camera, which enables extraction of raw output data. This data is linearly related to the scene radiance (i.e. the camera has no gamma correction in this mode) without automatic white balancing. We used a 20 mm wide angle Nikkor lens. The distance I0 to the dome was 80 mm. Preliminary experiments revealed that stray light had entered the housing from its back viewing ports and then reflected into the lens. We thus blocked the stray light by slightly modifying the internal structure of the commercial housing.
We performed experiments by scuba diving and taking images using the system. At each scene, two photographs were taken to capture Imax and Imin. For example, we scuba-dived at the Red-Sea (off Eilat, Israel) to a depth of 26 meters in an area containing coral reefs. The images acquired had a very low contrast, yet their slight mutual difference provided the key for substantial visibility improvement by a mathematical algorithm, described hereinafter.
The algorithm for improving underwater visibility overcomes the “veiling” effect caused by scattered light. For this reason, the adjective unveiled is used to describe the image resulting from the algorithm. In addition to compensating for the effect that arise from the distance of the object to the camera, it also compensates for the effect that arise from the distance of the object from the illumination source. A longer path from the illuminant to the object results not only in loss of brightness, but also in a color bias. The reason is attenuation of the illumination is a function of spectral bands. For example, in natural illumination in the open sea, as we go deeper underwater, the red portion of the illumination spectrum is absorbed by the water faster than the green-blue band. Compensating for this color bias is thus a very important step, that makes this problem unique, relative to methods that attempt improvement of visibility in the atmosphere (where the illumination color bias is usually marginal).
As mentioned hereinabove, space-invariant enhancement methods do not model the spatially varying distance dependencies of visibility problems. As a result, they are of limited utility. As an example of a naive space-invariant enhancement, let us consider a simplistic method for compensating the illumination blue color bias when photographing in natural lighting. Suppose we know that the sand in the diving site is rather white. If the scene does not include such a natural object, then a white target can be introduced to the scenery. Suppose that we normalize the color of the best-polarized raw image by the color of a sandy patch in the FOV:
Iredmodified−Iredraw/Isand
Igreenmodified−Igreenraw/Isand
Ibluemodified−Iblueraw/Isand
This multiplicative normalization does not compensate for the additive spatially varying veiling light. For this reason, the measured color values of a white patch Isand
We now describe a mathematical algorithm, which solves the problem for the case of natural illumination and horizontal photography. The method inverts the visibility degradation effects including the color bias. Assume for a moment that we have an estimate of the global parameters B4 and p. From Eqs. (22,23,25), the veiling light is estimated as
{circumflex over (B)}(x,y)=[Imax(x, y)−Imin(x y)]/p (27)
Inserting this estimate into Eqs. (13,15,24), we obtain an estimate for the object radiance
{circumflex over (L)}objecteffective(x, y)=[Itotal(x, y)−{circumflex over (B)}(x, y)]/{circumflex over (t)}(x, y) (28)
where
{circumflex over (t)}(x,y)=1−{circumflex over (B)}(x,y)/B∞ (29)
is the estimated water transmittance. The transmittance is related to the object distance z by
{circumflex over (t)}=e−cz (30)
Recall that all the variables in these equations are functions of the light wavelength λ. To account for the wavelength dependence, it is best to analyze the images with a fine spectral resolution. Each wavelength band can be analyzed independently. In our experiments, though, we used the traditional coarse wideband red, green, and blue (RGB) sampling of the spectrum. Therefore, the method is applied to every color channel separately.
The result of Eq.(28) is an estimate Lobjecteffective. Thus, it does not compensate for image blur, but only for the veiling effect of scattered light and for attenuation. If we wish, we can do make do with this estimate. The reason stems from the discussion above: veiling light is the prime reason for image contrast degradation, hence overcoming veiling light, rather than blur, is the prime step for recovering visibility. Alternatively, image “deblurring” can be performed (at the price of noise amplification). For deblurring, the distance-dependent blur kernel should be known. This may be done by the aid of the distance estimation described in the following.
Eqs. (27-29) invert the spatially varying visibility degradation effects that are associated with the object's position relative to the camera. This enables proper compensation for the color bias of the illumination. Assume that the illumination is uniform. Again, we use a patch of the scene, which we know should be white, if it were not for the illumination color (typically blue). However, we now perform the compensation over the image {circumflex over (L)}objecteffective:
{circumflex over (L)}redmodified={circumflex over (L)}object
{circumflex over (L)}greenmodified={circumflex over (L)}object
{circumflex over (L)}bluemodified={circumflex over (L)}object
The result is scene recovery. The recovered image has a much improved contrast and color.
In addition to visual assessment of the results, we seek a quantitative measure of the performance of this method. We use the contrast criterion described in Eq. (17), and apply it to Imodified defined in Eq. (26), to {circumflex over (L)}modified. defined in Eq. (31), and to the best-polarized image Imin. Before applying Eq. (17), we contrast-stretched the brightness of each of these images, to occupy the full dynamic range. For better insight, we calculated the contrast at distinct image segments, corresponding to different ranges of object distances. For each distance, the contrast resulting from naive white balancing is similar to that of the raw image. This is contrary to the great improvement demonstrated in the recovered image.
We now analyze the sensitivity of the method to noise. Let the noise in the raw images be independent and having zero mean. Denote the STD of the noise in the raw images Imin and Imax as σI
Differentiating Eq. (28) and making use of Eqs. (27) and (29),
Consider the case where the system noise is roughly determined by system quantization of the image irradiance to n bits. Then the noise STD of the raw images is equal to
σI
where the dynamic range of the imaging system is normalized to 1. Inserting Eqs. (33-35) in Eq. (32), the STD of the recovered image is
This noise depends on the recovered scene radiance {circumflex over (L)}objecteffective, thus somewhat complicating its analysis. Nevertheless, interesting insights can be drawn from this result. First, note that if the polarization of the veiling light is very low, i.e., p→0, then the recovery is very unstable. This is understandable, since low polarization does not convey significant information beyond a simple unpolarized image, and hence is not useful for reliable recovery. To contain the output noise (36), we may thus need to use in the recovery equations (27-29) a value of p which is slightly biased relative to the true one. Moreover, instability always occurs at pixels that correspond to large distances, where {circumflex over (t)}→0, since there the signal S may be lower than the system noise. Hence, {circumflex over (t)} may need to be biased as well.
Accounting for the distance dependence of {circumflex over (t)} in Eq. (30) and the dependence of noise on quantization bits (See Eq. 35),
σrecovered ∝ ecz−nln2 (37)
The signal to noise ratio (SNR) decreases exponentially with the distance z. This is not a distinctive shortcoming of our polarization-based method. Rather, it is due to attenuation of the signal with the distance from the object, as is the case for other underwater imaging modalities. In methods based on artificial illumination, the SNR decreases even faster as a function of object distance. The reason is that the radiation is attenuated for approximately twice the object distance, beyond the 1/z2 falloff of the artificial illumination incident on the object, caused by free space propagation. Eq. (37) indicates that for a given level of output noise, the visibility range is proportional to the number of reliable camera bits. Thus, information from a 12-bit camera can recover objects that are 50% farther than those recoverable from data taken by an 8-bit camera, if quantization is an effective indication for the overall noise.
Noise can be reduced by spatial filtering, e.g., local averaging. Performing this directly on the raw images or on {circumflex over (L)}effectiveobject(x, y) would result in image blur, however there is a way to bypass this problem. Note that {circumflex over (B)}(x,y) and {circumflex over (t)}(x,y) do not vary rapidly across the FOV. The veiling light is typically a smooth function of spatial location, contrary to the rapid spatial variations of intensity and texture of the object Lobject. Thus spatial noise filtering of {circumflex over (B)}(x,y) and {circumflex over (t)}(x,y) does not blur {circumflex over (L)}objecteffective(x,y), yet it decreases the noise in {circumflex over (L)}objecteffective(x, y).
To perform the a polarization-based scene recovery, estimates of the global parameters B28 and p are needed. These are intrinsic parameters of the water and lighting. These estimates are obtained by measuring a groups of pixels in the raw images. For example, they can be pixels corresponding to the water background, i.e., of objects so distant inside the water, that their signals are negligible due to attenuation. Let us denote these measured values by {circumflex over (p)} and {circumflex over (B)}28 .
The value {circumflex over (p)} obtained from direct measurement of image pixels should be slightly modified, for two reasons. The first is to limit instability. The second reason is the need to minimize the instances of negative pixel values. Negative pixel values stem from inaccurate measurement of {circumflex over (p)} and {circumflex over (B)}∞ and due to image noise. From Eqs. (15,27),
is the DOP of the measured scene (i.e., the signal combined with the veiling light). If our estimate of the veiling light DOP is too low ({circumflex over (p)}<p), then negative values can appear in the image of the signal. This is especially relevant to distant objects, because P(x,y)→p when z→∞.
Both of these problems can be considerably alleviated with ease if we slightly bias {circumflex over (p)} by multiplying it by a factor ε, such that 1≦ε≦1/{circumflex over (p)}.
{circumflex over (p)}→ε{circumflex over (p)} (40)
The increased {circumflex over (p)} makes it less likely to encounter negative values in Eq. (38). In addition, the noise amplification becomes bearable. While it is clear from Eq. (36) that using a larger {circumflex over (p)} decreases the noise, this is not the major contribution of this move, since we advocate using only a small bias. Rather, conditioning is obtained mainly thanks to the influence of this step on {circumflex over (t)}. As seen from Eqs. (27,29), increasing p leads to an increase of {circumflex over (t)}. This eliminates the instances of {circumflex over (t)}→0 in Eq. (36). The noise in pixels corresponding to distant objects is then.
σrecovered(z=∞)=√{square root over (2)}σ(1−1/ε)−1 (41)
Thus the unveiling process amplifies noise in pixels corresponding to distant objects, but this amplification decreases with the increase of ε.
How does this bias affect the recovered intensity and color? It can be shown that when ε>1, the estimated value of objects at infinity is
{circumflex over (L)}objecteffective(z=∞)=B∞ (42)
Thus, intensity and color of the raw unpolarized image are automatically retained in the recovered pixels at infinite distance. Thus, the result will have a bluish background, as is perceptually expected from an underwater image. In addition, the original raw value is retained at z=0 where
{circumflex over (L)}objecteffective(x, y)=Îtotal(x, y) (43)
At all other distances, the inverse solution (Eq. 28) is moderated by weighting it with the “best polarized” raw image. By using a small value of ε>1 the result is very similar to Eq. (28), but stabilized at the problematic pixels. This parameter offers the user a degree of freedom in deciding how aggressive the scene recovery should be. This is analogous to parameters used in standard regularization techniques, where typically a user-selected parameter trades a data-fitting term against an image-smoothness term. A benefit of this simple technique is that it does not impose blur for noise control, contrary to standard regularization. However, we do not claim this to be the optimal method for limiting noise amplification. We currently pursue further development of this issue. In our experiments we used ε=1.13.
We now compare our results to prior methods that rely on natural illumination. First we discuss the standard image processing procedures of histogram equalization and unsharp masking. Then, we discuss prior polarization based methods. Both histogram equalization and unsharp masking are space invariant methods, and are therefore useful when contrast degradation is uniform across the raw image. As previously described, underwater the contrast loss is strongly affected by the distance and is therefore spatially variant. As shown here, the effectiveness of the standard image processing method is limited. Unsharp masking is a linear enhancement method in which the image Laplacian is added to the original image. This method sharpens the image, but only slightly improves the long-range visibility, while increasing noise. It is noted that unsharp masking can be applied to our recovered image, should we want to sharpen it in trade of noise. Histogram equalization of the raw image may improve the parts of scene, but on the other hand decrease the image quality in other parts. Furthermore, because the histogram of an image depends on the objects in the scene, the histogram of a partial frame image is different than that of the whole frame, leading to inconsistent results. Moreover, histogram equalization of color images is ill defined, and therefore colors are generally distorted by this operation, rather than being recovered.
The DOP and polarization difference imaging (PDI) methods are based on imaging through a polarizer. DOP and DPI use Imax and Imin differently than our method, and implicitly assume that polarization is associated with the signal rather than the veiling light, contrary to our method. In our experiment, these prior methods perform poorly, and do not recover the scenes at all, especially at long distances. Recently, a different equation was suggested to enhance underwater polarization images (see P. C. Y. Chang, J. C. Flitton, K. I. Hopcraft, E. Jakeman, D. L. Jordan, and J. G. Walker, “Improving visibility depth in passive underwater imaging by use of polarization,” Applied Optics, vol. 42, pp. 2794-2802, 2003). That method is formulated for grayscale images, using ad-hoc formulation that does not resemble inversion of the image degradation process. The results of this formation have been unsuccessful in our experiments.
As a by-product of the radiance recovery process, we get an estimate of the distance map of the scene. From Eq. (30) the distance z is estimated as a function of (x, y) up to a global scale factor c, which is the attenuation coefficient. It is given by
c{circumflex over (z)}(x, y)=−ln[1−{circumflex over (B)}(x, y)/B∞] (44)
If a transmissiometer is used, we can know what the attenuation coefficient c is. Nevertheless, even if c is unknown, we can quantitatively determine the relative distances in the scene. Thanks to the relative distance map, we can compare two image regions and determine that one of them is, say, three times as distant from the camera as the other one. Note that if we know the distance to a single point, we can set the scale of the relative distance map. We can then derive the absolute distance to any point in the scene, as well as the attenuation coefficient.
Analyzing the noise-sensitivity yields the noise STD of the recovered distance as
Eq. (45) indicates that this noise is greatly amplified if B∞→0. This understanding helps us know which color bands are most reliable for the distance estimation. Recall that the illumination of scenes underwater typically suffers from low energy in the red portion of the spectrum, thus B∞red<<B∞blue. Hence, the distance recovery at the blue portion of the spectrum can be expected to be more reliable. Therefore, in our experiments, we derived the relative distance maps using Eq. (44), based on the blue channel.
The recovered range map can be used to render the scene (recovered, raw or processed otherwise) from viewpoints other than the ones used during acquisition.
In our experiments, we used the estimated range map to create a 3D surface, and changed the viewpoint.
To summarize, we presented a method that can compensate for degradation effects occurring in underwater vision. The method is based on a simple analysis of one or more (in the example we used a pair on images from which information on the relative distances of different parts of the scene can be obtained (in the example we acquired this information through the use of a polarizing filter). Since it is physics-based, the method also recovers information about the scene structure (distances). The algorithm can exploits natural lighting.
The practical ability to see farther into the water with high clarity may change some of the common practices of underwater photography. Underwater photography has focused on very short distances using wide angle lenses or macro lenses. However, the practical ability to see farther through water turbidity enables zooming on distant objects. This aspect is particularly important in marine biology research, where close encounters may affect animal behavior.
We describe hereinafter the experimental set up we used. Several considerations have dictated the configuration of our underwater imaging system. They arise since we make quantitative photometric measurements with the polarizer. For reduced noise in the scene radiance measurements, the camera should preferably have a linear radiometric response, besides being of low noise. For flexibility in acquisition, we used a system (camera and housing) having full control of the camera parameters (exposure time, aperture, etc.), while it is placed in a watertight housing. A polarizing filter can placed inside the housing, or outside. If the polarizer is inserted in the housing (
Typical ports for underwater photography are either of dome shape, or flat. Consider
Note that the above description of the experimental set up and technical considerations involved serves as an example, and in no way limits the scope of the present invention, as defined by the claims or their equivalents.
It should also be clear that a person skilled in the art, after reading the present specification could make adjustments or amendments to the attached Figures and above described embodiments that would still be covered by the following claims.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IL05/00180 | 2/13/2005 | WO | 7/31/2007 |
Number | Date | Country | |
---|---|---|---|
60544194 | Feb 2004 | US |