The present application claims priority from Australian provisional patent application No. 2010901672 filed 21 Apr. 2010 the content of which is incorporated herein by reference.
The disclosure concerns processing of electronic images, such as hyperspectral, or multispectral images. In particular, but is not limited to, a method, software and computer for estimating shape information or a photometric invariant of a location of image of a scene. Examples of processing the image using the estimates include material recognition, shape recovery, recolouring and re-shading of objects represented in the image.
The light source 110 is shown as a wave that oscillates in the two-dimensional plane of the drawing of
The wave of the light source 110 may also be a superposition of a parallel and a perpendicular polarised wave.
In
However, the wave may also be unpolarised light, which means that two components are completely uncorrelated and a representation as in
In this technical field transmitted is understood to mean through the surface of the object. According to the Fresnel Equations [1], parallel- and perpendicular-polarised light is reflected and transmitted differently when hitting a surface. With regards to reflection, perpendicular-polarised light is reflected better than parallel-polarised light. With regards to transmission, parallel-polarised light is transmitted better than perpendicular-polarised light. As a result, a larger portion of the parallel-polarised light 110 in
As indicated by the arrows within the structure of object 101, the transmitted light is reflected several times internally before it is finally emitted through surface 102 to the outside of object 101. Due to the irregular structure of the object, the light emitted from the object 101 is distributed over all directions. This is referred to as diffuse reflection 111.
In one example, the light source 110 emits unpolarised light and the surface 102 only shows diffuse reflection 111 and no specular reflection 112 as the surface is completely matte. When the unpolarised light from light source 110 is transmitted through the surface 102, the light within the structure of the object 101 is still unpolarised. Just before the light of the diffuse reflection is emitted, the light passes through the surface 102 of the object 101. At this point, according to the Fresnel Equations more parallel-polarised light reaches the outside of the object than perpendicular-polarised light. As a result, the light that reaches camera 120 is partially polarised in the direction parallel to the two-dimensional plane of the drawing.
If the object surface 102 is tilted as indicated by the azimuth angle 131, the direction of polarisation of the light that reaches the camera not anymore parallel to the two-dimensional plane of the drawing but rotated by a polarisation angle 153. In fact, the polarisation angle 153 is equal to the azimuth 131 of the surface. As a result, the azimuth 131 of the surface 102 can be determined by determining the angle of polarisation 153.
As mentioned above, the degree of polarisation changes with the zenith angle 130. As a result, the zenith angle 130 can be determined by determining the degree of polarisation.
The azimuth 131 and zenith 130 angles completely define the normal vector 103 of the surface 102. Knowing the normal vector 103 at all points of the surface 102 allows the reconstruction of the shape of the object 101. However, direct measurement of the polarisation angle 153 is inaccurate and the Fresnel Equations are also dependent on the refractive indexon coefficient of the surface, which is unknown in most real world applications.
Any discussion of documents, acts, materials, devices, articles or the like which has been included in the present specification is solely for the purpose of providing a context for the present disclosure. It is not to be taken as an admission that any or all of these matters form part of the prior art base or were common general knowledge in the field relevant to the present disclosure as it existed before the priority date of each claim of this application.
In a first aspect there is provided a computer implemented method to estimate shape information or one or more photometric invariants of a location of a scene as represented in a hyperspectral or multispectral image data, the image data indexed by wavelength and polarisation filter angle, the method comprising:
(a) for each wavelength index, estimating a polarisation angle from the image data;
(b) estimating the shape information or photometric invariants based on the estimated polarisation angle for each wavelength index.
It an advantage of this method that greater accuracies can be achieved in the estimated shape information and/or photometric invariants by using wavelength-indexed data.
It is a further advantage of the method that surface information or photometric invariant can be estimated based upon polarisation in a single-view multi-spectral imagery. Further, by relying on the polarisation angle for the estimation, the method is insensitive to changes in illumination power and direction.
Step (a) may comprise fitting intensity information of the image data to a function of the filter angle, for example cosine, to estimate the polarisation angle.
Step (b) may further comprise estimating the azimuth of a surface normal of the scene at the location based on the estimated polarisation angle for each wavelength index, and estimating the shape information based on the estimated azimuth. That is, the spectral variation of the phase of polarisation is used to estimate the azimuth.
Step (a) may comprise optimising a difference between intensity information of the image data and a function of the filter angle to estimate the polarisation angle for each wavelength index and determine a residual error for each wavelength index, and step (b) may further comprise estimating the azimuth based on the polarisation angles for each wavelength index weighted by the residual error for each wavelength index.
Step (a) may further comprise estimating a degree of polarisation for each wavelength index based on intensity values, for example maximum and minimum, of the function with reference to the estimated polarisation angle for that wavelength index. That is the wavelength indexed degree of polarisation is based on the Fresnel equation.
Step (b) may further comprise estimating the zenith of a surface normal of the scene at the location based on the estimated degree of polarisation for each wavelength index.
Step (b) may further comprise estimating the photometric invariant of the scene at the location based on the estimated degree of polarisation for each wavelength index.
Step (b) may further comprise optimising a difference between a model and the estimated degree of polarisation for each wavelength index, wherein the model parameters include the zenith and photometric invariant parameters. It is an advantage that the zenith and photometric invariant can be determined simultaneously and that the method does not rely on a predefined value of the photometric invariant.
Optimising the difference between a model and the estimated degree of polarisation for each wavelength index may comprise determining co-efficient of a dispersion equation.
Estimating the photometric invariants may be further based on a constraint over the wavelength index for the photometric invariant. That is, enforcing a wavelength dependent dispersion equation on the photometric invariant makes the problem well posed.
The photometric invariant may be the refractive index.
The method is repeated from each location in the scene represented in the image data. A location is represented in the image data as one or more pixels. The image data may taken from a single view point and/or using an unpolarised light source.
The shape information may comprise a surface normal.
The wavelength index may be comprised of one or more wavelengths used in the image.
In a second aspect there is provided a computer system to estimate shape information or one or more photometric invariants of a location of a scene as represented in a hyperspectral or multispectral image data, the image data indexed by wavelength and polarisation filter angle, the system comprising:
(a) a processor to estimate for each wavelength index a polarisation angle from the image data and to estimate the shape information or photometric invariants based on the estimated polarisation angle for each wavelength index.
The computer may be local to a sensor that captures the image data.
The optional features of the method described above similarly apply to the computer system.
In a third aspect there is provided software, that when installed on a computer causes the computer to perform the method described above.
The filter 200 may be seen as an arrangement of vertical slots. When the light wave 201 enters the filter, only a component with polarisation parallel to the slots will be transmitted through filter 200. In this example, the parallel-polarised light wave 203 is transmitted through filter 200 and exits the filter 200 as parallel-polarised output light wave 220.
When the filter 200 is gradually rotated by filter angle 211 the parallel-polarised light wave 203 is gradually attenuated by the filter 200 and the perpendicular-polarised light wave 204 will be gradually transmitted. At a filter angle of 90 degrees, the parallel-polarised light wave 203 is completely blocked while the perpendicular-polarised light wave 204 is fully transmitted.
This effect is used in the following where a surface of an object is illuminated by unpolarised light. As described with reference to
When the output light wave 220 is captured by camera 230, the only information available from the camera is the intensity of the light wave 220. If the parallel-polarised light wave 203 of the input light wave 201 has a higher amplitude than the perpendicular-polarised light wave 204 then a maximal intensity Imax will be observed by the camera for a filter angle of θ=0°. Accordingly, a minimal intensity Imin will be observed for a filter angle of θ=90°. The intensity between the maximum and the minimum can be described by a cosine function as shown in
The aim is to determine the three unknowns Imin, Imax and φ from multiple measurements of I at known filter angles θ. There are several alternative methods of determining Imin, Imax and φ in (1) from a successive sequence of images captured at several filter angles θ. By capturing the same scene with three or more filter orientations, the cosine function can be fitted to the measured intensities using a numerical nonlinear least square fitting algorithm. This method is, however, not efficient and especially in case of a large number of pixels this leads to a high computational cost since the optimisation has to be performed per pixel.
Alternatively, these parameters can be obtained making use of the method in [2], where three images at 0°, 45° and 90° with respect to a reference axis are acquired so as to compute the phase shift φ, intensity I and degree of polarisation ρ. However, the method in [2] is susceptible to noise corruption since it employs only three images. Here, we employ an alternative that yields a stable estimation of the phase shift φ, intensity I and degree of polarization ρ by solving an overdetermined linear system of equations. We commence by rewriting (1) in the following vector form:
After collecting N≧3 measurements (i.e. per pixel) at three or more filter angles θ, one arrives at the following overdetermined linear system
I=Ax (3)
where
Equation (3) is well-constrained since the number of equations N≧3 is no less than the number of unknowns. Moreover, the coefficient matrix A depends solely on the filter angle which allows for an efficient solution of (3). Solving (3) means minimising the difference between the measured intensity I of the image data and the cosine function
Having obtained the solution for x=[x1; x2; x3]T, one can recover the maximal intensity Imax and minimal intensity Imin on the cosine and the phase of polarisation as
As discussed above, the maximal intensity on the cosine curve is reached at the polarisation angle φ. Therefore, the azimuth angle α is either the same as the polarisation angle φ or differs from it by π [1], i.e. α=φ or α=φ±π. To disambiguate the azimuth angle between the two possibilities, and φ±π, we assume convexity on the surface under observation. Under this assumption, surface normal vectors point in the opposite direction to that of image gradients. One can determine the azimuth angle as the one between φ and φ±π that yields the closer orientation to the negative gradient direction.
However, due to weak polarisation that causes a drift in polarisation angle φ estimation, the result of the above method depends on the wavelength λ of the light. In order to improve accuracy, the same image is captured at the same filter angle θ for multiple wavelengths in a multispectral image.
Each pixel has an associated pixel location u within each sub-image. One pixel 311 at location u1 in the top sub-image of image 310 is highlighted. At the same location u1 in the top sub-image of image 320 and 330 are highlighted pixels 321 and 331 respectively.
The following method determines the azimuth α or zenith θ of the surface normal vector or the refractive index for one pixel location u. The method will be explained in one example of pixel location u1. In order to determine the azimuth α or zenith θ of the surface normal vector or the refractive index for all pixel locations, the method is simply repeated independently for all other pixel locations.
A cosine function 401 is fitted to the measurements 411, 421 and 431. The fitting usually results in differences 412, 422 and 432 for each of the measurements 411, 421 and 431 respectively. Based on these differences, a residual error ε is determined. This error is quantified as the L2-norm of the residual error ε of (3), ε=∥I−Ax∥2. From the cosine function 401 a maximal intensity Imax 402, a minimal intensity Imin 403 and the polarisation angle φ 404 are determined. Note that these values cannot be determined directly from the measurements 411, 421 and 431 since the measurements most likely do not fall exactly onto the maximum or minimum points.
In order to determine a single azimuth angle α per pixel, for each pixel a weighted average of the wavelength-indexed polarisation angles φ is computed, such that a
polarisation angle associated with a larger residual error ε is assigned a smaller weight ω than a polarisation angle φ associated with a smaller residual error ε. This is achieved by using the Epanechnikov kernel such that
where
and h is a bandwidth parameter.
Since the azimuth angle α is a directional quantity, instead of averaging the polarisation angles φ directly, we estimate the mean of the sines and cosines of the polarisation angles φ for each pixel location as follows
Where <•>λ denotes the mean value across wavelengths.
The estimated azimuth angle α at pixel u then becomes
The result is an azimuth angle α that reflects measurements for a wide range of different wavelengths and is therefore less prone to inaccuracies.
Joint Estimation of the Zenith Angle and Refractive Index
As mentioned above, the zenith θ angle can be determined by the degree of polarisation detected by the camera. The Fresnel Equation of the degree of polarisation is
where θ (u) is the zenith θ angle of the surface normal at location u, n(u, λ) is the material dependent refractive index of the same location at wavelength λ, and F⊥(u, λ) and F∥(u, λ) are the perpendicular and parallel Fresnel reflection coefficients respectively. Note that the wavelength-indexed degree of polarisation ρ=Imin/Imax has been determined as described with reference to
The only unknowns in (6) are the azimuth α and the refractive index n(u, λ). Since the refractive index n(u, λ) is different for each wavelength, θ and n cannot be determined by fitting (6) to the image data captured for different wavelengths. Each layer of data sub-structure 503 in
where r(u,λi) is the shorthand notation for (Imin(u,λi)/Imax(u,λi))1/2
In order to overcome the problem of too many unknown variables, a further constraint can be provided on the refractive coefficient n. Cauchy's dispersion equation provides a good model for the wavelength dependence of the refractive index:
This dispersion equation relates the refractive index to the wavelength with coefficients Ck(u), kε{1, . . . , M} that are characteristic for the material. With this representation of n(u, λ) the cost function E(u) in (7) can be reparameterised as a function of M+1 variables, including the zenith θ and the dispersion coefficients Ck(u), kε{1, . . . , M}. If the number of M dispersion coefficients is chosen such that M+1≦K, where K is the number of wavelengths, then the nonlinear least square problem can be solved numerically by standard line-search or trust-region methods in order to minimise the difference in equation (7) between the model of equation (6) and the estimated wavelength-indexed degree of polarisation.
Having obtained azimuth angle α and zenith angle θ of the surface normal vectors, the shape under study can be recovered by means of a surface integration method such as in [3]. On the other hand, the refractive index can be recovered from the dispersion equation using the dispersion coefficients Ck(u) and (8).
With the result of this fitting, an azimuth is estimated 704 by computing a weighted average of the wavelength-indexed polarisation angles for that point. The weight associated with each polarisation angle α is based on the respective residual error ε.
As an alternative or in addition to estimating the azimuth α the degree of polarisation p is estimated 705. This is also performed for each wavelength resulting in a wavelength-indexed degree of polarisation ρ(λ). A model function is then fitted 706 to the wavelength indexed degree of polarisation ρ(λ). The model function is a Fresnel Equation where the refractive index n is replaced by Cauchy's Dispersion Equation.
The method 700 then combines 706 the azimuth α and zenith θ to form a surface normal. The surface normal of all pixel locations is then integrated to recover 707 the shape.
In the example of
A computer system shown on
The received images and filter angles are stored in local memory 808(b) by the processor 810. The images may be stored and used in this method in the compact representation form as described in WO 2009/152583.
The processor 810 uses the software stored in memory 808(a) to perform the method shown in
The software provides a user interface that can be presented to the user on a monitor 812. The user interface is able to accept input from the user (i.e. touch screen), such as the number of images to be taken for different filter angles or the number of parameters of the model for the refractive index. The monitor 812 displays to the user the shape information recovered by the processor performing the method. In one example, the shape information is displayed as a three-dimensional model. The monitor 812 also displays information about the refractive index.
The user input is provided to the input/out port 806 by the monitor 812. The selected image is stored in memory 808(b) by the processor 810. In this example the memory 808(b) is local to the computer 802, but alternatively could be remote to the computer 802.
The estimated surface information and photometric invariants of an image can be used in pre and post processing of the image data.
Processing of the image includes, but is not limited to:
Image editing
re-illumination, such as colour manipulation to change an image from warm to cold, or change illumination spectrum of an image to change the illumination from daylight to sunlight, or sunlight to tungsten
re-shading
re-colouring, for example to change a black and white image to colour based on the properties of a known colour set or applying the reflective properties of one image to another
light source re-positioning
material modification
highlight removal
surface rendering
material modification
Shape estimation
recovery of shape an object or scene captured in the image
Material recognition or classification
material, pattern or object recognition and classification
Hardware calibration
improve photometric calibration, such as of sensor 800 that captured the image
Although in
The method for estimating the shape information or photometric invariants may also be directly implemented in hardware, such as an application specific integrated circuit or a field programmable gate array. This hardware may even be integrated with the sensor 800 into a single integrated circuit.
Applications of the methods described here include to the fields of:
digital photography, such as image editing
manufacturing, such as quality and production control.
product analysis, such as determining whether a vehicle had been in an accident, and
surveillance, such as face identification and tracking.
The spectral image can be converted to a colour band representation, such as RGB (Red, Green, Blue), and in that sense the methods described here can be used to generate colour images.
It will be appreciated by persons skilled in the art that numerous variations and/or modifications may be made to the specific embodiments without departing from the scope as defined in the claims.
It should be understood that the techniques of the present disclosure might be implemented using a variety of technologies. For example, the methods described herein may be implemented by a series of computer executable instructions residing on a suitable computer readable medium. Suitable computer readable media may include volatile (e.g. RAM) and/or non-volatile (e.g. ROM, disk) memory, carrier waves and transmission media. Exemplary carrier waves may take the form of electrical, electromagnetic or optical signals conveying digital data steams along a local network or a publically accessible network such as the interne.
It should also be understood that, unless specifically stated otherwise as apparent from the following discussion, it is appreciated that throughout the description, discussions utilizing terms such as “estimating” or “processing” or “computing” or “calculating”, “optimizing” or “determining” or “displaying” or the like, refer to the action and processes of a computer system, or similar electronic computing device, that processes and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
The present embodiments are, therefore, to be considered in all respects as illustrative and not restrictive.
[1] Hecht, Eugene, Optics, 2nd Ed, Addison Wesley, 1987 [2] L. B. Wolff. Polarization vision: a new sensory approach to image understanding. Image Vision Computing, 15(2):81-93, 1997. [3] Frankot, Robert T. and Chellappa, Rama. A Method for Enforcing Integrability in Shape from Shading Algorithms. IEEE Trans. Pattern Anal. Mach. Intell., 10(4):439-451, 1988.
Number | Date | Country | Kind |
---|---|---|---|
2010901672 | Apr 2010 | AU | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/AU2011/000458 | 4/20/2011 | WO | 00 | 11/7/2012 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2011/130793 | 10/27/2011 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5028138 | Wolff | Jul 1991 | A |
6987566 | Smith | Jan 2006 | B1 |
7742168 | Elton | Jun 2010 | B2 |
20070222781 | Kondo | Sep 2007 | A1 |
20080158550 | Arieli et al. | Jul 2008 | A1 |
20100303344 | Sato et al. | Dec 2010 | A1 |
Number | Date | Country |
---|---|---|
WO 2009152583 | Dec 2009 | WO |
Entry |
---|
Vimal Thilak, David G. Voelz, and Charles D. Creusere, “Polarization-based index of refraction and reflection angle estimation for remote sensing applications”, Applied Optics, vol. 46, No. 30, Oct. 2007, pp. 7527-7536. |
Firooz A. Sadjadi, “Passive three-dimensional imaging using polarimetric diversity”, Optical Society of America, Optics Letters, vol. 32, No. 3, Feb. 2007, pp. 229-231. |
Cong Phuoc Huynh and Antonio Robles-Kelly, “Simultaneous Photometric Invariance and Shape Recovery”, IEEE, 12th International Conference on Computer Vision, Oct. 2009, pp. 1757-1764. |
Gary A. Atkinson and Edwin R. Hancock, “Shape Estimation Using Polarization and Shading from Two Views”, IEEE, Transactions on Pattern Analysis and Machine Intelligence, vol. 29, No. 11, Nov. 2007, pp. 2001-2017. |
Th. Gevers and H. M. G. Stokman, “Robust Photometric Invariant Region Detection in Multispectral Images”, International Journal of Computer Vision, vol. 53, Issue 2, Jul. 2003, pp. 135-151. |
Shree K. Nayar, Xi-Sheng Fang and Terrance Boult, “Separation of Reflection Components Using Color and Polarization”, International Journal of Computer Vision, vol. 21, Issue 3, Feb. 1997, pp. 163-186. |
Lawrence B. Wolff, “Polarization-Based Material Classification from Specular Reflection”, IEEE, Transactions on Pattern Analysis and Machine Intelligence, vol. 12, No. 11, Nov. 1990, pp. 1059-1071. |
Daisuke Miyazaki, Robby T. Tan, Kenji Hara, and Katsushi Ikeuchi, “Polarization-based Inverse Rendering from a Single View”, IEEE, Proceedings of the Ninth IEEE International Conference on Computer Vision, vol. 2, Oct. 2003, pp. 982-987. |
Wolff, Lawrence B., “Polarization Vision: a new sensory approach to image understanding”, 1997, pp. 81-93. |
Frankot, Robert T. et al., “A Method for Enforcing Integrability in Shape from Shading Algorithms”, 1988, pp. 439-451. |
International Search Report, PCT/AU2011/000458; completion date Jun. 3, 2011; 4 pages. |
Written Opinion, PCT/AU2011/000458; completion date Jun. 3, 2011; 4 pages. |
Barrois, B. et al., 3D Pose Estimation Based on Multiple Monocular Cues; Computer Vision and Pattern Recognition, 2007; 9 pages. |
Atkinson, Gary et al., Recovery of Surface Orientation from Diffuse Polarization; IEEE Transactions on Image Processing 2006; 14 pages. |
Thilak, V. et al., Polarization-Based Index of Refraction and Reflection Angle Estimation for Remote Sensing Applications; 2007 Optical Society of America, vol. 46, No. 30; pp. 7527-7536. |
Atkinson, Gary et al., Recovery of Surface Orientation From Diffuse Polarization; IEEE Transactions on Image Processing, vol. 15, No. 6; Jun. 2006; pp. 1653-1664. |
Number | Date | Country | |
---|---|---|---|
20130202214 A1 | Aug 2013 | US |