This application claims the benefit, under 35 U.S.C. §365 of International Application PCT/EP2009/067651, filed on Dec. 21, 2009, which was published in accordance with PCT Article 21(2) on Jul. 1, 2010 in English and which claims the benefit of European patent application No. 08305992.3, filed on Dec. 22, 2008.
This invention applies to image capture devices, especially camcorder devices. A well-known issue in video capture is the apparition of visible flicker effects, for instance when filming with fluorescent lights or aiming at computer screens. When using a video capture device with a shutter, shutter synchronization can usually remove such artifacts. When using digital camcorders that generally do not comprise mechanical shutter, the abrupt temporal truncation of incoming photons in their light sensors made by electronic shutters may cause artifacts. This behavior is also responsible for increased motion blur artifacts.
As technology improves, electronic control of video camera shutters has been widely developed. Shutter control allows for instance to control the amount of motion that is recorded in a picture or in a video frame, by limiting exposure time prior to acquisition. Exposure time is determined by limiting the sampling time of the electronic sensor of the camera. As such control is completely electronic with no mechanical movement, this control can be very precise and reliable. However, from a digital signal processing theory point of view, because of their ultra-fast, “binary” (fully opened/fully closed) behavior, electronic shutters can present nonlinearity-related time-instability issues: the abrupt truncation of sensor illumination in time domain happens to generate a temporal ringing artifact known in signal processing theory as the “Gibbs” phenomenon.
In space domain, the effects of such abrupt truncations are well known, as, for instance, the geometrical limits of a lens or of an optical filter. Such effect are well known spatial artifacts as, for instance, sidelobes of diffraction pattern. Such artifacts can be avoided, for instance, by using so-called “apodized” optical filters. Apodization literally means “removing the foot”. It is the technical term for changing the shape of a mathematical function. It is also called tapering. In optical design jargon, an apodization function is used to purposely change the input intensity profile of an optical system. This technique is classically used in optics to remove the sidelobes of a diffraction pattern.
The sharpness of an image capture device or of a component of an optical system (lens, film, image sensor, scanner, enlarging lens, etc.) is characterized by a parameter called Modulation Transfer Function (MTF), also known as spatial frequency response. It is well known to add apodized filters to get apodized lens aperture (or apodized stop), in order to improve the MTF of lenses: according to US2007/139792, “a perfectly apodized aperture is an aperture for which light transmission T varies along its radius x as a Gaussian curve”.
For the capture of still pictures, it is known to use a long exposure. Long exposures are not adapted for video sequences. It is also known to distribute such long exposure over a plurality of successive smaller exposures, the integration of the plurality of exposures giving the long exposure. This integration or averaging also permits to be a time-windowing function, such as a Gaussian, that weights the different exposures near the center of the long exposure time more strongly. See for instance http://en.wikipedia.orq/wiki/Multiple exposure.
In time domain, specifically for the capture of video sequences, i.e. for the capture of successive images of a scene notably with moving objects, at least two types of temporal ringing artifacts may occur:
An object of the invention is to avoid or at least limit these temporal ringing artifacts that are generally generated when using digital video image capture devices, notably for the capture of a scene with moving objects.
For this purpose, the subject of the invention is a device for video capturing successive images of a scene comprising moving objects, said device comprising at least one image sensing device having an active surface made of an array of pixels and able to sense each of these pixels by providing a readout signal representative of the combination of intensity and duration of illumination of said pixel, an optical system able to form successively said images of said scene on said active surfacesuch as to illuminate said active surface, and driving means able to drive said image sensing device and to assign a value to each pixel of each formed successive image for said video capture, wherein, an image sensing period of illumination of said active surface being predetermined for said video capture, said driving means are adapted:
The image sensing period corresponds to the exposure time of each image of the scene to capture. Preferably, said image sensing period is inferior to 60 ms, which is a common requirement for the capture of video sequences.
The image sensing device may be for instance a CCD imager or a CMOS imager. Generally, each pixel of the imager is a light sensor.
The readout signals provided by the image sensing device may an electric voltage or a digital signal. Electric voltages are generally converted into digital values before being summed and weighted.
The optical system generally comprises an objective lens and may comprise prisms, notably when the video image capture device comprises a plurality of image sensing device, generally one for each primary color, red, green and blue.
Thank to the invention, temporal aliasing issues and temporal ringing artifacts are prevented or at least limited for the capture of video sequences.
In this invention, a new image sensor device matrix readout scheme with time-sequenced gain control is then proposed, thus providing a better truncation (or “apodization”) operation of the electronic shutter. Such a readout scheme that is implemented by the driving means of the video image capture device of the invention allows advantageously to simulate or emulate a mechanical shutter. As a matter of fact, as illustrated on
Digital video image capture devices have “electronic shutter”, corresponding generally to the frame integration period of the CCD imager used as image sensing device. Such an electronic shutter opens and closes abruptly the flux of light, as illustrated by the rectangle function shown on
Knowing that the sin c(.) function is polynomial (it has a rapidly convergent power series representation) and that a coarse approximation of the root of the above equations is fc=1/2T, this equation can be numerically solved using the Newton-Raphson iterative method for finding roots:
As an example, a shutter speed of T=1/50 (corresponding to a standard configuration in PAL camcorders) generates a low pass filter with cutoff frequency at fc=22.15 Hz, thus preventing aliasing issues, since in this case harmonics over the
threshold are filtered, with an ensured infinite rejection (−∞dB) over 50 Hz (1/T).
By using a apodization function to apodize the response of the video image sensing device during the exposure time or image sensing period of the capture of each image of a video sequence, the low pass filtering effect will still be improved and side lobe effects avoided or at least limited. Several apodization functions may be used to embody the invention. Unlike the rect(•) function (see above), these apodization functions go smoothly to zero at the beginning and the end of an integration of the captured light, thus preventing nonlinearity-related instabilities.
Preferably, said readout signal is proportional to the integral of intensity of illumination of a pixel over the duration of illumination of this pixel. Consequently, the combination of intensity and duration of illumination is represented by this integral.
Preferably, said driving means are also adapted to vary the number of successive illumination periods according to said predetermined image sensing period. The readout speed of the image sensing device may be a limitation for this number of successive illumination periods. In this situation, for very short image sensing periods, this number may be as low as four, and for longer image sensing periods, this number may be eight or higher.
Preferably, said apodization function is chosen in the group consisting of the Hanning apodization function, the Blackman apodization function, and the raised cosine apodization function.
The subject of the invention is also a method to capture successive images of a scene comprising moving objects, wherein each image is captured during an image sensing period, comprising the followings steps:
The invention will be more clearly understood on reading the description which follows, given by way of non-limiting example and with reference to the appended figures in which:
The present description illustrates the principles of the invention. It will thus be appreciated that those skilled in the art will be able to devise various arrangements that, although not explicitly described or shown herein, embody the same present principles and are included within its spirit and scope.
The video image capture device according to a main embodiment of the invention comprises:
The driving means are adapted to capture successive images according to the following method. An important feature of the invention is notably adding at least an apodization step to the usual digital shutter process. In this specific embodiment, the apodization function that is used is a Hanning function, which presents very little sidelobe effects (the gap over 30 dB is between the main lobe and the first sidelobe). Any other apodization function or window can be used to implement the invention. The instrument function for Hanning apodization can be algebraically written as follows:
The algebraical expression for an apodization-enabled shutter process of an incoming video signal x(t) of a given pixel will be expressed as follows:
To implement equation 2, we have to consider the fact that, for each pixel of any successive image, a CCD imager (as a CMOS imager anyway) will only provide portions of the incoming illumination signals that are integrated over the duration of illumination of this pixel. This limitation comes from the usual readout electronics of these imagers. In the current embodiment of this invention, the exposure time of each of the successive images, i.e. the image sensing period, is split into eight equal portions, i.e. eight successive illumination periods. The number of successive illumination periods is generally limited by the imager readout speed and by the noise robustness possibilities. A weighting constant coefficient is extracted to approximate the Hanning function for each of these successive illumination periods, which leads to the following decomposition of the value assigned to each pixel by the driving means:
where, as illustrated on
for a given illumination period n and term 2 is the value provided by the CCD imager for this pixel, that is the integral of the illumination signal of this pixel over this given illumination period. Therefore, to each pixel, is assigned the sum of the readout signals (term 2) provided by this pixel for each of said successive illumination periods, this sum being weighted by the Hanning apodization function (term 1).
Table 1 shows the values (and a rational approximation) of the approximation of the Hanning function for each illumination period, referenced 0 to 7, to be applied during shutter operation as a time-sequenced gain control.
Each approximation of the Hanning function is then a weighting coefficient, also called an apodization coefficient.
In reference to
The above readout scheme is defined for a readout-with-discharge behavior of the imager, and therefore can apply to both CCD imagers and CMOS imagers.
In reference to
Thank to these specific readout schemes using apodization coefficients, temporal aliasing issues and temporal ringing artifacts are prevented or at least limited in the captured scene comprising the successive images.
These and other features and advantages of the present principles may be readily ascertained by one of ordinary skill in the pertinent art based on the teachings herein. It is to be understood that the teachings of the present principles may be implemented in various forms of hardware, software, firmware, special purpose processors, or combinations thereof.
Although the illustrative embodiment above have been described herein with reference to the accompanying drawings, it is to be understood that the present principles are not limited to this precise embodiment, and that various changes and modifications may be effected therein by one of ordinary skill in the pertinent art. All such changes and modifications are intended to be included within the scope of the present principles as set forth in the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
08305992 | Dec 2008 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2009/067651 | 12/21/2009 | WO | 00 | 6/21/2011 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2010/072718 | 7/1/2010 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
4774588 | Noda et al. | Sep 1988 | A |
5198902 | Richards et al. | Mar 1993 | A |
5585848 | Hieda et al. | Dec 1996 | A |
6778210 | Sugahara et al. | Aug 2004 | B1 |
7498555 | Koh et al. | Mar 2009 | B2 |
20030030744 | Baer | Feb 2003 | A1 |
20040016919 | Daiku et al. | Jan 2004 | A1 |
20040109069 | Kaplinsky et al. | Jun 2004 | A1 |
20060012830 | Aiso | Jan 2006 | A1 |
20060158399 | Kimura et al. | Jul 2006 | A1 |
20060187328 | Lim | Aug 2006 | A1 |
20070139792 | Sayag | Jun 2007 | A1 |
20080063294 | Burt et al. | Mar 2008 | A1 |
20080143840 | Corkum et al. | Jun 2008 | A1 |
Number | Date | Country |
---|---|---|
0479213 | Apr 1992 | EP |
0506031 | Sep 1992 | EP |
0912047 | Apr 1999 | EP |
1067778 | Jan 2001 | EP |
1176807 | Jan 2002 | EP |
1764736 | Mar 2007 | EP |
1814073 | Aug 2007 | EP |
1924085 | May 2008 | EP |
88484 | Jun 1993 | IL |
2252371 | Oct 1990 | JP |
888794 | Apr 1996 | JP |
100660869 | Dec 2006 | KR |
1779931 | Dec 1992 | RU |
W09908388 | Feb 1999 | WO |
WO-9962249 | Dec 1999 | WO |
WO-0135657 | May 2001 | WO |
WO-2005059832 | Jun 2005 | WO |
WO-2006088722 | Aug 2006 | WO |
WO-2006094871 | Sep 2006 | WO |
WO-2008004230 | Jan 2008 | WO |
Entry |
---|
Dachille et al., “High-Degree Temporal Antialiasing”, Computer Animation 2000, May 3, 2000, pp. 49-54. |
Wikipedia: “Multiple Exposure”, Internet Citation, [Online], Sep. 29, 2008. |
PCT Search Report Dated Mar. 17, 2010 and EPO Search Report Dated Feb. 20, 2009. |
Weijer et al., “Learning Color Names for Real World Application”, IEEE Transactions on Image Processing, vol. 18, No. 7, 2009, pp. 1512-1523. |
Number | Date | Country | |
---|---|---|---|
20110254998 A1 | Oct 2011 | US |