Method related to noise reduction

Information

  • Patent Grant
  • 8310596
  • Patent Number
    8,310,596
  • Date Filed
    Friday, July 13, 2007
    16 years ago
  • Date Issued
    Tuesday, November 13, 2012
    11 years ago
Abstract
A pre-processing of video data before coding and transmission, for a method for noise reduction in High-Density video images. A low implementation-complex filter filters raw pixel data output of a video camera with a low pass property for both spatial and temporal noise and the filter is adapted based on the high frequency content of the picture, leaving most of the image content unaffected by the filtering process.
Description
FIELD OF THE INVENTION

The invention is related to noise reduction of pixels in video images captured by light sensors in a digital video camera.


BACKGROUND OF THE INVENTION

Transmission of moving pictures in real-time is employed in several applications like e.g. video conferencing, net meetings, TV broadcasting and video telephony and are becoming more widespread.


These applications require digital cameras and digital camcorders containing electronic image sensors that capture light for processing into a still or video image, respectively. The quality specifications for the image sensors are increasing as the conventional image resolution is becoming larger. The general factor influencing the image quality the most is the image noise. A large level of noise will generally degrade the quality of the images. Noise may result from different processes taking part in the formation of the image. The characteristics of the noise may also be different. There may be certain patterns due to physical processes or the noise may take the form of a more statistical character often somewhat vaguely named as “Gaussian”.


Most of the noise is created in the image sensors. The image sensors usually include an array of photosensitive elements in series with switching elements. Each photosensitive element receives an image of a portion of the scene being imaged. That portion is called a picture element or pixel. The image obtaining elements produce an electrical signal indicative of the image plus a noise component. Various techniques have been used in the art to minimize the noise, to thereby produce an output signal that closely follows the image. There are two primary types of electronic image sensors, charge coupled devices (CCDs) and complimentary metal oxide semiconductor (CMOS) sensors. CCD image sensors have relatively high signal to noise ratios (SNR) that provide quality images. Additionally, CCDs can be fabricated to have pixel arrays that are relatively small while conforming with most camera and video resolution requirements. A pixel is the smallest discrete element of an image. For these reasons, CCDs are used in most commercially available cameras and camcorders.


CMOS sensors are faster and consume less power than CCD devices. Additionally, CMOS fabrication processes are used to make many types of integrated circuits. Consequently, there is a greater abundance of manufacturing capacity for CMOS sensors than CCD sensors.


To date there has not been developed a CMOS sensor that has the same SNR and pixel pitch requirements as commercially available CCD sensors. Pixel pitch is the space between the centers of adjacent pixels. It would be desirable to provide a CMOS sensor that has relatively high SNR while providing a commercially acceptable pixel pitch.


CCD sensors contain pixel arrays that have multiple rows and columns. When capturing first and second images a CCD must read every row from the array for the first image and then every row in the array for the second image. This is a relatively inefficient approach that contains inherent delays in data retrieval. It would be desirable to decrease the time required to retrieve data from the pixel array.


All light sensors in arrays are exposed to noise due to their spatial and discrete nature. Some light photons simply do not hit the sensors, and consequently contribute to the noise picture. Another noise contributor is the quantization of the light captured by the respective sensors. The representation of the pixel values is digital, i.e. discrete, and conversion from analog to digital data also introduce quantization errors. The amounts of these errors depend on the quantization intervals, which in turn depend on the number of digits representing each pixel. In digital cameras this number has to be limited due to limited processing power and memory space.


The generated noise is often large enough to visually degrade the image. It is therefore desirable to be able to reduce the noise level to obtain a subjectively better image.


At the same time, the rate of picture elements (pixels) in a high quality video signal is very large. For a high definition (HD) video signal this may be in the order of 100 mega pixels/s. This means that a noise reduction method must take computational complexity into account and try to make the procedure simple enough to be feasible for real time implementation.


SUMMARY OF THE INVENTION

The present invention provides a method of noise reduction of a video image being composed of succeeding pictures, including the steps of consecutively calculating a respective delta value (Δ) for each pixel in a present picture of the succeeding pictures from a function of weighted pixels in said present picture and/or in a reference picture relative to said present picture, adjusting said respective delta value to a corresponding respective adjusted delta value (Δ′) according to an amount of high frequency content in said present picture and/or in said reference picture, and adding said respective adjusted delta value for each respective pixel in said present picture.





BRIEF DESCRIPTION OF THE DRAWINGS

In order to make the invention more readily understandable, the discussion that follows will refer to the accompanying drawing.



FIG. 1 shows a basic block diagram of a CMOS active pixel circuit,



FIG. 2 is illustrates a noise filter according to the present invention used in a video conference terminal,



FIG. 3 shows the notation of the pixels in a present and a reference picture used in the description,



FIG. 4 shows a relation between the delta value and the adjusted delta value according to one embodiment of the invention,





DETAILED DESCRIPTION OF THE PRESENT INVENTION

A block diagram of an example of a CMOS active pixel circuit as mentioned in the background section is shown in FIG. 1. The device has a pixel circuit 150, and a column circuit 155. Incident photons pass through the photogate (“PG”) 100 in the pixel circuit 150 and generate electrons which are integrated and stored under PG 100. A number of the pixel circuits are arranged in each row of the circuit. One of the rows is selected for readout by enabling the row selection transistor 102 (“RS”).


The floating diffusion output node 104 (“FD”) is first reset by pulsing reset transistor (“RST”) 106. The resultant voltage on FD 104 is read out from the pixel circuitry onto the column bus 112 using the source follower 110 within the pixel. The voltage on the column bus 112 is sampled onto a first holding capacitor 114 by pulsing transistor SHR 116. This initial charge is used as the baseline.


The signal charge is then transferred to FD 104 by pulsing PG 100 low. The voltage on FD 104 drops in proportion to the number of photoelectrons and the capacitance of FD. The new voltage on the column bus 112 is sampled onto a second capacitor 118 by pulsing SHR 120. The difference between the voltages on first capacitor 114 and second capacitor 118 is therefore indicative of the number of photoelectrons that were allowed to enter the floating diffusion. This voltage difference forms the basis for the quantized output pixel value, one for luminance (luma), and one for chrominance (chroma).


However, the voltage difference is derived from the photons hitting the photogate, but because of the discrete nature of the photogate arrays some photons will hit the space between the photogates, and will not affect the circuit deriving the voltage difference. Due to this fact and the quantization of the analogue voltage, there will always be a noise contribution to the pixel-made image. In video images, the noise will also be temporal, implying that the noise is seen to influence both the temporal and spatial resolution of the video signal.


The main object of the present invention is to reduce as much of the noise as possible and at the same time leave as much of the signal as possible untouched. This is realized by a low implementation-complex filter e.g. arranged between the camera and the codec in a video conference terminal as shown in FIG. 2, with low pass property for both spatial and temporal noise combined with an adaptation of the filter based on the high frequency content of the picture, leaving most of the image content unaffected by the filtering process.


The present invention will in the following be exemplified by describing an embodiment of the invention.


In FIG. 3 is shown notations of an extraction of pixels in a present (new) and a former (old) reference picture of a video image. Pixels ar to ir belong to “Old”, and it is assumed that all pixels in “old” have already been filtered. Then there is a present picture called “New” with the notation ap to ip. In the present definition it is assumed that pixels ap to dp have already undergone filtering.


The filtering process (both for luma and one for chroma) according to the present invention is defined by how pixel ep is modified to ep′. In the presently described embodiment this process is defined as follows:

Δ=(4er+2fr+gr+2hr+ir+ap+2bp+cp+2dp)/16−ep  (1)


A simple modification of n would then be to add the delta value calculated in (1), i.e. ep′=ep


This first iteration of a modification of ep would have a low pass filtering effect both temporally and spatially, since the filtering taps correspond to pixels in both “old” and “new”. Unfortunately this would affect high frequency content of the image too much. Thus, there should preferably be an assumption as to how large the noise contribution to ep may be. If Δ is larger than this contribution, it should be assumed that this is due to real picture content and consequently, ep should not be modified.


One solution is then first to modify Δ to Δ′ in a way similar to the illustration below so that Δ′≈Δ for small Δ and Δ=0 for large Δ. This is illustrated in FIG. 4 as Δ′ as a function of Δ.


This function can analytically be expressed as follows:

Δ′=Δ*max(0;(k0−k1*|Δ|)/k0)  (2)


Where k0 and k1 are constants, and ep′ is still defined as ep′=ep+Δ′.


Note that the filtering described so far is more temporal than spatial since the largest contribution of Δ comes from “Old”.


Hence, a first aspect of the embodiment described above would be to define two versions of Δ, one weighted towards temporal filtering, and one towards spatial filtering. The one to use depends on the difference compared with the pixel value in the same pixel position in the former picture.


To exemplify, Δt is defined as in (1) since it is weighted towards temporal filtering:

Δt=(4er+2fr+gr+2hr+ir+ap+2bp+cp+2dp)/16−ep  (3)


Further, Δs which is the weighted towards spatial and is defined as:

Δs=(4er+2fp+gp+2hp+ip+ap+2bp+cp+2dp)/16−ep  (4)


If |ep−er| is small it is assumed that it is desirable to do more temporal filtering, whereas if |ep−er| is large we want to do primarily spatial filtering. Consequently, the combination of these two expressions can be defined as in the following:

Δ=w*Δs+(1−w)*Δt  (5)

where w depends on the pixel value difference |ep−er|.


In this first aspect of the above described embodiment of the present invention, the filter adapts to temporal or spatial filtering depending on the picture content.


A second aspect of the above-described embodiment takes into account that noise tends to be more annoying in dark areas than in light areas of a picture. In this aspect, the filter strength, which is determined by k1 in (2), varies according to the luminance value, which will be equal to ep for luma filtering and a luma value near ep for chroma filtering. According to this aspect k1 is inversely proportional with ep so as to increase as the luminance decrease and vice versa. Then the filtering process is stronger at darker parts of the picture and more noise is thus removed.


Calculation of Δ above is a typical example, but the invention also covers different weighting factors on the different pixels as well as different selection of pixels. For example, it covers more implementation friendly calculations of Δ and ep′, e.g as the following for Δ:

Δ=2er+(ap+cp+1)>>1+2[(bp+dp+1)>>1]+(gr+ir+1)>>1+2[(fr+hr+1)>>1]−8ep

where >>1 means right shift by 1.


Note that the embodiment described above does not limit the scope of the present invention, and a person skilled in the art might realize many other aspects of the embodiment. The invention is not limited to videoconferencing only, but is applicable in all applications where digital video image capturing is involved.

Claims
  • 1. A method of noise reduction of a video image being composed of succeeding pictures, the method comprising: consecutively calculating a respective delta value (Δ) for each pixel in a present picture of the succeeding pictures from a function of weighted pixels in said present picture and in a reference picture relative to said present picture;adjusting said respective delta value to a corresponding respective adjusted delta value (Δ′) according to an amount of high frequency content in said present picture and in said reference picture; andadding said respective adjusted delta value for each respective pixel in said present picture, the consecutively calculating said delta value comprising: providing a first version of said function of weighted pixels, weighted towards temporal filtering,providing a second version of said function of weighted pixels, weighted towards spatial filtering, andselecting the version to use in dependency on the difference between each version of the delta value and the pixel value in the same pixel position in the former picture.
  • 2. A method according to claim 1, wherein said second version of said function of weighted pixels, weighted towards temporal filtering and calculating said respective delta value for a pixel (ep) in the present picture, is defined as: Δ=(4er+2fr+gr+2hr+ir+ap+2bp+cp+2dp)/16−ep,
  • 3. A method according to claim 1, wherein said first version of said function of weighted pixels, weighted towards temporal filtering and calculating said respective delta value for a pixel (ep) in the present picture, is defined as: Δ=2er+(ap+cp+1)>>1+2[(bp+dp+1)>>1]+(gr+ir+1)>>1+2[(fr+hr+1)>>1]−8ep
  • 4. A method according to claim 1, wherein said function of weighted pixels calculating said respective delta value for a first pixel (ep) in the present picture is defined as: Δ=w*Δs+(1−w)*Δt,
  • 5. A method according to claim 1, wherein said respective adjusted delta value is defined as: Δ′=Δ*max(0;(k0−k1*|Δ|)/k0),
  • 6. A method according to claim 1, wherein the video image is a video image captured by a camera used for video conferencing.
Priority Claims (2)
Number Date Country Kind
20063345 Jul 2006 NO national
20073201 Jun 2007 NO national
PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/NO2007/000270 7/13/2007 WO 00 4/29/2009
Publishing Document Publishing Date Country Kind
WO2008/010724 1/24/2008 WO A
US Referenced Citations (12)
Number Name Date Kind
4694342 Klees Sep 1987 A
5585859 Ward et al. Dec 1996 A
6067125 May May 2000 A
6259489 Flannaghan et al. Jul 2001 B1
6335990 Chen et al. Jan 2002 B1
6847408 Webb Jan 2005 B1
6907143 Ferguson Jun 2005 B2
7034892 Ojo Apr 2006 B2
7199838 Lin et al. Apr 2007 B2
20020101543 Ojo Aug 2002 A1
20050248687 Lee et al. Nov 2005 A1
20050280739 Lin et al. Dec 2005 A1
Foreign Referenced Citations (4)
Number Date Country
97 39572 Oct 1997 WO
98 46009 Oct 1998 WO
01 35657 May 2001 WO
WO 2005122588 Dec 2005 WO
Related Publications (1)
Number Date Country
20090323827 A1 Dec 2009 US