The invention relates generally to the field of digital image processing, and in particular to noise reduction in color digital images.
There are many approaches to reducing the presence of noise in digital images. In order to minimize the required amount of computing resources needed, most commercial approaches to noise reduction in digital images are based on using very small regions of the image for each set of noise reduction computations. One of the simplest approaches with practical value is to use a 3×3 square region of pixels centered on the pixel to be noise reduced. This region is known in the literature as a pixel neighborhood or a support region. With only nine pixel values in the neighborhood available for computational purposes, a variety of algorithms can be used for producing a noise-reduced pixel value without undue burden on most computational resources. If the predominant nature of the noise signal being reduced consists of isolated erroneous pixel values, this pixel neighborhood size is generally sufficient for most noise reduction schemes. If the noise signal is more complex, however, and has clumps of erroneous pixel values that are several pixels wide in extent, the 3×3 square region of pixels will probably be insufficient for adequate noise reduction. The 3×3 square region will still permit the reduction of isolated erroneous pixel values, but will be largely ineffective on larger clumps of noise, especially if said clumps are larger than the 3×3 square region itself. The obvious solution is to increase the size of the noise reduction support region. This is generally a viable alternative up to the point at which the additional burden on the computational resources becomes unacceptable. Unfortunately, it is common for the computational limits of the system to be reached before the support region has been allowed to expand to the desired dimensions for adequate noise reduction.
A solution to these computational limitations is to decompose the image to be noise reduced into a series of images with varying spatial resolutions. In the literature this is described as performing a Laplacian pyramid decomposition. The process is simple. The starting image at its initial spatial resolution is referred to as a base image. A copy of the base image is blurred and then subsampled to a lower resolution. This lower resolution image is a new base image at the next level of the pyramid. This process can be repeated until there are insufficient pixels for any subsequent subsampling operation. A residual image is associated with each base image. In order to create the residual image for a given level of the pyramid, the lower resolution base image from the adjacent level of the pyramid is upsampled and subtracted from the base image at the given level of the pyramid. This difference image is called a residual image. A fully decomposed image consists of a set of base images and corresponding residual images. The advantage of this representation of the image is that small support region image processing operations can be applied to each of the base and/or residual images so as to produce the same results as using a very large support region operation at the original image resolution. In the case of noise reduction, this permits the use of, for example, 3×3 square regions at each level of the pyramid to effectively noise reduce larger and larger clumps of noise. Once the individual images of the pyramid have been processed, the image decomposition process is essentially run in reverse order to reconstitute the full resolution image.
There are many examples of related prior art in this field. U.S. Pat. No. 5,488,374 (Frankot, et al.) discloses a pyramid decomposition-based noise reduction method that uses simple linear noise filters tuned for each pyramid level. U.S. Pat. No. 5,526,446 (Adelson, et al.) teaches using steerable noise filters within a pyramid decomposition-based architecture. U.S. Pat. No. 5,729,631 (Wober, et al.) reveals using Wiener filters and discrete cosine transforms to noise reduce in the spatial frequency domain within the framework of a pyramid decomposition. U.S. Pat. No. 5,963,676 (Wu, et al.) describes using wavelet decomposition to accomplish its pyramid decomposition and then using an edge-preserving smoothing filter to perform noise reduction at each pyramid level. U.S. Patent Application Publication No. 2002/0118887 (Gindele) discloses a pyramid decomposition-based noise reduction method that uses modified sigma filters tuned for each pyramid level.
A significant problem with existing noise reduction methods is that they are still very computationally intensive when either the image to be noise reduced is very large (e.g. 14 million pixels) or the noise signal is very large with respect to the genuine image signal. When the image consists of a large number of pixels, the required computing resources scales directly with the number of pixels to noise reduce. Using a pyramid decomposition architecture addresses this liability to some extent. When the amount of noise present is large, then generally more complex noise reduction algorithms, such as median filters, must be imbedded into the pyramid architecture to avoid seriously degrading genuine image information. A double jeopardy situation can exist when both circumstances are present.
What is needed is a method that provides the noise reduction capability of a pyramid decomposition approach without relying on complex noise reduction operations at each level of the pyramid so as to keep the required computational intensity to a minimum. This method must still noise reduce in an effective manner when dealing with images with high levels of noise.
The object of the present invention is to provide a pyramid noise decomposition method which efficiently reduces noise in large digital color images.
This object is achieved in a method of removing noise from a color digital image, comprising the steps of:
(a) receiving an original color digital image including a plurality of pixels represented in a primary-color space;
(b) producing at least one residual digital image and at least one base digital image from the original color digital image, the base digital image having a lower spatial resolution then the original color digital image;
(c) producing a noise reduced base digital image by removing noise from the residual image and the base digital image with a noise reduction filter and combining the noise reduced base digital image with the noise reduced residual image to produce a reconstructed digital image having reduced noise;
(d) transforming the reconstructed digital image into a luminance-chrominance color space;
(e) repeating steps (b) and (c) on the luminance-chrominance digital image to produce a reconstructed luminance-chrominance digital image; and
(f) transforming the reconstructed luminance-chrominance digital image into a noise reduced digital image in the original primary color space.
It is a feature of the present invention to provide an improved computationally efficient way to reduce noise in color digital images that are noisy.
It has been found that by decomposing original color digital into base and residual images and then noise-reducing each of them significant improvements can be made in noise reduction.
In the following description, a preferred embodiment of the present invention will be described in terms that would ordinarily be implemented as a software program. Those skilled in the art will readily recognize that the equivalent of such software can also be constructed in hardware. Because image manipulation algorithms and systems are well known, the present description will be directed in particular to algorithms and systems forming part of, or cooperating more directly with, the system and method in accordance with the present invention. Other aspects of such algorithms and systems, and hardware and/or software for producing and otherwise processing the image signals involved therewith, not specifically shown or described herein, can be selected from such systems, algorithms, components and elements known in the art. Given the system as described according to the invention in the following materials, software not specifically shown, suggested or described herein that is useful for implementation of the invention is conventional and within the ordinary skill in such arts.
Still further, as used herein, the computer program can be stored in a computer readable storage medium, which can include, for example; magnetic storage media such as a magnetic disk (such as a hard drive or a floppy disk) or magnetic tape; optical storage media such as an optical disc, optical tape, or machine readable bar code; solid state electronic storage devices such as random access memory (RAM), or read only memory (ROM); or any other physical device or medium employed to store a computer program.
Before describing the present invention, it facilitates understanding to note that the present invention is preferably utilized on any well-known computer system, such a personal computer. Consequently, the computer system will not be discussed in detail herein. It is also instructive to note that the images are either directly input into the computer system (for example by a digital camera) or digitized before input into the computer system (for example by scanning an original, such as a silver halide film).
Referring to
A compact disk-read only memory (CD-ROM) 124, which typically includes software programs, is inserted into the microprocessor based unit 112 for providing a means of inputting the software programs and other information to the microprocessor based unit 112. In addition, a floppy disk 126 can also include a software program, and is inserted into the microprocessor-based unit 112 for inputting the software program. The compact disk-read only memory (CD-ROM) 124 or the floppy disk 126 can alternatively be inserted into externally located disk drive unit 122 which is connected to the microprocessor-based unit 112. Still further, the microprocessor-based unit 112 can be programmed, as is well known in the art, for storing the software program internally. The microprocessor-based unit 112 can also have a network connection 127, such as a telephone line, to an external network, such as a local area network or the Internet. A printer 128 can also be connected to the microprocessor-based unit 112 for printing a hardcopy of the output from the computer system 110.
Images can also be displayed on the display 114 via a personal computer card (PC card) 130, such as, as it was formerly known, a PCMCIA card (based on the specifications of the Personal Computer Memory Card International Association) which contains digitized images electronically embodied in the card 130. The PC card 130 is ultimately inserted into the microprocessor based unit 112 for permitting visual display of the image on the display 114. Alternatively, the PC card 130 can be inserted into an externally located PC card reader 132 connected to the microprocessor-based unit 112. Images can also be input via the compact disk 124, the floppy disk 126, or the network connection 127. Any images stored in the PC card 130, the floppy disk 126 or the compact disk 124, or input through the network connection 127, can have been obtained from a variety of sources, such as a digital camera (not shown) or a scanner (not shown). Images can also be input directly from a digital camera 134 via a camera docking port 136 connected to the microprocessor-based unit 112 or directly from the digital camera 134 via a cable connection 138 to the microprocessor-based unit 112 or via a wireless connection 140 to the microprocessor-based unit 112.
In accordance with the invention, an algorithm can be stored in any of the storage devices heretofore mentioned and applied to images in order to noise reduce the images.
Referring to
Referring to
The resulting blurred image is then downsampled by a factor of two in both the horizontal and the vertical directions 224. The preferred embodiment performs the downsampling 224 by subsampling the image in a manner that discards every other row and every other column. The resulting image of this downsampling operation 224 becomes the base image for the next level of the pyramid decomposition process 232. Returning to the pyramid decomposition process 210, a copy of the output of the downsampling operation 224 is next upsampled by a factor of two in both the horizontal and vertical directions 222. In the preferred embodiment the up sampling 222 is accomplished by inserting a new row of zeros between each existing row in the output of the downsampling operation 224 and by inserting a new column of zeros between each existing column in the output of the downsampling operation 224. In order to determine the preferred pixel values within these new rows and columns, a convolution is performed with the following convolution kernel:
This kernel will leave the pixel values in the original rows and columns unaltered and will interpolate the pixel values in the new rows and columns using bicubic interpolation. The resulting upsampled image is then subtracted from the original base image 220. This results in the residual image 230 associated with the current level of the pyramid decomposition.
Returning to
The noise reducing directional blur operation 212 in the preferred embodiment is based on a 13×13 cross pixel neighborhood as shown in
For the green channel the classifiers would be
For the blue channel the classifiers would be
The two classifiers are then compared and the direction with the smaller classifier value is chosen as the preferred direction of blurring. In the preferred direction of blurring a one-dimensional blur kernel is used for producing the appropriate blurred pixel value. For the red channel the blur kernel is
For the green channel the blur kernel is
For the blue channel the blur kernel is
Each blurred pixel value is mixed with the original pixel value to produce the final result of the direction blur 212. The mixing method is a weighted average with the following weights:
In this expression, R=red, G=green, and B=blue. The subscript 0=original, subscript B=blurred, and subscript M=mixed.
y=2500[log10(x+100)−2]
wherein
where R=red, G=green, B=blue, Y=luma, C1=the first chroma channel, and C2=the second chroma channel. The image, now in a logarithm luminance-chrominance space, is now noise reduced within a pyramid noise reduction decomposition/reconstruction framework 254 to be described below. Once noise reduced, the image is converted back to RGB space 256 with the following transform:
Finally, the RGB image is converted back to linear space 258 with the following transform:
It is noted that the transform of block 258 is the inverse of the transform of block 250. Similarly, the transform of block 256 is the inverse of the transform of block 252.
The details of the pyramid noise reduction 254 are given in
Returning to
wherein
The blurred value is mixed with the original value using the following expression:
Once block 270 is complete, the chroma channels of the resulting image are noise reduced with a sigma filter 272a. (The luma channel is left unchanged.) The sigma filter 272a uses a 7×7 square pixel neighborhood. In a well-known variant of the sigma filter, the scaled difference in pixel values between each pixel in the neighborhood and the central pixel is computed and then weighted by a gaussian weighting function. In the preferred embodiment, which assumes 12-bit pixel values, the weighting function w(y) is
This weighting function evaluates to {1000, 607, 135, 11, 0 . . . }, i.e., only the first four values are nonzero. The scaled difference pixel values are computed with the following expressions:
where i refers to the central pixel in the neighborhood, j refers to pixel within the neighborhood, C1 and C2 are the pixel chroma values, TC1 and TC2 are thresholds that have been determined ahead of time to tune the overall aggressiveness of the noise reduction process, and ΔC1j and ΔC2j are the resulting scaled pixel differences. In order to preserve the overall hue of the pixel being noise reduced, the maximum scaled pixel difference is determined:
δj=max{ΔC
where δj is the maximum scaled pixel difference for the jth pixel in the neighborhood. Once δj has been determined for each pixel in the neighborhood, a weighted average is performed:
where C1i′ and C2i′ are the noise reduced chroma values of the resulting central pixel in the neighborhood. These noise reduced chroma channels in combination with the corresponding unaltered luma channel become the noise reduced R0 residual 274a. Blocks 272b-f are identical to block 272a and are used to noise reduce residual images 266b-e and base image 268. The resulting noise reduced residuals images are blocks 274b-e and the resulting base image is block 276.
Returning to
The noise reduction algorithm disclosed in the preferred embodiment of the present invention can be employed in a variety of user contexts and environments. Exemplary contexts and environments include, without limitation, wholesale digital photo finishing (which involves exemplary process steps or stages such as film in, digital processing, prints out), retail digital photo finishing (film in, digital processing, prints out), home printing (home scanned film or digital images, digital processing, prints out), desktop software (software that applies algorithms to digital prints to make them better—or even just to change them), digital fulfillment (digital images in—from media or over the web, digital processing, with images out—in digital form on media, digital form over the web, or printed on hard-copy prints), kiosks (digital or scanned input, digital processing, digital or scanned output), mobile devices (e.g., PDA or cell phone that can be used as a processing unit, a display unit, or a unit to give processing instructions), and as a service offered via the World Wide Web.
In each case, the algorithm can stand alone or can be a component of a larger system solution. Furthermore, the interfaces with the algorithm, e.g., the scanning or input, the digital processing, the display to a user (if needed), the input of user requests or processing instructions (if needed), the output, can each be on the same or different devices and physical locations, and communication between the devices and locations can be via public or private network connections, or media based communication. Where consistent with the foregoing disclosure of the present invention, the algorithm itself can be fully automatic, can have user input (be fully or partially manual), can have user or operator review to accept/reject the result, or can be assisted by metadata (metadata that can be user supplied, supplied by a measuring device (e.g. in a camera), or determined by an algorithm). Moreover, the algorithm can interface with a variety of workflow user interface schemes.
The algorithm disclosed herein in accordance with the invention can have interior components that utilize various data detection and reduction techniques (e.g., face detection, eye detection, skin detection, flash detection).
The invention has been described in detail with particular reference to certain preferred embodiments thereof, but it will be understood that variations and modifications can be effected within the spirit and scope of the invention.
Number | Name | Date | Kind |
---|---|---|---|
4982277 | Katoh et al. | Jan 1991 | A |
5049985 | Outa | Sep 1991 | A |
5146328 | Yamasaki et al. | Sep 1992 | A |
5189511 | Parulski et al. | Feb 1993 | A |
5408342 | Kumada et al. | Apr 1995 | A |
5461655 | Vuylsteke et al. | Oct 1995 | A |
5488374 | Frankot et al. | Jan 1996 | A |
5526446 | Adelson et al. | Jun 1996 | A |
5572329 | Otsubo et al. | Nov 1996 | A |
5579090 | Sasanuma et al. | Nov 1996 | A |
5644662 | Vuylsteke | Jul 1997 | A |
5717791 | Labaere et al. | Feb 1998 | A |
5729631 | Wober et al. | Mar 1998 | A |
5831744 | Kataoka | Nov 1998 | A |
5963676 | Wu et al. | Oct 1999 | A |
6480610 | Fang et al. | Nov 2002 | B1 |
6563946 | Kato | May 2003 | B2 |
6731797 | Prentice | May 2004 | B2 |
6807300 | Gindele et al. | Oct 2004 | B1 |
7181086 | Gindele et al. | Feb 2007 | B2 |
7181091 | Yoda | Feb 2007 | B2 |
7184171 | Hara | Feb 2007 | B2 |
20010045988 | Yamauchi et al. | Nov 2001 | A1 |
20020037101 | Aihara | Mar 2002 | A1 |
20020118887 | Gindele | Aug 2002 | A1 |
20020186390 | Hara | Dec 2002 | A1 |
20030044061 | Prempraneerach et al. | Mar 2003 | A1 |
20030095197 | Wheeler et al. | May 2003 | A1 |
20030228064 | Gindele et al. | Dec 2003 | A1 |
20040096103 | Gallagher et al. | May 2004 | A1 |
20060038899 | Tamaru et al. | Feb 2006 | A1 |
Number | Date | Country | |
---|---|---|---|
20050134734 A1 | Jun 2005 | US |