Multi layered image enhancement technique

Information

  • Patent Grant
  • 9053531
  • Patent Number
    9,053,531
  • Date Filed
    Thursday, February 28, 2013
    11 years ago
  • Date Issued
    Tuesday, June 9, 2015
    9 years ago
  • CPC
  • Field of Search
    • US
    • 382 260000
    • 382 266000
    • 382 252000
    • 382 254000
    • CPC
    • G06T2207/200192
  • International Classifications
    • G06K9/40
    • G06T5/00
    • Term Extension
      92
Abstract
A system for enhancing an input image including receiving an input image and filtering the input image with a plurality of non-linear smoothing filters providing a respective plurality of filtered outputs. The system processes a plurality of the filtered outputs with respect to at least one of another of the filtered outputs and the input image to determine a plurality of detail layers. The system filters the plurality of detail layers with a plurality of non-linear smoothing filters providing a respective plurality of smoothed layers. The system adjusts the plurality of smoothed layers in such a manner that regions closer to an edge are enhanced to a lesser extent than regions farther from an edge and combining the adjusted the smoothed layers to provide an enhanced output image.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

None.


BACKGROUND OF THE INVENTION

The present invention relates generally to a multi-layered image enhancement technique.


A multitude of different techniques have been developed for processing and filtering two-dimensional images and two-dimensional video sequences. In particular, many image processing techniques have been developed to modify a relatively low resolution image to a relatively high resolution image or otherwise enhancing the existing resolution, while maintaining image clarity without the introduction of excessive artifacts.


If an image is blurred or degraded by a well-understood process, such as shot noise occurring during transmission, the image can usually be enhanced by developing a model of the source of degradation, then reconstructing the original image using the model. However, in many circumstances, a source of degradation of the image cannot be modeled and, hence, the image cannot be faithfully reconstructed. Thus, generic image enhancement is problematic since the type of noise content therein may be hard to determine or otherwise characterize.


What is desired therefore is a multi-stage non-linear enhancement technique that is suitable for a variety of different types of noise content.


The foregoing and other objectives, features, and advantages of the invention may be more readily understood upon consideration of the following detailed description of the invention, taken in conjunction with the accompanying drawings.





BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS


FIG. 1 illustrates an overshoot at an edge region.



FIG. 2 illustrates an exemplary multi-stage non-linear enhancement technique.



FIG. 3 illustrates another multi-stage non-linear enhancement technique.





DETAILED DESCRIPTION OF PREFERRED EMBODIMENT

Referring to FIG. 1, when an image enhancement technique processes an edge based region illustrated by the step like function often the sharp transitions of the edges results in an overshoot in the resulting enhanced image. The overshoot tends to become even more pronounced when the image enhancement technique attempts to significantly enhance the image. When using non-linear enhancement techniques the likelihood of introducing significant overshoot tends to become more likely. Avoiding the introduction of substantial overshoot, particularly along edge regions of the image, permits the image enhancement technique to introduce additional enhancement that would have otherwise resulted in significant image artifacts. Moreover, the technique should be computationally efficient so that it may be applied to image content in real time without significant computational resources, even when using a non-linear enhancement technique.


Referring to FIG. 2, an exemplary multi-channel image detail enhancement technique is illustrated. A series of video input frames 200 are provided to the image detail enhancement technique. If desired, the system is likewise applicable for the enhancement of a single input image rather than a series of images.


The input frame 200 is provided to a non-linear smoothing filter λ1202. As a general matter, non-linear filters tend to be more effective than linear filters, but likewise non-linear filters tend to be more destructive than linear filters if not suitably controlled. The non-linear smoothing filter λ1202 is suitable to reduce the amount of noise in the image. Preferably, the non-linear smoothing filter λ1 is suitable to reduce limited levels of noise in the image, which is suitable for image content that has quite limited amounts of noise. In this manner, the image content is not significantly degraded as a result of the application of the non-linear smoothing filter λ1 for image content with quite limited amounts of noise. The non-linear smoothing filter λ1 provides an output base layer λ1204 which is the result of the smoothing filter.


The input frame 200 is provided to a non-linear smoothing filter λ2210. The non-linear smoothing filter λ2210 is suitable to reduce the amount of noise in the image. Preferably, the non-linear smoothing filter λ2 is suitable to reduce medium levels of noise in the image, which is suitable for image content that has texture based amounts of noise which is a greater amount of noise than the quite limited amounts of noise. In this manner, the image content is not significantly degraded as a result of the application of the non-linear smoothing filter λ2 for image content with texture based amounts of noise. The non-linear smoothing filter λ2 provides an output base layer λ2212 which is the result of the smoothing filter.


The input frame 200 is provided to a non-linear smoothing filter λ3220. The non-linear smoothing filter λ3220 is suitable to reduce the amount of noise in the image. Preferably, the non-linear smoothing filter λ3 is suitable to reduce significant levels of noise in the image, which is suitable for image content that has significant amounts of noise which is a greater amount of noise than the quite limited amounts of noise and a greater amount of noise than the texture based amounts of noise. In this manner, the image content is not as significantly degraded as a result of the application of the non-linear smoothing filter λ3 for image content with such significant amounts of noise. The non-linear smoothing filter λ3 provides an output base layer λ3222 which is the result of the smoothing filter. For example, the output base layer λ3222 may resemble a cartoon image.


The input frame 200 may bypass 224 the non-linear smoothing filters 202, 210, 220. In this manner, a reference image is maintained that does not have any (or an insignificant amount) of filtering applied thereto. In this manner, the image content is not degraded (or insignificantly degraded) as a result of the application of one of the non-linear smoothing filters.


Additional non-linear smoothing filters λn may be included that have different amounts of noise removal, as desired. One or more linear filters may be included, if desired. As it may be observed, each of the different non-linear smoothing filters provides a different amount of smoothing to the same input frame 200. In this manner, the system has available to it a set of output images that have different amounts of smoothing applied thereto, without having to pre-process the image content to determine which particular smoothing filter should be applied at the exclusion of other smoothing filters. In addition, the arrangement of a plurality of different non-linear smoothing filters, each processing the same video frame during overlapping time periods, is suitable for parallel processing.


It is desirable to determine the amount of smoothing that was applied by each smoothing filter. In particular, it is desirable to determine the amount of smoothing that was applied by a particular non-linear smoothing filter relative to one or more of the other non-linear smoothing filters and/or the original input image.


To determine the amount of smoothing applied to the input frame 224 with respect to the output 204 of the non-linear smoothing filter λ1202 may be determined by using a subtraction operation 230. The output of the subtraction operation 230 is a detail layer 1 image 240. The detail layer 1 image 240 includes the differences between the input frame 224 and the output 204 of the non-linear smoothing filter λ1202.


To determine the amount of smoothing applied to the input frame 224 with respect to the output 212 of the non-linear smoothing filter λ2210 may be determined by using a subtraction operation 232. The output of the subtraction operation 232 is a detail layer 2 image 242. The detail layer 2 image 242 includes the differences between the output 204 of the non-linear smoothing filter λ1202 and the output 212 of the non-linear smoothing filter λ2210.


To determine the amount of smoothing applied to the input frame 224 with respect to the output 222 of the non-linear smoothing filter λ3220 may be determined by using a subtraction operation 234. The output of the subtraction operation 234 is a detail layer 3 image 244. The detail layer 3 image 244 includes the differences between the output 222 of the non-linear smoothing filter λ3220 and the output 212 of the non-linear smoothing filter λ2210.


Additional comparisons may be included relative to different images, combinations of images, filtered and non-filtered images, as desired. As it may be observed, each of the different subtraction operations 230, 232, 235, identifies a different amount of noise that was removed as a result of one or more non-linear smoothing filters. Thus, the technique indicates different amounts of noise that exists as a result of different amounts of applied smoothing. In this manner, the system has available to it a set of output images indicating the effects of the different amounts of smoothing applied thereto, without having to pre-process the image content to determine a particular smoothing filter to be applied at the exclusion of other smoothing filters. In addition, the arrangement of a plurality of different non-linear smoothing filters, each using the same input frame, is suitable for parallel processing. Also, the arrangement of a plurality of different subtraction operations, each using a different sets of input, are suitable for parallel processing, and at least partially temporally overlapping processing for the same input frame. Additional subtraction operations (or otherwise) may be included, if desired.


The detail layer 1 image 240 may be provided to a non-linear smoothing filter σ1250. The non-linear smoothing filter al 250 reduces the noise in the resulting detail layer 1 based upon the image content, noise levels, etc. The output of the non-linear smoothing filter σ1250 is a smoothed layer 1260.


The detail layer 2 image 242 may be provided to a non-linear smoothing filter σ2252. The non-linear smoothing filter σ2252 reduces the noise in the resulting detail layer 2 based upon the image content, noise levels, etc. The output of the non-linear smoothing filter σ2252 is a smoothed layer 2262.


The detail layer 3 image 244 may be provided to a non-linear smoothing filter σ3254. The non-linear smoothing filter σ3254 reduces the noise in the resulting detail layer 3 based upon the image content, noise levels, etc. The output of the non-linear smoothing filter σ3254 is a smoothed layer 3264.


The non-linear smoothing filters 250, 252, and 254 are preferably different filters suitable for the corresponding non-linear smoothing filters 202, 210, and 220. In some cases, the non-linear smoothing filters 250, 252, and 254 are the same filter. The non-linear smoothing filters an are suitable for parallel processing, where the same input image may be at least partially be temporally processed in parallel. Additional non-linear smoothing filters may be included, if desired.


The output of the non-linear smoothing filters 250, 252, 254 may be provided to an adjustment process 270. The adjustment process 270 may be, for example, a gamma correction. The adjustment process 270 may result in a noise reduction because primarily (or substantially) only the residual noise content is being adjusted. For example, the adjustment process 270 may selectively not amplify smaller values to the same extent as larger values. In addition, adjustment of the residual noise facilitates selective limiting the amount of the enhancement so that a suitable multiplier can be limited for stronger details to reduce the effects of clipping. The adjustment process 270 may provide the same adjustment to each of the outputs of the non-linear smoothing filters 250, 252, and 254 or it may be different for one or more of such outputs.


The output of the adjustment process 270 may be corresponding adjusted images 272, 274, 276 for each of the input images 260, 262, 264. With each of the adjusted images 272, 274, 276 being a portion of the original input frame 200 they may be summed together to provide the noise related image content that was subtracted from the original input frame 200. In addition, the output of the non-linear smoothing filter λ1204 is also added to the output of the adjustment process 270 to provide an enhanced output image 280.


Any suitable filer or a selection of different filters may be used for the non-linear smoothing filters 202, 210, 220. For example, the non-linear smoothing filters may be a bilateral filter, a weighted least square filter, and/or a guided filter. In general, the filters preferably reduce the modification of image content near an edge relative to areas that are not as near to such an edge. In this manner, the output of the adjustment process 270 near an edge is small or near zero since there is limited need for changes in regions proximate an edge. Thus for regions near an edge the output of the adjustment process 270 is zero or otherwise near zero and therefore the adjustment process 270 will not significantly boost such edges. Without excessively boosted edges the overshooting problem near edges will not substantially occur.


The preferred implementation of the system uses a 1-dimensional operation on the image content.


Referring to FIG. 3, a modified technique includes receiving input video frames 400. An edge preserving filter 410 may receive each input frame 400. Preferably, the edge preserving filter 410 is a single channel filter. The edge preserving filter 410 attenuates the image to a greater extent in regions that are not proximate an edge than in the regions proximate the edge. Therefore, the non-edge regions of the image tend to be small or otherwise zero. The modified technique determines the difference 420 between the input frame 400 and the edge preserving filter 410. The result of the difference 420 is to provide a difference frame where the edge regions of the image are small or otherwise zero. Since the edge regions are typically those that would otherwise result in undesirable artifacts if boosted, the attenuation of the edge regions reduces the likelihood of such artifacts during subsequent processing. The output of the difference 420 is enhanced by an enhancement process 430. For example, the enhancement process may multiple the values by a value or by a linear and/or non-linear function. The enhancement process 430 may further be modified by a user input, if desired. The enhancement process 430 may be based upon the content of the image, if desired. Preferably, the enhancement process 430 is not based upon the content of the image to determine the filtering to be applied. The output of the enhancement process 430 is combined 440 with the input frame 400. The combination 440 may be a summation process, if desired. The output of the combination 440 is the enhanced output image 450.


Preferably, the edge preserving filter 410 is implemented as follows:

















filter[x] = weight_left*input[x−1] + 4 * input[x] +



      weight_right*input[x+1];











filter[x]=filter[x]/(weight_left+weight_right+4), where the weights are calculated as:

















diff_left = input[x]−input[x−1];



  weight_left = LUT( abs(diff_left) );



  and



  diff_right = input[x+1]−input[x];



  weight_right = LUT( abs(diff_right) );










Preferably, filter[x] may be the output of the process at location x. Preferably, input [x] is the input to the process at location x.


LUT(x) may be 2 when x is less than a threshold, 1 when x is less than 2*Threshold (but not less than said Threshold) and 0 if x is greater than or equal to 2*Threshold.


Preferably in one embodiment the enhancement process 430 is computed a multiplication process. In another embodiment the enhancement process 430 is computed as a multiplication and a clip


Preferably, the combining process 440 may be in addition to the output of the enhancement process 430, which may include additional enhancements, where the combining process 440 includes an additional clipping operation.


The terms and expressions which have been employed in the foregoing specification are used therein as terms of description and not of limitation, and there is no intention, in the use of such terms and expressions, of excluding equivalents of the features shown and described or portions thereof, it being recognized that the scope of the invention is defined and limited only by the claims which follow.

Claims
  • 1. A method for enhancing an input image comprising: (a) receiving an input image;(b) filtering said input image with a first non-linear filter that attenuates regions of input image proximate edges to an extent less than regions of said input image not proximate said edges to provide a first filtered image; (i) where said first non-linear filter filters a first pixel of said input image by calculating a first value as the summation of (1) a left pixel weighting multiplied by a left pixel input, (2) a first pixel weighting multiplied by a first pixel input and (3) a right pixel weighting multiplied by a right pixel input, wherein the left pixel weighting corresponds to a pixel to the left of said first pixel of said input image, the first pixel weighting corresponds to said first pixel of said input image, and a right pixel weighting corresponds to a pixel to the right of said first pixel of said input image,(ii) calculating a second value as said first value divided by the summation of (1) said left weighting of said left pixel, (2) said first pixel weighting of said first pixel, and (3) said right weighting of said right pixel,(iii) said left weighting of said left pixel is determined using a look up table based upon the absolute value of the difference between said first pixel and said left pixel,(iv) said right weighting of said right pixel is determined using said look up table based upon the absolute value of the difference between said right pixel and said first pixel,(v) where said look up table has a first value when said absolute value of the difference between said first pixel and said left pixel is less than a first threshold, where said look up table has a second value when said absolute value of the difference between said first pixel and said left pixel is greater than said first threshold and less than a second threshold; and where said look up table has a third value when said absolute value of the difference between said first pixel and said left pixel is greater than said second threshold,(vi) where said look up table has a third value when said absolute value of the difference between said first pixel and said right pixel is less than a fourth threshold, where said look up table has a fourth value when said absolute value of the difference between said first pixel and said right pixel is greater than said fourth threshold and less than a fifth threshold; and where said look up table has a fifth value when said absolute value of the difference between said first pixel and said right pixel is greater than said sixth threshold;(c) processing said first filtered image by determining a comparison between said first filtered image and said input image that has not been filtered by said first non-linear filter to provide a difference filtered image;(d) filtering said difference filtered image with a second non-linear filter that reduces the amount of noise in said difference filtered image in a manner not dependent on the content of said difference filtered image to provide a second filtered image;(e) processing said second filtered image by determining a comparison between said second filtered image and said input image to provide an enhanced output image.
  • 2. The method of claim 1 wherein said first non-linear filter is an edge preserving filter.
  • 3. The method of claim 1 wherein said comparison between said first filtered image and said input image is a difference operation.
  • 4. The method of claim 1 wherein said second non-linear filter is an enhancement process.
  • 5. The method of claim 1 wherein said comparison between said second filtered image and said input image is a difference operation.
US Referenced Citations (15)
Number Name Date Kind
7529425 Kitamura et al. May 2009 B2
7898583 Wakahara et al. Mar 2011 B2
RE43360 Demos May 2012 E
8265409 Katagiri et al. Sep 2012 B2
8369640 Ishiga Feb 2013 B2
8374457 Wang et al. Feb 2013 B1
20010007478 Kim Jul 2001 A1
20070071355 Imai Mar 2007 A1
20080019605 Yea et al. Jan 2008 A1
20080239090 Ohwaki et al. Oct 2008 A1
20090245679 Ohwaki et al. Oct 2009 A1
20110135217 Su et al. Jun 2011 A1
20120062797 Segall Mar 2012 A1
20120093432 Akiyama Apr 2012 A1
20140241646 Seo et al. Aug 2014 A1
Non-Patent Literature Citations (6)
Entry
E. Wharton, K. Panetta. and S. Agaian, “Edge preserving image enhancement using anisotropic diffusion”, Proc. SPIE, vol. 6812, 2008.
Z. Farbman et al., “Edge-Preserving Decompositions for Multi-Scale Tone and Detail Manipulation,” ACM Transactions on Graphics 27(3) (Proc. ACM SIGGRAPH 2008), Aug. 2008, 10 pgs.
K. He et al., “Guided Image Filtering,” The 11th European Conference on Computer Vision (ECCV), 2010, 14 pgs.
C. Tomasi et al., “Bilateral Filtering for Gray and Color Images,” IEEE International Conference on Computer Vision, 1998, 8 pgs.
H. Takeda et al., “Kernel Regression for Image Processing and Reconstructions,” IEEE Transactions on Image Processing, vol. 16, No. 2, Feb. 2007, pp. 349-366.
A. Buades et al., “A Review of Image Denoising Algorithms, with a New One,” Multiscale Modeling and Simulation (SIAM Interdisciplinary journal), vol. 4, No. 2, 2005, pp. 490-530.
Related Publications (1)
Number Date Country
20140241642 A1 Aug 2014 US