A preview image typically represents a reduced or downsampled version of a larger image. Imaging applications typically present an image preview on a display device for a viewer to approximate image composition, quality, and/or other aspects of the larger image from which the preview image was derived. However, because preview images are the results of downsampling operations, preview images are created with only a subset of the information present in the corresponding larger images. As a result, a user may not be able to ascertain significant perceptual features that are present in the larger image (e.g., noise, blur, depth of field, white balance, bloom effects, etc.) merely by viewing the preview image. In such scenarios, preview images may not adequately address a user's image previewing and browsing needs.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
In view of the above, systems and methods for perceptual image preview are described. In one aspect, a naïve preview image is generated by downsampling a larger image. Perceptual features of the larger image are then detected. Information associated with the detected perceptual features is then incorporated into the naïve preview image to create a perceptual preview image. Since the perceptual preview image incorporates information associated with the detected perceptual features, a viewer of the perceptual preview image will easily detect the presence or absence of such perceptual features in the larger image.
In the Figures, the left-most digit of a component reference number identifies the particular Figure in which the component first appears.
a) illustrates an exemplary arbitrary naïve preview image.
b) shows an exemplary gradient field inside a region of arbitrary naïve preview image of
c) shows exemplary arbitrary structurally enhanced perceptual preview image generated from the arbitrary naïve preview image of
d) shows an exemplary gradient field inside a region of the arbitrary structurally enhanced perceptual preview image of
a) through 3(e) illustrate various aspects of exemplary noise preservation for synthesized incorporation into a preview image to generate a perceptual preview image, according to one embodiment.
Although not required, the systems and methods for perceptual image preview are described in the general context of computer-executable instructions (program modules) being executed by a computing device such as a personal computer. Program modules generally include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types while the systems and methods are described in the foregoing context, acts and operations described hereinafter may also be implemented in hardware.
Perceptual image preview module 112 (“preview module 112”) downsamples a larger image 116 to create a conventional naïve preview image. Preview module 112 then enhances this naïve preview image with additional information associated with the larger image 116 to better show one or more of structure and or perceptual features of the larger image 116. This enhanced naïve preview image is a perceptual preview image 114. More particularly, to generate the perceptual preview image 114, preview module 112 detects and evaluates structure and arbitrary perceptual features (e.g., one or more of noise, blur, depth of field, blooming, white balance, and/or so on) from the larger image 116. Such detected structure and arbitrary perceptual features are shown as respective portions of detected perceptual features 120. Preview module 112 enhances the naïve preview image with information associated with one or more of these detected aspects to allow a viewer to more accurately ascertain the existence or absence of these aspects in the larger image 116.
These and other aspects of the systems and methods for perceptual image preview are now discussed in greater detail.
Preview module 112 augments a naïve preview image of a larger image 116 with structural enhancements to structurally enhanced preview image 114. These augmentation operations strengthen salient edges and flatten weak details in the resulting image. Since salient edges often separate different objects, this approach typically increases the inter-object contrast and reduces intra-object contrast. As a result, image structure becomes more visually apparent and attractive to a users' attention. The more apparent structure enables a user to more readily detect any information associated with other detected perceptual features 120 that are subsequently incorporated into a structurally enhanced perceptual preview image 114.
In this implementation, preview module 112 implements structure enhancement via non-linear modulation in the image gradient domain. Because salient edges in spatial domain have large gradient magnitudes, while weak details correspond to small gradient magnitudes, the problem becomes to increase the large gradients and reduce the small gradients. The increase of large gradients strengthens the important edges, and the reduction of small gradients suppresses the redundant details. By solving a Poisson equation, the image can be reconstructed given the gradient field and the image boundary. The adjustment in the gradient field is thus reflected in the resultant image.
In this implementation, preview module 112 creates a structurally enhanced perceptual preview image 114 by first converting the generated naïve preview image to YUV color space. Preview module 112 then computes the gradient field for the luminance component. The chrominance components are kept intact to guarantee that the image color will not change. A modified sigmoid function is used to modulate the gradient magnitudes:
where Gi is the gradient magnitude for pixel i and G′i is the adjusted gradient magnitude. The gradient direction is not changed since local direction adjustment may destroy the consistency of gradient field. Parameter α controls the maximum adjustment magnitude. When α is set to a value smaller than 1, all the gradients will be suppressed. Parameter k controls the modulation slope. The larger k, the more large gradient magnitudes are magnified, and the more small gradient magnitudes are suppressed. Parameter β defines the threshold to differentiate large and small gradient magnitudes. Parameter β has a larger impact on the result than parameters α and k do. To preserve locally salient tiny details, β is chosen adaptively as follows:
where βg is a global threshold, and βl is a local threshold. βl is evaluated as the average gradient magnitudes in the neighborhood of the pixel, weighted by a Gaussian. βg is evaluated in the same way on the whole image. Equation (2) respects strong edges and favors weak yet salient edges.
a) illustrates an exemplary arbitrary naïve preview image, according to one embodiment.
There are many arbitrary types of perceptual features that may be detected from an image 116. Such perceptual features include, for example, white balance, depth of field, blooming, blocky artifacts caused by image compression, image quality, etc. (Blooming is the affect that a dynamic range pixel is overexposed so that the neighboring pixels are brightened and overexposed). To present perceptual features exhibited by a larger image 116 in a perceptual preview image 114, preview module 112 detects such perceptual features (i.e., shown as detected perceptual features 120) from the larger image 116. The number and types of perceptual features represented by detected perceptual features 120 are arbitrary and a function a particular implementation.
Next, preview module 112 incorporates information (e.g., synthesized/derived information 122) associated with at least a subset of the detected perceptual features 120 into a naïve preview image or a structurally enhanced perceptual preview image 114. When the information is incorporated into a naïve preview image, a new perceptual preview image 114 results. Although the incorporated information can be information that was extracted from the larger image 116, the incorporated information is not necessarily extracted information. Rather, in some or all instances, such information is synthesized or derived from the information provided by the detected perceptual features 120. In view of this, and in this implementation, exact accuracy of detecting perceptual features 120 exhibited by the apparent image 116 is not necessary. Inclusion of information corresponding to the detected perceptual features 100 into the perceptual preview image 114 allows the perceptual preview image 114 to convey more information to a viewer about the larger parent image 116. For instance, by showing that the detected perceptual features exist to some degree in the parent image 116.
Noise is often introduced into an image 116 when the image is captured using a high ISO mode (e.g. ISO 400 or greater), or when a short exposure is used to capture the image. Noise typically appears like color grains that are distributed across the image. Noise usually appears in dark regions. It is assumed that noise has an additive property and complies with uniform distribution. Though this assumption does not always hold if considering various noise generation schemes, it is sufficient for reflecting noise on a structurally enhanced perceptual preview image 114.
To represent noise present in a larger image 116 in a corresponding perceptual preview image 114, preview module 112 detects and stores the noise from the parent image 116 as source noise image using any of multiple well-known noise detection techniques. Based on the assumption of uniform distribution, preview module 112 generates a destination noise image (with dimensions equivalent to the structurally enhanced preview perceptual preview image 114) by randomly sampling from the source noise image. Such source and destination noise images are shown as respective portions of “other program data” 126 (
In one implementation, noise is not detected from a large image 116, but rather only from a uniform smaller region of the larger image 116, for example, from a region that includes few salient edges. The objective is to present noise in the resulting perceptual preview image 114 that looks similar to that one that would be viewed on the larger image 116. It is unnecessary to extract noise from the whole original image 116. For instance, suppose the uniform region Ω is expected to have a size of M×M pixels, where M is of a mild value, say 10d, in which d is the rate of size reduction. Too large M will make the desired uniform region non-existent. Too small M will not provide accurate enough noise.
In one implementation, instead of detecting Ω in the larger image 116 directly, preview module 112 searches a corresponding region Ωd of size (M/d)×(M/d) pixels in the naïve preview image, then maps it to Ω of the original image 116. The naïve preview image will exhibit less noise than present in the larger image from which it was derived. Additionally, because the naïve preview image has a much smaller dimension, the search for Ωd is efficient and reliable. More specifically, preview module 112 divides the naïve preview image into non-overlapped blocks, and selects the block Ωd that has the smallest variance. In most cases, the region Ω that corresponds to Ωd contains no salient edges. Therefore, its high frequency components compose the noise map Ns. Preview module 112 applies discrete stationary wavelet transform to reliably estimate noise Ns. Because the region size is quite small, the computation is very fast.
In one implementation, to produce the noise map Nd with the reduced dimension, preview module 112 utilizes texture synthesis methods to keep the noise spatial distribution in Ns. In practice, preview module 112 randomly generates a noise map Nd from Ns based on the assumption that noise has a uniform distribution. Though Nd does not exactly match the noise distribution in the original image 116, Nd conveys users a similar visual experience as the original one (see
If=Id+γ·Nd,
where Id is the reduced image, and γ is a parameter to control how salient the noise need to be visualized.
Blur in an image 116 occurs when a relative motion happens between an image acquisition device such as a camera and subject(s) of the image or when the subjects are out of focus. In different situations blur may appear over the whole image or just around some regions. Blur reduces image 116 sharpness. However, a naively reduced image may still look sharp, especially when blur in the corresponding parent image from which the naïve preview image is generated is not serious. Preview module 112 detects image blur for presentation in a perceptual preview image 114. This is accomplished by evaluating blur degree locally in the original image 116. In view of the evaluation, preview module 112 blurs a corresponding region in the perceptual preview image 114 to present a corresponding amount of blur. Edge region blur is generally more visible to a user when the whole image 116 is blurred. In view of this, when the whole image 116 is blurred, preview module 112 performs blur manipulation in the perceptual preview image 114 only along its edges.
Preview module 112 performs edge detection on a structurally enhanced perceptual preview image 114. For each edge pixel Ei in the image, preview module 112 estimates its respective blur degree. Blur estimation is based on the observation that blurry edge regions usually have a smaller deviation among gradient angles than sharp edge regions do. More specially, for the edge pixel Ei, the region Ri is located in the original image 116 that is shrunk to this pixel Ei. The gradient angles in the region Ri are then denoted as Ai(k). The amount of blurring Bi at edge pixel Ei is computed as follows:
where D(Ai) is the variance of the gradient angles in the region Ri. The parameters η and α control the estimated amount of blur. These are set empirically to η=3, and α=1.2. These parameters can also be determined subjectively.
Preview module 112 then synthesizes blur in the perceptual preview image 114 in the neighborhood of pixel Ei according to the estimated amount of blur degree associated with the pixel. More particularly, preview module 112 implements blur operations using a Gaussian kernel according to the amount Bi of blur, where Bi is actually as the sigma of the Gaussian. The neighborhood of Ei is then weighted using the Gaussian kernel.
An Exemplary Procedure
At block 410, perceptual image preview module 112 incorporates information associated with at least a subset of the detected perceptual features 120 into a preview image to create a perceptual preview image 114. The particular preview image into which such information is incorporated can either be a naïve preview image or a structurally enhanced perceptual preview image 114. At block 412, perceptual image preview module 112 or a different application that leverages output of the perceptual image preview module 112, presents the perceptual preview image 114 to a user to represent the larger image 116 from which the perceptual preview image 114 was derived. Since the perceptual preview image 114 presents information associated with detected perceptual features of the larger image 116, viewing the perceptual preview image 114 allows a user to make educated determinations as to the quality or other aspects of the larger image 116.
Although the systems and methods for perceptual image preview have been described in language specific to structural features and/or methodological operations or actions, it is understood that the implementations defined in the appended claims are not necessarily limited to the specific features or actions described. For example, information associated with detected perceptual features 120 that have been incorporated into a perceptual preview image 114 can be used to allow a user to assess aspects of the larger image in addition to image quality (e.g., depth of field, etc.). Accordingly, the specific features and operations of system 100 are disclosed as exemplary forms of implementing the claimed subject matter.
Number | Name | Date | Kind |
---|---|---|---|
5641596 | Gray et al. | Jun 1997 | A |
6181834 | Li et al. | Jan 2001 | B1 |
6404901 | Itokawa | Jun 2002 | B1 |
6548800 | Chen et al. | Apr 2003 | B2 |
20030026495 | Gondek et al. | Feb 2003 | A1 |
20040205629 | Rosenholtz et al. | Oct 2004 | A1 |
Number | Date | Country |
---|---|---|
WO0207013 | Jan 2002 | WO |
Number | Date | Country | |
---|---|---|---|
20070196027 A1 | Aug 2007 | US |