This invention relates to computer-aided synthesis of visual images. More particularly, this invention relates to a system and method for synthesizing image pairs from a single high dynamic range image used to produce a single percept.
An additional image domain is introduced when extending from monocular display to binocular display. Existing binocular display systems only utilize this additional image domain for stereopsis. Human vision is not only able to fuse two displaced images, but also fuse two images with difference in detail, contrast and luminance, up to a certain limit. This phenomenon is known as binocular single vision. Humans can perceive more visual content via binocular fusion than just a linear blending of two views.
The popularity of 3D movies leads to the wide availability of low-cost binocular display devices. While the dual display domains (one for the left eye and the other for the right eye) double the space of visualization, existing binocular displays only serve for one type of binocular vision, stereopsis. Another commonly experienced binocular vision phenomenon in daily life is binocular single vision (or singleness of vision), i.e. images from two eyes are fused and perceived as a single percept, even though these two images are different (
Unlike binocular display, high-dynamic range (HDR) display is less accessible to the general public. Even though tone mapping can be adapted to present the HDR content on a low-dynamic range (LDR) display, there is a tension between displaying large-scale contrast and fine-scale details. Striking a good balance is always challenging.
It is known how to combine different images, such as stereopsis [O'Shea 1983], which differs from combining different images from two eyes into a single vision or percept [von Helmholtz 1962]. While binocular single vision occurs only in a small volume of retinal area around where eyes are fixating, stereopsis occurs at places even where eyes are not fixating (when images of object appear double). It was discovered that such a fusion process is a non-linear combination of luminance, contrast, and color. To prove this, MacMillan et al. [2007] measured the interocular brightness response using asymmetrical neutral density filters and Baker et al. [2007] measured the interocular contrast response using sine-wave gratings.
This non-linear fusion is a complicated neurophysiological procedure and is generally regarded as a combination of binocular fusion and suppression [Ono et al. 1977; Steinman et al. 2000]. Binocular fusion is a process of superimposing and combining similar content from the two views into one unified and stable percept, which happens when the two views are similar or identical (
The above discomforts can greatly impede or destroy the visual experience. Hence, what is needed is an assessment tool for binocular viewing comfort. The need for an assessment tool seems to suggest an image similarity metric. There are several existing metrics, including mean squared error (MSE), structural similarity (SSIM) [Wang et al. 2004], perception-oriented metrics Visible Difference Predictor (VDP) [Daly 1993] and its extension High Dynamic Range Visible Difference Predictor (HDR-VDP, HDR-VDP-2) [Mantiuk et al. 2005; Mantiuk et al. 2011]. Known metrics consider the visible difference between two images when the observer looks at these images with both eyes. However, these existing metrics do not consider the binocular vision in which the left eye and right eye of observers are presented with two different images. An obvious shortcoming of existing metrics can be illustrated by binocular suppression (
Several sophisticated tone mapping techniques have been proposed to generate LDR images from HDR images. Reinhard [2006] provided a comprehensive survey on tone mapping techniques, ranging from sigmoidal compression to image appearance model, and to perception and engineering-based methods. Tone mapping methods can be roughly classified into global and local operators. Histogram adjustment methods and adaptive logarithmic mapping such as [Larson et al. 1997; Drago et al. 2003] are two main categories of global operators. On the other hand, there are also several prevalent local operators, such as bilateral filtering approach [Durand and Dorsey 2002], gradient domain optimization [Fattal et al. 2002] and perceptual-based contrast processing [Mantiuk et al. 2006].
Known tone mapping operators can be used as building blocks for a binocular tone mapping framework to generate two LDR images that optimally increase the human-perceivable visual content without triggering discomfort.
It desirable to produce two views that are as different as possible in order to retain more visual content from the source HDR. However, there is a limit on the difference between two views. When such limit is exceeded, binocular viewing discomfort appears, and even worse, binocular single vision may fail. Such viewing discomfort [Lambooij et al. 2009] is an important health issue receiving much attention due to the wide availability of 3D displays.
The following references provide background to the present invention:
According to the invention, in a computer-based graphics imaging system, binocular tone mapping is employed with an optimization-based binocular tone mapping framework to produce low-dynamic range (LDR) image pairs to generate a binocular single vision image preserving more human-perceivable visual content than an arbitrary single image while simultaneously avoiding viewing discomfort. More specifically, given a tone-mapped LDR image (left, without loss of generality) of a source HDR image, the inventive framework optimally synthesizes its counterpart (right) of the image pair from the same source HDR image. The extra image space of the binocular framework is employed to enhance the image content (e.g. enlarge the contrast, boost the brightness, or enhance more details), while visual information of the rest of the image is preserved in another image space, the binocular tone mapping being guided by a binocular viewing comfort predictor (BVCP) metric in order to assure stable formation of binocular single vision and avoid viewing discomfort. The initial BVCP metric is designed to avoid visual discomfort for most individuals. The optimization-based framework generates the correct tone-mapped image with the goal of maximizing the overall visual information content under the guidance of a BVCP.
The invention is believed to be the first attempt in the graphics area to enrich the visual experience with binocular single vision. Thus the invention provides image improvement for existing tone mapping techniques.
The disclosed metric is believed to be a novel metric for measurement of the viewing discomfort due to binocular content difference.
The utilization of an optimization-based binocular tone mapping framework to produce low-dynamic range (LDR) image pairs that preserve more human-perceivable visual content than an arbitrary single image while simultaneously avoiding viewing discomfort is believed to be unique.
Further according to the invention, it is disclosed how to utilize an existing LDR binocular display to simultaneously present the contrast and details in HDR images by a binocular tone mapping framework. Given a source HDR image and its tone-mapped LDR image (such as a left-eye view) that is generated by known tone mapping techniques, the inventive framework optimally synthesizes its counterpart (right) LDR image in the binocular image pair. Thus, through the phenomenon of binocular single vision, the two LDR images aggregate to present more human-perceivable visual content than any arbitrary single LDR image.
The inventive framework has been evaluated with four state-of-the-art tone mapping methods, including bilateral filtering approach [Durand and Dorsey 2002], gradient domain HDR compression [Fattal et al. 2002], adaptive logarithmic mapping [Drago et al. 2003] and perceptual-based contrast processing [Mantiuk et al. 2006].
The invention will be better understood by reference to the following detailed description with accompanying drawings.
The file of this patent contains at least one drawing/photograph executed in color. Copies of this patent with color drawing(s)/photograph(s) will be provided by the Office upon request and payment of the necessary fees.
According to an embodiment of the invention, a novel technique in computer graphics utilizes a human vision phenomenon and proposes a binocular tone mapping framework. The proposed framework generates a binocular low-dynamic range (LDR) image pair that preserves more human-perceivable visual content than a single LDR image using the additional image domain. Given a tone-mapped LDR image (left, without loss of generality), the inventive framework optimally synthesizes its counterpart (right) in the image pair from the same source HDR image. The two LDR images are different, so that they can aggregate to present more human-perceivable visual richness than a single arbitrary LDR image without triggering visual discomfort. For example, in
An overview of the system 10 implemented on a computer processor 11 with associated memory 13 according to the invention is illustrated in
In operation, and referring to
A pseudo-code listing of the operation is given below:
In operation, a first LDR image is first tone-mapped from the source HDR image using a selected tone mapping operator in the first channel
The goal of the system is to generate the optimal right LDR image using the same or same type of tone mapping operator as the left channel, by maximizing the visual difference between two views while avoiding any visual discomfort. The optimization framework processor 22 generates the optimal right-side view by iteratively adjusting the tone mapping parameters in a gradient ascent fashion. The iteration continues until the objective value converges to produce the second or right image as in the second channel 20 on the lower path in
The objective function comprises two metrics, a visible difference predictor (VDP) 32 such as taught by Daley [Daly 1993] (the content of which is incorporated by reference for all purposes and a binocular viewing comfort predictor (BVCP) 30, as hereinafter explained (Section 4), corresponding to the total amount of visual content and the viewing comfortability, respectively. The BVCP function is set as a hard constraint such that any LDR image pair leading to discomfort is rejected. During each iteration the tone mapping parameters are adjusted (in an iterative feedback process) to follow the gradient ascent direction, based on the VDP of the current LDR image pair. If the LDR image pair cannot pass the BVCP metric test, the step size of gradient ascent is reduced.
Since the gradient ascent approach is sensitive to initial values, it can be easily trapped by a local optimum. To increase the chance of finding the global optimum, n random seeds are distributed in the search space and the search is for n paths accordingly. The final output is the best among the n pairs. In current configurations, n=10.
While tone mapping limits the visual content for display, binocular tone mapping offers one additional domain for visualization. An obvious way to exploit the dual image domain is to ensure the two images span different dynamic ranges in order to maximize the total amount of visual content. However, there is a limit on the difference between the two images. When such limit is exceeded, rivalry and high level suppression appear and lead to visual displeasures, such as flicker, nausea and fatigue.
To avoid such viewing discomfort, the BVCP metric is used. The BVCP metric was determined by examination of results of known psychophysical studies [Levelt 1965; Liu et al. 1992; Steinman et al. 2000; Kooi and Toet 2004]. In particular, the inventive BVCP measures the difference between the left and right images, in terms of contour density, contrast, luminance, and color. The limit of contour density can be relaxed when the contrast is small. This phenomenon is called failure of rivalry, found by Liu et al. [1992]. Such relaxation allows more visual information to be represented. This phenomenon is used in the inventive BVCP design.
Whether two corresponding points in the left (L) and right (R) images can be fused into a single percept or not requires a complex consideration. The decision is not solely based on the local colors of the two points, but based on the visual agreement of neighborhoods surrounding the two points. This neighborhood is called Panum's fusional area in the literature of vision science. Fusional area is an area on the retina. In graphics terminology, fusional area occupies a constant solid angle subtended at the eye. When it is stimulated together with a given single retinal point in the other eye, a single binocular percept is formed [Steinman et al. 2000]. When both of an observer's eyes fixate at a pixel with position (i,j) in both left and right images, the whole fusional areas (both left and right) surrounding the position (i,j) must be considered for fusion stability, in terms of contour, contrast, luminance and color differences.
By projecting the fusional area to the screen via the pupil (
rƒ=γ tan θ·d (1)
where θ is the maximal retinal disparity, which is around 60 to 70 arcmin [Wopking 1995] for most people. Hence, the fusional area ξ in the image is a circular neighborhood of radius rƒ. To simplify computation, approximate the neighborhood by a rectangle of 2rƒ×2rƒ pixels instead. Note that rƒ is a function of viewing distance d, which means the viewing distance affects the viewing comfort.
As an observer's eyes may fixate at an arbitrary pixel position in the image pair, all pixel pairs from the LDR image pair must be considered in their corresponding fusional areas in order to measure the fusion stability. All following assessments consider the fusional area ξ at an arbitrary position (i,j). The corresponding fusional areas in the left and right images are denoted as L and R.
It has been found that the contour (edge) difference is more important than the contrast or color differences in determining binocular fusion [Treisman 1962].
Contour fusion does not require contours in the two corresponding fusional areas L and R to be exactly the same, because our eyes have two mechanisms, motor fusion and sensory fusion. Motor fusion superimposes corresponding points or similar contour by the movement of our eyes. After the alignment by motor fusion, a neurophysiological process, sensory fusion, combines the two views into one unified percept. As illustrated in
Contour (edge) has different definitions in different domains. Herein, a meaningful contour segment is defined as obvious color difference expanding to, or beyond a specified visual degree. To figure out as many contour segments as possible, a scale space representation is applied to the fusional area. Constructing a pyramid from the original fusional area, first Fourier transform the fusional area to the frequency domain. Then apply a pyramid of low-pass filters in this frequency domain (
Define the contour based on the concept of visual acuity (VA). A well-designed diameter of letter ‘C’ or height of letter ‘E’ in Landolt C and Tumbling E Chart is five times the visual acuity. Hence, one defines contour segment to be meaningful when its occupied visual degree reaches or exceeds 5·VA (a typical human eye has the visual acuity to separate 1-2 arcmin). Thus the lowest level of mesa pyramid involved in the computation should be level S with a width of less than or equal to (tan θ/tan(VA/24)). With such a mesa pyramid, obvious color differences between two adjacent pixels in each level is regarded as a segment of visible contour. For each pair of fusional areas L and R, two mesa pyramids are set up respectively. The k-th level of the pyramids are denoted as Lk and Rk, where kε[5, K−1].
To identify a contour, we measure the color difference ΔEc between the adjacent pixels. It is defined as a 2-norm distance of their colors in LAB color space. Consider
ΔLk(p1,p2)=ΔEc(Lk(p1),Lk(p2)) (2)
Similarly, one can obtain ΔRk(p1, p2) for the right image. Now one can predict their fusion state Sk(p1, p2) (i.e. whether a contour is recognized by an observer) by looking up a decision table (Table 1), with ΔLk(p1, p2) and ΔRk(p1, p2) as query. In this table, JND stands for a constant called just noticeable color difference and OCD stands for another constant called obvious color difference. According to existing studies [Chen and Wang 2004; Lin and Jane 2009], we set JND=2.3 and OCD=6.0. If both ΔLk(p1,p2) and ΔRk(p1, p2) are less than OCD, no contour is recognized. So, the fusion state is set to be 0 (stands for ‘no contour’). If both of them reach OCD, two obvious contour segments are recognized and fused together. Hence the fusion state is set to be 1 (stands for ‘match’). Confusion appears only when one of the color differences reaches OCD, while the other falls below JND, in that case the fusion state is set to be −1 (stands for ‘not match’). If one reaches OCD and the other falls between JND and OCD, it can still be regarded as support for existence of contour. Hence the state is 1 too.
The k-th level of two corresponding fusional areas (Lk and Rk) are regarded as fusible if the count of ‘1’ (match) is equal to or larger than the count of ‘−1’ (not match). On the other hand, if the count of ‘1’ is smaller than that of ‘−1’, Lk and Rk are not fusible. If both Lk and Rk contain no contour, contour fusion takes no effect, and the final fusion state is determined by other levels. Here, record the contour fusion state of two fusional areas at the k-th level in the pyramid with a state variable Bk as follows (0 stands for ‘no contour’, 1 stands for ‘fused’, and −1 stands for ‘not fused’),
Note that the contour fusion states at higher levels override the lower ones, so
B
cƒ
k-1
=B
cƒ
k if Bcƒk≠0 (4)
The final fusion state of two fusional areas is,
B
cƒ
=B
cƒ
S (5)
The differences in luminance can be generally referred to as contrast. Two types of contrast can influence the binocular single vision. They are contour contrast and regional contrast. Contour contrast coexists with contour if it can be detected by human eye. Matched contour pair generally helps the fusion except when their contrasts are obviously inversed. So, revise Sk(p1,p2) before evaluating Eq.3 as follows. When a pair of matched contour has obviously inversed contrast, their fusion state changes to −1 (‘not match’). The revised function is:
S
k(p1,p2)=−1, if |C(Lk(p1),Lk(p2))|>OCD
and |C(Rk(p1),Rk(p2))|>OCD
and C(Lk(p1),Lk(p2))·C(Rk(p1),Rk(p2))<0 (6)
where C(c1, c2) computes the lightness difference between the pixel pair c1 and c2.
Regional contrast refers to the contrast between two regions (in our case, the two corresponding fusional areas). It has a relatively smaller impact on the viewing comfort (compared to contour fusion), unless the two regions differ too much [Kooi and Toet 2004]. A restrictive constraint for regional contrast is preferred, such that the average color difference between two fusional areas L and R must be less than a distinct color difference (DCD). According to the existing study [Carter and Huertas 2010], we set DCD=34. The viewing comfort due to the regional contrast (Brc), between two corresponding fusional areas, is assessed by
where L(p) and R(p) are two corresponding pixels located at position p in L and R, respectively.
Normally when contour fusion fails, rivalry or over-suppression occurs and leads to discomfort. However, it was discovered that the occurrence of rivalry depends also on contrast. When contrast is below a certain threshold, a stable single percept always forms regardless of the contour fusion criteria mentioned above. This phenomenon is called failure of rivalry and more likely to happen in low-contrast regions.
where Y1 and Y2 are the normalized luma of c1 and c2, respectively. Γ(Y) maps the normalized luma Yin [0, 1] to the physical measurement in the unit of cd/m2.
Whenever the contrast value is below the contrast threshold, a stable single percept always forms. However, known reports in the literature are either of measurement of the contrast threshold against luminance or of the contrast threshold against spatial frequency. However, the present invention formulates contrast threshold as a function of both luminance and spatial frequency. The procedure is as follows: Construct the contrast function based on the existing findings in the literature. Then fit the plots (blue dots) in
T(l)=min(2, 1.999−0.362 log10(l)+0.026 log102(l)) (9)
T(ƒ)=min(2, log10(3.557−1.334ƒ+1.881ƒ2−0.108ƒ3)) (10)
where l is average luminance measured in cd/m2, ƒ is the spatial frequency measured in cycles per degree, and T represents log percentage contrast threshold. When l=40 cd/m2 and ƒ=5 cy/deg, Eq. 9 and 10 fixate at the same percentage contrast threshold. By assuming the curvature is constant, extrapolate a function of both luminance and spatial frequency as
T(l,ƒ)=min(2, log10(3.557−1.334ƒ+1.881ƒ2−0.108ƒ3)+0.514−0.362 log10(l)+0.026 log102(l)+δ) (11)
where δ is user parameter and is set to a value in [−0.15, 0.15]. Its default value is 0.
Including the failure of rivalry in the metric of the invention allows acceptance of more pairs of LDR images and further maximizes the information content of the optimal pair. To incorporate the failure of rivalry into the metric, the procedure is as follows: First compute the average luminance lL in the left fusional area L as Γ(
T
L
k
=T(lL,ƒk) (12)
Similarly, one can obtain TRk. So their common threshold Tk is min(TLk, TRk). When the contrast is below this threshold Tk, stable fusion always forms and hence Sk (p1, p2) can never be negative (‘not match’). Thus, take a second revision to the fusion state variable Sk (p1, p2), as follows:
S
k(p1,p2)=0, if Sk(p1,p2)<0
and P(Lk(p1),Lk(p2))≦Tk
and P(Rk(p1),Rk(p2))≦Tk (13)
With the contour fusion predictor Bcƒ and the regional contrast predictor Brc defined above, one can now predict the ultimate viewing comfort for an arbitrary pair of fusional areas by:
When B=−1 (‘not fused’), viewing discomfort exists.
So far, only the BVCP test of two fusional areas have been described. An image pair passes the BVCP test only when substantially all pixels pass the BVCP test, i.e. no pixel in the image triggers visual discomfort (B=−1). While it is preferred that all pixels pass the BVCP test, human observers may be able to tolerate certain amount of pixels violating the BVCP without undue discomfort.
In general, the greater the differences between the left and right images, the more visual content they can preserve in aggregate. To measure the visual difference, the visible difference predictor (VDP) [Daly 1993] can be adopted. The VDP has been utilized in various applications [Myszkowski 1998]. Given a pair of images, the output of VDP is a probability of detecting visible difference, V(i,j), at each pixel location (i,j). The overall binocular visual difference E is defined by:
where τ is a user-defined probability threshold and generally set as 75%. H is the Heaviside step function. Ω is the total number of pixels in the image.
With the VDP and the proposed BVCP, the present optimization framework maximizes E without violating the BVCP criteria (
The present framework is independent of the adopted tone mapping operators. In the current implementation, four state-of-the-art tone mapping operators are supported, including bilateral filtering approach [Durand and Dorsey 2002], gradient domain HDR compression [Fattal et al. 2002], adaptive logarithmic mapping [Drago et al. 2003] and perceptual-based contrast processing [Mantiuk et al. 2006]. Both global and local tone mappings are included in these operators. The bilateral filtering approach [Durand and Dorsey 2002] is a two-scale decomposition of the image into a base layer, encoding large-scale variations, and a detail layer. Gradient domain HDR compression [Fattal et al. 2002] manipulates the gradient field of the luminance image by attenuating the magnitudes of large gradients. Adaptive logarithmic mapping [Drago et al. 2003] displays high-contrast scenes by logarithmic compressing the luminance values and imitating the human response to light. Perceptual-based contrast processing [Mantiuk et al. 2006] enhances or compresses the contrast in a visual response space, in which the contrast values directly correlate with their visibility in an image.
To demonstrate the effectiveness of the framework, all four operators with the invention have been tested on a rich variety of test images.
To validate the inventive method, user studies have been conducted. The experiment set-up used a Zalman ZM-M215W 21.5″ 3D LCD display with the highest luminance of around 300 cd/m2 for displaying binocular images. The pixel density of the display is 102 PPI. The observer was asked to sit at a distance of 0.5 meters from the display and to wear a pair of 3D polarized glasses (with transmittance of around 44%). All experiments were conducted indoor with an ambient illumination of around 200 lux.
Visual Richness The first user study evaluated the effectiveness of binocular tone mapping in terms of visual richness. It compared bioptic image pairs (both views are identical) to dichoptic image pairs (the two views are different). To fairly compare, the image of bioptic pair was generated using the “best” parameter values, instead of the default parameter values which may not be satisfactory for certain HDR input. The “best” parameters were determined as follows. For each HDR input, first randomly generate 10 LDR images with 10 sets of parameter values in the recommended parameter range of the particular tone mapping operator. Then, 8 participants were asked to choose the best (in terms of details and contrast) LDR image among them. The top-rated one was then referred to as the image of bioptic pair in the user study. The same image was also referred to as one view of the dichoptic image pair, while the other view was optimally determined by the inventive framework processor system.
All four tone mapping operators (Durand, Fattal, Drago and Mantiuk) and eight different HDR images (32 sets of image pairs in total) were the subject of the experiments in the user study. Thirty-five participants were invited to evaluate these 32 sets of randomly displayed image pairs. In each round, a dichoptic image pair and the corresponding bioptic image pair were chosen for comparison. Note that the bioptic image pair was presented to the participants via the same 3D glasses. These two image pairs were shown in random order, i.e., the participants did not know which one was the dichoptic image pair. Each participant was then asked to select the one he/she preferred. The testers briefly explained to them that he/she could consider the visual richness and/or visual content clarity during selection. However, it was up to the participants' decision in selecting the preferred image pair. To allow the participant to better inspect the image pairs, he/she could press a button to toggle between these two image pairs during the selection. Once the participant made the decision, he/she could press a ‘select’ button.
Table 2 shows the statistics for four tone mapping operators. It is clear that most participants preferred binocularly tone-mapped results (80.4% for Drago, 71.7% for Durand, 75.3% for Fattal and 72.1% for Mantiuk). It also shows that the binocular tone mapping can effectively preserve more visual richness than a single one, especially for the case with Drago's operator. One possible explanation is that global operators, like Drago's, are generally more constrained than the local ones, leading to either more severe loss in details or loss in contrast.
Binocular Symmetry The second user study evaluated whether the effectiveness of binocular tone mapping is symmetric to the left and right eyes. The previous visual richness experiment was again conducted on 20 participants with the same data set (four tone-mapping operators for eight different HDR images, 32 sets of image pairs in total). However, this time these 32 sets of image pairs were evaluated twice. In one of these two evaluations, the LDR image generated by the framework processor was shown to the participants' left eyes. While in the other evaluation, the generated LDR image was shown to participants' right eyes. Participants do not know which of their eyes were shown the generated LDR images.
Table 3 compares the statistics from these two sets of evaluations. It is clear that, no matter which optimal image is shown to which eye, binocular tone mapping can always effectively preserve more visual richness than a single image. From the statistics, the left and right eyes are slightly asymmetric. This conforms to a prior study on ocular dominance [Ehrenstein et al. 2005].
Predictability of BVCP The third user study evaluated how well the BVCP predicts the discomfort limit of binocular vision. The experiment was conducted based on the classical psychophysical methodology, method of adjustment [Norton et al. 2002]. Given an HDR image and a specific tone mapping operator, a sequence of 10 LDR image pairs was presented labeled from 0 to 9. The left and right LDR images of the 0th image pair were equivalent and were tone-mapped from the source HDR using a random parameter set. As the label number increased, the left LDR image remained unchanged throughout the whole sequence, while the right LDR images were generated by linearly increasing/decreasing the values of the most influential parameter of that particular tone mapping operator. The actual increasing/decreasing step size of the parameter value did not matter, provided that the sequence contained discomfort image pairs.
Twenty-two participants were invited to take part in the experiment. They were asked to determine the discomfort limits of all 20 sequences. For each sequence, the participants were asked to search among the 10 image pairs to find the image pair of discomfort threshold, i.e. the image pair they begin to feel discomfort, and any image pairs with lower label numbers that were still acceptable. The participants could move along the sequence by increasing or decreasing the label number. The discomfort thresholds identified by participants were then recorded. To avoid any bias, the first image pair within a sequence shown to the participants was randomly selected among the 10 pairs.
From the statistics, the BVCP metric does not perform equally well with all tested tone-mapping operators. Due to the current conservative design, its prediction is overly conservative for certain observers (
Note that the binocular fusion of color, luminance or contrast difference does not prevent the simultaneous perception of stereopsis [Steinman et al. 2000]. In other words, it is feasible to extend the inventive binocular tone mapping to incorporate the stereopsis.
The current BVCP design as implemented is very conservative. It rejects an image pair if any test of the contour fusion, contour contrast, or regional contrast fails. It rejects the image pair if there is obvious contour error at any level of the mesa pyramid. An image pair passes the BVCP test only if all pixels pass the BVCP test. All of these add up to give a conservative metric. Although the current prediction can effectively avoid visual discomfort for most individuals, it sometimes may not be very close to the user statistics (
The current framework treats all pixels in the image equally. But in reality, human vision is not equally sensitive to every pixel, due to visual attention.
A binocular tone mapping framework system with associated method has been described that is useful to generate a binocular LDR image pair that presents more visual richness than just a single tone-mapped image. The invention employs a novel BVCP metric that can conservatively predict the discomfort threshold. It guides generation of binocularly tone-mapped image pairs, so that visual information content of the image pair can be maximized without triggering visual discomfort. User studies demonstrate the effectiveness of the framework on a wide variety of images. Multiple tone mapping operators have been tested using the framework. Other than the current tone mapping application, the BVCP metric may also be applied in any other applications requiring binocular display. The work serves as the first attempt in graphics applications to maximize the utilization of stereo display system for binocular single vision.
The invention has been explained with respect to specific embodiments. Other embodiments are contemplated without departing from the scope of the invention. Therefore, it is not intended that the invention be limited, except as indicated by the appended claims.
This application claims benefit under 35 USC §119 of U.S. Provisional Patent Application Ser. No. 61/678,732 filed Aug. 2, 2012.
Number | Date | Country | |
---|---|---|---|
61678732 | Aug 2012 | US |