The invention relates to a method for re-targeting a three-dimensional image signal, a method for generating a three-dimensional image signal, a system for re-targeting a three dimensional image signal, a system for generating a three-dimensional image signal, a three-dimensional image signal, a data carrier holding a three-dimensional image signal, a computer program for re-targeting a three-dimensional image signal, a computer program for generating a three-dimensional image signal, a computer program product for storing a computer program for retargeting a three-dimensional image signal and a computer program product for storing a computer program for generating a three-dimensional image signal.
Over the last two decades three-dimensional display technology has matured further. Three-dimensional (3D) display devices add a third dimension (depth) to the viewing experience by providing each of the viewer's eyes with different views of the scene that is being watched.
As a result we now have various ways to view three-dimensional image/video signals. On the one hand we have glasses based three-dimensional display system where a user is presented with distinct images for her/his left eye and right eye. On the other hand we have auto-stereoscopic three-dimensional display systems that provide the un-aided eye of a viewer with a three-dimensional view of a scene.
In glasses-based system the active/passive glasses provide a filter for separating the distinct images as presented on the screen for the respective eye of the viewer. In glasses-free or auto-stereoscopic systems the devices quite often are multi-view displays that use light directing means, e.g. in the form of a barrier or lenticular, to direct a left image to the left eye and a right image to the right eye.
In standard 3D cinema glasses are used to separate left and right views for the viewer's eyes. Instead of, for example 24 frames (of image data) per second, in a stereo system 48 frames per second may be provided; i.e. 24 for the left eye, and 24 for the right eye. The left frames of a stereo pair comprise a slightly different view of the same scene than the right frames. This information is combined by the human brain and results in a perceived three-dimensional image or video.
One of the cues used to perceive a three-dimensional image, is parallax, i.e. the apparent displacement in apparent position of an object between the two views. A problem that exists with systems that provide three-dimensional image signals is that the parallax information comprised therein is typically optimized for a particular three-dimensional display system/device. As a result the parallax information contained within the signal may not be appropriate for display on another type of three-dimensional display system/device.
Therefore there is a need for adaptation of an input three-dimensional signal to the capabilities of a three-dimensional display device used for displaying the three-dimensional input signal.
The inventors have observed that in particular for auto-stereoscopic three-dimensional display systems/devices that display a significant depth effect, i.e. when objects between left and right have more pronounced disparities this may result in blurring of the image.
A first aspect of the present invention relates to a method of retargeting three-dimensional video signal data for use on a target three-dimensional display device, the method comprising: receiving the three-dimensional video signal, the three-dimensional video signal in turn comprising three-dimensional video data suitable for driving a multi-view display device, and a reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device. The method further comprises deriving a target disparity mapping based on the reference disparity mapping and characteristics of the reference three-dimensional display device and corresponding characteristics of the target three-dimensional display device and retargeting the three-dimensional video data using the target disparity mapping data into a target three-dimensional video data.
The invention effectively enables an efficient retargeting of the three dimensional video data to a target three-dimensional display device. To this end the present invention includes a reference disparity mapping of a reference three-dimensional display, which may be used by a target three-dimensional display device for retargeting. The reference disparity mapping is dependent on the three-dimensional video data and is preferably time-variable adapting the mapping over time to the image content. The reference mapping may be provided by the content author, whereby it is generated using known characteristics of the reference three-dimensional display device.
In turn at the target three-dimensional display device the corresponding characteristics of the target three-dimensional display device are considered known. These corresponding characteristics may e.g. be available embedded in flash memory, or another non-volatile storage in the target display device, as these characteristics are stable over the display life-time. The display manufacturer, assembler, or another party that has measured/determined such characteristics of the target three-dimensional display device may subsequently store these in the display device.
The target three-dimensional display device may retrieve this data and compare the retrieved data with the characteristics of the reference three-dimensional signal. The present invention allows the use of either one or more standardized display devices, the characteristics of which may also be placed in the non-volatile storage. Using this approach it may suffice to indicate in the three-dimensional video signal which particular standardized reference display the reference disparity mapping is associated with. In case of a single standardized display device, such an indicator may even be redundant.
Alternatively the three-dimensional video signal may comprise the characteristics of the reference three-dimensional display device, thereby providing the system with flexibility and allowing content authors to provide a reference disparity mapping for one or more target display devices.
The target three-dimensional display device may compare the characteristics of the reference three-dimensional display device and its own corresponding characteristics and based thereon decide to generate a target disparity mapping based on the outcome of this comparison and the reference disparity mapping.
As historically three-dimensional display devices over time have achieved better quality 3D output, it is expected that over time the disparity range in which three-dimensional display devices can display content without substantially blurring (e.g. beyond a predetermine blur-metric) will increase. As a result content generated for a current reference display device may not provide the best disparity mapping for a future improved display device, the rationale begin that such content comprises a reference disparity mapping that from an artistic viewpoint best maps the content depth onto the currently available display depth/disparity.
The present invention effectively allows an elegant enhancement, also for future display devices, in that the target three-dimensional display device is provided with the required information to generate a three-dimensional disparity mapping that better suits the target three-dimensional display device.
Notably within the context of the present application depth and disparity are used within the same context. Simplified depth is inversely proportional with disparity and as a result operations in the depth-domain typically have an equivalent in the disparity-domain. Both depth and disparity provide information that relates to image parallax between e.g. a left-eye and a right-eye view of a multi-view (2 or more) video signal.
The three-dimensional image data comprised within the three-dimensional video signal is preferably three-dimensional image data that may be used to generate multi-view output. The three-dimensional image data may e.g. relate to a single view with an associated depth and/or disparity map suitable for image-based rendering, or more preferably comprises a stereo signal with one or more associated depth maps. The latter signal is beneficial because it allows reproduction of stereo content without loss, but also provides the respective views of the stereo signal information that is useful for filling in de-occluded regions when rendering content for n-view displays where n>2.
Furthermore the present invention may also be used in situations where no explicit depth or disparity information is provided, but where the three-dimensional image signal may provide sufficient context to derive a depth/disparity map therefrom in real-time. For example in case of stereo content without depth map depth/disparity maps can be estimated using well known techniques. However, mileage may vary, in that for certain scenes it may be difficult to accurately derive such a depth/disparity map based on stereo content only.
In a preferred embodiment, the reference disparity mapping is generated using characteristics of the reference three-dimensional display device, such that the reference disparity mapping maps the disparity of salient elements in the three-dimensional video data onto a predetermined disparity range of the reference three-dimensional display device suitable for representing said salient elements.
The advantage of this approach is that for those elements in the three-dimensional image signal that typically require a sharp representation; i.e. with a blur below a predetermined blur threshold, this can actually be provided.
In a more preferred embodiment, provided that the target three-dimensional display device has a larger disparity range for displaying salient elements than the reference three-dimensional display device, the deriving of the target disparity mapping comprises a weighted addition of the reference disparity mapping data and a linear disparity mapping within the larger disparity range.
Using this approach the disparity mapping is brought closer to the ideal unity mapping; i.e. input is output, however the amount of negative and positive disparity used are substantially preserved. In addition it was experimentally confirmed that this method of enhancing provides perceptively pleasing results.
Even more preferably the target disparity mapping corresponds to
The above approach has shown to give good results because it substantially preserves linearity present in the disparity mapping and moreover preserves symmetry present in the input disparity to the output disparity.
The characteristics of the reference three-dimensional display device may preferably comprise one of a blur-disparity range in the form of a curve indicating blur of the display as a function of disparity. Ideally the blur is based on a standardized blur metric so as to improve re-mapping. Alternatively an inverse sharpness disparity curve may be provided. In a simplified system a similar type of information may be provided in the form of a blur-disparity range indicating a preferred disparity range where blur remains minimal. Likewise a sharpness-disparity range may be provided. On account of the generally symmetric nature the blur-range or sharpness range may be characterized using a single disparity value.
A second aspect of the present invention relates to a method of generating a three-dimensional video signal, the three-dimensional video signal suitable for retargeting the three-dimensional video signal to a target three-dimensional display device, the method comprising obtaining three-dimensional video data suitable for driving a multi-view display device, obtaining characteristics of a reference three-dimensional display device, generating reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device and generating the three-dimensional video signal, wherein generating the three-dimensional video signal comprises combining the three-dimensional video data and the reference disparity mapping data.
The thus generated three-dimensional image signal may function as input to a method according to the first aspect of the invention.
Preferably generating the reference disparity mapping data comprises using the characteristics of the reference three-dimensional display device, such that the reference disparity mapping maps the disparity of salient elements in the three-dimensional video data onto a predetermined disparity range of the reference three-dimensional display device suitable for representing said salient element, this results in the ability of the target three-dimensional display device being capable of better handling disparity for salient elements in the three-dimensional video data.
Notably salient elements correspond to elements such as e.g. faces that in general draw the focus of attention of the viewer, use of artistic blurring aside, such salient objects are preferably represented in a sharp; i.e. un-blurred, manner on a display device.
More preferably generating the three-dimensional video signal comprises combining the three-dimensional video data, the reference disparity mapping data and the characteristics of the reference three-dimensional display device. As a result of this feature there is no longer a need for defining characteristics of one or more reference display devices in target devices, and in case of multiple reference devices there is no longer a need for selecting a reference display device.
A third aspect of the present invention relates to a system for retargeting a three-dimensional video signal for use on a target three-dimensional display device, the system comprising: a receiver arranged to receive the three-dimensional video signal, the three-dimensional video signal comprising: three-dimensional video data suitable for driving a multi-view display device, and reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device. The system further comprising a deriver arranged to derive a target disparity mapping based on the reference disparity mapping and characteristics of the reference three-dimensional display device and corresponding characteristics of the target three-dimensional display device; and a converter arranged to retarget the three-dimensional video data using the target disparity mapping data into a target three-dimensional video data.
A fourth aspect of the present invention relates to a three-dimensional display device comprising a system for retargeting a three-dimensional video signal as described herein above.
A fifth aspect of the present invention relates to a system for generating a three-dimensional video signal, the three-dimensional video signal suitable for retargeting the three-dimensional video signal to a target three-dimensional display device, the method comprising: a receiver arranged to receive three-dimensional video data suitable for driving a multi-view display device, an acquisition unit arranged to acquire characteristics of a reference three-dimensional display device, a first generator arranged to generate reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device and a second generator arranged to generate the three-dimensional video signal, wherein generating the three-dimensional video signal comprises combining the three-dimensional video data and the reference disparity mapping data.
A sixth aspect of the present invention relates to a three-dimensional video signal, the three-dimensional video signal suitable for retargeting the three-dimensional video signal to a target three-dimensional display device, the three-dimensional video signal comprising: three-dimensional video data suitable for driving a multi-view display device and reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device.
The present invention further relates to a computer program for retargeting a three-dimensional video signal, the computer program comprising instructions for executing a retargeting method as described above.
The present invention further relates to a computer program for generating a three-dimensional video signal, the computer program comprising instructions for executing a generation method as described above.
The present invention further relates to a data carrier comprising the three-dimensional video signal described above, a data carrier comprising the computer program for retargeting described above or a data carrier comprising the computer program for generating as described above. Examples of suitable data carriers are data carriers that provide non-volatile storage capabilities such as an optical disc, a hard-disc or a solid-state storage device.
These and other aspects, features and advantages of the invention will be apparent from and elucidated with reference to the embodiment(s) described hereinafter.
Embodiments of the invention will be described, by way of example only, with reference to the drawings, in which
The inventors of the present invention propose to add metadata to a three-dimensional video signal in order to enable the re-targeting or re-mapping of the three-dimensional video data contained therein for rendering on a target three-dimensional display device.
The inventors realized that content typically is authored by content authors for an ideal display device that can render the three-dimensional video signal as it was intended from an artistic point of view. However in practice when such content is to be rendered on a practical display device this is not always feasible. As a result the disparity in the three-dimensional video may have to be remapped. The manner in which this should be done may not be apparent or easily determined at a target display device.
To address this issue the inventors propose to add metadata for a reference three-dimensional display device to the three-dimensional image signal. This metadata is based on the one hand on the actual three-dimensional video data contained in the image signal. On the other hand it is based on characteristics of the display device.
A target 3D display device arranged for using the present invention will need to be aware of its own capabilities. These capabilities will typically be made available by the manufacturer of the device to the retargeting software. The information may be made available in a non-volatile storage, or using a ROM or other permanent storage. Using this information the target display device can relate its characteristics to that of the reference 3D display device.
When the target 3D display device receives a three-dimensional signal according to the present invention, the 3D display device will need to relate the characteristics of the reference 3D device to its own characteristics. To this end all target display devices may be provided with characteristics of one or more reference display devices, or alternatively such characteristics may be included in the three-dimensional image signal.
Once the target three-dimensional display device has both characteristics available it can make an assessment as to whether or not it is possible to obtain an improved disparity mapping based on the reference disparity mapping.
For example, consider the situation wherein the target 3D display device knows its own blur-disparity relation; i.e. the target 3D display device knows the range of display disparity in luminance pixels that can be displayed with a blur measure remaining below a predetermined threshold. Next consider the situation that the target 3D display device receives a three-dimensional image signal that comprises three-dimensional data and an accompanying reference disparity mapping for a reference 3D display device together with the blur-disparity relation of the reference 3D display device.
Next the target 3D display device may compare its blur-disparity relation with that of the reference 3D display device.
In one scenario the blur-disparity relation may be condensed in a single symmetrical disparity range; i.e. into a Salient disparity range. The salient disparity range corresponds with the disparity range where ideally salient elements of the three-dimensional video data are mapped. When salient objects are mapped such that their disparity falls within this disparity range, the salient object will remain sharp.
When the salient disparity range of the target 3D display device is bigger than that of the reference 3D display device, and the reference disparity mapping shows significant “companding”, there may be room to derive a better quality disparity mapping.
It is an insight of the inventors that it is beneficial to more closely approximate the original identity curve (input disparity=output disparity), but to preserve linearity of the reference disparity mapping.
To this end the inventors propose a target disparity mapping that comprises a weighted addition of the reference disparity mapping data and a linear disparity mapping within the larger disparity range.
In the present example the ratio between the saliency range of the target 3D display device and the reference 3D display device is a factor 2.
It is an insight of the inventors that when remapping preferably the linearity of the reference map should be preserved as much as possible. However, if there is room available the curve should be brought closer to the ideal mapping; i.e. towards the identity curve.
A particularly advantageous manner to do so is to derive an improvement factor based on the salience disparity ranges of the reference and target display devices and to use this improvement factor to perform a weighted addition of the reference disparity mapping and a linear approximation of the reference disparity mapping. This improvement factor is referred to as alpha.
alpha=Star/Sref
For values of alpha larger than 1 we preferably enhance the reference disparity mapping as follows:
and wherein
reference_disparity_mapping(x) corresponds to the reference disparity mapping for x and linear_approximation_ref_disparity(x) corresponds to a linear disparity mapping through the points (dmin, reference_disparity_mapping(dmin)) and (dmax, reference_disparity_mapping(dmax)).
The advantage of this way of working is that the display disparities remain symmetrically distributed (vertically) for the whole input content disparity range. This is important because the blur relation is symmetrical for positive and negative display disparities.
For values of alpha that are smaller than one, mixing should be handled such that the variation around the then downscaled linear approximation linear_approximation_ref disparity(x) should remain symmetrical.
Alternatively, mixing may be performed with a zero display disparity curve, in other words the target disparity mapping is adapted in vertical direction, by scaling using alpha.
Notably further improvements are possible by deriving a smooth curve; e.g. the target device could derive a smooth continuous curve from the resulting disparity mapping by applying cubic Hermite splines using three-point difference for internal points and one-sided difference at the end points and preserve monotonicity by applying the Fritch-Carlson method.
As an aside it is noted that saliency detection in video has received considerable attention and a wide range of efficient saliency detection mechanism are known. For the sake of completeness a reference to “Salient Object Detection: A Benchmark, by Ali Borji et al”, is provided which was downloaded from
http://ilab.usc.edu/˜borji/papers/paperBorji.pdf
this document comprises a comparison of 35 state-of-the-art saliency detection models.
The method further comprises a step 403 of retargeting the three-dimensional video data using the target disparity mapping data into a target three-dimensional video data. In the event the input three-dimensional video data comprises content in a so-called image and depth format, the retargeting may comprise generating an image and accompanying on-screen disparity. Alternatively when image based rendering is included this may also involve image-based rendering in order to generate a multi-view image signal having the proper on-screen disparity.
The method further comprises a step 503 for generating reference disparity mapping data, the reference disparity mapping data indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device and a step 504 for generating the three-dimensional video signal, wherein generating the three-dimensional video signal comprises combining the three-dimensional video data and the reference disparity mapping data.
The three-dimensional video signal 74 comprises three-dimensional video data suitable for driving a multi-view display device, and reference disparity mapping data. The reference disparity mapping data is indicative of a disparity mapping for the three-dimensional video data for a reference three-dimensional display device (not shown). In the block diagram the information is de-multiplexed and transmitted to a deriver 40. The deriver is arranged to derive a target disparity mapping based on the reference disparity mapping and characteristics of the reference three-dimensional display device and corresponding characteristics of the target three-dimensional display device. To this end the deriver may retrieve part or all of the characteristic data 78 from non-volatile storage 70, or may alternatively obtain the characteristics of the reference device from the de-multiplexed three-dimensional image signal 75.
Once the target disparity mapping has been computed the target disparity mapping data together with the three-dimensional video data is transmitted to a converter 50 arranged to retarget the three-dimensional video data using the target disparity mapping data into a target three-dimensional video data 77. The target three-dimensional video data 77 in turn may be transmitted to a target display 60.
Notably the invention also envisages an integrated target three-dimensional display device 20 comprising the target display 60 and the integrated retargeting system 10.
Although in the exemplary devices shown, the three-dimensional video data is transmitted in conjunction with the reference disparity mapping data, it will be clear to those skilled in the art, that alternatively the three-dimensional video data can also bypass the deriver 60 and be transferred from the receiver 30 to the converter 50 directly.
The system 110 further comprises a first generator 140 arranged to generate reference disparity mapping data 104, the reference disparity mapping data 104 is indicative of a disparity mapping for the three-dimensional video data 105 for a reference three-dimensional display device. The first generator 140 is arranged to receive the characteristics 103 as transmitted by the acquisition unit 130 as well as the three-dimensional video data 105 as received by the receiver 120.
The system 110 further comprises a second generator 150 arranged to generate the three-dimensional video signal 74, wherein generating the three-dimensional video signal 74 comprises combining the three-dimensional video data 105 and the reference disparity mapping data 104 as received from the first generator.
When generating a video image signal, it is possible to encode additional information in so-called SEI messages. SEI messages may for example be added in to AVC (ISO/IEC 14496-10:2012—Information technology—Coding of audio-visual objects—Part 2 10: Advanced Video Coding) or HEVC (ISO/IEC 23008-2:201x/FDIS—Information technology—High efficiency coding and media 13 delivery in heterogeneous environments Part 2: High efficiency video coding) based video sequences. Such metadata may be used to provide additional data to re-target the video image signal. Transmission of metadata elements across device interconnection links (such as in HDMI) can also be accommodated for.
Preferably metadata is comprised in user data unregistered SEI messages. More preferably these SEI messages are included in the elementary stream that contains the depth or disparity information that is provided together with the three-dimensional video data.
Dependent on the manner of transport the depth can be transmitted in a single elementary stream with video and depth or disparity or alternatively as a separate disparity/depth elementary stream depending on the preferred manner of transport.
An example of the SEI metadata message container syntax is provided below.
Notably, here uuid_xxxx can be used as an identifier to indicate the type of metadata, so as to allow an application parsing the incoming data to recognize the metadata and take appropriate action so as to retarget the three-dimensional video image signal based on the information provided. In turn the metadata_type field can be used to provide additional information on the type of metadata comprised in the SEI message container.
Advantageously we can use the SEI metadata to provide additional information related to the reference disparity mapping, or for the characteristics of the reference three dimensional display device.
SEI messages can be inserted in the stream at frame accurate positions and thus can be used to accurately adjust the disparity mapping when needed. This information may also be referred to as depth/disparity processing information.
The depth processing information message, when present in the coded video sequence, is recommended to be at least included with every IDR access unit (=Instantaneous Decoding Refresh access unit. This way it may be updated as needed with any access unit and is valid until the next access unit.
The depth processing information message may be used to provide depth format information and information on how depth should be remapped to displays with limited depth. Off course care should be taken so that information contained in these messages is kept consistent with possibly overlapping information in other SEI messages that may be included for consistency with other functions.
The depth format information includes parameters that enable the display device to convert depth map data to disparity values in pixels. By including the depth processing information with the coded video, the content owner/creator has the possibility to optimize the mapping from depth map values to disparity values for the receiving display without the need to have detailed knowledge about the characteristics of all possible target displays.
Part of the depth processing information may be updated with frame accuracy if desired, e.g. when salient objects are moving towards or away from the viewer. Remapping information may be included in the signal/bitstream by means of a curve that defines the relation between input disparity values and output values that relate to disparity on target displays with a limited depth range.
Of course several types of remapping may be used side-by-side. For example for simple devices a simple fixed disparity mapping such as a linear scaling may be defined. However in more advanced systems and where such data is available a mapping in-line with the present invention may be utilized.
In the case of fixed blur model remapping, it is assumed that the target three-dimensional display device the perceived horizontal resolution degrades with increasing disparity following a more or less similar degradation model.
For these displays the horizontal resolution degradation as a function of disparity can be characterized by a single parameter alpha. Alpha values of the displays are required to be determined by the manufacturer according to a standardized method.
The alpha value of the reference display that was used to determine the curve is included with the metadata. The target display device applies the curve and scales the output values in a linear fashion, depending on the ratio between the alpha of the reference display and the target display. As a result the sharpness degradation on the target device more or less resembles that on the reference device.
As the disparity scaling is linear, any curve applied to the original disparities by the content author, that is any mapping applied in the studio representing the artistic intent, is not affected.
An example of the syntax of the coding of a blur curve is provided in Table 2 below.
i(12)
i(12)
i(12)
i(12)
In table 2, the depth_coding_format indicates the manner in which depth is included in the bitstream. The depth_source_type in turn may specify the source of the content; which may e.g. indicate whether or not the content is CGI content, professional movie content, off line product broadcast content, or real-time broadcast content. In this manner the nature of the input content is clear and allows conclusions to be drawn based on the possible quality of the depth and/or e.g. whether scene changes are accurately detectable/encoded.
The depth_representation value defines the meaning of how depth information is encoded; e.g. the depth_representation can indicate that the depth is uniformly quantized 0-255 inclusive in case of an 8-bit value. Alternatively it may be indicated that disparity values are being used instead of depth values.
In case depth values are being used in the input format, then preferably d_min defines the input disparity value in luminance pixels of the texture view that corresponds with decoded depth value 0. Notably here input disparity corresponds to the disparity as envisioned by the content author, so this disparity may also be referred to as content disparity. Preferably negative values of d_min correspond to virtual locations in front of the screen. At the same time preferably d_max defines the input disparity value in luminance pixels of the texture view that corresponds with decoded depth value 255. So logically d_max is greater than d_min and together d_min and d_max define the disparity range to which depth value should be uniformly mapped/.
Alternatively when disparity values are being used there is no need for providing information on mapping depth to input/content disparity
The remapping type in turn may be used to indicate whether a fixed blur model remapping should be used or whether another remapping method should be used.
The disparity remapping may be represented using a number of nodes as indicated in Table 2. When the num_curve_nodes is unequal to 0 this indicates that a (non-linear) remapping should be applied to the input/content disparity values according to a remapping curve to optimize presentation of the 3D video content on an autostereoscopic display with a limited depth range.
This field specifies the number of nodes that is included in the metadata to define the remapping curve. This field should be set to 0 if remapping_type is set to fixed blur. In turn disparity_in[j] defines the input disparity value of the j-th node of the remapping curve. The nodes included in the metadata have increasing input disparity values. The first input disparity value has to be greater than or equal to the value of d_min. The last input disparity value has to be smaller than or equal to d_max.
curve_out[j] in turn defines the output value of the j-th node of the curve. The curve out values should increase with increasing j.
The target device should derive a smooth continuous curve from the nodes by applying cubic Hermite splines using three-point difference for internal points and one-sided difference at the end points and preserve monotonicity by applying the Fritch-Carlson method.
Although the present invention has been described with reference to a three-dimensional image signal, wherein depth and/or disparity input was explicitly provided, it is noted that the present invention may also be used for retargeting stereoscopic of multi-view content in the absence of depth or disparity data, provided that a reference disparity mapping and characteristics of the reference and target 3D display devices are available.
In the case of missing depth or disparity data, depth or disparity data can be estimated based on the stereo and/or multi-view content and one or more new views taking into account the remapped disparity may be rendered using well known image based rendering techniques.
In short the present invention may be used to map input depth and/or disparity data onto output or display disparity data in an advantageous manner. The present invention is particularly useful for autostereoscopic multi-view display devices as these typically have a more limited disparity range that can be utilized.
The invention can be implemented in any suitable form including hardware, software, firmware or any combination of these. The invention may optionally be implemented at least partly as computer software running on one or more data processors and/or digital signal processors. The elements and components of an embodiment of the invention may be physically, functionally and logically implemented in any suitable way. Indeed the functionality may be implemented in a single unit, in a plurality of units or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units, circuits and processors.
Although the present invention has been described in connection with some embodiments, it is not intended to be limited to the specific form set forth herein. Rather, the scope of the present invention is limited only by the accompanying claims. Additionally, although a feature may appear to be described in connection with particular embodiments, one skilled in the art would recognize that various features of the described embodiments may be combined in accordance with the invention. In the claims, the term comprising does not exclude the presence of other elements or steps.
Furthermore, although individually listed, a plurality of means, elements, circuits or method steps may be implemented by e.g. a single circuit, unit or processor. Additionally, although individual features may be included in different claims, these may possibly be advantageously combined, and the inclusion in different claims does not imply that a combination of features is not feasible and/or advantageous. Also the inclusion of a feature in one category of claims does not imply a limitation to this category but rather indicates that the feature is equally applicable to other claim categories as appropriate. Furthermore, the order of features in the claims do not imply any specific order in which the features must be worked and in particular the order of individual steps in a method claim does not imply that the steps must be performed in this order. Rather, the steps may be performed in any suitable order. In addition, singular references do not exclude a plurality.
Thus references to “a”, “an”, “first”, “second” etc. do not preclude a plurality. Reference signs in the claims are provided merely as a clarifying example shall not be construed as limiting the scope of the claims in any way.
Number | Date | Country | Kind |
---|---|---|---|
13162505.5 | Apr 2013 | EP | regional |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2014/032955 | 4/4/2014 | WO | 00 |