The present invention relates to an image conversion apparatus that converts a two-dimensional image (2D image) into a three-dimensional stereoscopic image (3D image).
Various systems are planned and realized as devices for reproducing 3D images. A reproducing apparatus for reproducing a 3D image reads a left-eye image signal and a right-eye image signal from, for example, a disk, to alternately output the read left-eye image signal and the read right-eye image signal to a display. When a display is used in combination with glasses with a liquid crystal shutter as described in Patent Document 1, the display alternately displays a left-eye image indicated by a left-eye image signal input from a reproducing apparatus and a right-eye image indicated by a right-eye image signal input from the reproducing apparatus on a screen in a predetermined cycle. The display controls the glasses with liquid crystal shutter such that a left-eye shutter of the glasses with liquid-crystal shutter opens when the left-eye image indicated by the left-eye image signal is displayed and a right-eye shutter of the glasses with liquid crystal shutter opens when the right-eye image indicated by the right-eye image signal is displayed. With this configuration, only the left-eye image reaches the left eye of a user who wears the glasses with liquid crystal shutter, and only the right-eye image reaches the right eye. Thus, the user can visually recognize a three-dimensional stereoscopic image.
Meanwhile, contents including 3D images are not sufficiently provided at the present. For this reason, there is proposed an image conversion apparatus that converts existing 2D images into 3D images. With this, a user can view the existing 2D image as the 3D image.
Patent Document 1: JP-A-2002-82307
However, a 3D image provided by a conventional image conversion apparatus, for example, is an image that is recognized by a user as if the image entirely protrudes from a display surface of a display apparatus to the user's side. In this case, there has been a problem in that, although the user obtains a feeling of protrusion, the user feels the display surface of the display apparatus according to visual characteristics of a human being.
It is an object of the present invention to provide an image conversion apparatus that can generate, from a 2D image, a 3D image on which a user can visually recognize a sufficient spatial effect.
An image conversion apparatus according to a first aspect of the present invention converts non-stereoscopic image data into stereoscopic image data configured by left-eye image data and right-eye image data. The image conversion apparatus includes an input unit that inputs a non-stereoscopic image, and a conversion unit that generates and outputs the left-eye image data and the right-eye image data based on the non-stereoscopic image data input through the input unit. When a stereoscopic image configured by the left-eye image and the right-eye image is displayed on a display apparatus capable of displaying a stereoscopic image, the conversion unit generates the left-eye image data and the right-eye image data to cause a user to visually recognize the stereoscopic image so that a predetermined portion in a horizontal direction in the displayed stereoscopic image is present at a position farthest from the user in a direction vertical to a display surface of the display apparatus, and a portion other than the predetermined portion is present at a position closer to the user toward left and right ends of the stereoscopic image.
An image conversion apparatus according to a second aspect of the present invention converts non-stereoscopic image data into stereoscopic image data configured by left-eye image data and right-eye image data. The image conversion apparatus includes an input unit that inputs a non-stereoscopic image, and a conversion unit that generates and outputs the left-eye image data and the right-eye image data based on the non-stereoscopic image data input through the input unit. When a stereoscopic image configured by the left-eye image and the right-eye image is displayed on a display apparatus capable of displaying a stereoscopic image, the conversion unit generates the left-eye image data and the right-eye image data to cause a user to visually recognize the stereoscopic image so that the entire displayed stereoscopic image is present at a position farther than the display apparatus when viewed from the user in a direction vertical to a display surface of the display apparatus, a predetermined portion in the horizontal direction in a display region of the display apparatus is present at a closest position, and a portion other than the predetermined portion is present at a position farther from the user toward left and right ends of a stereoscopic image.
An image conversion apparatus according to a third aspect of the present invention processes stereoscopic image data. The image conversion apparatus includes an input unit that inputs a non-stereoscopic image, and a conversion unit that provides different moving distances to a left-eye image and a right-eye image of the stereoscopic image based on the stereoscopic image data input through the input unit to generate and output left-eye image data and right-eye image data. When differences between moving distances provided to left-eye image data and right-eye image data generated from the identical stereoscopic image data are compared with each other, the conversion unit generates the left-eye image data and the right-eye image data to make a difference between moving distances provided to a first pixel position of the stereoscopic image different from a difference between moving distances provided to a second pixel position different from the first pixel position.
According to a first aspect of the present invention, when a stereoscopic image configured by the left-eye image and the right-eye image is displayed on a display apparatus capable of displaying a stereoscopic image, left-eye image data and right-eye image data are generated such that a user visually recognizes the stereoscopic image so that a predetermined portion in a horizontal direction in the displayed stereoscopic image is present at a position farthest from the user in a direction vertical to a display surface of the display apparatus and a portion other than the predetermined portion is present at a position closer to the user toward left and right ends of the stereoscopic image. Thus, a stereoscopic image (3D image) can be generated from a non-stereoscopic image (2D image), which can cause a user to feel sufficient depth perception and sufficient spatial perception and which can cause the user to feel the display surface of the display apparatus larger, because of the visual characteristics of a human being.
According to a second aspect of the present invention, when a stereoscopic image configured by the left-eye image and the right-eye image is displayed on a display apparatus capable of displaying a stereoscopic image, left-eye image data and right-eye image data are generated such that the user visually recognizes the stereoscopic image so that the entire displayed stereoscopic image is present at a position farther than the display apparatus when viewed from the user in a direction vertical to the display surface of the display apparatus, a predetermined portion in a horizontal direction in a display region of the display apparatus is present at a position closest to the user, and a portion other than the predetermined portion is present at a position farther from the user toward left and right ends of a 3D image. Thus, a stereoscopic image (3D image) can be generated from a non-stereoscopic image (2D image), which can cause the user to feel sufficient depth perception and sufficient spatial perception which can cause the user to feel a feeling of protrusion to user's side with respect to the predetermined portion, and which can cause the user to feel the display surface of the display apparatus larger, because of the visual characteristics of the human being.
According to a third aspect of the present invention, when differences between moving distances provided to left-eye image data and right-eye image data generated from the identical stereoscopic image data are compared with each other, left-eye image data and right-eye image data are generated such that a difference between moving distances provided to a first pixel position of the stereoscopic image is different from a difference between moving distances provided to a second pixel position different from the first pixel position. Thus, a 3D effect obtained in consideration of the visual characteristics of the human being can be provided.
1. Configuration
1. 1. Three-dimensional Stereoscopic Image Reproducing Display System
When a disk inserted into the reproducing apparatus 101 includes contents of a 2D image, the 2D image is converted into a 3D image to make it possible to output the 3D image. Details related to conversion from the 2D image to the 3D image will be described later.
1. 2. Reproducing Apparatus
When the contents recorded on the disk 201 are 2D contents configured by a 2D image, the signal processor 203 can convert the 2D contents into 3D contents configured by a 3D image and output the 3D contents. The details of the converting process will be described later.
1. 3. Configuration of Signal Processor
The CPU 305 receives a reproducing start instruction by a user through the remote-control receiver 205 and causes the disk reproducing unit 202 to reproduce the disk 201. The stream separating unit 301 separates an image (video), sound, graphics, additional data including ID data, or the like included in the data stream output from the disk 201 in the disk reproducing unit 202. The audio decoder 302 decodes audio data read from the disk 201 and transfers the audio data to the memory 204. The video decoder 303 decodes video data read from the disk 201 and transfers the video data to the memory 204. The graphics decoder 304 decodes the graphics data read from the disk 201 and transfers the graphics data to the memory 204.
The CPU 305 reads GUI data of the device main body from the program storing memory 207 and generates and transfers the GUI data to the memory 204. The video signal processor 306 generates an L image and an R image by using the various types of data according to the determination by the CPU 305 and outputs the L image and the R image in a 3D image format.
2. 1. Converting Operation from 2D Image to 3D Image
When the contents recorded on the disk 201 are 2D contents configured by a 2D image, an operation in which the signal processor 203 converts the 2D image of the 2D contents into a 3D image and outputs the 3D image will be described. A stream including video data is input to the stream separating unit 301. The stream separating unit 301 outputs the video data of the 2D image to the video decoder 303. The video decoder 303 decodes the video data of the 2D image and transfers the video data to the memory 204. A video signal output from the video decoder 303 is a 2D video signal. The memory 204 records the video signal.
When the remote-control receiver 205 receives an instruction to convert a 2D image into a 3D image, the CPU 305 provides to the memory 204 and the video signal processor 306 an instruction to convert the 2D image into the 3D image and to output the 3D image. At this time, in order to generate a 3D image, the memory 204 outputs video signals of 2 frames representing the same 2D image for generating an L image and an R image of the 3D image. On the other hand, in order to convert the 2D image into the 3D image based on the instruction from the CPU 305, the video signal processor 306 performs different processings to image signals representing the same 2D image of the two frames output from the memory 204, generates image signals representing the L image and the R image configuring the 3D image, and outputs the generated image signals to the output unit 206.
As indicated by a broken line in
When a magnifying process is performed to the L image and the R image with the horizontal magnification shown in
First, a case in which a magnifying process is performed to the L image will be described. As shown in
Next, a case in which a magnifying process is performed to the R image will be described. As shown in
A difference between the output horizontal pixel position of the L image and the output horizontal pixel position of the R image, i.e., a parallax amount is shown in
As described above, in the reproducing apparatus 101 according to this embodiment, moving distances of pixels are changed depending on input horizontal pixel positions. In this embodiment, the L image data and the R image data are generated to cause a user to visually recognize a stereoscopic image configured by an L image and an R image, when the stereoscopic image is displayed on the display apparatus 102 capable of displaying a 3D image, so that the central portion in the horizontal direction in the displayed 3D image is present at a position farthest from the user in a direction vertical to the display surface of the display apparatus 102, and a portion other than the horizontal central portion is present at a position farther from the user toward both the left and right ends of the stereoscopic image. The parallax amount may be changed stepwise instead of being continuously changed.
When the L image and the R image are generated from the 2D image by the image converting method, the converted L and R images look like a pseudo 3D image according to the visual characteristics of the human being. In the image converting method, since the 2D image is only extended or reduced in the horizontal direction, the image can be prevented from being broken down.
Depending on a protruding amount correcting instruction received by the remote-control receiver 205 and an instruction of converting a 2D image into a 3D image, the CPU 305 may adjust a protruding amount of the 3D image generated by the video signal processor 306.
When the protruding amount of the entire image is adjusted, the characteristics shown in
As a method of adjusting a protruding amount of a part of the image, there is a method of adjusting a horizontal magnification while maintaining an average value of the horizontal magnifications in
3. Conclusion
In this embodiment, the reproducing apparatus 101 includes the stream separating unit 301 that receives a 3D image and the video signal processor 306 that generates and outputs L image data and R image data based on 2D image data input from the stream separating unit 301. The video signal processor 306 generates L image data and R image data to cause a user to visually recognize the image when a 3D image configured by an L image and an R image is displayed on the display apparatus 102 capable of displaying a 3D image so that the central portion in the horizontal direction in the displayed 3D image is present at a position farthest from the user in a direction vertical to the display surface of the display apparatus 102, and a portion other than the central portion is present at a position farther from the user toward both the left and right ends of the stereoscopic image.
With such a simple configuration, there can be generated a 3D image from a 2D image, which can cause a user to feel sufficient depth perception and sufficient spatial perception according to the visual characteristics of the human being and which can cause the user to feel as if the display surface of the display apparatus 102 is larger.
The video signal processor 306 generates L image data and R image data to cause a user to recognize a 3D stereoscopic image configured by an L image and an R image when the 3D stereoscopic image is displayed on the display apparatus 102 capable of displaying a 3D image so that the entire displayed 3D image is present at a position farther from the display surface of the display apparatus 102 when viewed from the user in a direction vertical to the display surface of the display apparatus 102.
According to the stereoscopic image realized with the above configuration, the user is caused to more strongly feel depth perception and spatial perception according to the visual characteristics of the human being.
In this embodiment, although the image is configured to be recognized by a user so that the substantial central portion in the horizontal direction is present at the farthest position, the image may be configured to be recognized by the user so that a portion other than the central portion is present at the farthest position. Also in this case, the same effect can be obtained.
A position at which a 3D image is recognized by the user in a direction vertical to the display surface of the display apparatus 102 may be configured to be adjustable with a remote controller. A signal from the remote controller is received by the remote-control receiver 205 and processed by the signal processor 203. With this configuration, a 3D image according to the user's preferences can be generated.
In Embodiment 1, L image data and R image data are generated to cause a user to visually recognize the image so that a central portion in a horizontal direction in a displayed 3D image is present at a position farthest from the user (most rear side), and a portion other than the central portion is present at a position closer to the user (on the front) toward left and right ends. In contrast, in Embodiment 2, an image is displayed to cause a user to visually recognize an image so that a displayed entire 3D image is present at a position farther than the display surface of the display apparatus 102 when viewed from the user, a central portion in a horizontal direction in a display region of the display apparatus 102 is present at the closest position, and present at a position closer to the user toward left and right ends of the stereoscopic image at the central position. The configuration of the reproducing apparatus 101 is the same as that in Embodiment 1. A configuration of Embodiment 2 will be described below in detail.
Embodiment 2 is different from Embodiment 1 in that a region in which a horizontal magnification is changed in
In Embodiment 2, based on the horizontal magnification shown in
When the 3D image has a parallax at both the ends in the horizontal direction, a part of the L image or the R image is out of the display area and lacks, thereby making a user uncomfortable. In this embodiment, in order to reduce the uncomfortable feeling, as shown in
In this embodiment, although the parallax amount between the L image and the R image, as shown in
As described above, in Embodiment 2, the reproducing apparatus 101 includes the stream separating unit 301 that receives a 3D image and the video signal processor 306 that generates and outputs L image data and R image data based on 2D image data input from the stream separating unit 301. The video signal processor 306 generates L image data and R image data to cause a user to visually recognize a 3D image configured by an L image and an R image when the 3D image is displayed on the display apparatus 102 capable of displaying a 3D image, so that the entire displayed 3D image is present at a position farther than the display apparatus 102 when viewed from the user in a direction vertical to the display surface of the display apparatus 102, a central portion in the horizontal direction in the display region of the display apparatus 102 is present at the closest position, and a portion other than the central portion is present at a position farther from the user toward both the left and right ends of the stereoscopic image.
With the stereoscopic image realized by the above configuration, there can be generated a 3D image that can cause a user to feel sufficient depth perception and sufficient spatial perception according to the visual characteristics of the human being, can cause the user to feel a feeling of protrusion to the user's side with respect to a central portion, and can cause the user to feel as if the display surface of the display apparatus 102 larger.
In this embodiment, the video signal processor 306 reduces image amplitudes at ends of the L image and the R image. With this manner, the uncomfortable feeling occurring when a part of the L image or the R image lacks in the both the horizontal end portions of the 3D image can be reduced. This technical idea and a technical idea of another embodiment (to be described later) related thereto can also be applied to Embodiment 1.
In Embodiment 3, a 2D image is converted into a 3D image based on the same characteristics as those in Embodiment 1, and graphics data is 3-dimensionalized based on the same characteristics as those in Embodiment 1, the 3-dimensionalized data is superposed to the 3D image to display. The configuration of the reproducing apparatus 101 is the same as that in Embodiment 1.
Embodiment 3 is different from Embodiment 1 and Embodiment 2 in that not only processing for a video signal but also processing for a graphics signal are performed as processing contents in the video signal processor 306. The processing contents in the video signal processor 306 can be independently performed to the video signal and the graphics signal so that a front-and-back positional relationship between the generated 3D image and the graphics can be changed.
In Embodiment 3, in the video signal processor 306, the same signal processing as that in Embodiment 1 is performed to the 3D image and the graphics.
When a 3D converting process is performed based on the characteristics shown in
As described above, according to the reproducing apparatus 101 of Embodiment 3, not only the 2D image, but also the graphics data can be 3-dimensionalized and displayed. Thus, a 3D effect can also be obtained with respect to the graphics data. In particular, in this embodiment, since the 3D conversion characteristics of the graphics data are the same as those of the 2D image, there can be obtained a 3D image that appears to be obtained by sticking the graphics and the image to each other. As in Embodiments 1 and 2, both the graphics and the image can cause a user to feel sufficient depth perception and sufficient spatial perception according to the visual characteristics of the human being.
In Embodiment 4, although a 2D image is converted into a 3D image based on the same characteristics as those in Embodiment 1, the graphics data is 3-dimensionalized so as not to be curved as in Embodiment 3, and is superimposed and displayed. The configuration of the reproducing apparatus 101 is the same as that in Embodiment 1.
In Embodiment 4, there will be described a case in which processing to the graphics data performed by the video signal processor 306 in Embodiment 3 is changed from processing based on the characteristics shown in
In this embodiment, as shown in
On the other hand, in the graphics, the 300th pixel configuring the left end of the input graphics is moved to the 290th pixel in the L image and moved to the 310th pixel in the R image. As a result, a 20-pixel parallax is generated between the R image and the L image. On the other hand, the 1620th pixel configuring the right end of the input graphics is moved to the 1610th pixel in the L image and moved to the 1630th pixel in the R image. As a result, similarly to the L image, a 20-pixel parallax is generated between the R image and the L image. Thus, as shown in
As described above, according to the reproducing apparatus 101 Embodiment 4, similarly to Embodiment 3, not only the 2D image but also the graphics data can be 3-dimensionalized and displayed. Thus, a 3D effect can also be obtained with respect to the graphics data. In particular, in this embodiment, conversion is performed such that an offset between the L image data and the graphics data combined with the R image data different, and conversion is performed such that an offset between the R image data and the graphics data combined with the R image data different. With this manner, independent 3D effects can be obtained in the graphics data and the L and R image data. In particular, in Embodiment 4, the same effects as those in Embodiments 1 and 2 can be obtained for an image, and an effect that raises planar graphics with respect to a curved image can also be obtained. When the graphics are raised, an effect of causing the graphics to be easily recognized can be obtained.
Embodiments 1 to 4 have been illustrated as the embodiments of the present invention. However, the present invention is not limited to these embodiments. Other embodiments of the present invention will be collectively described below. Note that the present invention is not limited thereto, and can also be applied to an embodiment that is appropriately corrected.
In Embodiments 1 to 4, the case in which the present invention is applied to a 2D image has been described. However, the present invention may also be applied to a 3D image. In this case, a 3D effect such as a protruding amount can be adjusted to adjust a parallax amount in the 3D image.
In each of the embodiments, although the image is configured to be recognized by a user so that the horizontal central portion of the generated 3D image protrudes to the forefront or is present on the most rear side, this forefront position or the most rear position may be an arbitrary position on the left or right of the central portion instead of the horizontal central portion. For example, when a 2D image serving as a 3D image source includes a person or the like therein, a position where the person or the like is present may be detected, and the image may be configured to be recognized by a user so that the position protrudes to the forefront.
In the embodiment, although the horizontal magnification is changed based on only the horizontal pixel position, the horizontal magnification may be change in consideration of a vertical pixel position. For example, a change rate of a horizontal magnification of an upper portion of an input image may be set to be large, and the change rate of the horizontal magnification may be reduced toward the lower portion. In this case, the lower portion of the image is recognized by a user so that the lower portion relatively protrudes to the front of the upper portion.
A horizontal magnification may be changed based on a state of an image. For example, in a dark scene in which a field of view of the human being becomes narrow, setting is performed such that a parallax amount decreases. In a bright scene, setting is performed such that a parallax amount increases. For example, a brightness (average value) of an entire image is obtained, and a parallax amount is determined based on the brightness.
An output amplitude of both horizontal ends (both the left and right ends) in a 3D image is reduced. However, output amplitudes of not only both the horizontal ends (both the left and right ends) but also output amplitudes (gains) of both the vertical ends (both the upper and lower ends) may be reduced. With this manner, uncomfortable feeling occurring by cutting an image having a parallax with a television frame without a parallax can be reduced.
A reduction in image amplitude to reduce the uncomfortable feelings at both the screen ends is realized by making an output gain of an image variable. However, a combination ratio (a value) to the graphics (OSD screen) is set to OSD 100% and video 0% at both the horizontal ends, the combination ratio is set to OSD 0% and video 100% in a region in a gain of 1 in
A region in which output amplitudes at both the horizontal ends (and both the vertical ends) in the 3D image are reduced may be made variable depending on parallax information of an image input to the video signal processor 306.
A region in which output amplitudes at both the horizontal ends (and both the vertical ends) in the 3D image are reduced may be made variable depending on a parallax amount that is increased or decreased by processing performed to an image input to the video signal processor 306.
In Embodiment 1 and Embodiment 2, similarly to Embodiment 3, a 2D image and graphics input to the video signal processor 306 may be subjected to different processings and then combined with each other. With this manner, for example, the graphics can be displayed while always being raised from the image.
The image processing may be performed in combination with audio processing. For example, conversion may be performed such that acoustic fields are formed at the rear when the center in the horizontal direction is recessed. With this manner, the effect of image conversion can be further enhanced.
In Embodiment 3, after different processings are performed to image data and graphics data, respectively, the image data and the graphics data are combined with each other. However, after processing serving as a difference between the image data and the graphics data may be performed to the graphics data, the graphics data may be combined with the image data, and processing in the horizontal direction may be performed to the combined image.
In Embodiment 1, the display apparatus 102 displays the left-eye image and the right-eye image such that the images are alternately switched, and in synchronization with the switching, the left and right shutters of the 3D glasses 103 are alternately switched. However, the following configuration may also be used. That is, the display apparatus 102 displays the left-eye image and the right-eye image such that odd-number lines and even-number lines are separated from each other with respect to each of the lines, and different polarizing films are respectively stuck to the odd-number lines and the even-number lines in the display unit. The 3D glasses 103 do not use a liquid crystal shutter system. Polarizing filters having different directions are stuck to the left-eye lens and the right-eye lens to separate the left-eye image from the right-eye image with the polarizing filters. The display apparatus may be configured such that a left-eye image and a right-eye image are alternately displayed in a lateral direction per each pixel, and polarizing films having different planes of polarization are alternately stuck to the display unit per each pixel. In short, left-eye and right-eye image data may be configured to be caused to reach the left eye and the right eye of the user, respectively.
The reproducing apparatus 101 is configured to reproduce data on the disk 201 by the disk reproducing unit 202. However, a 2D image serving as a source may be a stream input obtained through a broadcasting station or a network or data recorded on a recording medium such as a blue-ray disk, a DVD disk, a memory card, or a USE memory.
In Embodiments to 4, conversion of video images is exemplified. However, the present invention is not limited thereto. Specifically, the present invention can also be applied to a still image such as a JEPG image.
In each of the embodiments, although conversion of a 2D image into a 3D image is performed in the signal processor 203 of the reproducing apparatus 101, means having the same conversion function may be arranged in the display apparatus 102 to perform the conversion in the display apparatus 102.
The present invention can be applied to an image conversion apparatus that converts a 2D image into a 3D image. For example, the present invention can be applied to, in particular, 3D image compatible devices such as a 3D blue-ray disk player, a 3D blue-ray disk recorder, a 3D DVD player, a 3D DVD recorder, a 3D broadcast receiving device, a 3D television set, a 3D image display terminal, a 3D mobile phone terminal, a 3D car navigation system, a 3D digital still camera, a 3D digital movie, a 3D network player, a 3D-compatible computer, and a 3D-compatible game player.
Number | Date | Country | Kind |
---|---|---|---|
2010-103327 | Apr 2010 | JP | national |
2010-229433 | Oct 2010 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2011/002472 | 4/27/2011 | WO | 00 | 10/26/2012 |