The present application relates to a single-lens 3D image capturing technology for generating multi-viewpoint images using a single optical system and a single image sensor.
Recently, the performance and functionality of digital cameras and digital movie cameras that use some solid-state image sensor such as a CCD and a CMOS (which will be sometimes simply referred to herein as an “image sensor”) have been enhanced to an astonishing degree. In particular, the size of a pixel structure for use in a solid-state image sensor has been further reduced these days thanks to rapid development of semiconductor device processing technologies, thus getting an even greater number of pixels and drivers integrated together in a solid-state image sensor. As a result, the resolution of an image sensor has lately increased rapidly from around one million pixels to ten million or more pixels in a matter of few years. On top of that, the quality of an image captured has also been improved significantly as well. As for display devices, on the other hand, LCD and plasma displays with a reduced depth now provide high-resolution and high-contrast images, thus realizing high performance without taking up too much space. And such video quality improvement trends are now spreading from 2D images to 3D images. In fact, 3D display devices that achieve high image quality, although they require the viewer to wear a pair of polarization glasses, have been developed just recently.
As for the 3D image capturing technology, a typical 3D image capture device with a simple arrangement uses an image capturing system with two cameras to capture a right-eye image and a left-eye image. According to the so-called “two-lens image capturing” technique, however, two cameras need to be used, thus increasing not only the overall size of the image capture device but also the manufacturing cost as well. To overcome such a problem, methods for capturing multiple images with parallax (which will be sometimes referred to herein as “multi-viewpoint images”) by using a single camera have been researched and developed.
For example, Patent Document No. 1 discloses a scheme that uses two polarizers, of which the transmission axes cross each other at right angles, and a rotating polarization filter.
According to such a scheme, however, images at mutually different positions are captured time-sequentially by rotating the circular polarization filter 15, and therefore, two images with parallax cannot be captured at the same time, which is a problem. In addition, the durability of such a system is also a question mark because the system uses mechanical driving. On top of that, since the incoming light passes through the polarizers 11, 12 and the polarization filter 15, the quantity of the light received eventually by the image capture device 9 decreases by as much as 50%, which is non-negligible, either.
To overcome these problems, Patent Document No. 2 discloses a scheme for capturing two images with parallax at the same time without using such mechanical driving. An image capture device that adopts such a scheme gets the two incoming light beams, which have come from two different directions, condensed by a reflective mirror, and then received by an image sensor in which two different kinds of polarization filters are arranged alternately, thereby capturing two images with parallax without using a mechanical driving section.
With such an arrangement, the incoming light beams are transmitted through the polarizers 11 and 12, reflected from the reflective mirrors 13, passed through the optical lens 3 and then incident on the imaging area of the image sensor 1. Those light beams to be transmitted through the polarizers 11 and 12, respectively, and then incident on the image sensor 1 are transmitted through the polarization filters 17 and 18 and then photoelectrically converted by the pixels that are located right under those polarization filters 17 and 18. If the images to be produced by those light beams that have been transmitted through the polarizers 11 and 12 and then incident on the image sensor 1 are called a “right-eye image” and a “left-eye image”, respectively, then the right-eye image and the left-eye images are generated by a group of pixels that face the polarization filters 17 and a group of pixels that face the polarization filter 18, respectively.
As can be seen, according to the scheme disclosed in Patent Document No. 2, two kinds of polarization filters, of which the transmission axes are arranged so as to cross each other at right angles, are arranged alternately over the pixels of the image sensor, instead of using the circular polarization filter disclosed in Patent Document No. 1. As a result, although the resolution decreases to a half compared to the method of Patent Document No. 1, a right-eye image and a left-eye image with parallax can be obtained at the same time by using a single image sensor. According to such a technique, however, the incoming light has its quantity decreased considerably when being transmitted through the polarizers and the polarization filters, and therefore, the quantity of the light received by the image sensor decreases as significantly as in Patent Document No. 1.
To cope with such a problem of the decreased quantity of light received by the image sensor, Patent Document No. 3 discloses a technique for obtaining two images with parallax and a normal image with a single image sensor. According to such a technique, those two images with parallax and the normal image can be obtained by a single image sensor by changing mechanically some components that have been used to capture two images with parallax with alternative components for use to capture a normal image, and vice versa. When two images with parallax are going to be obtained, two polarization filters are arranged on the optical path as disclosed in Patent Document No. 2. On the other hand, when a normal image is going to be obtained, those polarization filters are mechanically removed from the optical path. By introducing such a mechanism, those images with parallax and a normal image that uses the incoming light highly efficiently can be obtained.
Although a polarizer or a polarization filter is used according to the techniques disclosed in Patent Document Nos. 1 to 3, color filters may also be used according to another approach. For example, Patent Document No. 4 discloses a technique for obtaining two images with parallax at the same time using color filters.
In such an arrangement, the incoming light passes through the lens 3, the lens diaphragm 19 and the light beam confining plate 20 and produces an image on the photosensitive film. In the meantime, only red- and blue-based light rays are respectively transmitted through the two color filters 20a and 20b of the light beam confining plate 20. As a result, a magenta-based color image is produced on the photosensitive film by the light rays that have been transmitted through the two color filters. In this case, since the color filters 20a and 20b are arranged at mutually different positions, the image produced on the photosensitive film comes to have parallax. Thus, if a photograph is developed with the photosensitive film and viewed with a pair of glasses, in which red and blue films are attached to its right- and left-eye lenses, the viewer can view an image with depth. In this manner, according to the technique disclosed in Patent Document No. 4, multi-viewpoint images can be produced using the two color filters.
According to the technique disclosed in Patent Document No. 4, the light rays are imaged on the photosensitive film, thereby producing images with parallax there. Meanwhile, Patent Document No. 5 discloses a technique for producing images with parallax by transforming incoming light into electrical signals.
Patent Document No. 6 also discloses a technique for obtaining images with parallax using a similar configuration to the one used in Patent Document No. 5.
Patent Document No. 7 also discloses a technique for generating multiple images with parallax using a pair of filters with mutually different colors, which are arranged symmetrically to each other with respect to an optical axis. By using red and blue filters as the pair of filters, an R pixel that senses a red ray observes the light that has been transmitted through the red filter, while a B pixel that senses a blue ray observes the light that has been transmitted through the blue filter. Since the red and blue filters are arranged at two different positions, the light received by the R pixel and the light received by the B pixel have come from mutually different directions. Consequently, the image observed by the R pixel and the image observed by the B pixel are ones viewed from two different viewpoints. By defining corresponding points between those images on a pixel-by-pixel basis, the magnitude of parallax can be calculated. And based on the magnitude of parallax calculated and information about the focal length of the camera, the distance from the camera to the subject can be obtained.
Patent Document No. 8 discloses a technique for obtaining information about a subject distance based on two images that have been generated using either a diaphragm to which two color filters with mutually different aperture sizes (e.g., red and blue color filters) are attached or a diaphragm to which two color filters in two different colors are attached horizontally symmetrically with respect to the optical axis. According to such a technique, if light rays that have been transmitted through the red and blue color filters with mutually different aperture sizes are observed, the degrees of blur observed vary from one color to another. That is why the degrees of blur of the two images that are associated with the red and blue color filters vary according to the subject distance. By defining corresponding points with respect to those images and comparing their degrees of blur to each other, information about the distance from the camera to the subject can be obtained. On the other hand, if light rays that have been transmitted through two color filters in two different colors that are attached horizontally symmetrically with respect to the optical axis are observed, the direction from which the light observed has come changes from one color to another. As a result, two images that are associated with the red and blue color filters become images with parallax. And by defining corresponding points with respect to those images and calculating the distance between those corresponding points, information about the distance from the camera to the subject can be obtained.
According to the techniques disclosed in Patent Documents Nos. 4 to 8 mentioned above, images with parallax can be produced by arranging RGB color filters on a light beam confining plate. However, since a light beam confining plate is used, the percentage of the incoming light that can be used decreases significantly. In addition, to increase the effect of parallax, those RGB color filters should be arranged at distant positions and should have decreased areas. In that case, however, the percentage of the incoming light that can be used further decreases.
Unlike these techniques, Patent Document No. 9 discloses a technique for obtaining multiple images with parallax and a normal image that is free from the light quantity problem by using a diaphragm in which RGB color filters are arranged. According to that technique, when the diaphragm is closed, only the light rays that have been transmitted through the RGB color filters are received. On the other hand, when the diaphragm is opened, the RGB color filter areas are outside of the optical path, and therefore, the incoming light can be received entirely. Consequently, images with parallax can be obtained when the diaphragm is closed and a normal image that uses the incoming light highly efficiently can be obtained when the diaphragm is opened.
According to any of these techniques of the related art, multi-viewpoint images can be certainly generated, but the quantity of the light received by the image sensor is smaller than usual because a polarizer or color filters are used. In order to receive a sufficient quantity of incoming light, some mechanism that removes the polarizing portion or color filter areas from the optical path needs to be used. That is to say, according to none of these techniques of the related art, multi-viewpoint images and an image that uses the incoming light highly efficiently can be obtained at the same time without using such a mechanism.
An embodiment of the present invention provides an image capturing technique by which multi-viewpoint images that use incoming light highly efficiently can be obtained without using such mechanical driving.
To overcome these problems, a 3D image capture device as an embodiment of the present invention includes: a light transmitting section with first and second light transmitting areas; an image sensor which is arranged to receive light that has been transmitted through the light transmitting section and which has an imaging area where a plurality of unit blocks, each including a first type of pixel and a second type of pixel, are arranged; an imaging section which produces an image on the imaging area of the image sensor; and an image processing section which generates multi-viewpoint images based on photoelectrically converted signals supplied from the first and second types of pixels. If functions representing the respective spectral transmittances of the first light transmitting area, the second light transmitting area, the first type of pixel, and the second type of pixel with respect to the wavelength λ of visible radiation are identified by TL(λ), TR(λ), T1(λ), and T2(λ), respectively, TL(λ)≠TR(λ) and T1(λ)≠T2(λ) are satisfied, and each of TL(λ), TR(λ), T1(λ) and T2(λ) has at least one local maximum value and at least one local minimum value within each of red, green and blue wavelength ranges.
An image sensor as an embodiment of the present invention is used in the 3D image capture device described above.
A light transmitting section as an embodiment of the present invention is used in the 3D image capture device described above.
An image processor as an embodiment of the present invention generates an image based on a signal supplied from a 3D image capture device. The device includes: a light transmitting section with first and second light transmitting areas; an image sensor which is arranged to receive light that has been transmitted through the light transmitting section and which has an imaging area where a plurality of unit blocks, each including a first type of pixel and a second type of pixel, are arranged; and an imaging section which produces an image on the imaging area of the image sensor. If functions representing the respective spectral transmittances of the first light transmitting area, the second light transmitting area, the first type of pixel, and the second type of pixel with respect to the wavelength λ of visible radiation are identified by TL(λ), TR(λ), T1(λ), and T2(λ), respectively, TL(λ)≠TR(λ) and T1(λ)≠T2(λ) are satisfied, and each of TL(λ), TR(λ), T1(λ) and T2(λ) has at least one local maximum value and at least one local minimum value within each of red, green and blue wavelength ranges. The image processor generates multi-viewpoint images based on photoelectrically converted signals supplied from the first and second types of pixels.
These general and particular embodiments can be implemented as a system, a method, a computer program or a combination thereof.
According to an embodiment of the present invention, by using a light transmitting section with a light transmitting area that has a higher transmittance than an ordinary color filter, multi-viewpoint images can be generated with the incoming light used highly efficiently.
Exemplary embodiments of the present invention are outlined as follows:
(1) A 3D image capture device as an embodiment of the present invention includes: a light transmitting section with first and second light transmitting areas; an image sensor which is arranged to receive light that has been transmitted through the light transmitting section and which has an imaging area where a plurality of unit blocks, each including a first type of pixel and a second type of pixel, are arranged; an imaging section which produces an image on the imaging area of the image sensor; and an image processing section which generates multi-viewpoint images based on photoelectrically converted signals supplied from the first and second types of pixels. If functions representing the respective spectral transmittances of the first light transmitting area, the second light transmitting area, the first type of pixel, and the second type of pixel with respect to the wavelength λ of visible radiation are identified by TL(λ), TR(λ), T1(λ), and T2(λ), respectively, TL(λ)≠TR(λ) and T1(λ)#T2(λ) are satisfied, and each of TL(λ), TR(λ), T1(λ) and T2(λ) has at least one local maximum value and at least one local minimum value within each of red, green and blue wavelength ranges.
(2) In one embodiment, TL(λ)=T1(λ) and TR(λ)=T2(λ) are satisfied.
(3) In one embodiment of the 3D image capture device of (2), TL(λ), TR(λ), T1(λ) and T2(λ) are periodic functions.
(4) In one embodiment of the 3D image capture device of one of (1) to (3), the first type of pixel includes a first photosensitive cell and a first transmitting filter which is arranged so as to face the first photosensitive cell. The second type of pixel includes a second photosensitive cell and a second transmitting filter which is arranged so as to face the second photosensitive cell. T1(λ) is a function representing the spectral transmittance of the first transmitting filter. And T2(λ) is a function representing the spectral transmittance of the second transmitting filter:
(5) In one embodiment of the 3D image capture device of (4), each unit block further includes a third type of pixel and a fourth type of pixel. The third type of pixel includes a third photosensitive cell and a third transmitting filter which is arranged to face the third photosensitive cell and which selectively transmits a light ray with a first color component. The fourth type of pixel includes a fourth photosensitive cell and a fourth transmitting filter which is arranged to face the fourth photosensitive cell and which selectively transmits a light ray with a second color component. The image processing section generates color information based on photoelectrically converted signals supplied from the third and fourth types of pixels.
(6) In one embodiment of the 3D image capture device of one of (1) to (3), the first type of pixel includes a plurality of photosensitive cells and a plurality of transmitting filters, each of which is arranged so as to face an associated one of the photosensitive cells and which have mutually different spectral transmittance characteristics. The second type of pixel includes a single photosensitive cell and a single transmitting filter which is arranged so as to face the single photosensitive cell. TL(λ) is a function representing the sum of the spectral transmittances of the transmitting filters included in the first type of pixel. And T2(λ) is a function representing the spectral transmittance of the single transmitting filter included in the second type of pixel.
(7) In one embodiment of the 3D image capture device of (6), the transmitting filters included in the first type of pixel are configured to selectively transmit light rays with mutually different color components.
(8) In one embodiment of the 3D image capture device of (7), the number of photosensitive cells and the number of transmitting filters included in the first type of pixel are both three. First, second and third ones of those three transmitting filters are configured to selectively transmit a light ray with a color red component, a light ray with a color green component, and a light ray with a color blue component, respectively.
(9) In one embodiment of the 3D image capture device of one of (1) to (3), the first type of pixel includes a plurality of photosensitive cells and a plurality of transmitting filters, each of which is arranged so as to face an associated one of the photosensitive cells and which have mutually different spectral transmittance characteristics. The second type of pixel includes a plurality of photosensitive cells and a plurality of transmitting filters, each of which is arranged so as to face an associated one of the photosensitive cells and which have mutually different spectral transmittance characteristics. T1(λ) is a function representing the sum of the spectral transmittances of the transmitting filters included in the first type of pixel. And T2(λ) is a function representing the sum of the spectral transmittances of the transmitting filters included in the second type of pixel.
(10) In one embodiment of the 3D image capture device of (9), each of the first and second types of pixels includes first through Nth (where N is an integer that is equal to or greater than three) photosensitive cells and first through Nth transmitting filters, each of which is arranged so as to face an associated one of the first through Nth photosensitive cells. The first through Nth transmitting filters included in the first type of pixel are configured to selectively transmit light rays with first through Nth color components, respectively. The first through Nth transmitting filters included in the second type of pixel are configured to selectively transmit light rays with the first through Nth color components, respectively. And the spectral transmittance characteristic of an ith (where i is an integer that fails within the range of one through N) transmitting filter, included in the first type of pixel, with respect to the ith color component is different from that of an ith transmitting filter, included in the second type of pixel, with respect to the ith color component.
(11) In one embodiment of the 3D image capture device of (10), N=3, and the first through third color components are red, green and blue components, respectively.
(12) In one embodiment of the 3D image capture device of one of (1) to (11), the image processing section generates color multi-viewpoint images based on the signals supplied from the first and second types of pixels.
(13) In one embodiment of the 3D image capture device of one of (1) to (12), the device further includes a storage section which stores an output signal of the image sensor, and the image processing section generates the multi-viewpoint images based on the signal that is stored in the storage section.
(14) An image sensor as an embodiment of the present invention is used in the 3D image capture device of one of (1) to (13).
(15) A light transmitting section as an embodiment of the present invention is used in the 3D image capture device of one of (1) to (13).
(16) An image processor as an embodiment of the present invention generates an image based on a signal supplied from a 3D image capture device. The device includes: a light transmitting section with first and second light transmitting areas; an image sensor which is arranged to receive light that has been transmitted through the light transmitting section and which has an imaging area where a plurality of unit blocks, each including a first type of pixel and a second type of pixel, are arranged; and an imaging section which produces an image on the imaging area of the image sensor. If functions representing the respective spectral transmittances of the first light transmitting area, the second light transmitting area, the first type of pixel, and the second type of pixel with respect to the wavelength λ of visible radiation are identified by TL(λ), TR(λ), TL(λ), and T2(λ), respectively, TL(λ)≠TR(λ) and T1(λ)≠T2(λ) are satisfied, and each of TL(λ), TR(λ), T1(λ) and T2(λ) has at least one local maximum value and at least one local minimum value within each of red, green and blue wavelength ranges. The image processor generates multi-viewpoint images based on photoelectrically converted signals supplied from the first and second types of pixels.
Before specific embodiments of the present invention are described, first of all, the basic principle of an embodiment of the present invention will be described briefly. In the following description, a signal or information representing an image will be sometimes referred to herein as just an “image”.
A 3D image capture device as an embodiment of the present invention (which will be simply referred to herein as an “image capture device”) includes a light transmitting section with two light transmitting areas, an image sensor which has an imaging area where a plurality of unit blocks, each including at least two kinds of pixels, are arranged, an imaging section which produces an image on the imaging area of the image sensor, and an image processing section which generates multi-viewpoint images based on the output signal of the image sensor.
The light transmitting section 1 shown in
The two photosensitive cells 60-1 and 60-2 each receive the incident light and output an electrical signal representing the intensity of the light received (which will be referred to herein as a “photoelectrically converted signal” or a “pixel signal”). The transmitting filters 40-1 and 40-2 that face the photosensitive cells 60-1 and 60-2, respectively, have the property of cutting the incoming light partially. The transmitting filters 40-1 and 40-2 also have mutually different spectral transmittance characteristics, are not completely transparent, but have the property of transmitting a light ray falling within any of the red (R), green (G) and blue (B) wavelength ranges.
As shown in
As long as TL≠TR and T1≠T2 are satisfied and as long as each of TL, TR, T1 and T2 has a local maximum value and a local minimum value within each of the red, green and blue wavelength ranges, the spectral transmittance characteristics do not have to be the ones shown in
With such an arrangement adopted, multi-viewpoint images are generated based on the photoelectrically converted signals supplied from the two photosensitive cells 60-1 and 60-2, which will be identified herein by W1s and W2s, respectively, in the following description. Meanwhile, the integrals of the transmittances of the light that has been transmitted through the light transmitting area 1L and then incident on the photosensitive cells 60-1 and 60-2 will be identified herein by W1Ls and W2Ls, respectively. In the same way, the integrals of the transmittances of the light that has been transmitted through the light transmitting area 1R and then incident on the photosensitive cells 60-1 and 60-2 will be identified herein by W1Rs and W2Rs, respectively. That is to say, W1Ls, W2Ls, W1Rs and W2Rs are given by the following Equations (1) through (4), respectively:
W1Ls=k∫TL(λ)T1(λ)dλ (1)
W1Rs=k∫TR(λ)T1(λ)dλ (2)
W2Ls=k∫TL(λ)T2(λ)dλ (3)
W2Rs=k∫TR(λ)T2(λ)dλ (4)
In this case, the incoming light is supposed to have an achromatic color. That is to say, the wavelength dependence of the intensity of the incoming light is supposed to be a negligible one and components of the light other than visible radiation are supposed to have been filtered out by an infrared cut filter, for example. Also, k is a factor of proportionality and the interval of integration is supposed to be the visible radiation wavelength range (e.g., from 400 nm through 700 nm).
Also, supposing that the light transmitting areas 1L and 1R and the transmitting filters 40-1 and 40-2 are completely transparent, signals representing the intensities of light rays that pass through the light transmitting areas 1L and 1R and are incident on a single photosensitive cell are identified by IMG(L) and IMG(R), respectively. In this case, as the spatial sizes of the photosensitive cells 60-1 and 60-2 are very small and as the photosensitive cells 60-1 and 60-2 are located close to each other, the intensities of the light rays incident on the photosensitive cells 60-1 and 60-2 are supposed to be substantially equal to each other. Then, IMG(L) and IMG(R) correspond to respective portions of an image when the subject is viewed from the positions of the light transmitting areas 1L and 1R, respectively. That is why if IMG(L) and IMG(R) are obtained for every pixel that is used to generate an image, multi-viewpoint images can be obtained.
If the factor of proportionality is neglected, the photoelectrically converted signals W1s and W2s can be represented by the following Equation (5) using a matrix:
As can be seen from this Equation (5), the signals IMG(L) and IMG(R) representing the multi-viewpoint images can be calculated by the following Equation (6) using the inverse matrix of the matrix represented by Equation (5):
The respective elements of the matrix represented by this Equation (6) are known numbers to be derived from the spectral transmittance characteristics of the light transmitting areas 1L and 1R and the transmitting filters 40-1 and 40-2. That is why by making the calculation represented by this Equation (6), the image processing section can generate multi-viewpoint images. According to an embodiment of the present invention, as the light transmitting areas 1L and 1R and the transmitting filters 40-1 and 40-2, not ordinary color filters but filters with high transmittance such as the ones shown in
In the example described above, the transmitting filters 40-1 and 40-2 are each supposed to have the characteristic shown in
And the characteristic of the transmitting filters 40-1 and 40-3 combined is supposed to agree with the characteristic represented by TL(λ) in
In the example shown in
Hereinafter, specific embodiments of the present invention will be described with reference to
First of all, an image capture device as a first embodiment of the present invention will be described.
The image capturing section 100 includes a color solid-state image sensor 2a (which will be simply referred to herein as an “image sensor”) with a number of photosensitive cells that are arranged on its imaging area, a light-transmitting plate (light-transmitting section) 1, which has two transmitting areas, with mutually different spectral transmittances, an optical lens 3 for producing an image on the imaging area of the image sensor 1, and an infrared cut filter 4. The image capturing section 100 further includes a signal generating and receiving section 5, which not only generates a fundamental signal to drive the image sensor 2a but also receives the output signal of the image sensor 2a and sends it to the signal processing section 200, and a sensor driving section 6 for driving the image sensor 2a in accordance with the fundamental signal generated by the signal generating and receiving section 5. The image sensor 2a is typically a CCD or CMOS sensor, which may be fabricated by known semiconductor device processing technologies. The signal generating and receiving section 5 and the sensor driving section 30 may be implemented as an LSI such as a CCD driver.
The signal processing section 200 includes an image processing section 7 for generating multi-viewpoint images by processing the signal supplied from the image capturing section 100, a memory 30 for storing various kinds of data for use to generate the image signal, and an interface (I/F) section 8 for sending out the image signal thus generated and depth information to an external device. The image processing section 7 is suitably a combination of a hardware component such as a known digital signal processor (DSP) and a software program for use to perform image processing involving the image signal generation. The memory 30 may be a DRAM, for example. And the memory 30 not only stores the signal supplied from the image capturing section 100 but also temporarily retains the image data that has been generated by the image processing section 7 or compressed image data. These image data are then output to either a storage medium or a display section (neither is shown) by way of the interface section 8.
The image capture device of this embodiment actually further includes an electronic shutter, a viewfinder, a power supply (or battery), a flashlight and other known components. However, description thereof will be omitted herein because none of them are essential components that would make it difficult to understand how the present invention works unless they were described in detail. Also, this configuration is only an example. Thus, in this embodiment, additional components other than the light-transmitting plate 1, the image sensor 2a and the image processing section 7 may be implemented as an appropriate combination of known elements.
Hereinafter, the configuration of the image capturing section 100 will be described in further detail. In the following description, the xy coordinate system shown on the drawings will be used.
The light transmitting areas 1L and 1R of the light-transmitting plate 1 and the transparent elements W1 and W2 of the image sensor 1 may be made of a multilayer dielectric film. By using such a multilayer dielectric film, a transmitting filter, of which the transmittance becomes either local maximum or local minimum with respect to light with a predetermined wavelength, can be made. Meanwhile, the red (R) and blue (B) elements of the image sensor 1 may be made of a known pigment, for example.
According to such an arrangement, the light that has entered this image capture device during an exposure process is transmitted through the light-transmitting plate 1, the lens 3, and the infrared cut filter 4, imaged on the imaging area 2b of the image sensor 2a, and then photoelectrically converted by the photosensitive cells 60. The photoelectrically converted signal that has been output from each photosensitive cell 60 is sent to the signal processing section 200 by way of the signal generating and receiving section 5. In the signal processing section 200, the image processing section 7 generates two multi-viewpoint images based on the signals supplied. According to this embodiment, color information is further superposed on these multi-viewpoint images, thereby generating color multi-viewpoint images.
Hereinafter, it will be described how the image capture device operates when a subject is captured through the light-transmitting plate 1. In the following description, as for respective pixel signals of the image sensor 2a, a signal representing the light that has been transmitted through the R element and then photoelectrically converted will be identified herein by Rs, a signal representing the light that has been transmitted through the B element and then photoelectrically converted by Bs, a signal representing the light that has been transmitted through the W1 element and then photoelectrically converted by W1s, and a signal representing the light that has been transmitted through the W2 element and then photoelectrically converted by W2s.
The color red and color blue components of the subject are transmitted through the light transmitting areas 1L and 1R of the light-transmitting plate 1, the lens, the infrared cut filter and the red and blue elements and then photoelectrically converted by the image sensor 2a, thereby generating Rs and Bs signals. The color white component of the subject (i.e., including every one of the color red, color green and color blue components) is also photoelectrically converted by the W1 and W2 elements of the image sensor but its signal level varies according to the magnitude of difference in the quantity of light incident between the light transmitting areas 1L and 1R. The reason is that the W1 and W2 elements not only have mutually different spectral transmittance characteristics but also have the same characteristics as the light transmitting areas 1L and 1R. In other words, the W1 element passes more light that has been transmitted through the light transmitting area 1L than the light that has been transmitted through the light transmitting area 1R, and the W2 element passes more light that has been transmitted through the light transmitting area 1R than the light that has been transmitted through the light transmitting area 1L. As a result, the W1 and W2 elements come to have mutually different signal levels.
Hereinafter, the photoelectrically converted signals W1s and W2s of this embodiment will be described. First of all, the signal W1Ls representing the light that passes through the light transmitting area 1L of the light-transmitting plate 1 and the W1 element and then is photoelectrically converted and the signal W1Rs representing the light that passes through the light transmitting area 1R and the W1 element and then is photoelectrically converted will be described. As described above, the transmittance of the light transmitting area 1L and the W1 element is proportional to the square of the trigonometric function Sin, the transmittance of the light transmitting area 1R and the W2 element is proportional to the square of the trigonometric function Cos and a photoelectric conversion efficiency including those factors of proportionality is supposed to be k.
The signal W1Ls is proportional to the integral of the respective transmittances of the light transmitting area 1L and the W1 element, and therefore, is represented by the following Equation (1). On the other hand, the signal W1Rs is proportional to the integral of the respective transmittances of the light transmitting area 1R and the W1 element, and therefore, is represented by the following Equation (2). In this case, the integration operation is performed on the entire visible radiation wavelength range.
W1Ls=k∫sin2[(λ−400)π/100] sin2[(λ−400)π/100]dλ (7)
W1Rs=k∫cos2[(λ−400)π/100] sin2[(λ−400)π/100]dλ (8)
If the interval of integration is set to be the visible radiation range of 400 nm to 700 nm and if the data is replaced so that the angle θ is 0 radians when the wavelength λ is 400 nm and is n radians when the wavelength λ is 500 nm by reference to
W1Ls=6k=∫0π/2 sin2θ×2θdθ (9)
W1Rs=6k∫0π/2 cos2θ×2θdθ (10)
Equations (9) and (10) may be further modified into the following Equations (11) and (12), respectively:
In the same way, the signal W2Ls is proportional to the integral of the respective transmittances of the light transmitting area 1L and the W2 element, and therefore, is represented by the following Equation (13). On the other hand, the signal W2Rs is proportional to the integral of the respective transmittances of the light transmitting area 1R and the W2 element, and therefore, is represented by the following Equation (14):
W2Ls=k∫ cos2[(λ−400)π/100] sin2[(λ−400)π/100]dλ (13)
W2Rs=k∫ cos2[(λ−400)π/100] cos2[(λ−400)π/100]dλ (14)
In this case, since the right side of Equation (13) is equal to the right side of Equation (8), the same result of calculation is obtained by these two Equations (13) and (8). If the variable of integration λ is converted into θ as in the example described above with respect to Equation (14), the following Equation (15) is obtained:
Consequently, W1Ls: W1Rs==3:1 and W2Ls: W2Rs:=1:3. This means that as for a signal representing the light ray that has been transmitted through the W1 element and then photoelectrically converted, its component that has been transmitted through the light transmitting area 1L and its component that has been transmitted through the light transmitting area 1R have mutually different signal levels, of which the ratio is 3:1. On the other hand, this means that as for a signal representing the light ray that has been transmitted through the W2 element and then photoelectrically converted, its component that has been transmitted through the light transmitting area 1L and its component that has been transmitted through the light transmitting area 1R have a signal ratio is 1:3.
If the images generated by the light beams that have come through the light transmitting areas 1L and 1R are identified by IMG(L) and IMG(R), respectively, then the relation between those images and the pixel signals W1s and W2s is represented by the following Equation (16):
Furthermore, by multiplying both sides of this Equation (16) by the inverse matrix of the 2×2 matrix on the right side of Equation (16), the following Equation (17) is obtained:
As can be seen from the foregoing description, by using the light-transmitting plate 1 and the transparent elements W1 and W2 of this embodiment, the image processing section 7 can generate multi-viewpoint images. According to this embodiment, the transparent elements W1 and W2 have the property of transmitting light rays in all of the three primary colors of RGB, and therefore, signals representing the light rays that have been transmitted through those transparent elements W1 and W2 and then photoelectrically converted can be used as luminance signals.
Next, the color signal processing of this embodiment will be described. By adopting the configuration described above, pixel signals W1s, W2s, Rs and Bs are obtained from the image sensor 2a. The image processing section 7 processes W1s and W2s as luminance signals, and processes Rs and Bs as color signals. The respective pixel signals obtained by the image sensor 2a are passed to the image processing section 7, where calculation based on Equation (17) is performed on the pixel signals W1s and W2s, thereby generating two images with parallax (i.e., images represented by the signals IMG(L) and IMG(R)). However, these multi-viewpoint images have only luminance information and do not have any color information. That is why a luminance signal Yl is obtained by YL=W1s+W2s, color difference signals (Rs−YL) and (Bs−YL) are generated, and then signals obtained by turning these color difference signals into low frequency ones through a band-pass filter are superposed on the multi-viewpoint images obtained. After that, the image processing section 7 performs the processing of coloring the multi-viewpoint images in the following manner. If one of the colored multi-viewpoint images is identified by cIMG(L) and its RGB components are identified by cIMG(L)r, cIMG(L)g and cIMG(L)b, respectively, then cIMG(L)r, cIMG(L)g and cIMG(L)b are given by the following Equation (18):
Equation (18) is used to transform the luminance signal and two color difference signals into RGB signals and the elements of its conversion matrix are identified by M11 through M33, respectively.
In the same way, if the other of the colored multi-viewpoint images is identified by cIMG(R) and its RGB components are identified by cIMG(R)r, cIMG(R)g and cIMG(R)b, respectively, then cIMG(R)r, cIMG(R)g and cIMG(R)b are given by the following Equation (19):
The matrix elements M11 through M33 of Equations (18) and (19) are obtained in advance by simulation. By performing computations based on these Equations (18) and (19), the image processing section 7 transforms the luminance signal and two color difference signals into color signals cIMG(L)r, cIMG(L)g and cIMG(L)b.
In this embodiment, multi-viewpoint images are generated based on the luminance information and the same color signal is used in common for the two multi-viewpoint images when the coloring processing is performed after that. As far as a human being's visual sensation is concerned, he or she should be able to recognize colors less sensitively than luminance by nature. That is why even if color multi-viewpoint images are generated by such processing, no problem will arise as far as the human visual sensation is concerned.
As can be seen, according to this embodiment, a light-transmitting plate 1 with two light transmitting areas 1L and 1R that have mutually different spectral transmittance characteristics but that transmit any light ray falling within the visible radiation range and an image sensor 2a, in which two kinds of transparent elements W1 and W2 with the same optical properties as the light transmitting areas 1L and 1R are arranged, are used. Thus, multi-viewpoint images can be generated based on two pieces of luminance information. In addition, by using signals of pixels with R and B elements, the multi-viewpoint images can be colored. Particularly, according to this embodiment, the light-transmitting plate 1 has no opaque portion and can transmit a light ray in any of the three primary colors of RGB, and therefore, the incoming light can be used highly efficiently, which is beneficial.
Although red and blue elements R and B are used in the embodiment described above as two color elements in addition to transparent elements W1 and W2, those two color elements do not have to be red and blue elements. But the two color elements may have any combination of colors as long as the two colors are different. In addition, those color elements do not have to be arranged as shown in
Hereinafter, a second embodiment of the present invention will be described. In this second embodiment, the image sensor 2a has a different basic color scheme from the first embodiment. But other than that, this embodiment is the same as the first embodiment. Thus, the following description of the second embodiment will be focused on those differences and their common features will not be described all over again to avoid redundancies.
Hereinafter, the signal processing of this embodiment will be described in comparison with the signal processing of the first embodiment. Specifically, the image sensor 2a of the first embodiment includes a W1 element and generates an image using its signal. In this embodiment, on the other hand, the sum of the signals supplied from the R, G and B elements, not the W1 element, is used as a W1s signal. As a result, the computation performed on the W1 and W2 elements becomes the same as that of the first embodiment. The images IMG(L) and IMG(R) produced by the light rays that have come through the light transmitting areas 1L and 1R are also generated by Equation (17) as in the first embodiment described above.
The multi-viewpoint images can also be colored by performing the same processing as in the first embodiment described above. First, by using the signals Rs, Bs, W1s and W2s, the image processing section 7 generates a luminance signal YL=W1s+W2s and two color difference signals (Rs−YL) and (Bs−YL). Next, these color difference signals are turned into low frequency ones by a band-pass filter and then those low-frequency signals are superposed on the two multi-viewpoint images. After that, the same processing is carried out based on Equations (18) and (19) as in the first embodiment, thereby generating color multi-viewpoint images.
As can be seen, according to this embodiment, by using an image sensor 2a in which R, G, B and W2 elements are arranged, color multi-viewpoint images can be generated. And if the sum of the respective spectral transmittance characteristics of the respective color R, G and B elements is the same as the characteristic of the light transmitting area 1L shown in
In this embodiment, the basic color scheme of the image sensor 2a is supposed to consist of R, G, B and W2 elements as shown in
Hereinafter, a third embodiment of the present invention will be described. In the image capture device of this embodiment, the two light transmitting areas 1L and 1R of the light-transmitting plate 1 have different spectral transmittance characteristics, and the image sensor 2a has a different filter arrangement, from those of the image capture device of the first embodiment. Thus, the following description of the third embodiment will be focused on those differences and their common features will not be described all over again to avoid redundancies.
In this embodiment, the basic color scheme of a single unit block of the image sensor 2a is supposed to consist of three rows by two columns of elements.
The spectral transmittance characteristics of the light-transmitting plate 1 are also different from those of the first and second embodiments described above.
Look at the respective phases of the spectral transmittance curves of respective color elements of the image sensor 2a and those of the spectral transmittance curves of respective light transmitting areas of the light-transmitting plate 1, and it can be seen that the characteristics of the R1, G1 and B1 elements have substantially the same phase as the characteristic of the light transmitting area 1L but almost the inverse phase to the characteristic of the light transmitting section 1R and that the characteristics of the R2, G2 and B2 elements have substantially the same phase as the characteristic of the light transmitting area 1R but have almost the inverse phase to the characteristic of the light transmitting section 1L. In the following description, the R1, G1 and B1 elements will be collectively referred to herein as a “G1 element”, the R2, G2 and B2 elements will be collectively referred to herein as a “C2 element” and photoelectrically converted signals generated by converting the light rays that have been transmitted through those C1 and C2 elements will be identified herein by generalized signs C1s and C2s, respectively. According to the spectral transmittance characteristics described above, the light ray transmitted through the light transmitting area 1L and the C1 element has a larger quantity of light than the light ray transmitted through the light transmitting area 1R and the C1 element. In the same way, the light ray transmitted through the light transmitting area 1R and the C2 element has a larger quantity of light than the light ray transmitted through the light transmitting area 1L and the C2 element.
With such a configuration adopted, the relation between the signals C1s and C2s, the image IMG(L) generated by the light that has come through the light transmitting area 1L and the image IMG(R) generated by the light that has come through the light transmitting area 1R is represented by the following Equation (20):
These matrix elements C11 through C22 satisfy C11>C12 and C22>C21. Since this relation is satisfied, the inverse matrix of the matrix consisting of these elements C11 through C22 can always be obtained, and the images IMG(L) and IMG(R) can be represented by the signals C1s and C2s, respectively. Even though signals representing respective colors are given by C1s and C2s in this example, the computation of this Equation (20) is actually carried out based on the signals representing the respective colors of RGB. That is to say, the images IMG(L) and IMG(R) can be calculated based on those signals representing the respective colors of RGB.
As can be seen, according to this embodiment, an image sensor 2a with a set of R elements (R1, R2), a set of G elements (G1, G2) and a set of B elements (B1, B2), each set of which has mutually different spectral transmittance characteristics, and a light-transmitting plate 1 with two light transmitting areas 1L and 1R, which also have mutually different spectral transmittance characteristics, are used. The characteristics of the R1, G1 and B1 elements and the characteristic of the light transmitting area 1L have almost the same phase, so do the characteristics of the R2, G2 and B2 elements and the characteristic of the light transmitting area 1R. And the characteristics of the R1, G1 and B1 elements and the characteristic of the light transmitting area 1R have almost inverse phases, so do the characteristics of the R2, G2 and B2 elements and the characteristic of the light transmitting area 1L. By adopting such a configuration, parallax images IMG(L) and IMG(R) can be obtained based on the respective color signals. Furthermore, since the light-transmitting plate 1 transmits light rays representing the respective colors of RGB, the incoming light can be used more efficiently as well.
In the embodiment described above, the basic color scheme of the image sensor 2a is supposed to consist of three rows by two columns of elements. However, this is only an example and there is no problem at all even if the basic color scheme consists of two rows by three columns of elements. Also, the color elements to use do not have to be R, G and B elements but may be any other combination of color elements because as long as at least three different kinds of color elements are used, the signals can be converted into RGB signals by performing matrix computations during the signal processing. For example, the R, G and B elements may be replaced with the color filters in cyan (Cy), yellow (Ye) and magenta (Mg) as well.
In the embodiment described above, three different pairs of color filters are included in a single unit block. However, more than three pairs of color filters may be included as well. If the configuration of this embodiment is generalized, each of the first and second types of pixels may include first through Nth (where N is an integer that is equal to or greater than three) photosensitive cells and first through Nth transmitting filters which are arranged so as to face those photosensitive cells. In that case, in any type of pixels, the first through Nth transmitting filters are configured to selectively transmit light rays with first through Nth color components. And those transmitting filters are designed so that the spectral transmittance characteristic of the ith (where i is an integer that falls within the range of 1 through N) transmitting filter, which is included in the first type of pixels, with respect to the ith color component is different from the spectral transmittance characteristic of the ith transmitting filter, which is included in the second type of pixels, with respect to the ith color component. With such a configuration adopted, if the sum of the respective outputs of the N photosensitive cells is treated as a single output, the effects of this embodiment can also be achieved.
In the embodiments described above, the light-transmitting plate 1 is supposed to be split into two and the respective areas are supposed to have mutually different spectral transmittance characteristics. However, this is just an example of the present invention and such a configuration is not necessarily adopted. Alternatively, the light-transmitting plate 1 may be divided into three areas, two of which may have the same characteristics as the two light transmitting areas according to any of the embodiments described above and the other of which may be completely transparent. Even so, multi-viewpoint images can also be obtained. For example, if the basic color scheme of the image sensor 2a is as shown in
By multiplying this Equation (21) by the inverse one of the matrix of Equation (21) from the left to the right, signals IMG(L) and IMG(R), representing multi-viewpoint images, can be obtained. In this example, however, the ratios (C11: C12) and (C22: C21) with respect to the multi-viewpoint image change into (C11+W: C12+W) and (C22+W: C21+W), respectively. As a result, the magnitude of parallax obtained decreases compared to a situation where there is no transparent component W.
In the embodiments described above, the output signals of the image sensor 2a are supposed to be directly processed to generate signals with parallax. However, the output signals of the image sensor 2a may be once stored on a storage medium and then multi-viewpoint images may be generated while the data stored on the storage medium is being read out.
Furthermore, in the embodiments described above, the image processing section 7 generates color multi-viewpoint images. If no color information is needed, however, the image processing section 7 may also be configured to generate only monochrome multi-viewpoint images and perform no coloring processing. In that case, there is no need to use any color image sensor and the image sensor may have two transmitting filters such as the ones shown in
In the embodiments described above, the image processing is supposed to be carried out by the image processing section that is built in the image capture device. However, such image processing may also be carried out by another device that is provided independently of that image capture device. For example, even if a signal that has been obtained by an image capture device including the image capturing section 100 of the embodiment described above is loaded into another device (image processor) to get a program defining the signal arithmetic processing described above executed by a computer built in the image processor, the effects of the embodiments described above can also be achieved.
A 3D image capture device according to an embodiment of the present invention can be used effectively in any camera that ever uses a solid-state image sensor. Examples of those cameras include consumer electronic cameras such as digital cameras and digital camcorders and solid-state surveillance cameras for industrial use.
Number | Date | Country | Kind |
---|---|---|---|
2011-112101 | May 2011 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2012/003027 | 5/9/2012 | WO | 00 | 1/17/2013 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2012/157209 | 11/22/2012 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6807295 | Ono | Oct 2004 | B1 |
20020171740 | Seo | Nov 2002 | A1 |
20070102622 | Olsen et al. | May 2007 | A1 |
20070268377 | Nagano et al. | Nov 2007 | A1 |
20090284627 | Bando et al. | Nov 2009 | A1 |
20100066854 | Mather et al. | Mar 2010 | A1 |
20110254985 | Hiramoto et al. | Oct 2011 | A1 |
20120133743 | Hiramoto et al. | May 2012 | A1 |
Number | Date | Country |
---|---|---|
101076126 | May 2007 | CN |
101427372 | May 2009 | CN |
62-217790 | Sep 1987 | JP |
62-291292 | Dec 1987 | JP |
02-171737 | Jul 1990 | JP |
2001-016611 | Jan 2001 | JP |
2002-344999 | Nov 2002 | JP |
2003-134533 | May 2003 | JP |
2009-276294 | Nov 2009 | JP |
2010-038788 | Feb 2010 | JP |
2010-079298 | Apr 2010 | JP |
WO 2006026354 | Mar 2006 | WO |
Entry |
---|
International Search Report for corresponding International Application No. PCT/JP2012/003027 mailed Jul. 24, 2012. |
Chinese Search report for corresponding Chinese Application No. 201280002061.0 (with English translation), dated Apr. 23, 2015. |
Number | Date | Country | |
---|---|---|---|
20130120541 A1 | May 2013 | US |