This application is based on Japanese Patent Application No. 2004-353957 filed in Japan on 7 Dec. 2004, the entire content of which is hereby incorporated by reference.
The present invention relates to the technical field of an image sensor comprising a plurality of pixels arranged in a matrix, and more particularly, to an image sensor having color pixels where color filters are disposed and monochrome pixels where no color filter is disposed, an image capturing apparatus provided with the image sensor, and an image processing method using the pixel signals obtained from the image sensor.
Generally, in an image sensor of a Bayer arrangement in which color filters of, for example, R (red), G (green) and B(blue) having different spectral characteristics are disposed at a ratio of 1:2:1, since the light directed to the photoelectrically conversion portions of the pixels is attenuated by the color filters, the effective sensitivity is low compared to an image sensor in which no color filter is disposed. In recent years, as image sensors have decreased in size and increased in the number of pixels, the size of one pixel has been reduced and the light reception amount per pixel is further reduced, so that the effective sensitivity of image sensors is further reduced and the dynamic range tends to be small.
Consequently, it is frequently necessary to emit flash light to ensure a necessary light reception amount, so that various problems arise such that power consumption increases to decrease the number of images that can be taken, that when a so-called camera shake compensation function is provided, the effect of compensation is small even though the camera shake compensation is performed and that the S/N ratio deteriorates (decreases) due to an increase in the amplification factor of the pixel signal obtained from each pixel.
Japanese Laid-Open Patent Application No. H09-116913 discloses an art such that in an image sensor, color filters of R (red) or B (blue) are disposed at a half of the pixels and no color filter is disposed at the remaining half of the pixels in order to improve the effective sensitivity of the image sensor.
However, according to the art of the patent document 1, since the pixels having no color filter are only a half of all the pixels of the image sensor, a significant improvement in effective sensitivity cannot be expected.
The present invention is made in view of the above-mentioned circumstances, and an object thereof is to provide an image sensor, an image capturing apparatus and an image processing method with high effective sensitivity.
The above-mentioned object is attained by providing the following structure:
According to an image sensor of the present invention, in an image sensor comprising a plurality of pixels arranged in a matrix and having pixels where at least three kinds of color filters are disposed, color pixels where the color filters are disposed and monochrome pixels where no color filter is disposed are provided, the sum total of the monochrome pixels is larger than the sum total of the color pixels, and when pixels including a predetermined number of color pixels for each kind of color filter constitute one group, the color pixels or the pixels of the groups are dispersedly disposed with the monochrome pixels in between.
According this aspect of the invention, the effective sensitivity of the image sensor can be improved compared to the image sensors in which only color pixels where color filters are disposed are arranged and the conventional image sensors in which the sum total of the monochrome pixels is equal to or smaller than the sum total of the color pixels.
Consequently, since the effective sensitivity of the image sensor is improved, photographing with high sensitivity can be performed, so that a beautiful image with an excellent (high) S/N ratio can be obtained.
These and other objects and features of the present invention will become clear from the following description taken in conjunction with the preferred embodiments thereof with reference to the accompanying drawings, in which:
FIGS. 4(a) to 4(c) are views showing an example of the arrangement of color pixels and monochrome pixels;
FIGS. 6(a) to 6(d) are views for explaining an example of a method of interpolation of the color data in the positions of monochrome pixels when a live view image or a moving image is generated;
FIGS. 7(a) to 7(c) are views for explaining an example of a method of interpolation of the brightness data in the positions of the color pixels when a still image is generated;
FIGS. 12(a) to 12(c) are views showing a modification of the interpolation of the brightness data in the positions of the color pixels;
FIGS. 14(a) and 14(b) are views for explaining a method of generating the brightness data by use of the pixel signals obtained from the color pixels;
FIGS. 30(a) and 30(b) are views showing another color pixel arrangement.
A first embodiment of the present invention will be described.
As shown in
The power button 2 is for turning on and off the image capturing apparatus 1. The optical system 3 comprises a zoom lens and a non-illustrated mechanical shutter, and forms an optical image of the subject on the image capturing surface of an image sensor 10 (see
The LCD 4 is for displaying a live view image and images recorded in an image storage portion 17 described later (see
The live view image is a series of images displayed on the LCD 4 so as to be switched at predetermined intervals ( 1/30 second) in a period up to the recoding of the image of the subject. By the live view image, the condition of the subject is displayed substantially in real time on the LCD 4, so that the user can confirm the condition of the subject on the LCD 4.
The optical viewfinder 5 is for enabling the photographed area of the subject to be viewed optically. The built-in flash 6 applies illumination light to the subject by causing a non-illustrated discharge lamp to discharge, for example, when the amount of exposure to the image sensor 10 is insufficient.
The mode setting switch 7 is for switching the mode among a “still image photographing mode” to take still images of the subject, a “moving image photographing mode” to take moving images of the subject and a “playback mode” to play back the taken images recorded in the image storage portion 17 (see
The quadruple switch 8 is, although not described in detail, for setting a menu mode to make the setting of various functions, moving the zoom lens in the direction of the optical axis, performing exposure compensation and advancing the frame of the recorded images played back on the LCD 4.
The shutter button 9 is a button depressed in two strokes (a half depression and a full depression), and for providing the timing of the exposure control. The image capturing apparatus 1 has the still image photographing mode to take still images and the moving image photographing mode to take moving images. When the still image photographing mode or the moving image photographing mode are set, under a condition where the shutter button 9 is not operated, an optical image of the subject is captured every 1/30 (second), and the live view image is displayed on the LCD 4.
In the still image photographing mode, by the LCD 4 being half depressed, the image capturing apparatus 1 is set in a photographing standby state in which the exposure control values (the shutter speed and the aperture value) and the like are set, and by the LCD 4 being fully depressed, the exposure operation (exposure operation for recording) by the image sensor 10 to generate an subject image to be recorded in the image storage portion 17 (see
In the moving image photographing mode, by the LCD 4 being fully depressed, the exposure operation for recording is started, pixel signals are periodically obtained and images are successively generated by the pixel signals, and by the LCD 4 being fully depressed again, the exposure operation for recording is stopped.
The image capturing apparatus 1 is provided with the optical system 3, the LCD 4, the image sensor 10, a timing generator 11, a signal processor 12, an A/D converter 13, an image memory 14, a VRAM (video random access memory) 15, an operation portion 16, the image storage portion 17 and a controller 18.
The optical system 3 corresponds to the optical system 3 shown in
The image sensor 10 is a CCD color area sensor in which a plurality of photoelectrical conversion elements comprising, for example, photodiodes (hereinafter, referred to as pixels) are two-dimensionally arranged in a matrix.
In the case of the conventional color area sensors of the Bayer arrangement in which color filters of, for example, R (red), G (green) and B (blue) having different spectral characteristics are disposed at a ratio of 1:2:1, since the color filters attenuate the light directed to the photodiode of the pixels, the effective sensitivity of the image sensor is low. In particular, in image sensors reduced in size and increased in the number of pixels, since the size (light reception area) of each pixel is small and the light reception amount of each pixel is small, the reduction in effective sensitivity is larger.
To resolve this deficiency, as shown in
That is, in the example shown in
The sensitivity of the monochrome pixels is, for example, three times the sensitivity of the color pixels of G (green) and five times the sensitivities of the color pixels of R (red) and B (blue).
The image sensor 10 converts the light image of the subject formed by the optical system 3 into analog electric signals, and outputs the electric signals as pixel signals. From the pixel signals outputted from the color pixels, the analog color data and brightness data of the color components of R (red), G (green) and B (blue) are obtained, and from the pixel signals outputted from the monochrome signals, the brightness data is obtained.
The image sensor 10 is, for example, an interline image sensor provided with light reception portions each comprising a photodiode or the like, vertical transfer portions, horizontal transfer portions and the like, and the charges of the pixels are taken out by a progressive transfer method. That is, the charges accumulated in the light reception portions are transferred to the vertical transfer portions by a vertical synchronizing signal and the charges transferred to the vertical transfer portions are transferred to a horizontal transfer path from the pixels closer to the horizontal transfer path by a horizontal synchronizing signal, whereby the charges are taken out as pixel signals. Image capturing operations such as the readout of the output signals of the pixels at the image sensor 10 (horizontal synchronization and vertical synchronization) and the timing of the start and end of the exposure operation by the image sensor 10 are controlled by the timing generator 11 and the like described later.
As described later, in the present embodiment, the image generation method is different between when the still image photographing mode is set, and when the live view image is generated and displayed on the LCD 4 (photographing preparation period) or when the moving image photographing mode is set.
The timing generator 11 generates driving control signals of the image sensor 10, for example, clock signals such as timing signals to start/end the integration (start/end the exposure) and readout control signals (a horizontal synchronizing signal, a vertical synchronizing signal, etc.) of the light reception signals of the pixels based on a reference clock CLK0 transmitted from the controller 18, and outputs them to the image sensor 10.
The signal processor 12 performs predetermined analog signal processings on the analog pixel signals outputted from the image sensor 10. The signal processor 12 having a CDS (correlated double sampling) circuit and an AGC (automatic gain control) circuit reduces the noise of the pixel signals by the CDS circuit and adjusts the levels of the pixel signals by the AGC circuit.
The A/D converter 13 converts the analog pixel signals outputted from the signal processor 12 into digital pixel signals of a plurality of bits.
The image memory 14 temporarily stores the pixel signals outputted from the A/D converter 13, and is used as the work space for performing subsequently-described processings on the image signals by the controller 18.
The VRAM 15 is a buffer memory for the pixel signals of the image played back on the LCD 4, and has a pixel signal storage capacity corresponding to the number of pixels of the LCD 4. The operation portion 16 includes switches such as a switch that detects the release operation of the shutter button 9, the mode setting switch 7 and the quadruple switch 8.
The controller 18 comprises a microcomputer incorporating a non-illustrated storage portion comprising, for example, a ROM that stores a control program and a RAM that temporarily stores data, and controls the drivings of the above-described members so as to be associated with one another.
The brightness data in the positions of the pixels is obtained also by using the pixel signals outputted from the color pixels. However, it is considered that a bright and beautiful image can be generated while the deterioration of the S/N ratio is avoided when the brightness data in the positions of the monochrome pixels is derived by use of the pixel signals outputted from the monochrome pixels and, since the sensitivity of the monochrome pixels is higher than that of the color pixels as mentioned above, the brightness data in the positions of the color pixels is derived by an interpolation processing using the brightness data of the monochrome pixels situated around the color pixels compared to when the brightness data in the positions of the color pixels and the monochrome pixels is derived by use of the brightness data obtained by use of the pixel signals of the color pixels.
Therefore, according to the present embodiment, as described above, the brightness data in the positions of the pixels is derived by use of the brightness data obtained from the monochrome pixels.
Moreover, since no color filter is disposed at the monochrome pixels, the color data of R (red), G (green) and B (blue) in the positions of the monochrome pixels is not obtained from the monochrome pixels. Therefore, according to the present embodiment, the color data in the positions of the monochrome pixels is derived by an interpolation processing using the color data obtained from the color pixels situated around the monochrome pixels.
Since the above-mentioned interpolation processings are different between when the live view image or moving images in the moving image photographing mode are generated and when still images in the still image photographing mode are generated, these processings will be described with respect to each of these cases.
To realize this function, as shown in
The live view image/moving image generator 19 causes the image sensor 10 to perform the exposure operation at predetermined intervals during the photographing preparation mode and when a moving image setting mode is set, thereby generating the live view image displayed on the LCD 4 or a series of images (moving image) to be stored in the image storage portion 17. The live view image/moving image generator 19 has a first thinning out processor 20, a first brightness data interpolator 21, a first color data interpolator 22 and a second thinning out processor 23.
Since it is necessary for the live view image and the moving image only to be enough for the user to confirm the angle of view of the taken image and the like on the LCD 4 and these images are not required to have very high resolution, the first thinning out processor 20 selects some horizontal pixel rows including both color pixels and monochrome pixels from among a plurality of pixels of the image sensor 10, further selects some horizontal pixel rows from the horizontal pixel rows, and extracts the brightness data or the color data of the pixels belonging to the selected horizontal pixel rows. For example, as shown in FIGS. 4(a) to 4(c), the first, second, seventh and eighth horizontal pixel rows in the vertical direction are selected as the pixel rows which are the objects of the brightness data or color data extraction.
The first brightness data interpolator 21 derives the brightness data in the positions of, of the pixels selected by the first thinning out processor 20, the color pixels where the color filters of R (red), G (green) and B (blue) are disposed, by an interpolation processing using the brightness data obtained from the monochrome pixels situated around the color pixels.
For example, in FIGS. 4(a), 4(b), 4(c) and 5, as shown by the arrow A, the pixels selected by the first thinning out processor 20 are sectioned into large blocks each comprising, for example, two pixel rows in the longitudinal direction and four pixel rows in the lateral direction including adjoining color pixels of R (red), G (green) and B (blue). At this time, the pixels of each large block is sectioned so that a small block comprising two pixel rows in the longitudinal direction and two pixel rows in the lateral direction including the color pixels of R (red), G (green) and B (blue) is situated in the center of the large block.
Then, the first brightness data interpolator 21 sets, as the brightness data in the positions of the color pixels belonging to the large block, the brightness data obtained from the monochrome pixels adjoining the color pixels in the large block.
That is, as shown in
The arrows in
The first color data interpolator 22 derives the color data in the positions of the pixels selected by the first thinning out processor 20 by the interpolation processing using the color data obtained from the color pixels situated around the pixels.
For example, in FIGS. 4(a), 4(b), 4(c) and 6(a), as shown by the arrow B, the first color data interpolator 22 sections the pixels selected by the first thinning out processor 20 into large blocks each comprising two pixel rows in the longitudinal direction and six pixel rows in the lateral direction including adjoining color pixels of R (red), G (green) and B (blue). At this time, the pixels of each large block is sectioned so that a small block comprising two pixel rows in the longitudinal direction and two pixel rows in the lateral direction including the color pixels of R (red), G (green) and B (blue) is situated in the center of the large block.
Then, the first color data interpolator 22 interpolates, in each large block, the color data of the colors in the positions of the monochrome pixels by use of the color data obtained from the color pixels of the colors included in the block.
That is, as shown in
Further, as shown in
Moreover, as shown in
The arrows in FIGS. 6(b) to 6(d) indicate that the color data in the positions of the color pixels in the block is substituted for the color data in the positions of the monochrome pixels.
The second thinning out processor 23 thins out the pixels in the horizontal direction at the same thinning out rate as that of the first thinning out processor 20. For example, describing this with reference to the example shown in FIGS. 4(a) to 4(c), since the first thinning out processor 20 regularly thins out the horizontal pixel rows to 2/6 in the vertical direction, the second thinning out processor 23 regularly thins out the vertical pixel rows to 2/6 in the horizontal direction.
The still image generator 24 causes the image sensor 10 to perform the exposure operation with a preset exposure time (shutter speed) when the still image photographing mode is set, and generates an image (still image) by use of the pixel signals obtained from substantially all the pixels in order to generate a high-resolution image. The still image generator 24 has a second brightness data interpolator 25 and a second color data interpolator 26.
The second brightness data interpolator 25 derives the brightness data in the positions of the color pixels where the color filters of R (red), G (green) and B (blue) are disposed, by the interpolation processing using the brightness data of the monochrome pixels situated around the pixels. A method of calculating the brightness data will be described with the brightness data in the position of the color pixel of R (red) as an example.
For example, when in FIGS. 4(a) and 4(b), attention is paid to pixels in four rows in the longitudinal direction and four rows in the lateral direction including adjoining color pixels of R (red), G (green) and B (blue) as shown by the arrow C and these pixels are numbered P1 to P16 as shown in
That is, in the present embodiment, in this block, a pair of monochrome pixel combinations (P2, P10) and (P3, P9) sandwiching the color pixel P6 to be interpolated are derived in any of the vertical, horizontal and slanting directions.
Then, the difference in brightness between the two monochrome pixels of each combination is calculated, and whether the brightness difference is higher or lower than a threshold value α is determined. When the brightness difference of one combination is larger than the threshold value α and the brightness difference of the other combination is smaller than the threshold value α (patterns 1 and 2), since it is considered that the brightness data in the position of the color pixel P6 to be interpolated is approximate to the brightness data of the two monochrome pixels of the combination whose brightness difference is smaller, the average value of the brightness values of the two monochrome pixels of the combination whose brightness difference is smaller than the threshold value α is calculated, and the average value is set as the brightness value (brightness data) in the position of the color pixel.
For example, when the brightness values of the pixels P1 to P16 are w1 to w16, respectively, as shown in
When the brightness differences of both of the combinations are smaller than the threshold value α (pattern 3), the average value of the brightness values of the two monochrome pixels of the combination whose brightness difference is smaller is calculated, and the average value is set as the brightness value (brightness data) of the color pixels.
When the brightness differences of both of the combinations are larger than the threshold value α (pattern 4), the average value of the brightness values of all the monochromes pixels P1 to P3, P5, P9 and P10 in the block comprising pixels in three rows in the longitudinal direction and three rows in the lateral direction, and the average value is set as the brightness value (brightness data) in the position of the color pixel P6 to be interpolated.
For example, describing this with reference to the above-described example, in the case of |w2−w10|<α and |w3−w9|<α, when |w2−w10|<|w3−w9|, the brightness value w6 of the color pixel P6 is (w2+w10)/2, and when |w2−w10|>|w3−w9|, the brightness value w6 is (w3+w9)/2. Moreover, in the case of |w2−w10|≧α and |w3−w9|≧α, the brightness value w6 of the color pixel P6 is (w1+w2+w3+w5+w9+w10)/6.
Likewise, to derive the brightness data in the position of the color pixel of G (green), as shown in
The second color data interpolator 26 derives the color data in the positions of the monochrome pixels from the interpolation processing using the color data of the color pixels situated around the pixels. A method of calculating the color data of R (red) in the positions of the monochrome pixels will be described as an example.
As shown in
In the present embodiment, the color data of R (red) in the position of the monochrome pixel P13 is the average value of the color data of the color pixels P1, P10, P16 and P25. That is, when the values represented by the color data of R (red) of the pixels P1 to P25 are denoted by r1 to r25, respectively, the value r13 represented by the color data of R (red) in the position of the monochrome pixel P13 is (r1+r10+r16+r25)/4.
Then, this rhombus is divided into four triangular areas at the diagonal lines, and the color data of R (red) in the positions of the monochrome pixels other than the monochrome pixel P13 is derived by an interpolation processing using the color data of the pixels situated at the vertices of the triangles to which the monochrome pixels belong (any of the color pixels P1, P10, P16 and P25 and the monochrome pixel P13). Hereinafter, the pixels situated at the vertices of the triangles will be referred to as vertex pixels.
In that case, when the monochrome pixel to be interpolated is situated on a side of the triangle, the vertex pixel situated on the side is derived, and a weighting factor corresponding to the distance between each of the derived vertex pixels and the monochrome pixel to be interpolated is calculated. Then, the weighted average of the color data of the derived vertex pixels is obtained by use of the weighting factors, and the average value is set as the color data of the monochrome pixel to be interpolated.
For example, as shown in
Moreover, the reciprocal of the distance between the monochrome pixel P2 and the color pixel P1 and the reciprocal of the distance between the monochrome pixel P2 and the color pixel P10 are calculated, and the ratios “2/3” and “1/3” of the reciprocals to the sum total of the reciprocals are used as the weighting factors. This is done because it is considered that the color data in the position of the monochrome pixel to be interpolated is approximate to the color data of the vertex pixel close to the monochrome pixel, and in the present embodiment, this is done on the assumption that the color data is approximate in proportion to the reciprocal of the distance.
Then, the values obtained by multiplying the pieces of color data r1 and r10 of the vertex pixels corresponding to the distances by the corresponding weighting factors, that is, “(2/3)·r1” and “(1/3)·r10” are added together, and the sum is set as the color data r1 in the position of the monochrome pixel P2.
The color data of R (red) in the positions of the other monochrome pixels P3 to P5, P7, P9, P11 to P15 (excluding P13), P17, P19 and P21 to P24 situated on the sides of the rhombus can be calculated in a like manner.
Moreover, with respect to the monochrome pixels P6, P8, P18 and P20 not situated on the sides of the rhombus, the three vertex pixels constituting each of the triangles to which these monochrome pixels belong are derived, and the color data is derived by an interpolation processing using the color data of the vertex pixels.
For example, as shown in
Here, in the present embodiment, the color data in the position of the monochrome pixel P6 is regarded as situated in the center of the triangle, and is the average value (r1+r10+r13)/3 of the color data of the vertex pixels P1, P10 and P13. The color data in the position of the monochrome pixel P6 to be interpolated may be derived in accordance with the actual distance between each of the vertex pixels, the color pixels P1 and P10 and the monochrome pixel P13, and the monochrome pixel P6. The color data of R (red) in the positions of the other monochrome pixels P8, P18 and P20 not situated on the sides of the rhombus can be derived in a like manner.
Further, the color data of G (green) and B (blue) in the positions of the monochrome pixels can be calculated by a like derivation method. The live view image/moving image generator 19 and the still image generator 24 correspond to the image generator as claimed in claims.
An image processor 27 performs a black level correction to correct the black level to the reference black level, a white balance adjustment to convert the levels of the digital signals of the color components of R (red), G (green) and B (blue) based on the reference of white corresponding to the light source and a gamma correction to correct the gamma characteristics of the digitals signals of R (red), G (green) and B (blue), on the images generated by the live view image/moving image generator 19 and the still image generator 24.
A display controller 28 transfers the pixel data of the image outputted from the live view image/moving image generator 19, to the VRAM 15 in order to display the image on the LCD 4. By this, the condition of the subject can be displayed on the LCD 4 in real time as the live view image until the exposure operation for recording is started.
An image compressor 29 generates compressed image data by performing a predetermined compression processing by the JPEG (Joint Picture Experts Group) method such as the two-dimensional DCT (discrete cosine transform) or the Huffman coding on the pixel data of the recorded image having undergone the above-mentioned processings by the image processor 27, and an image file comprising the compressed image data to which information related to the taken image (information such as the compression rate) is added is recorded in the image storage portion.
In the image storage portion 17, the pieces of image data are recorded in a condition of being arranged in time sequence, and in each frame, a compressed image compressed by the JPEG method is recorded together with the index information related to the taken image (information such as the frame number, the exposure value, the shutter speed, the compression rate, the date of photographing, data as to whether the flash is on or off at the time of photographing, and scene information).
Next, a series of image capturing processings by the image capturing apparatus 1 of the present embodiment will be described with reference to the flowchart of
As shown in
Then, it is determined whether a half depression of the shutter button 9 is detected by a non-illustrated switch Si or not (step #4). When no half depression is performed (No of step #4), the process returns to the processing of step #2, and the processings of steps #2 and #3 are performed. When a half depression is performed (YES of step #4), the focusing operation is performed (step #5).
Then, it is determined whether a full depression of the shutter button 9 is detected by a non-illustrated switch S2 or not (step #6). When no full depression is performed (NO of step #6), the process returns to the processing of step #2, and the processings of steps #2 to #5 are performed. When a full depression is performed (YES of step #6), after settings for the exposure operation for recording such as a change of the exposure control values set at step #2 are performed (step #7), the pixel signals for recording are generated and stored (step #8).
As shown in
Next, the controller 18 interpolates the color data of R (red), G (green) and B (blue) in the positions of the monochrome pixels by use of the color data of the color pixels of R (red), G (green) and B (blue) situated therearound (step #34). As the interpolation processing method, the interpolation processing method shown in
As shown in
Then, the controller 18 interpolates the color data of R (red), G (green) and B (blue) in the positions of the monochrome pixels by use of the color data of the color pixels of R (red), G (green) and B (blue) situated therearound (step #84). As the interpolation processing method, in the case of the still image photographing mode, the interpolation processing method shown in
Then, the controller 18 performs the above-described compression processing and the like on the image for recording (step #86), and then, stores the compressed image into the image storage portion 17 (step #87). Then, when the set photographing mode is the still image photographing mode (YES of step #88), the process returns to the processing of step #2 of the flowchart shown in
On the other hand, in the case of the moving image photographing mode (NO of step #88), the controller 18 determines whether a full depression of the shutter button 9 is again detected by the non-illustrated switch S2 or not (step #89). When no full-depression is performed again (NO of step #89), the process returns to the processing of step #81, and the processings of steps #81 to #88 are repeated. When a full depression is performed again (YES of step #89), the process returns to the processing of step #2 of the flowchart shown in
As described above, since the image sensor 10 has color pixels where color filters of R (red), G (green) and B (blue) having different spectral characteristics are disposed and monochrome pixels where no color filter is disposed and a plurality of color pixels of R (red), G (green) and B (blue) are dispersedly disposed among a plurality of monochrome pixels so that the number of monochrome pixels is larger than the number of color pixels, the effective sensitivity of the image sensor 10 can be improved.
Although the number of color pixels is small compared to the number of monochrome pixels, since the sensitivity, to colors (hues and chromas), of the human eye is low, even when an image is generated from the color data in the positions of the monochrome pixels by interpolating the color data of the color pixels situated therearound, a taken image recognized as having high image quality can be generated.
Further, when the live view image and a moving image are generated, since the pixels belonging to the pixel row in the horizontal direction where both monochrome pixels and color pixels are present are selected as the pixels for generating the live view image and the moving image, a color live view image and moving image can be generated.
The present invention is not limited to the above-described embodiment, but the following modifications [1] to [8] are adoptable:
[1] The interpolation of the brightness data in the positions of the color pixels is not limited to that of the above-described embodiment, but the following mode is adoptable:
As shown in FIGS. 4(a) to 4(c) and 12(a) to 12(c), when pixels in four rows in the longitudinal direction and four rows in the lateral direction including adjoining color pixels of R (red), G (green) and B (blue) are numbered P1 to P16, the brightness data in the position of the color pixel P6 of R (red) may be interpolated by use of the brightness data of all the monochrome pixels belonging to a block comprising the pixels in three rows in the longitudinal direction and three rows in the lateral direction with the color pixel P6 in the center.
For example, as shown in
Likewise, when the brightness data in the position of the color pixel of G (green) is derived, as shown in
When the brightness data in the position of the color pixel of B (blue) is derived, as shown in
As described above, as the brightness data in the position of the color pixel to be interpolated, the average of the brightness data of all the monochrome pixels adjoining the color pixel may be set.
[2] Since the brightness data can be obtained not only from the monochrome pixels but also from the color pixels, the gradation of the brightness can be increased by using the brightness data obtained from both the monochrome pixels and the color pixels.
As shown in
On the other hand, the color pixels have a characteristic such that when the light reception amount P is in a range of 0<P<P3 (P3>P2), the output value increases at a constant rate lower than the increase rate of the output value of the monochrome pixels in the range of 0<P<P2, when the light reception amount P is in a range of P3<P<P4, the increase rate is lower than that in the range of 0<P<P3 and when the light reception amount P becomes P4 (>P3), the output is saturated.
As described above, there is a range of the light reception amount where the output values of the color pixels are not saturated although the output values of the monochrome pixels are saturated and in
Therefore, with respect to the pixels where, for example, with the output value S1 of the color pixels or the output value S2 of the monochrome pixels corresponding to the light reception amount P1 as the boundary, the brightness data of the monochrome pixels or the brightness data obtained by the interpolation from the monochrome pixels in the positions of the pixels is in a range of 0<S<S2 or the brightness data obtained from only the color pixels or the brightness data obtained by the interpolation processing using the brightness data in the positions of the pixels is in a range of 0<S<S1, an image is generated by use of only the brightness data obtained from the monochrome pixels.
Moreover, with respect to pixels with comparatively high brightness where the brightness data of the monochrome pixels or the brightness data obtained by the interpolation from the monochrome pixels is in a range of S>S2 or the brightness data obtained from only the color pixels or the brightness data obtained by the interpolation processing using the brightness data is in a range of S>S1, an image is generated by combining together, adding together in the present embodiment, the brightness data obtained from the monochrome pixels and the brightness data obtained from the color pixels.
By this, compared to when an image is generated from only the brightness data obtained from the monochrome pixels, the dynamic range is increased from the range of 0<P<P2 to the range of 0<P<P3, gradation can be expressed also for a subject image (equivalent) that is high in brightness by the range of P1<P<P3 of the light reception amount P corresponding to the range shown by the arrow A of
As a method of generating the brightness data by use of the pixel signals obtained from the color pixels, the following method, for example, is adopted:
For example, in the arrangement of the color pixels and the monochrome pixels shown in FIGS. 4(a) to 4(c), attention is paid to the four color pixels of G (green) arranged in a rhombus and the color pixels of R (red) and B (blue) adjoining these color pixels as shown in
After the brightness data in the position of each color pixel of G (green) is calculated in this manner, as shown in
The brightness values at the boundary to determine whether the brightness data obtained from the monochrome pixels and the brightness data obtained from the color pixels are combined together or not are not limited to the brightness values S1 and S2, but may be set as appropriate within a range where the monochrome pixels are not saturated.
[3] The color pixel arrangement is not limited to that of the above-described embodiment (see FIGS. 4(a) to 4(c)), but color pixel arrangements as shown in FIGS. 15 to 29 described in the following are adoptable: When pixels including a predetermined number of color pixels for each kind of color filter constitute a group, the color pixels or the pixels of the groups are dispersedly disposed with monochrome pixels in between.
The color pixel arrangements shown in
The color pixel arrangement shown in
In this case, a processing system that processes pixel signals of the conventional image sensors where only color pixels are Bayer-arranged (image sensors having no monochrome pixel) can be adopted.
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In this case, since color pixels of R (red), G (green) and B (blue) are disposed in one pixel row extending in the horizontal direction, when the live view image is generated, only the pixel rows where color pixels are disposed are selected and the pixel data is extracted from the pixels belonging to the selected pixel rows to thereby generate the live view image.
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In this case, since the color pixels of R (red), G (green) and B (blue) in one pixel group are disposed so as to adjoin (gather) together, compared to when the color pixels of R (red), G (green) and B (blue) in one pixel group are disposed so as to be separate from one another, the brightness data and the color data can be more accurately interpolated, so that false colors are less frequently generated.
That is, for example, in a case where color pixels where color filters of the same color are disposed (hereinafter, referred to as first and second color pixels) are arranged so as to be separate from one another, when the pixel signal (color signal) in the position of a color pixel where a color filter of a different color is disposed which color pixel is situated between the first and second color pixels is interpolated by use of the pixel signals of the first and second color pixels, there are cases where the color signal derived by the interpolation is significantly different between when the color boundary is present on the first color pixel side of the pixel to be interpolated and when it is present on the second color pixel side thereof, and consequently, there are cases where the color in the position of the monochrome pixel to be interpolated cannot be accurately reproduced.
On the contrary, according to the present invention, since color pixels where color filter of different colors are disposed are arranged so as to adjoin together, the color signals of colors different from those of the color pixels in the positions of the color pixels can be interpolated by the color pixels adjoining the color pixels, so that the above-mentioned problem can be avoided or suppressed.
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
In the color pixel arrangement shown in
[4] The color data interpolation processing is not limited to the above-described one; for example, the following may be adopted:
As shown in FIGS. 30(a) and 30(b), attention is paid to four color pixels of R (red) arranged in a rhombus like in
In that case, according to the present embodiment, for the color data of R (red) in the position of the monochrome pixel P13, a pair of combinations of color pixels of R (red) (P1, P25) and (P10, P16) situated at the vertices of the rhombus and opposed with the monochrome pixel P13 in between are derived.
Then, in the combinations (P1, P25) and (P10, P16), the difference in color data between the two color pixels is calculated, and whether the color data difference is larger or smaller than a threshold value β is determined. When one color data difference is larger than the threshold value β and the other color data difference is smaller than the threshold value β, the average value of the color data of the two color pixels of the combination whose color data difference is smaller is calculated, and the average value is set as the color data in the position of the monochrome pixel P13.
This is because it can be considered that as shown in FIGS. 30 (a) and 30(b), the possibility is high that the color boundary passes through any position between the two color pixels belonging to the combination whose color difference is larger than the threshold value β and the possibility is low that the color boundary passes between the two color pixels belonging to the combination whose color difference is smaller than the threshold value β.
Moreover, when the color data differences are both smaller than the threshold value β, the average value of the color data of the two color pixels belonging to the combination whose color data difference is smaller is calculated, and the average value is set as the color data in the position of the monochrome pixel P13. Moreover, when the color data differences are both larger than the threshold value β, the average value of the color data of all the color pixels P1, P10, P16 and P25 is calculated, and the average value is set as the color data in the position of the monochrome pixel P13.
The interpolation method for the color data in the positions of the other monochrome pixels P2 to P9, P11, P12, P14, P15 and P17 to P24 will not be described because it is similar to the interpolation method described with reference to
The color boundary can be estimated based on brightness (density) changes. For example, in the above-described example, when the difference in brightness data between the positions of the two color pixels in each of the combinations (P1, P25) and (P10, P16) is calculated, it can be considered that the possibility is high that the color boundary passes through any position between the two color pixels belonging to the combination whose brightness data difference is larger and the possibility is low that the color boundary passes between the two color pixels belonging to the combination whose brightness data difference is smaller.
[5] When the exposure control is performed by use of only the brightness data obtained from the monochrome pixels, the exposure control can be accurately performed (the shutter speed, the aperture value and the like can be accurately set) compared to when the exposure control is performed based on the brightness data obtained from the color pixels.
That is, in the conventional image sensors comprising only color pixels of R (red), G (green) and B (blue), the brightness data is generated from the pixel data obtained from the color pixels, and the exposure control is performed based on the brightness data. However, there is a possibility that an error with respect to the actual brightness of the subject is caused because of the brightness data generation processing.
Moreover, when most of the pixels for obtaining images for recording are monochrome pixels like the present invention, since there is a large brightness difference between the pieces of brightness data obtained from the monochrome pixels and the color pixels that are significantly different in sensitivity, using these in mixture for the exposure control also becomes a cause of the error.
On the contrary, in the present embodiment, since the brightness data obtained from only the monochrome pixels is used, the brightness data generation processing as described above is unnecessary, so that the above-mentioned error that can be caused because of the generation processing is never caused. By this, the exposure control can be accurately performed. Moreover, since the effective sensitivity of the image sensor 10 is high because of the provision of the monochrome pixels, the exposure control can be accurately performed even for a dark image. The exposure control is performed by an exposure condition determiner (corresponding to the exposure condition determiner as claimed in claims) in the controller 18.
[6] When an image sensor of a type that specifies a given pixel from among a plurality of pixels and causes the selected pixel to output a pixel signal is used instead of the above-described image sensor, the pixel signal readout from the monochrome pixels and the pixel signal readout from the color pixels can be separately performed such as reading out the pixel signals from the monochrome pixels first and then, reading out the pixel signals from the color pixels, so that compared to when the pixel signals from the monochrome pixels and the pixel signals from the color pixels are present in mixture, the pixel signal processing is easy. Consequently, the processing time can be reduced and the structure of the signal processing system can be simplified.
[7] While a mode provided with one each of the signal processor 12, the A/D converter 13 and the image memory 14 is described as the first embodiment of the present invention, the present invention is not limited thereto; two each of the signal processor 12, the A/D converter 13 and the image memory 14 may be provided.
With this, for example, the processing of the pixel signals obtained from the color pixels and the processing of the pixel signals obtained from the monochrome pixels can be performed in parallel, and in another signal processor 12, by amplifying the pixel signals obtained from the monochrome pixels and the pixel signals obtained from the color pixels at different amplification factors, the SIN ratio can be improved.
Further, since the pixel signal processing can be performed with the pixel signals from the monochrome pixels and the pixel signals from the color pixels being separated from each other, the pixel signal processing is easy compared to when the pixel signals are present in mixture, so that the processing time can be reduced and the structure of the signal processing system can be simplified. Consequently, for example, the white balance adjustment performed only by the pixel signals obtained from the color pixels can be easily performed.
[8] While the colors of the color filters disposed at the color pixels are R (red), G (green) and B (blue) in the above-described embodiment, the present invention is not limited thereto; the colors may be C (cyan), M (magenta), Y (yellow) and G (green). In this case, for example, the color pixel arrangement is such that, like those of FIGS. 15 to 17, 21 and 25, for the pixel group including one color pixel of R (red), two color pixels of G (green) and one color pixel of B (blue), instead of these color pixels, color pixels where color filters of C (cyan), M (magenta), Y (yellow) and G (green) are disposed.
As described above, according to an image sensor of the present invention, in an image sensor comprising a plurality of pixels arranged in a matrix and having pixels where at least three kinds of color filters are disposed, color pixels where the color filters are disposed and monochrome pixels where no color filter is disposed are provided, the sum total of the monochrome pixels is larger than the sum total of the color pixels, and when pixels including a predetermined number of color pixels for each kind of color filter constitute one group, the color pixels or the pixels of the groups are dispersedly disposed with the monochrome pixels in between.
According this aspect of the invention, the effective sensitivity of the image sensor can be improved compared to the image sensors in which only color pixels where color filters are disposed are arranged and the conventional image sensors in which the sum total of the monochrome pixels is equal to or smaller than the sum total of the color pixels.
Consequently, since the effective sensitivity of the image sensor is improved, photographing with high sensitivity can be performed, so that a beautiful image with an excellent (high) S/N ratio can be obtained.
Moreover, in the above-described image sensor, color pixels where different color filters are disposed are arranged so as to adjoin in each pixel group.
According to this aspect of the invention, since color pixels where different color filters are disposed are arranged so as to adjoin in each pixel group, the generation of false colors (color moire) can be prevented or suppressed, so that a beautiful image can be obtained.
As the mode where color pixels are arranged so as to adjoin, for example, a mode where color pixels are arranged so as to adjoin each other and a mode where color pixels are arranged in a line (continuously) are considered.
Moreover, an image capturing apparatus of the present invention is provided with: a taking optical system that forms a light image of the subject; the above-described image sensor whose image capturing surface is disposed on an image forming surface of the taking optical system; an input operation portion for inputting instructions to start and end an exposure operation to the image sensor; an image generator that generates an image from a pixel signal obtained by the exposure operation by the image sensor; and an image display that displays the image generated by the image generator.
According to this aspect of the invention, since the effective sensitivity of the image sensor is improved, an image capturing apparatus with high photographing sensitivity can be obtained, so that a bright and beautiful image can be obtained.
Moreover, in the above-described image capturing apparatus, the image generator generates first brightness data in the positions of the monochrome pixels based on pixel signals obtained from the monochrome pixels, generates second brightness data in the positions of the color pixels by an interpolation processing using the first brightness data, generates first color data in the positions of the color pixels based on pixel signals obtained from the color pixels, and generates second color data in the positions of the monochrome pixels by an interpolation processing using the first color data.
According to this aspect of the invention, since the first brightness data in the positions of the monochrome pixels is generated based on the pixel signals obtained from the monochrome pixels with comparatively high sensitivity and the second brightness data in the positions of the color pixels are generated by the interpolation processing using the first brightness data, the apparent effective sensitivity of the image sensor can be improved, so that a bright and beautiful image can be obtained. Moreover, since the first color data in the positions of the color pixels is generated based on the pixel signals obtained from the color pixels and the second color data in the positions of the monochrome pixels is generated by the interpolation processing using the first color data, a color image can be generated even in the case of a pixel structure where the majority of the pixels are monochrome pixels having no color data. Since the sensitivity (resolution), to colors (hues and chromas), of the human eye is low, a taken image recognized as having high image quality can be generated even in the case of a color image generated in the above-described manner.
Moreover, in the above-described image capturing apparatus, the image generator further generates third brightness data in the positions of the color pixels based on the pixel signals obtained from the color pixels, generates fourth brightness data in the positions of the monochrome pixels by an interpolation processing using the third brightness data, and generates an image of the monochrome pixels whose brightness exceeds a predetermined threshold value by combining the first brightness data and the fourth brightness data.
According to this aspect of the invention, the third brightness data in the positions of the color pixels is generated based on the pixel signals obtained from the color pixels whose sensitivity range is shifted with respect to that of the monochrome pixels, the fourth brightness data in the positions of the monochrome pixels is generated by the interpolation processing using the third brightness data and an image of the monochrome pixels whose brightness exceeds a predetermined threshold value is generated by combining the first brightness data and the fourth brightness data, so that the gradation of the brightness can be easily increased. Consequently, an image with rich gradation can be obtained.
Moreover, in the above-described image capturing apparatus, a mode to cause the image sensor to perform the exposure operation a plurality of times at predetermined intervals is provided, and in the mode, the image generator selects a pixel row where both the color pixels and the monochrome pixels are present, from among a plurality of pixel rows where a plurality of pixels are arranged in one direction, and generates an image by use of the brightness data and the color data in the position of each pixel belonging to the selected pixel row.
According to this aspect of the invention, since in the mode to cause the image sensor to perform the exposure operation a plurality of times at predetermined intervals, a pixel row where both the color pixels and the monochrome pixels are present is selected from among a plurality of pixel rows where a plurality of pixels are arranged in one direction and an image is generated by use of the brightness data and the color data in the position of each pixel belonging to the selected pixel row, the color data can be obtained from only the selected pixel row. Consequently, a color image (moving image) can be generated. The moving image referred to here is a series of images obtained by causing the image sensor to perform the exposure operation a plurality of times at predetermined intervals to thereby generate an image from the pixel signals obtained by each exposure operation and displaying the images so as to be switched at predetermined intervals in an updated manner.
Moreover, in the above-described image capturing apparatus, an exposure condition determiner that determines the exposure condition of the image sensor is provided, and the exposure condition determiner determines the exposure condition by use of only the brightness data in the positions of the monochrome pixels.
According to this aspect of the invention, since the exposure condition is determined by use of only the brightness data of the monochrome pixels, compared to when the exposure condition is determined by use of the brightness data of the color pixels, the exposure control can be accurately performed even for a subject that is dark because of the sensitivity of the monochrome pixels being high.
Moreover, while in the conventional image sensors where color filters of, for example, R (red), G (green) and B (blue) are Bayer-arranged, the brightness data under a condition where no color filter is present in each pixel is generated from the pixel signals obtained from these color pixels and the exposure control is performed based on the brightness data, there is a possibility that an error with respect to the actual brightness of the subject is caused in the brightness data generation processing.
On the contrary, according to the present invention, since the brightness data under a condition where no color filter is present in each pixel is obtained from the monochrome pixels, the brightness data generation processing as described above is unnecessary, so that the above-mentioned error that can be caused because of the generation processing is never caused. The exposure control can be accurately performed also form this respect.
Although the resent invention has been fully described in connection with the preferred embodiments thereof with reference to the accompanying drawings, it is to be noted that various changes and modifications are apparent to those skilled in the art. Such changes and modifications are to be understood as included within the scope of the present invention as defined by the appended claims unless they depart therefrom.
Number | Date | Country | Kind |
---|---|---|---|
2004-353957 | Dec 2004 | JP | national |