This nonprovisional application claims priority under 35 U.S.C. §119(a) on Patent Application No. 2011-049128 filed in Japan on Mar. 7, 2011, the entire contents of which are hereby incorporated by reference.
1. Field of the Invention
The present invention relates to image-shooting devices such as digital cameras.
2. Description of Related Art
There have been proposed methods of producing an output image by use of only photoreceptive pixel signals within a region that is part of the entire photoreceptive pixel region of an image sensor. These methods are by and large like the one shown in
The method shown in
In
The maximum spatial frequency that can be expressed in the 2-megapixel output image 902 is lower than that in the 8-megapixel input image 901. On the other hand, the maximum spatial frequency that can be expressed in the 2-megapixel output image 912 is similar to that in the 2-megapixel input image 911. In one conventional method, however, irrespective of the ratio (DOUT/DIN), that is, irrespective of the RAW zoom magnification, the same signal processing (for example, demosaicing processing) is performed.
For the purpose of noise elimination, there have been proposed technologies of applying filtering to the input image (RAW data).
In a case where the signal processing performed on the input image or the output image is of a kind suitable for a state where √(DOUT/DIN)=1, when √(DOUT/DIN) is actually equal to 0.5, the high-frequency spatial frequency components that can be expressed in the 8-megapixel input image 901 but that cannot be expressed in the 2-megapixel output image 902 may mix with the 2-megapixel output image 902, causing aliasing in the 2-megapixel output image 902. Aliasing appears as false color or noise.
Aliasing can be suppressed by incorporating smoothing (low-pass filtering) in the signal processing. Incorporating uniform smoothing in the signal processing, however, results in unnecessarily smoothing signals when √(DOUT/DIN)=1, producing an output image with lack in resolution (resolving power).
Needless to say, it is beneficial to suppress aliasing on one hand and suppress lack in resolution on the other hand with a good balance.
Expectations are high for achieving both suppression of aliasing and suppression of lack in resolution with a good balance.
According to the present invention, an image-shooting device is provided with: an image sensor having a plurality of photoreceptive pixels; and a signal processing section which generates the image data of an output image from the photoreceptive pixel signals within an extraction region on the image sensor. Here, the signal processing section controls the spatial frequency characteristic of the output image according to an input pixel number, which is the number of photoreceptive pixels within the extraction region, and an output pixel number, which is the number of pixels of the output image.
Hereinafter, examples of how the present invention is embodied will be described specifically with reference to the accompanying drawings. Among the different drawings referred to in the course, the same parts are identified by the same reference signs, and in principle no overlapping description of the same parts will be repeated. Throughout the present specification, for the sake of simple notation, particular data, physical quantities, states, members, etc. are often referred to by their respective reference symbols or signs alone, with their full designations omitted, or in combination with abbreviated designations. For example, while the RAW zoom magnification is identified by the reference symbol ZFRAW, the RAW zoom magnification ZFRAW may also be referred to as the magnification ZFRAW or, simply, ZFRAW.
An image-sensing section 11 shoots a subject by use of an image sensor.
The image sensor 33 is composed of a plurality of photoreceptive pixels arrayed both in the horizontal and vertical directions. The photoreceptive pixels of the image sensor 33 photoelectrically convert the optical image of a subject incoming through the optical system 35 and the aperture stop 32, and outputs the resulting electrical signal to an AFE (analog front end) 12.
The AFE 12 amplifies the analog signal output from the image sensor 33 (photoreceptive pixels), converts the amplified analog signal into a digital signal, and outputs the digital signal to a video signal processing section 13. The amplification factor of signal amplification in the AFE 12 is controlled by a CPU (central processing unit) 23. The video signal processing section 13 applies necessary image processing to the image represented by the output signal of the AFE 12, and generates a video signal representing the image having undergone the image processing. A microphone 14 coverts the ambient sound around the image-shooting device 1 into an analog audio signal, and an audio signal processing section 15 convers the analog audio signal into a digital audio signal.
A compression processing section 16 compresses the video signal from the video signal processing section 13 and the audio signal from the audio signal processing section 15 by use of a predetermined compression method. An internal memory 17 is a DRAM (dynamic random-access memory) or the like, and temporarily stores various kinds of data. An external memory 18 as a recording medium is a non-volatile memory such as semiconductor memory or a magnetic disk, and records the video and audio signals having undergone the compression by the compression processing section 16.
A decompression processing section 19 decompresses the compressed video and audio signal read out from the external memory 18. The video signal having undergone the decompression by the decompression processing section 19, or the video signal from the video signal processing section 13, is fed via a display processing section 20 to a display section 27, which is a liquid crystal display or the like, to be displayed as an image. The audio signal having undergone the decompression by the decompression processing section 19 is fed via an audio output circuit 21 to a loudspeaker 28 to be output as sounds.
A TG (timing generator) 22 generates timing control signals for controlling the timing of different operations in the entire image-shooting device 1, and feeds the generated control signals to the relevant blocks within the image-shooting device 1. The timing control signals include a vertical synchronizing signal Vsync and a horizontal synchronizing signal Hsync. A CPU 23 comprehensively controls the operation of different blocks within the image-shooting device 1. An operation section 26 includes, among others, a record button 26a for entering a command to start and end the shooting and recording of a moving image, a shutter-release button 26b for entering a command to shoot and record a still image, and a zoom button 26c for specifying the zoom magnification, and accepts various operations by the user. How the operation section 26 is operated is communicated to the CPU 23. The operation section 26 may include a touch screen.
The image-shooting device 1 operates in different modes including a shooting mode in which it can shoot and record images (still or moving images) and a playback mode in which it can play back and display on the display section 27 images (still or moving images) recorded on the external memory 18. According to operation on operation section 26, the different modes are switched. Unless otherwise stated, the following description deals with the operation of the image-shooting device 1 in shooting mode.
In shooting mode, a subject is shot periodically, at predetermined frame periods, so that shot images of the subject are acquired sequentially. A video signal representing an image is also referred to as image data. Image data corresponding to a given pixel may also be referred to as a pixel signal. The size of an image, or of an image region, is also referred to as an image size. The image size of an image of interest, or of an image region of interest, can be expressed in terms of the number of pixels constituting the image of interest, or belonging to the image region of interest.
In the present specification, the image data of a given image is occasionally referred to simply as an image. Accordingly, for example, generating, acquiring, recording, processing, modifying, editing, or storing a given image means doing so with the image data of that image. Compression and decompression of image data are not essential to the present invention; therefore compression and decompression of image data are disregarded in the following description. Accordingly, for example, recording compressed image data of a given image is referred to simply as recording image data, or recording an image.
Each photoreceptive pixel within the effective pixel region 33A is represented by PS[x, y]. Here, x and y are integers. In the image sensor 33, the up-down direction corresponds to the vertical direction, and the left-right direction corresponds to the horizontal direction. In the image sensor 33, the photoreceptive pixels adjacent to a photoreceptive pixel PS[x, y] at its right, left, top, and bottom are PS[x+1, y], PS[x−1, y], PS[x, y−1], PS[x, y+1] respectively. Each photoreceptive pixel photoelectrically converts the optical image of the subject incoming through the optical system 35 and the aperture stop 32, and outputs the resulting electrical signal as a photoreceptive pixel signal.
The image-shooting device 1 uses only one image sensor, thus adopting a so-called single-panel design. That is, the image sensor 33 is a single-panel image sensor.
Photoreceptive pixels having red, green, and blue filters arranged in front of them are also referred to as red, green, and blue photoreceptive pixels respectively. Red, green, and blue photoreceptive pixels react only to the red, green, and blue components, respectively, of the light incoming through the optical system. Each photoreceptive pixel photoelectrically converts the light incident on it through the color filter arranged in front of itself into an electrical signal, and outputs the thus obtained electrical signal as a photoreceptive pixel signal.
Photoreceptive pixel signals are amplified and also digitized by the AFE 12, and the amplified and digitized photoreceptive pixel signals are output as RAW data from the AFE 12. In the following description, however, for the sake of simple explanation, signal digitization and signal amplification in the AFE 12 are disregarded, and the photoreceptive pixel signals themselves that are output from photoreceptive pixels are also referred to as RAW data.
In the embodiment under discussion, a concept is introduced of RAW zooming that allows change of image size through change of the size of the extraction frame EF. The factor by which image size is changed by RAW zooming is referred to as the RAW zoom magnification.
A RAW zoom magnification is fed into the extraction frame setting section 50. As will be described in detail later, the RAW zoom magnification is set according to a user operation. A user operation denotes an operation performed on the operation section 26 by the user. According to the RAW zoom magnification, the extraction frame setting section 50 sets the size of the extraction frame EF. The number of photoreceptive pixels belonging to the extraction frame EF is expressed as (DIN×1,000,000) (where DIN is a positive real number). The extraction frame setting section 50 serves also as a reading control section, reading out RAW data worth DIN megapixels from photoreceptive pixels worth DIN megapixels that belong to the extraction frame EF. The DIN-megapixels-worth RAW data thus read out is fed to the color interpolation section 51. In other words, a RAW image having DIN-megapixels-worth RAW data as image data is fed to the color interpolation section 51.
A single piece of RAW data is a color signal of one of red, green, and blue. Accordingly, in a two-dimensional image represented by RAW data, red color signals are arranged in a mosaic pattern according to the color filter array (the same applies to green and blue). The color interpolation section 51 performs color interpolation (color interpolation processing) on the DIN-megapixels-worth RAW data to generate color-interpolated image composed of DIN megapixels (in other words, a color interpolation image having a DIN-megapixel image size). Well-known demosaicing processing can be used as color interpolation processing. The pixels of the color-interpolated image are each assigned R, G, and B signals as mutually different color signals, or luminance signal Y and color difference signals U and V. In the following description, it is assumed that, through color interpolation processing, R, G, and B signals are generated from RAW data, and image data expressed by R, G, and B signals is referred to as RGB data. Then, the color-interpolated image generated by the color interpolation section 51 has RGB data worth DIN megapixels. DIN-megapixels-worth RGB data is composed of DIN-megapixels-worth R signals, DIN-megapixels-worth G signals, and DIN-megapixels-worth B signals (the same applies to DOUT-megapixels-worth RGB data or YUV data, which will be discussed later).
The resolution conversion section 52 performs resolution conversion to convert the image size of the color-interpolated image from DIN megapixels to DOUT megapixels, and thereby generates, as a conversion result image, a color-interpolated image having undergone the resolution conversion (that is, a color-interpolated image having a DOUT-megapixel image size). The resolution conversion is achieved by well-known resampling. The conversion result image generated by the resolution conversion section 52 is composed of DOUT megapixels of pixels, and has RGB data worth DOUT megapixels. DOUT is a positive real number, and fulfills DIN≧DOUT. When DIN=DOUT, the conversion result image generated by the resolution conversion section 52 is identical with the color-interpolated image generated by the color interpolation section 51.
The value of DOUT is fed into the resolution conversion section 52. The user can specify the value of DOUT through a predetermined operation on the operation section 26. Instead, the value of DOUT may be constant. In the following description, unless otherwise indicated, it is assumed that DOUT=2. Then, DIN is 2 or more but 8 or less (because, as mentioned above, it is assumed that MH=4,000 and MV=2,000; see
Now, with reference to
The extraction frame setting section 50 determines the image size (dimensions) of the extraction frame EF from the RAW zoom magnification according to the following definition formula:
That is, the extraction frame setting section 50 determines the size of the extraction frame EF (in other words the image size of the extraction frame EF) such that the positive square root of (DOUT/DIN) equals (or approximately equals) the RAW zoom magnification. In the embodiment under discussion, since it is assumed that DOUT=2, the variable range of the RAW zoom magnification is between 0.5 times and 1 time.
When the RAW zoom magnification is 0.5 times, the definition formula above dictates that the image size of the extraction frame EF is 8 megapixels; thus, as shown in
When the RAW zoom magnification is 1 time, the definition formula above dictates that the image size of the extraction frame EF is 2 megapixels; thus, as shown in
As will be understood from the definition formula above and
Reducing the image size by resolution conversion based on the RAW zoom magnification accordingly alleviates the calculation load in signal processing (such as YUV conversion and signal compression) in later stages. Thus, during the shooting and recording of moving images, when temporal constraints in signal processing are comparatively strict, the use of RAW zooming is particularly beneficial.
The image-shooting device 1 is capable of, in addition to RAW zooming mentioned above, optical zooming and electronic zooming.
An operation of the zoom button 26c by the user is referred to as a zoom operation. According to a zoom operation, the zooming main control section 60 determines an overall zoom magnification and, from the overall zoom magnification, determines an optical zoom magnification, a RAW zoom magnification, and an electronic zoom magnification. According to the RAW zoom magnification set by the zooming main control section 60, the extraction frame setting section 50 sets the size of the extraction frame EF.
The optical zooming processing section 61 controls the position of the zoom lens 30 such that the angle of view of the image formed on the effective pixel region 33A is commensurate with the optical zoom magnification set by the zooming main control section 60. That is, the optical zooming processing section 61 controls the position of the zoom lens 30 according to the optical zoom magnification, and thereby sets the angle of view of the image formed on the effective pixel region 33A of the image sensor 33. As the optical zoom magnification increases to kC times from a given magnification, the angle of view of the image formed on the effective pixel region 33A diminishes to 1/kC times both in the horizontal and vertical directions of the image sensor 33 (where kC is a positive number, for example 2).
The YUV conversion section 53 converts, through YUV conversion, the data format of the image data of the conversion result image obtained at the resolution conversion section 52 into a YUV format, and thereby generates a YUV image. Specifically, the YUV conversion section 53 converts the R, G, and B signals of the conversion result image into luminance signals Y and color difference signals U and V, and thereby generates a YUV image composed of the luminance signal Y and color difference signals U and V thus obtained. Image data expressed by luminance signals Y and color difference signals U and V is also referred to as YUV data. Then, the YUV image generated at the YUV conversion section 53 has YUV data worth DOUT megapixels.
The electronic zooming processing section 54 applies electronic zooming processing according to the electronic zoom magnification set at the zooming main control section 60 to the YUV image, and thereby generates a final result image. Electronic zooming processing denotes processing whereby, as shown in
The overall zoom magnification, the optical zoom magnification, the electronic zoom magnification, and the RAW zoom magnification are represented by the symbols ZFTOT, ZFOPT, ZFEL, and ZFRAW respectively. Then, the formula
ZF
TOT
=ZF
OPT
×ZF
EL
×ZF
RAW×2
holds. Accordingly, the angle of view of the final result image decreases as the overall zoom magnification increases.
In the embodiment under discussion, it is assumed that the variable ranges of the optical zoom magnification and the electronic zoom magnification are each between 1 time and 10 times. Then, the variable range of the overall zoom magnification is between 1 time and 200 times.
In the range fulfilling 1≦ZFTOT≦20, while the magnification ZFEL is kept constant at 1 time, as the magnification ZFTOT increases from 1 time to 20 times, the magnification ZFOPT increases from 1 time to 10 times and also the magnification ZFRAW increases from 0.5 times to 1 times.
In the range fulfilling 20≦ZFTOT≦200, while the magnification ZFOPT is kept constant at 10 times and also the magnification ZFRAW is kept constant at 1 time, as the magnification ZFTOT increases from 20 times to 200 times, the magnification ZFEL increases from 1 time to 10 times.
In the range fulfilling 1≦ZFTOT≦20, as the magnification ZFTOT varies, the magnification ZFRAW varies together, and as the magnification ZFRAW varies, the size of the extraction frame EF (hence, the number of photoreceptive pixels inside the extraction frame EF) varies together.
Next, color interpolation processing will be described in detail. In color interpolation processing, as shown in
When photoreceptive pixel PS[p, q] is the pixel of interest, color interpolation processing can be performed by use of a filter FILA shown in
Instead, when photoreceptive pixel PS[p, q] is the pixel of interest, color interpolation processing can be performed by use of a filter FILB shown in
The color interpolation section 51 extracts the photoreceptive pixel signals of green photoreceptive pixels within a predetermined region centered around the pixel of interest, and mixes the extracted photoreceptive pixel signals to generate the G signal of the target pixel (in a case where only one photoreceptive pixel signal is extracted, the extracted photoreceptive pixel signal itself may be used as the G signal of the target pixel).
Similarly, the color interpolation section 51 extracts the photoreceptive pixel signals of red photoreceptive pixels within a predetermined region centered around the pixel of interest, and mixes the extracted photoreceptive pixel signals to generate the R signal of the target pixel (in a case where only one photoreceptive pixel signal is extracted, the extracted photoreceptive pixel signal itself may be used as the R signal of the target pixel).
Similarly, the color interpolation section 51 extracts the photoreceptive pixel signals of blue photoreceptive pixels within a predetermined region centered around the pixel of interest, and mixes the extracted photoreceptive pixel signals to generate the B signal of the target pixel (in a case where only one photoreceptive pixel signal is extracted, the extracted photoreceptive pixel signal itself may be used as the B signal of the target pixel).
To generate a G signal through basic color interpolation processing, the color interpolation section 51,
if, as shown in
if, as shown in
To generate an R signal through basic color interpolation processing, the color interpolation section 51,
if, as shown in
if, as shown in
if, as shown in
if, as shown in
As shown in
The filters 401 to 405 are each an example of the filter FILA.
Of the filter coefficients kA1 to kA25 of the filter 401, only kA13 is 1, and all the rest are 0.
Of the filter coefficients kA1 to kA25 of the filter 402, only kA8, kA12, kA14 and kA18 are 1, and all the rest are 0.
Of the filter coefficients kA1 to kA25 of the filter 403, only kA12 and kA14 are 1, and all the rest are 0.
Of the filter coefficients kA1 to kA25 of the filter 404, only kA8 and kA18 are 1, and all the rest are 0.
Of the filter coefficients kA1 to kA25 of the filter 405, only kA7, kA9, kA17, and kA19 are 1, and all the rest are 0.
When a G signal is generated through basic color interpolation processing with the pixel of interest being a green photoreceptive pixel as show in
In contrast, the filter 402 in
The same applies also when a G signal of the target pixel corresponding to a blue photoreceptive pixel is generated. A description similar to that given above may apply when the R and B signals of the target pixel are generated.
In view of the foregoing, as shown in
The color interpolation section 51 (and the resolution conversion section 52) can control the spatial frequency characteristic of the conversion result image according to the ratio DOUT/DIN of DOUT megapixels, which represents the number of pixels of the conversion result image, to DIN megapixels, which represents the number of photoreceptive pixels within the extraction frame EF (that is, the number of photoreceptive pixels belonging to the extraction frame EF). Here, the color interpolation section 51 (and the resolution conversion section 52) can change the spatial frequency characteristic of the conversion result image by changing the content of the color interpolation processing (the content of the filters used in the color interpolation processing) according to variation in the ratio DOUT/DIN. Since variation in the RAW zoom magnification causes the ratio DOUT/DIN to vary, the color interpolation section 51 (and the resolution conversion section 52) may be said to change the spatial frequency characteristic of the conversion result image in a manner interlocked with variation in the RAW zoom magnification or the overall zoom magnification.
In the following description, for the sake of simple reference, the control of the spatial frequency characteristic of the conversion result image is referred to simply as frequency characteristic control. Frequency characteristic control amounts to the control of the spatial frequency characteristic of the color-interpolated image, the YUV image, or the final result image. As specific methods of frequency characteristic control, or as specific examples of related methods, four practical examples will be presented below. Unless inconsistent, two or more of those practical examples may be combined, and any feature of one practical example may be applied to any other.
A first practical example (Example 1) of frequency characteristic control through color interpolation processing will now be described. Whereas in some later-described practical examples, it is assumed that the RAW image contains blur ascribable to camera shake or the like, in Example 1, and also in Example 2, which will be described next, it is assumed that the RAW image contains no blur.
Consider an input RAW image 451 and an output RAW image 452 shown in
The curves MTF451 and MTF452 in
When ZFRAW=0.5, the number of pixels of the output RAW image equals one-half of that of the input RAW image both in the vertical and horizontal directions. Therefore, the Nyquist frequency of the output RAW image 452 equals 0.5 FN. That is, the maximum spatial frequency that can be expressed in the output RAW image 452 equals one-half of the maximum spatial frequency that can be expressed in the input RAW image 451.
On the other hand, when ZFRAW=1.0, the number of pixels of the output RAW image equals that of the input RAW image both in the vertical and horizontal directions. Accordingly, the Nyquist frequency of the output RAW image 462 equals 1.0 FN. That is, the maximum spatial frequency that can be expressed in the output RAW image 462 equals the maximum spatial frequency that can be expressed in the input RAW image 461.
With consideration given to the above-discussed difference in frequency characteristic according to the RAW zoom magnification ZFRAW, in the color interpolation processing in Example 1, to suppress aliasing as well as lack in resolution (resolving power), filters as shown in
Specifically, when a G signal is generated under the condition ZFRAW=0.5, the color interpolation section 51,
if, as shown in
if, as shown in
On the other hand, when a G signal is generated under the condition ZFRAW=1.0, the color interpolation section 51,
if, as shown in
if, as shown in
The filters 501, 502, 511, and 512 are each an example of the filter FILA (see
Of the filter coefficients kA1 to kA25 of the filter 501, kA13 is 8, kA3, kA7, kA9, kA1l, kA15, kA17, kA19, and kA23 are 1, and all the rest are 0.
The filter coefficients of the filters 502 and 511 are the same as the filter coefficients of the filters 401 and 402, respectively, in
Of the filter coefficients kA1 to kA25 of the filter 512, kA8, kA12, kA14, and kA18 are 6, kA2, kA4, kA6, kA10, kA16, kA20, kA22, and kA24 are −1, and all the rest are 0.
Whereas the filter 501 has a function of smoothing the RAW image, the filter 502 does not have a function of smoothing the RAW image (smoothing of a RAW image is synonymous with smoothing of RAW data or photoreceptive pixel signals). Thus, the intensity of smoothing through color interpolation processing by use of the filter 501 can be said to be higher than the intensity (specifically, 0) of smoothing through color interpolation processing by use of the filter 502. Consequently, whereas when a G signal is generated by use of the filter 501, the high-frequency components of the spatial frequency of the G signal are attenuated, when a G signal is generated by use of the filter 502, no such attenuation occurs.
Whereas the filter 511 has a function of smoothing the RAW image, the filter 512 has a function of enhancing edges in the RAW image (edge enhancement of a RAW image is synonymous with edge enhancement of RAW data or photoreceptive pixel signals). Thus, the intensity of edge enhancement through color interpolation processing by use of the filter 512 can be said to be higher than the intensity (specifically, 0) of edge enhancement through color interpolation processing by use of the filter 511. Consequently, whereas when a G signal is generated by use of the filter 511, the high-frequency components of the spatial frequency of the G signal are attenuated, when a G signal is generated by use of the filter 512, either attenuation of the high-frequency components of the spatial frequency of the G signal does not occur too much or the same components are augmented. Alternatively, the degree of attenuation of the high-frequency components of the spatial frequency of the G signal through color interpolation processing is smaller when the filter 512 is used than when the filter 511 is used.
As described above, by controlling the content of color interpolation processing according to the RAW zoom magnification, the color interpolation section 51 achieves both suppression of aliasing and suppression of lack in resolution (resolving power). It should be noted that the spatial frequency here is the spatial frequency of a G signal. Specifically, when ZFRAW=0.5, the smoothing function of the filters 501 and 511 suppresses aliasing in the conversion result image. On the other hand, when ZFRAW=1.0, using the filters 502 and 512 eliminates or alleviates lack in resolution (resolving power) in the conversion result image.
It is merely as typical examples that filters for cases where ZFRAW=0.5 and ZFRAW=1.0 are discussed above; so long as 0.5≦ZFRAW≦1.0, including when ZFRAW=0.5 and ZFRAW=1.0, advisably, the intensity of smoothing by the filters is increased as ZFRAW decreases, or the intensity of edge enhancement by the filters is increased as ZFRAW increases.
For example, when a G signal is generated under the condition ZFRAW=0.7, the color interpolation section 51,
if, as shown in
if, as shown in
Of the filter coefficients kA1 to kA25 of the filter 503, kA13 is 10, kA7, kA9, kA17, and kA19 are 1, and all the rest are 0.
Of the filter coefficients kA1 to kA25 of the filter 513, kA8, kA12, kA14, and kA18 are 8, kA2, kA4, kA6, kA10, kA16, kA20, kA22, and kA24 are −1, and all the rest are 0.
The filters 501 and 503 both have a function of smoothing the RAW image, and the intensity of smoothing through color interpolation processing by use of the filter 501 is higher than the intensity of smoothing through color interpolation processing by use of the filter 503. The filters 512 and 513 both have a function of enhancing edges in the RAW image, and the intensity of edge enhancement through color interpolation processing by use of the filter 512 is higher than the intensity of edge enhancement through color interpolation processing by use of the filter 513.
Of R, G, and B signals, G signals are most visually affected by variation in spatial frequency characteristic. Accordingly, frequency characteristic control according to the RAW zoom magnification is applied only to G signals, and basic color interpolation processing is used for R and B signals.
Of course, changing of color interpolation processing according to the RAW zoom magnification may be applied also to the generation of R and B signals. A method of achieving that will now be described as a second practical example (Example 2). While the following description deals only with color interpolation processing with respect to R signals, color interpolation processing with respect to B signals can be performed in a similar manner to that with respect to R signals.
When an R signal is generated under the condition ZFRAW=0.5, the color interpolation section 51,
if, as shown in
if, as shown in
When an R signal is generated under the condition ZFRAW=1.0, the color interpolation section 51,
if, as shown in
if, as shown in
The filters 551, 552, and 561 are each an example of the filter FILA, and the filter 562 is an example of the filter FILB (see
Of the filter coefficients kA1 to kA25 of the filter 551, kA13 is 8, kA3, kA1l, kA15, and kA23 are 1, and all the rest are 0.
The filter coefficients of the filters 552 and 561 are the same as the filter coefficients of the filter 401 and 403, respectively, in
Of the filter coefficients kB1 to kB49 of the filter 562, kB24 and kB26 are 6, kB10, kB12, kB22, kB28, kB38, and kB40 are −1, and all the rest are 0.
Whereas the filter 551 has a function of smoothing the RAW image, the filter 552 does not have a function of smoothing the RAW image. Accordingly, the intensity of smoothing through color interpolation processing by use of the filter 551 can be said to be higher than the intensity (specifically, 0) of smoothing through color interpolation processing by use of the filter 552. Consequently, whereas when an R signal is generated by use of the filter 551, the high-frequency components of the spatial frequency of the R signal are attenuated, when an R signal is generated by use of the filter 552, no such attenuation occurs.
Whereas the filter 561 has a function of smoothing the RAW image, the filter 562 has a function of enhancing edges in the RAW image. Thus, the intensity of edge enhancement through color interpolation processing by use of the filter 562 can be said to be higher than the intensity (specifically, 0) of edge enhancement through color interpolation processing by use of the filter 561. Consequently, whereas when an R signal is generated by use of the filter 561, the high-frequency components of the spatial frequency of the R signal are attenuated, when an R signal is generated by use of the filter 562, either attenuation of the high-frequency components of the spatial frequency of the R signal does not occur too much or the same components are augmented. Alternatively, the degree of attenuation of the high-frequency components of the spatial frequency of the R signal through color interpolation processing is smaller when the filter 562 is used than when the filter 561 is used.
As described above, by controlling the content of color interpolation processing according to the RAW zoom magnification, the color interpolation section 51 achieves both suppression of aliasing and suppression of lack in resolution (resolving power). It should be noted that the spatial frequency here is the spatial frequency of an R signal. Specifically, when ZFRAW=0.5, the smoothing function of the filters 551 and 561 suppresses aliasing in the conversion result image. On the other hand, when ZFRAW=1.0, using the filters 552 and 562 eliminates or alleviates lack in resolution (resolving power) in the conversion result image.
It is merely as typical examples that filters for cases where ZFRAW=0.5 and ZFRAW=1.0 are discussed above; so long as 0.5≦ZFRAW≦1.0, including when ZFRAW=0.5 and ZFRAW=1.0, advisably, the intensity of smoothing by the filters is increased as ZFRAW decreases, or the intensity of edge enhancement by the filters is increased as ZFRAW increases. The same applies to the other practical examples described later.
No illustration or description is given of examples of filters used when the pixel of interest is a green photoreceptive pixel PS[2nA, 2nB] or a blue photoreceptive pixel; when the pixel of interest is a green photoreceptive pixel PS[2nA, 2nB] or a blue photoreceptive pixel, on a principle similar to that described above, filters according to the RAW zoom magnification can be used in color interpolation processing.
A third practical example (Example 3) of frequency characteristic control through color interpolation processing will now be described. In Example 3, it is assumed that, during the shooting of the RAW image, the image-shooting device 1 moves, with a result that the RAW image contains degradation due to blur.
Consider now an input RAW image 471 and an output RAW image 472 as shown in
The curves MTF471 and MTF472 in
Because of degradation due to blur, the maximum spatial frequency that can be included in the input RAW images 471 and 481 is lower than the Nyquist frequency FN, and is about (0.7×FN) in the examples shown in
When ZFRAW=0.5, the number of pixels of the output RAW image equals one-half of that of the input RAW image both in the vertical and horizontal directions. Thus, the Nyquist frequency of the output RAW image 472 equals 0.5FN.
On the other hand, when ZFRAW=1.0, the number of pixels of the output RAW image equals that of the input RAW image both in the vertical and horizontal directions. Thus, the Nyquist frequency of the output RAW image 482 equals 1.0FN. Even then, since the maximum spatial frequency that can be included in the input RAW image 481 is lower than the Nyquist frequency FN, the maximum spatial frequency that can be included in the output RAW image 482 also is lower than the Nyquist frequency FN.
Even in cases where degradation due to blur is involved, filters similar to those in Example 1 or 2 can be used in color interpolation processing, and this makes it possible to suppress aliasing and suppress lack in resolution (resolving power).
However, in a case where the RAW image contains degradation due to blur, in comparison with a case where the RAW image contains no degradation due to blur, the modulation transfer function is degraded, and the filter coefficients of filters can be determined with that degradation taken into consideration. Specifically, for example, the color interpolation section 51 may change the content of color interpolation processing between in a case (hereinafter referred to as case αBLUR) where the RAW image contains degradation due to blur and in a case (hereinafter referred to as case αNONBLUR) where the RAW image contains no degradation due to blur (that is, it may change the filter coefficients of the filters used in color interpolation processing between those cases). Between cases αBLUR and αNONBLUR, only part of the content of color interpolation processing may be changed, or the entire content of color interpolation processing may be changed.
To achieve that, in Example 3, as shown in
The motion detection section 62 may be realized, for example, with a motion sensor which detects the motion of the image-shooting device 1. The motion sensor is, for example, an angular acceleration sensor which detects the angular acceleration of the image-shooting device 1, or an acceleration sensor which detects the acceleration of the image-shooting device 1. In a case where the motion detection section 62 is realized with a motion sensor, the motion detection section 62 generates motion information that represents the motion of the image-shooting device 1 as detected by the motion sensor. The motion information based on the detection result of the motion sensor at least includes motion magnitude information that represents the magnitude of the motion of the image-shooting device 1, and may also include motion direction information that represents the direction of the motion of the image-shooting device 1.
Instead, the motion detection section 62 may generate motion information based on photoreceptive pixel signals from the image sensor 33. In that case, the motion detection section 62 can, for example, derive, from the image data of two images (RAW images, color-interpolated images, conversion result images, YUV images, or final result images) obtained by shooting at two temporally close time points, an optical flow between those two images and then, from the optical flow, generate motion information including motion magnitude information and motion direction information as mentioned above.
In Example 3, the color interpolation section 51 controls the content of the filters used in color interpolation processing according to the RAW zoom magnification and the motion information, and thereby controls the spatial frequency characteristic of the image having undergone color interpolation processing.
For the sake of concrete description, consider now a case where the RAW data of a RAW image 600 (not shown) is fed to the color interpolation section 51. Based on the motion information obtained for the RAW image 600, the color interpolation section 51 checks which of case αBLUR of case αNONBLUR applies to the RAW image 600. For example, if the magnitude of the motion of the image-shooting device 1 as indicated by the motion information is greater than a predetermined level, the color interpolation section 51 judges case αBLUR to apply to the RAW image 600 (that is, the RAW image 600 contains degradation due to blur); otherwise, the color interpolation section 51 judges case αNONBLUR to apply to the RAW image 600 (that is, the RAW image 600 contains no degradation due to blur).
When case αNONBLUR applies to the RAW image 600, the G signal of the target pixel is generated by the method described in connection with Example 1 (that is, through color interpolation processing using the filters 501 and 502 in
In case αBLUR, the filter 601 is used when ZFRAW=0.5 and in addition the pixel of interest is a green photoreceptive pixel, and the filter 602 is used when ZFRAW=1.0 and in addition the pixel of interest is a green photoreceptive pixel. The filters 601 and 602 are each an example of filter FILA (see
When the RAW image 600 obtained in cases αBLUR and αNONBLUR is identified by the symbols 600BLUR and 600NONBLUR respectively, then the modulation transfer functions of the RAW images 600BLUR and 600NONBLUR look like the curve MTF471 in
On the other hand, when ZFRAW=1.0, spatial frequency components equivalent to the spatial frequency components of the RAW image can be expressed in the conversion result image, and therefore priority is given to suppression of lack in resolution (resolving power), and the same filters are used in cases αBLUR and αNONBLUR (see the filter 502 in
As the magnitude of the motion of the image-shooting device 1 increases, the degree of degradation due to blur increases, and the RAW image 600 tends to contain less of high-frequency components. Conversely, even in case αBLUR, if the magnitude of the motion of the image-shooting device 1 is small, the RAW image 600 tends to contain high-frequency components in comparatively large amounts. Accordingly, in case αBLUR, the color interpolation section 51 may perform color interpolation processing according to motion magnitude information while taking the RAW zoom magnification into consideration. For example, the content of color interpolation processing may be changed (that is, the filter coefficients of the filters used in color interpolation processing may be made different) between in a case where the magnitude of the motion of the image-shooting device 1 as indicated by the motion magnitude information is a first magnitude and in a case where it is a second magnitude. Here, the first and second magnitudes differ from each other.
While the above description discusses the filters used to generate a G signal when the pixel of interest is a green photoreceptive pixel, also to generate a G signal when the pixel of interest is a red or blue photoreceptive pixel, and to generate an R or B signal when the pixel of interest is a green, red, or blue photoreceptive pixel, on a principle similar to that described above, filters according to the RAW zoom magnification and the motion information are used in color interpolation processing.
A fourth practical example (Example 4) will be described. The frequency characteristic control described above, including that discussed in connection with Examples 1 to 3, is realized through the control of the content of color interpolation processing. Frequency characteristic control equivalent to that described above may be realized through processing other than color interpolation processing. For example, configurations as shown in
In the configuration of
The color interpolation section 51 in
The filtering section 71 can control the spatial frequency characteristic of RAW data according to the RAW zoom magnification (in other words, according to the ratio DOUT DIN), or according to the RAW zoom magnification and the motion information. As the spatial frequency characteristic of RAW data is controlled, the spatial frequency characteristic of the conversion result image is controlled as well. Here, the filtering section 71 can, by changing the content of filtering according to variation in the ratio DOUT/DIN, change the spatial frequency characteristic of the conversion result image. Since variation in the RAW zoom magnification brings variation in the ratio DOUT/DIN, the filtering section 71 can be said to change the spatial frequency characteristic of the conversion result image in a manner interlocked with variation in the RAW zoom magnification or in the overall zoom magnification.
The filtering section 71 performs filtering according to the RAW zoom magnification, or filtering according to the RAW zoom magnification and the motion information, on the RAW image (that is, on the DIN-megapixel RAW data) in such a way that the spatial frequency characteristics of the color-interpolated image obtained from the color interpolation section 51 and the conversion result image obtained from the resolution conversion section 52 are similar between in the configuration of Example 4 and in the configuration of Example 1, 2, or 3. To achieve that, the filtering section 71 can operate as follows.
For example, only when ZFRAW<ZHTH1, the filtering section 71 performs filtering with a low-pass filter on the RAW data fed to the filtering section 71; when ZFRAW≧ZHTH1, the filtering section 71 does not perform filtering but feeds the RAW data fed to the filtering section 71 as it is to the color interpolation section 51. Here, ZHTH1 is a predetermined threshold value fulfilling 0.5<ZHTH1≦1.0, and for example ZHTH1=1.0.
Instead, for example, the filtering section 71 always performs filtering with a low-pass filter on the RAW data fed to the filtering section 71 irrespective of the value of ZFRAW, and increases the intensity of that low-pass filter as ZFRAW decreases from 1 to 0.5. For example, reducing the cut-off frequency of the low-pass filter belongs to increasing the intensity of the low-pass filter.
It is also possible to vary the intensity of the low-pass filter according to motion information. Specifically, for example, the filtering section 71 may check which of cases αBLUR and αNONBLUR applies to the RAW image based on the RAW data fed to the filtering section 71 according to motion information, and change the content of filtering between those cases. More specifically, for example, the filtering section 71 makes the intensity of the low-pass filter applied to the RAW image in case αBLUR lower than in case αNONBLUR so that, under the condition ZFRAW=0.5, an effect similar to that obtained in Example 3 is obtained.
The filtering by the filtering section 71 and the color interpolation processing by the color interpolation section 51 may be performed in the reversed order. That is, it is possible to first perform the color interpolation processing and then perform the filtering by the filtering section 71.
Example 4 offers benefits similar to those Examples 1, 2, or 3 offers. In Example 4, however, the filtering section 71 is needed separately from the color interpolation section 51. Accordingly, Examples 1 to 3, where frequency characteristic control can be performed according to the RAW zoom magnification etc. in color interpolation processing, are more advantageous in terms of processing speed and processing load.
The present invention may be carried out with whatever variations or modifications made within the scope of the technical idea presented in the appended claims. The embodiments described specifically above are merely examples of how the invention can be carried out, and the meanings of the terms used to describe the invention and its features are not to be limited to those in which they are used in the above description of the embodiments. All specific values appearing in the above description are merely examples and thus, needless to say, can be changed to any other values. Supplementary comments applicable to the embodiments described above are given in Notes 1 to 4 below. Unless inconsistent, any part of the comments can be combined freely with any other.
Note 1: In the configuration shown in
Note 2: In the configuration shown in
Note 3: The image-shooting device 1 shown in
Note 4: For example, the following interpretation is possible:
The image-shooting device 1 is provided with a specific signal processing section which, through specific signal processing, generates the image data of an output image from photoreceptive pixel signals within an extraction frame EF on the image sensor 33. A conversion result image, a YUV image, or a final result image is an example of the output image. Specific signal processing is processing performed on the photoreceptive pixel signals within the extraction frame EF, and on a signal based on the photoreceptive pixel signals within the extraction frame EF, to generate the image data of the output image from the photoreceptive pixel signals within the extraction frame EF.
The specific signal processing section includes a color interpolation section 51 and a resolution conversion section 52, or includes a filtering section 71, a color interpolation section 51, and a resolution conversion section 52, and may additionally include a YUV conversion section 53, an electronic zooming processing section 54, and a filtering section 71. Thus, in Examples 1 to 3, the specific signal processing includes color interpolation processing and resolution conversion, and in Example 4, the specific signal processing includes filtering (the filtering by the filtering section 71), color interpolation processing, and resolution conversion. Although not shown in
Number | Date | Country | Kind |
---|---|---|---|
2011-049128 | Mar 2011 | JP | national |