1. Field of the Invention
The present invention relates to an image processing technique for converting a first image (for example, RAW data) having color components arranged mixedly, to generate a second image having at least one kind of components arranged on each pixel.
2. Description of the Related Art
(Prior Art 1)
Conventionally, there have been known electronic cameras that perform spatial filtering such as edge enhancement and noise removal.
Spatial filtering of this type is typically applied to luminance and chrominance planes YCbCr (the luminance component Y in particular), after RAW data (for example, Bayer array data) output from a single-plate image sensor is subjected to color interpolation and the luminance and chrominance planes YCbCr are subjected to color system conversion processing. For example, one of known typical noise removal filters is a ε-filter.
However, this processing has had a problem in complicating the image processing since the color interpolation, the color system conversion processing, and the spatial filtering must be performed separately. There have thus been problems such as extended processing time for the RAW data. The complexity of the image processing has also produced another problem that image processing ICs to be mounted on the electronic cameras are to be complex in configuration.
Furthermore, these processings (the color interpolation, the color system conversion processing, and the spatial filtering) are applied to a single image step by step, which causes another problem that minute image information can be lost easily in the course of the cumulative processes.
(Prior Art 2)
When color interpolation is performed on RAW data of a single-plate image sensor, original signals in the RAW data and interpolation signals generated by averaging the original signals are usually arranged on a single plane. Here, the original signals and the interpolation signals slightly differ in spatial frequency characteristics.
In U.S. Pat. No. 5,596,367 (hereinafter, referred to as patent document 1), low-pass filter processing is applied to the original signals alone so as to reduce the differences in the spatial frequency characteristics between the original signals and the interpolation signals.
In this processing, however, the low-pass filter processing is applied to the original signals by using the interpolation signals adjoining to the original signals. In other words, this processing still involves the step-by-step application of color interpolation and spatial filtering. It is also disadvantageous that minute image information can be lost easily.
(Prior Art 3)
The inventors of the present invention previously filed an international application, International Patent Publication No. WO 02/21849 (hereinafter, referred to as patent document 2). The international application discloses an image processing apparatus which applies color system conversion processing directly to RAW data.
In this processing, color system conversion is performed by performing weighted addition of the RAW data according to a coefficient table. This coefficient table can contain in advance fixed coefficient terms intended for edge enhancement and noise removal. No particular mention has been made therein, however, of the technique for creating another groups of coefficient tables having different spatial frequency characteristics in advance and switching the groups of coefficient tables as needed, as in the embodiments to be described later.
(Problem with Imaging Sensitivity)
Typical electronic cameras can change the imaging sensitivity of their image sensor (for example, the amplifier gain of the image sensor output). This change in the imaging sensitivity causes large variations in the noise amplitude of the captured images. Patent documents 1 and 2 have not described the technique of changing a conversion filter for the first image in accordance with the imaging sensitivity, as in the embodiments to be described later. Therefore, over-blurring may occur for low sensitivity images with high S/N due to excessive smoothing by the conversion filter, whereas color artifacts may be closed up or granularity may be left for high sensitivity images with low S/N due to inadequate smoothing by the conversion filter.
In view of solving the foregoing problems, it is an object of the present invention to easily, efficiently perform sophisticated spatial filtering conforming to image structures.
Another object of the present invention is to provide an image processing technique for realizing appropriate noise removal while maintaining resolution and contrast regardless of changes in the imaging sensitivity.
Hereinafter, description will be given of the present invention.
(1) An image processing apparatus of the present invention is an image processing apparatus for converting a first image composed of any one of first to nth color components (n≧2) entirely arranged on each pixel, into a second image composed of all of the first to nth color components arranged on each pixel.
This image processing apparatus includes a smoothing unit. This smoothing unit smoothes a pixel position of the first color component in the first image, using the first color components of pixels adjacent to the pixel position. The smoothing unit outputs the first color component having been smoothed as the first color component in the pixel position of the second image. The smoothing unit further includes a control unit. The control unit changes a characteristic of a smoothing filter in accordance with an imaging sensitivity at which the first image is captured. Such processing makes it possible to obtain noise removal effects of high definition adaptable to changes in the imaging sensitivity.
(2) Preferably, among the first to nth color components, the first color component is a color component that carries a luminance signal.
(3) It is also preferable that the first to nth color components are red, green, and blue, and the first color component is green.
(4) The control unit preferably changes a size (a range of pixels to be referred) of the filter in accordance with the imaging sensitivity.
(5) It is also preferable that the control unit changes coefficient values (contribution ratios of pixel components to be referred among pixels around a smoothing target pixel) of the filter in accordance with the imaging sensitivity.
(6) The smoothing unit preferably includes a similarity judgment unit and a switching unit. This similarity judgment unit judges a magnitude of similarity among pixels in a plurality of directions. Meanwhile, the switching unit switchingly outputs the first color component of the first image simply and the first color component having been smoothed as the first color component of the second image, according to a result of the judgment.
(7) It is also preferable that the similarity judgment unit judges similarity by calculating similarity degrees among pixels at least in four directions.
(8) Another image processing apparatus of the present invention is an image processing apparatus for converting a first image composed of any one of first to nth color components (n≧2) arranged on each pixel, into a second image composed of at least one signal component arranged entirely on each pixel.
This image processing apparatus includes a signal generating unit. The signal generating unit generates a signal component of the second image by performing weighted addition of color components in the first image. This signal generating unit further includes a control unit. The control unit changes weighting coefficients for the weighted addition in accordance with an imaging sensitivity at which the first image is captured, the weighting coefficients being used for adding up the color components of the first image.
(9) The signal generating unit preferably generates a signal component different from the first to nth color components.
(10) It is also preferable that the signal generating unit generates a luminance component different from the first to nth color components.
(11) The control unit preferably changes the weighting coefficients for a pixel position of the first color component in the first image in accordance with the imaging sensitivity.
(12) It is also preferable that the control unit changes a range of the weighted addition in accordance with the imaging sensitivity.
(13) The control unit preferably changes the weighting coefficients within the identical range in accordance with the imaging sensitivity.
(14) It is also preferable that the signal generating unit has a similarity judgment unit. This similarity judgment unit judges a magnitude of similarity among pixels in a plurality of directions. The control unit also changes the weighting coefficients in accordance with the similarity judgment in addition to the imaging sensitivity.
(15) The control unit preferably executes weighted addition of a color component originally existing on a pixel to be processed in the first image and the same color component existing on the surrounding pixels when a result of the judgment indicates no distinctive similarity in any direction or higher similarity than a predetermined level in all of the directions.
(16) It is also preferable that the similarity judgment unit judges similarity by calculating similarity degrees among pixels at least in four directions.
(17) Another image processing apparatus of the present invention is an image processing apparatus for converting a first image composed of a plurality of kinds of color components mixedly arranged on a pixel array, to generate a second image composed of at least one kind of signal component (hereinafter, new component) arranged entirely on each pixel. The color components constitute a color system.
This image processing apparatus includes a similarity judgment unit, a coefficient selecting unit, and a conversion processing unit. Initially, the similarity judgment unit judges similarity of a pixel to be processed along a plurality of directions in the first image. The coefficient selecting unit selects a predetermined coefficient table in accordance with a result of the judgment on the similarity having been made by the similarity judgment unit. The conversion processing unit performs weighted addition of the color components in a local area including the pixel to be processed based on the coefficient table having been selected, thereby generating the new component. In particular, the coefficient selecting unit described above selects a different coefficient table having a different spatial frequency characteristic in accordance with an analysis of an image structure based on the similarity. Changing the coefficient table thus achieves an adjustment to a spatial frequency component of the new component.
As has been described, according to the present invention, the coefficient table is switched to one with a different spatial frequency characteristic in accordance with the similarity-based analysis of the image structure, so as to adjust the spatial frequency component of the new component to be generated.
Such an operation eliminates the need for step-by-step processing of generating the new component once before subjecting this new component to spatial filtering as in the prior art. Accordingly, the steps of the image processing can be simplified efficiently.
Moreover, the similarity used for generating the new component is also used to fulfill the analysis of the image structure, which makes the processing efficient to achieve sophisticated spatial filtering in consideration of the image structure easily.
Furthermore, in this processing, the generation of the new component and the adjustment to the spatial frequency component based on the analysis of the image structure are achieved by a single weighted addition. Minute image information is thus less likely to be lost as compared to the cases where the arithmetic processing is divided and repeated a plurality of times.
According to the present invention, weighting ratios of the color components are preferably associated with weighting ratios for color system conversion. This can eliminate the need for conventional color interpolation, and complete the color system conversion processing and the spatial filtering in consideration of the image structure by a single weighted addition. By such processing, it is possible to significantly simplify and accelerate the image processing on, for example, RAW data or the like which has taken a long time heretofore.
(18) The coefficient selecting unit preferably analyzes the image structure of pixels near the pixel to be processed, based on a result of judgment on a magnitude of the similarity. In accordance with the analysis, the coefficient selecting unit selects a different coefficient table having a different spatial frequency characteristic.
(19) It is also preferable that the coefficient selecting unit selects a different coefficient table having a different array size. Selecting one in the different array size is to select one with a different spatial frequency characteristic.
(20) The coefficient selecting unit preferably selects a different coefficient for a higher level of noise removal to suppress a high frequency component of the signal component greatly and/or over a wider bandlength, when the similarity is judged to be substantially uniform in the plurality of directions and judged to be high from the analysis of an image structure.
(21) It is also preferable that the coefficient selecting unit selects a different coefficient table for a higher level of noise removal to suppress a high frequency component of the signal component greatly and/or over a wider bandlength, when the similarity is judged to be substantially uniform in the plurality of directions and judged to be low from the analysis of an image structure.
(22) The coefficient selecting unit preferably selects a different coefficient table for a higher level of edge enhancement to enhance a high frequency component of the signal component in a direction of low similarity, when a difference in the magnitude of the similarity in the directions is judged to be large from the analysis of an image structure.
(23) It is also preferable that the coefficient selecting unit selects a different coefficient table for a higher level of edge enhancement to enhance a high frequency component of the signal component in a direction of low similarity, when a difference in the magnitude of the similarity in the directions is judged to be small from the analysis of the image structure.
(24) The coefficient selecting unit preferably selects a different coefficient table for a higher level of noise removal such that the higher the imaging sensitivity at which the first image is captured is, the higher the level of noise removal through the selected coefficient table is.
(25) It is also preferable that weighting ratios between the color components are to be substantially constant before and after selecting the different coefficient table.
(26) Preferably, the weighting ratios between the color components are intended for color system conversion.
(27) Another image processing apparatus of the present invention includes a smoothing unit and a control unit. The smoothing unit smoothes image data by performing weighted addition on a pixel to be processed and the surrounding pixels in the image data. Meanwhile, the control unit changes a referential range of the surrounding pixels in accordance with an imaging sensitivity at which this image data is captured.
(28) An image processing program of the present invention enables a computer to operate as an image processing apparatus according to any one of (1) to (27) above.
(29) An electronic camera of the present invention includes: an image processing apparatus according to any one of (1) to (27) above; and an image sensing unit capturing a subject and generating a first image. In this electronic camera, the image processing apparatus processes the first image captured by the image sensing unit to generate a second image.
(30) An image processing method of the present invention is for converting a first image composed of any one of first to nth color components (n≧2) arranged on each pixel, into a second image composed of at least one signal component arranged entirely on each pixel. This image processing method includes the step of generating the signal component of the second image by performing weighted addition of color components in the first image. In particular, the step of generating this signal component includes the step of changing weighting coefficients for the weighted addition in accordance with an imaging sensitivity at which the first image is captured. The weighting coefficients are used for adding up the color components in the first image with a weight.
(31) Another image processing method of the present invention is for converting a first image composed of a plurality of kinds of color components mixedly arranged on a pixel array, to generate a second image composed of at least one kind of signal component (hereinafter, as new component) arranged entirely on each pixel. The color components constitute a color system. This image processing method has the following steps:
[S1] the step of judging image similarity of a pixel to be processed along a plurality of directions in the first image;
[S2] the step of selecting a predetermined coefficient table in accordance with a result of the judgment on similarity in the step of judging similarity; and
[S3] the step of performing weighted addition of the color components in a local area including the pixel to be processed according to the coefficient table having been selected, thereby generating the new component.
In particular, in the foregoing coefficient table selecting step, a different coefficient table having a different spatial frequency characteristic is selected in accordance with an analysis of an image structure based on the similarity, thereby adjusting the spatial frequency component of the new component.
The nature, principle, and utility of the invention will become more apparent from the following detailed description when read in conjunction with the accompanying drawings in which like parts are designated by identical reference numbers, in which:
Hereinafter, a first embodiment according to the present invention will be described with reference to the drawings.
In
This memory 13 is connected with an image processing unit (for example, a single-chip microprocessor dedicated to image processing) 11, a control unit 12, a compression/decompression unit 14, an image display unit 15, a recording unit 17, and an external interface unit 19 via the bus.
The electronic camera 1 is also provided with an operating unit 24, a monitor 25, and a timing control unit 23. Moreover, the electronic camera 1 is loaded with a memory card 16. The recording unit 17 compresses and records processed images onto this memory card 16.
The electronic camera 1 can also be connected with an external computer 18 via the external interface unit 19 (USB or the like).
FIGS. 2 to 7 are operational flowcharts of the image processing unit 11.
Referring to
The image processing unit 11 makes a direction judgment on similarity in the horizontal and vertical directions around a pixel to be processed on the RAW data plane, thereby determining an index HV (steps S1). This index HV is set to “1” when the vertical similarity is higher than the horizontal, set to “−1” when the horizontal similarity is higher than the vertical, and set to “0” when the vertical and horizontal similarities are indistinguishable.
Moreover, the image processing unit 11 makes a direction judgment on similarity in the diagonal directions around the pixel to be processed on the RAW data plane, thereby determining an index DN (steps S2). This index DN is set to “1” when the similarity along a 45° diagonal direction is higher than that along a 135° diagonal direction, set to “−1” when the similarity along a 135° diagonal direction is higher than that along a 45° diagonal direction, and set to “0” when these similarities are indistinguishable.
Next, the image processing unit 11 performs luminance component generation processing (step S3) while performing chromaticity component generation processing (step S4).
Since the chromaticity (chrominance) component generation processing is detailed in the embodiment of the foregoing patent document 2, description thereof will be omitted here.
Hereinafter, concrete description will be given of the operations of the processing for setting the index HV, the processing for setting the index DN, and the luminance component generation processing in order.
<<Processing for Setting Index HV>>
Initially, the processing for calculating the index HV[i,j] will be described with reference to
Step S12: The image processing unit 111 initially calculates difference values between pixels in the horizontal and vertical directions at coordinates [i,j] on the RAW data as similarity degrees.
For example, the image processing unit 11 calculates a vertical similarity degree Cv[i,j] and a horizontal similarity degree Ch[i,j] by using the following equations 1 to 4. (The absolute values ∥ in the equations may be replaced with squares or other operations.)
(1) If the coordinates [i,j] fall on an R position or B position:
(2) If the coordinates [i,j] fall on a G position:
The smaller values the similarity degrees calculated thus have, the higher the similarities are.
Step S13: Next, the image processing unit 11 compares the similarity degrees in the horizontal and vertical directions.
Step S14: For example, when the following condition 2 holds, the image processing unit 11 judges the horizontal and vertical similarity degrees as being nearly equal, and sets the index HV[i,j] to 0.
|Cv[i,j]−Ch[i,j]|≦Th4 condition 2
In condition 2, the threshold Th4 functions to avoid either one of the similarities from being misjudged as being higher because of noise when the difference between the horizontal and vertical similarity degrees is small. For noisy color images, the threshold Th4 is thus preferably set to higher values.
Step S15: On the other hand, if condition 2 does not hold but the following condition 3 does, the image processing unit 11 judges the vertical similarity as being higher, and sets the index HV[i,j] to 1.
Cv[i,j]<Ch[i,j] condition 3
Step S16: Moreover, if neither of conditions 2 and 3 holds, the image processing unit 11 judges the horizontal similarity as being higher, and sets the index HV[i,j] to −1.
Note that the similarity degrees calculated here are for both R and B positions and G positions. For the sake of simplicity, however, it is possible to calculate the similarity degrees for R and B positions alone, and set the directional index HV at the R and B positions. The directional index at G positions may be determined by referring to HV values around. For example, the directional index at a G position may be determined by averaging the indices from four points adjoining the G position and converting the average into an integer.
<<Processing for Setting Index DN>>
Next, the processing for calculating the index DN[i,j] will be described with reference to
Step S31: Initially, the image processing unit 111 calculates difference values between pixels in the 45° diagonal direction and the 135° diagonal direction at coordinates [i,j] on the RAW data as similarity degrees.
For example, the image processing unit 11 determines a similarity degree C45[i,j] in the 45° diagonal direction and a similarity degree C135[i,j] in the 135° diagonal direction by using the following equations 5 to 8.
(1) If the coordinates [i,j] fall on an R position or B position:
(2) If the coordinates [i,j] fall on a G position:
The smaller values the similarity degrees calculated thus have, the higher the similarities are.
Step S32: Having thus calculated the similarity degrees in the 45° diagonal direction and the 135° diagonal direction, the image processing unit 11 judges from these similarity degrees whether or not the similarity degrees in the two diagonal directions are nearly equal.
For example, such a judgment can be made by judging if the following condition 5 holds.
|C45[i,j]−C135[i,j]|≦Th5 condition 5
The threshold Th5 functions to avoid either one of the similarities from being misjudged as being higher because of noise when the difference between the similarity degrees C45[i,j] and C135[i,j] in the two directions is small. For noisy color images, the threshold Th5 is thus preferably set to higher values.
Step S33: If such a judgment indicates that the diagonal similarities are nearly equal, the image processing unit 11 sets the index DN[i,j] to 0.
Step S34: On the other hand, if the direction of higher diagonal similarity is distinguishable, a judgment is made as to whether or not the similarity in the 45° diagonal direction is higher.
For example, such a judgment can be made by judging if the following condition 6 holds.
C45[i,j]<C135 [i,j] condition 6
Step S35: Then, if the judgment at step S34 indicates that the similarity in the 45° diagonal direction is higher (when condition 5 does not hold but condition 6 does), the image processing unit 11 sets the index DN[i,j] to 1.
Step S36: On the other hand, if the similarity in the 135° diagonal direction is higher (when neither of conditions 5 and 6 holds), the index DN[i,j] is set to −1.
Note that the similarity degrees calculated here are for both R and B positions and G positions. For the sake of simplicity, however, it is possible to calculate the similarity degrees for R and B positions alone, and set the directional index DN at the R and B positions. The directional index at G positions may be determined by referring to DN values around. For example, the directional index at a G position may be determined by averaging the indices from four points adjoining the G position and converting the average into an integer.
<<Luminance Component Generation Processing>>
Next, the operation of the luminance component generation processing will be described with reference to FIGS. 5 to 7.
Step S41: The image processing unit 11 judges whether or not the indices (HV,DN) of the pixel to be processed are (0,0).
Here, if the indices (HV,DN) are (0,0), it is possible to judge that the similarities are generally uniform both in the vertical and horizontal directions and in the diagonal directions, and the location indicates isotropic similarity. In this case, the image processing unit 11 moves the operation to step S42.
On the other hand, if the indices (HV,DN) are other than (0,0), it is possible to judge that the similarities in the horizontal and vertical directions or the diagonal directions are non-uniform and the location has directionality in the image structure as shown in
Step S42: The image processing unit 11 acquires, from the control unit 12, information on the imaging sensitivity (corresponding to the amplifier gain of the image sensor) at which the RAW data is captured.
If the imaging sensitivity is high (for example, equivalent to ISO 800 or above), the image processing unit 11 moves the operation to step S46.
On the other hand, if the imaging sensitivity is low, the image processing unit 11 moves the operation to step S43.
Step S43: The image processing unit 11 performs a judgment on the magnitudes of similarity. For example, such a magnitude judgment is made depending on whether or not any of the similarity degrees Cv[i,j] and Ch[i,j] used in calculating the index HV and the similarity degrees C45[i,j] and C135[i,j] used in calculating the index DN satisfies the following condition 7:
Similarity degree>threshold th6 condition 7
Here, the threshold th6 is a boundary value for determining whether the location having isotropic similarity is a flat area or a location having significant relief information, and is set in advance in accordance with the actual values of the RAW data.
If condition 7 holds, the image processing unit 11 moves the operation to step S44.
On the other hand, if condition 7 does not hold, the image processing unit 11 moves the operation to step S45.
Step S44: Here, since condition 7 holds, it is possible to determine that the pixel to be processed has low similarity to its surrounding pixels, i.e., is a location having significant relief information. To keep this significant relief information, the image processing unit 11 then selects a coefficient table 1 (see
Step S45: Here, since condition 7 does not hold, it is possible to determine that the pixel to be processed has high similarity to its surrounding pixels, i.e., is a flat area. In order to remove noise of small amplitudes noticeable in this flat area with reliability, the image processing unit 11 selects either one of coefficient tables 2 and 3 (see
After such a selecting operation, the image processing unit 11 moves the operation to step S51.
Step S46: Here, since the imaging sensitivity is high, it is possible to determine that the RAW data is low in S/N. Then, in order to remove the noise of the RAW data with reliability, the image processing unit 11 selects a coefficient table 4 (see
Step S47: Here, the pixel to be processed has anisotropic similarity. Then, the image processing unit 11 determines a difference in magnitude between the similarity in the direction of similarity and the similarity in the direction of non-similarity.
For example, such a difference in magnitude can be determined from a difference or ratio between the vertical similarity degree Cv[i,j] and the horizontal similarity degree Ch[i,j] which are used in calculating the index HV. In another example, it can also be determined from a difference or ratio between the similarity degree C45[i,j] in the 45° diagonal direction and the similarity degree C135[i,j] in the 135° diagonal direction which are used in calculating the index DN.
Step S48: The image processing unit 11 makes a threshold judgment on the determined difference in magnitude, in accordance with the following condition 8.
|Difference in magnitude|>threshold th7 condition 8
Note that the threshold th7 is a value for distinguishing whether or not the pixel to be processed has the image structure of an edge area, and is set in advance in accordance with the actual values of the RAW data.
If condition 8 holds, the image processing unit 11 moves the operation to step S50.
On the other hand, if condition 8 does not hold, the operation is moved to step S49.
Step S49: Here, since condition 8 does not hold, the pixel to be processed is estimated not to be an edge area of any image. The image processing unit 11 then selects a coefficient table from among a group of coefficient tables for low edge enhancement (coefficient tables 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, and 27 having a matrix size of 3×3, shown in FIGS. 9 to 13).
Specifically, the image processing unit 11 classifies the pixel to be processed among the following cases 1 to 12, based on the conditions of the judgment on the direction of the similarity by the indices (HV,DN) and the color component of the pixel to be processed in combination. “x” below may be any one of 1, 0, and −1.
<<R position or B position>>
case 1: (HV,DN)=(1,1): high similarity in the vertical and 45° diagonal directions;
case 2: (HV,DN)=(1,0): high similarity in the vertical direction;
case 3: (HV,DN)=(1,−1): high similarity in the vertical and 135° diagonal directions;
case 4: (HV,DN)=(0,1): high similarity in the 45° diagonal direction;
case 5: unused;
case 6: (HV,DN)=(0,−1): high similarity in the 135° diagonal direction;
case 7: (HV,DN)=(−1,1): high similarity in the horizontal and 45° diagonal directions;
case 8: (HV,DN)=(−1,0): high similarity in the horizontal direction; and
case 9: (HV,DN)=(−1,−1): high similarity in the horizontal and 135° diagonal directions.
<<G position>>
case 10: (HV,DN)=(1,x): high similarity at least in the vertical direction;
case 11_1: (HV,DN)=(0,1): high similarity in the 45° diagonal direction;
case 11_2: (HV,DN)=(0,−1): high similarity in the 135° diagonal direction; and
case 12: (HV,DN)=(−1,x): high similarity at least in the horizontal direction.
In accordance with this classification of cases 1 to 12, the image processing unit 11 selects the following coefficient tables from among the group of coefficient tables for low edge enhancement (the coefficient tables 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, and 27 shown in FIGS. 9 to 13).
In case 1: select the coefficient table 5;
In case 2: select the coefficient table 7;
In case 3: select the coefficient table 9;
In case 4: select the coefficient table 11;
In case 5: unused;
In case 6: select the coefficient table 13;
In case 7: select the coefficient table 15;
In case 8: select the coefficient table 17;
In case 9: select the coefficient table 19;
In case 10: select the coefficient table 21;
In case 11_1: select the coefficient table 23;
In case 11_2: select the coefficient table 25; and
In case 12: select the coefficient table 27.
The coefficient tables selected here contain coefficients that are arranged with priority given to the directions of relatively high similarities.
After a coefficient table is selected thus, the image processing unit 11 moves the operation to step S51.
Step S50: Here, since condition 8 holds, the pixel to be processed is estimated to be an edge area of an image. The image processing unit 11 then selects a coefficient table from among a group of coefficient tables for high edge enhancement (coefficient tables 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, and 28 having a matrix size of 5×5, shown in FIGS. 9 to 13).
Specifically, the image processing unit 11 classifies the pixel to be processed among cases 1 to 12 as in step S49.
In accordance with this classification of cases 1 to 12, the image processing unit 11 selects the following coefficient tables from among the group of coefficient tables for high edge enhancement (the coefficient tables 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, and 28 shown in FIGS. 9 to 13).
In case 1: select the coefficient table 6;
In case 2: select the coefficient table 8;
In case 3: select the coefficient table 10;
In case 4: select the coefficient table 12;
In case 5: unused;
In case 6: select the coefficient table 14;
In case 7: select the coefficient table 16;
In case 8: select the coefficient table 18;
In case 9: select the coefficient table 20;
In case 10: select the coefficient table 22;
In case 11_1: select the coefficient table 24;
In case 11_2: select the coefficient table 26; and
In case 12: select the coefficient table 28.
The coefficient tables selected here contain coefficients that are arranged with priority given to the directions of relatively high similarities. In addition, these coefficient tables contain negative coefficient terms which are arranged in directions generally perpendicular to the directions of similarities, thereby allowing edge enhancement on the image.
After a coefficient table is selected thus, the image processing unit 11 moves the operation to step S51.
Step S51: By the series of operations described above, coefficient tables are selected pixel by pixel. The image processing unit 11 adds the color components in the local area including the pixel to be processed of the RAW data, by multiplying the coefficient values of the coefficient table selected thus.
Here, whichever coefficient tables shown in FIGS. 9 to 13 are selected, the weighting ratios of the respective color components in this weighted addition shall always be kept as R:G:B=1:2:1. These weighting ratios are equal to weighting ratios for determining a luminance component Y from RGB color components. In the foregoing weighted addition, the luminance component Y is thus generated directly from the RAW data pixel by pixel.
As has been described, according to the first embodiment, the groups of coefficient tables having different spatial frequency characteristics are prepared in advance, and the groups of coefficient tables are switched for use in accordance with the analysis of the image structure (steps S43 and S48). As a result, the basically-separated image processes of color system conversion and spatial filtering in consideration of the image structure can be performed by a single weighted addition.
This eliminates the need to perform the spatial filtering and the color system conversion separately, thereby allowing a significant reduction in time necessary for processing the RAW data.
Besides, since what is necessary is a single weighted addition, it is also possible to reduce deterioration of image information as compared to the background art where color system conversion and spatial filtering are conducted step by step.
Moreover, according to the first embodiment, a type of coefficient tables having a higher level of noise removal will be selected when it is judged that similarities in a plurality of directions are isotropic and the similarities are high (steps 543 and S45). It is therefore possible to suppress noise noticeable in flat areas of the image strongly, while performing color system conversion.
On the other hand, according to the first embodiment, coefficient tables having low LPF characteristics will be selected for locations that have significant relief information (steps S43 and S44). It is therefore possible to generate high-quality image data that contains much image information.
Furthermore, according to the first embodiment, if the similarities among a plurality of directions are judged as having a large difference in magnitude, coefficient tables can be switched to a type of those having a higher level of edge enhancement for enhancing the high frequency components in the direction of non-similarity (steps S48 and S50). It is therefore possible to make images sharp in edge contrast, while performing color system conversion.
In addition, according to the first embodiment, the coefficient tables can be changed to ones having a higher level of noise removal as the imaging sensitivity increases (steps S42 and S46). This makes it possible to more strongly suppress noise which increases as the imaging sensitivity increases, while performing color system conversion.
Now, description will be given of another embodiment.
The electronic camera (including an image processing apparatus) according to a second embodiment performs color interpolation on RGB Bayer-array RAW data (corresponding to the first image), thereby generating image data that has RGB signal components arranged entirely on each pixel (corresponding to the second image).
The configuration of the electronic camera (
Step S61: The image processing unit 11 makes a similarity judgment on a G pixel [i,j] of RAW data to be processed, thereby determining whether or not the location has similarities indistinguishable in any direction, i.e., whether or not the location has high isotropy, having no significant directionality in its image structure.
For example, the image processing unit 11 determines the indices (HV,DN) of this G pixel [i,j]. Since this processing is the same as in the first embodiment (
Next, the image processing unit 11 judges whether or not the determined indices (HV,DN) are (0,0). If the indices (HV,DN) are (0,0), it is possible to judge that the similarities are generally uniform both in the vertical and horizontal directions and in the diagonal directions, and the G pixel [i,j] is a location having indistinguishable similarities. In this case, the image processing unit 11 moves the operation to step S63.
On the other hand, if the indices (HV,DN) are other than (0,0), the image structure has a significant directionality. In this case, the image processing unit 11 moves the operation to step S62.
Step S62: In this step, the image structure has a significant directionality. That is, it is highly possible that the G pixel [i,j] to be processed falls on an edge area, detailed area, or the like of an image and is an important image structure. Then, in order to maintain the important image structure with high fidelity, the image processing unit 11 skips smoothing processing (steps S63 and S64) to be described later. That is, the image processing unit 11 uses the value of the G pixel [i,j] in the RAW data simply as the G color component of the pixel [i,j] on a color interpolated plane.
After this processing, the image processing unit 11 moves the operation to step S65.
Step S63: In this step, in contrast, the image structure has no significant directionality. It is thus likely to be a flat area in the image or spot-like noise isolated from the periphery. The image processing unit 11 can perform smoothing on such locations alone, whereby noise in G pixels is suppressed without deteriorating important image structures. The image processing unit 11 determines the smoothing level by referring to the imaging sensitivity in capturing the RAW data, aside from the foregoing similarity judgment (judgment on image structures).
Hereinafter, description will be given of the selection of the coefficient tables shown in
Initially, if the imaging sensitivity is ISO 200, the image processing unit 11 selects the coefficient table shown in
If the imaging sensitivity is ISO 800, the image processing unit 11 selects the coefficient table shown in
If the imaging sensitivity is ISO 3200, the image processing unit 11 selects the coefficient table shown in
The coefficient tables shown in
Initially, if the imaging sensitivity is ISO 200, the coefficient table shown in
If the imaging sensitivity is ISO 800, the coefficient table shown in
If the imaging sensitivity is ISO 3200, the coefficient table shown in
Subsequently, description will be given of typical rules for changing the weighting coefficients here.
Initially, the lower the imaging sensitivity is, i.e., the lower noise the RAW data has, the greater the image processing unit 11 makes the weighting coefficient of the central G pixel relatively and/or the smaller it makes the size of the coefficient table. Such a change of the coefficient table can soften the smoothing.
On the contrary, the higher the imaging sensitivity is, i.e., the higher noise the RAW data has, the smaller the image processing unit 11 makes the weighting coefficient of the central G pixel relatively and/or the greater it makes the size of the coefficient table. Such a change of the coefficient table can intensify the smoothing.
Step S64: The image processing unit 11 adds the values of the surrounding G pixels to that of the G pixel [i,j] to be processed with weights in accordance with the weighting coefficients on the coefficient table selected. The image processing unit 11 uses the value of the G pixel [i,j] after the weighted addition as the G color component of the pixel [i,j] on a color interpolated plane.
After this processing, the image processing unit 11 moves the operation to step S65.
Step S65: The image processing unit 11 repeats the foregoing adaptive smoothing processing (steps S61 to S64) on G pixels of the RAW data.
If the image processing unit 11 completes this adaptive smoothing process on all the G pixels of the RAW data, it moves the operation to step S66.
Step S66: Subsequently, the image processing unit 11 performs interpolation on the R and B positions of the RAW data (vacant positions on the lattice of G color components), thereby generating interpolated G color components. For example, interpolation in consideration of the indices (HV,DN) as described below is performed here. “Z” in the equations generically represents either of the color components R and B.
Step S67: Subsequently, the image processing unit 11 performs interpolation on R color components. For example, pixels [i+1,j], [i,j+1], and [i+1,j+1] other than in R positions [i,j] are subjected to respective interpolations as follows:
R[i+1,j]=(R[i,j]+R[i+2,j])/2+(2·G[i+1,j]−G[i,j]−G[i+2,j])/2;
R[i,j+1]=(R[i,j]+R[i,j+2])/2+(2·G[i,j+1]−G[i,j]−G[i,j+2])/2; and
R[i+1,j+1]=(R[i,j]+R[i+2,j]+R[i,j+2]+R[i+2,j+2])/4
+(4·G[i+1,j+1]−G[i,j]−G[i+2,j]−G[i,j+2]−G[i+2,j+2])/4.
Step S68: Subsequently, the image processing unit 11 performs interpolation on B color components. For example, pixels [i+1,j], [i,j+1], and [i+1,j+1] other than in B positions [i,j] are subjected to respective interpolation processes as follows:
B[i+1,j]=(B[i,j]+B[i+2,j])/2+(2·G[i+1,j]−G[i,j]−G[i+2,j])/2;
B[i,j+1]=(B[i,j]+B[i,j+2])/2+(2·G[i,j+1]−G[i,j]−G[i,j+2])/2; and
B[i+1,j+1]=(B[i,j]+B[i+2,j]+B[i,j+2]+B[i+2,j+2])/4
+(4·G[i+1,j+1]−G[i,j]−G[i+2,j]−G[i,j+2]−G[i+2,j+2])/4.
By the series of processes described above, RGB color interpolation is completed.
The electronic camera (including an image processing apparatus) according to a third embodiment performs color interpolation on RGB Bayer-array RAW data (corresponding to the first image), thereby generating image data that has RGB signal components arranged on each pixel (corresponding to the second image).
The configuration of the electronic camera (
For example, the image processing unit 111 determines the similarity degrees Cv, Ch, C45, and C135 of this G pixel [i,j]. Since this processing is the same as in the first embodiment, description thereof will be omitted.
Next, the image processing unit 11 judges if all the similarity degrees Cv, Ch, C45, and C135 determined are lower than or equal to predetermined thresholds, based on the following conditional expression:
(Cv≦Thv) AND (Ch≦Thh) AND (C45≦Th45) AND (C135≦Th135).
The thresholds in the expression are values for judging if the similarity degrees show significant changes in pixel value. It is thus preferable that the higher the imaging sensitivity is, the higher the thresholds are made in consideration of increasing noise.
If this conditional expression is satisfied, the location is judged as being flat in the horizontal, vertical, and diagonal directions. In this case, the image processing unit 11 moves the operation to step S73.
On the other hand, if this conditional expression is not satisfied, the image structure has a significant directionality. In this case, the image processing unit 11 moves the operation to step S72.
Steps S72 to S78: The same as steps S62 to S68 of the second embodiment. Description thereof will thus be omitted.
By the series of processes described above, RGB color interpolation is completed.
At step S43 of the foregoing first embodiment, if it is judged that the similarities in a plurality of directions are isotropic and the similarities are low, then a type of coefficient tables having a higher level of noise removal may be selected. In this case, it is possible to consider locations of low similarity as being noise and remove them powerfully, while performing color system conversion. In such an operation, relief information on isotropic locations (locations that are obviously non-edges) can be removed powerfully as isolated noise points. That is, it becomes possible to remove grains of noise, mosaics of color noise, and the like appropriately without losing the image structures of the edge areas.
Moreover, at step S48 of the foregoing first embodiment, if it is judged that the difference in the magnitude of similarity between directions is small, coefficient tables of detail enhancement type for enhancing high frequency components of signal components may be selected. In this case, it is possible to enhance fine image structures that no directionality, while performing color system conversion.
In one of the foregoing embodiments, description has been given of the color system conversion into a luminance component. However, the present invention is not limited thereto. For example, the present invention may be applied to color system conversion into chrominance components. In this case, it becomes possible to perform spatial filtering (LPF processing in particular) in consideration of image structures, simultaneously with the generation of chrominance components. The occurrence of color artifacts ascribable to chrominance noise can thus be suppressed favorably.
Moreover, in one of the foregoing embodiments, description has been given of the case where the present invention is applied to color system conversion. However, the present invention is not limited thereto. For example, the coefficient tables for color system conversion may be replaced with coefficient tables for color interpolation, so that color interpolation and sophisticated spatial filtering in consideration of image structures can be performed at the same time.
More specifically, while the second embodiment has only dealt with the case of performing color interpolation and low-pass processing simultaneously, edge enhancement processing may also be included as in the first embodiment.
Moreover, the foregoing embodiments have dealt with the cases where the present invention is applied to the electronic camera 1. However, the present invention is not limited thereto. For example, an image processing program may be used to make the external computer 18 execute the operations shown in FIGS. 2 to 7.
Moreover, image processing services according to the present invention may be provided over communication lines such as the Internet.
Furthermore, the image processing function of the present invention may be added to electronic cameras afterwards by rewriting the firmware of the electronic cameras.
The invention is not limited to the above embodiments and various modifications may be made without departing from the spirit and scope of the invention. Any improvement may be made in part or all of the components.
Number | Date | Country | Kind |
---|---|---|---|
2003-186629 | Jun 2003 | JP | national |
This application is a continuation application of International Application PCT/JP 2004/009601, filed Jun. 30, 2004, designating the U.S., and claims the benefit of priority from Japanese Patent Application No. 2003-186629, filed on Jun. 30, 2003, the entire contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP04/09601 | Jun 2004 | US |
Child | 11320460 | Dec 2005 | US |