The present invention relates to a digital still camera, a camera mounted in a mobile phone, a camera mounted in a personal digital assistant, an image inspection system, an industrial camera for automatic control, or another imaging apparatus using an imaging element and provided with an optical system and to an image processing method.
In recent years, rapid advances have been made in digitalization of information. This has led to remarkable efforts to meet with this in the imaging field.
In particular, as symbolized by digital cameras, imaging surfaces are changing from the conventional film to solid-state imaging elements such as CCDs (charge coupled devices) or CMOS (complementary metal oxide semiconductor) sensors in the majority of cases.
An imaging lens device using a CCD or CMOS sensor for the imaging element in this way optically captures the image of an object by the optical system and extracts the image as an electric signal by the imaging element. Other than a digital still camera, this is used in a video camera, a digital video unit, a personal computer, a mobile phone, a personal digital assistant (PDA), an image inspection system, an industrial camera for automatic control, and so on.
This imaging lens device 1 has an optical system 2 and a CCD or CMOS sensor or other imaging element 3.
The optical system includes object side lenses 21 and 22, a stop 23, and an imaging lens 24 sequentially arranged from the object side (OBJS) toward the imaging element 3 side.
In the imaging lens device 1, as shown in
Further, imaging devices using phase plates (wavefront coding optical elements) to regularly diffuse the light beams, using digital processing to restore the image, and thereby enabling capture of an image having a deep depth of field and so on have been proposed (see for example Non-patent Documents 1 and 2 and Patent Documents 1 to 5).
Further, when capturing an image by a camera, for example, the imaging technique of setting the stop to the open side and focusing on an object while making the depth of the object shallow so as to intentionally blur parts other than the main object is known.
Further, to obtain an image blurred only at the background without being constrained by the distance relationship between the object and the background, the imaging technique of capturing the image at a plurality of focus positions and combining the images is known.
Further, an automatic exposure control system of a digital camera performing filter processing using a transfer function has been proposed (see for example Patent Document 6).
Non-patent Document 1: “Wavefront Coding; jointly optimized optical and digital imaging systems”, Edward R. Dowski, Jr., Robert H. Cormack, Scott D. Sarama.
Non-patent Document 2: “Wavefront Coding; A modern method of achieving high performance and/or low cost imaging systems”, Edward R. Dowski, Jr., Gregory E. Johnson.
Patent Document 1: U.S. Pat. No. 6,021,005
Patent Document 2: U.S. Pat. No. 6,642,504
Patent Document 3: U.S. Pat. No. 6,525,302
Patent Document 4: U.S. Pat. No. 6,069,738
Patent Document 5: Japanese Patent Publication (A) No. 2003-235794
Patent Document 6: Japanese Patent Publication (A) No. 2004-153497
All of the imaging apparatuses proposed in the documents explained above are predicated on a PSF (Point Spread Function) being constant when inserting the above phase plate in the usual optical system. If the PSF changes, it is extremely difficult to realize an image having a deep depth of field by convolution using the subsequent kernels.
Accordingly, leaving aside lenses with single focal points, in lenses of the zoom system, AF system, etc., the high level of precision of the optical design and the accompanying increase in costs cause a major problem in their use.
In other words, in a conventional imaging apparatuses, suitable convolution processing is not possible. An optical design eliminating astigmatism, coma aberration, zoom chromatic aberration, and other aberration causing deviation of the spot image at the time of the “wide” mode and at the time of the “tele” mode is required.
However, an optical design eliminating these aberrations increases the difficulty of the optical design and induces problems such as an increase of the amount of design work, an increase of the costs, and an increase in size of the lenses.
Further, in the imaging technique of capturing images at a plurality of focus positions and combining these in order to obtain an image blurred in only the background explained before, since the focus position is changed and the image captured a plurality number of times, there is the problem that a long time is taken until all of the images finish being captured. Further, in this imaging technique, there is the problem that the main object and an object located in the background move and change during the plurality of imaging operations and therefore the combined image ends up becoming unnatural.
Further, in the apparatuses disclosed in the documents explained above, in for example capturing an image in a dark place, when restoring the image by signal processing, noise is simultaneously amplified as well.
Accordingly, in an optical system including an optical system and signal processing for example using the phase plate or other optical wavefront modulation element as explained above and the signal processing after that, there is the disadvantage that noise is amplified when capturing an image in a dark place and ends up having an influence upon the restored image.
An object of the present invention is to provide an imaging apparatus and an image processing method able to simplify the optical system, able to reduce the costs, able to obtain an image blurred only in the background or a focused image from a single imaging operation, and able to obtain a restored image with little influence of noise.
An imaging apparatus according to a first aspect of the present invention is provided with an imaging element capturing a diffused image of an object passed through at least an optical system and an optical wavefront modulation element, a signal processing portion including a converting means for generating a diffusion-free image signal from a diffused image signal from the imaging element and performing predetermined processing on the image signal from the imaging element, and a generating means for combining the image before the processing of the signal processing portion and the image after the processing to form a new image.
Preferably, the generating means generates a plurality of images by blurred image processing for a background region and combines a focused image in an object region including a main object after the processing to generate a new image.
Preferably, the apparatus is further provided with a recording portion recording an image before processing by the signal processing portion, an image after the processing, and a combined new image.
Preferably, the apparatus is further provided with a recording portion recording a blurred image before the processing by the signal processing portion, a focused image after the processing, and/or a new image obtained by combining the blurred image after the processing and the focused image, a display portion displaying the image recorded in the recording portion or an image for recording, and an operation portion setting a range in the display portion and/or selecting the blurred image, and the generating means generates a focused image in the set range or out of the set range in the display portion by the operation portion combines this with the blurred image to generate a new image and/or combines one or more of the blurred images selected by the operation portion with the focused image to generate a new image.
Preferably, the apparatus is further provided with a recording portion recording a blurred image before processing by the signal processing portion, a focused image after the processing, or an intermediate image after the processing, and/or a new image obtained by combining the blurred image, focused image, or intermediate image, a display portion displaying an image recorded in the recording portion or an image for recording, and an operation portion setting a range in the display portion and/or selecting a blurred image, and the generating means generates a focused image in the set range or out of the set range in the display portion by the operation portion, combines a range other than for generation of the focused image with the blurred image or intermediate image to generate a new image and/or combines one or more of the blurred images selected by the operation portion or the intermediate image with the focused image to generate a new image.
Preferably, the optical system includes a zoom optical system and has a zoom information generating means for generating information corresponding to a zoom position or zoom amount of the zoom optical system, and the converting means generates a diffusion-free image signal from the diffused image signal based on the information generated by the zoom information generating means.
Preferably, the apparatus includes an object distance information generating means for generating information corresponding to a distance up to the object, and the converting means generates a diffusion-free image signal from the diffused image signal based on the information generated by the object distance information generating means.
Preferably, the apparatus includes an object distance information generating means for generating information corresponding to the distance up to the object and a conversion coefficient operation means for performing operation to obtain a conversion coefficient based on the information generated by the object distance information generating means, and the converting means converts the image signal according to the conversion coefficient obtained from the conversion coefficient operation means and generates a diffusion-free image signal.
Preferably, the apparatus includes an imaging mode setting means for setting the imaging mode of the object to be photographed, and the converting means performs different conversion processing in accordance with the imaging mode set by the imaging mode setting means.
Preferably, the imaging apparatus can be switched between a plurality of lenses, the imaging element can capture an object aberration image passed through at least one lens of the plurality of lenses and the optical wavefront modulation element and further includes a conversion coefficient acquiring means for acquiring a conversion coefficient in accordance with the above one lens, and the converting means converts the image signal according to the conversion coefficient obtained from the conversion coefficient acquiring means.
Preferably, the apparatus includes an exposure controlling means for controlling the exposure, and the signal processing portion performs filter processing with respect to an optical transfer function (OTF) in accordance with the exposure information from the exposure controlling means.
An image processing method according to a second aspect of the present invention has a first step of capturing a diffused image of an object passed through at least an optical system and an optical wavefront modulation element, a second step of performing predetermined signal processing on the diffused image signal obtained at the first step and generating a diffusion-free image signal from the diffused image signal, and a third step of combining the image before the processing at the second step and the image after the processing to form a new image.
Preferably, the third step includes a fourth step of recording a blurred image before the processing according to the second step, a focused image after the processing, and/or a new image obtained by combining the blurred image after the processing and the focused image, a fifth step of displaying the image recorded at the fourth step or the image for recording in a display portion, and a sixth step of setting a range in the display portion and/or selecting a blurred image, and the third step generates a focused image in the set range or out of the set range in the display portion according to the sixth step and combines it with a blurred image to generate a new image and/or combines one or more blurred images selected according to the sixth step and the focused image to generate a new image.
Preferably, the third step includes a fourth step of recording a blurred image before the processing according to the second step, a focused image after the processing, or an intermediate image after the processing and/or a new image obtained by combining the blurred image, focused image, or intermediate image, a fifth step of displaying the image recorded at the fourth step or the image for recording in a display portion, and a sixth step of setting a range in the display portion and/or selecting a blurred image, and the third step generates a focused image in the set range or out of the set range in the display portion according to the sixth step and combines a range other than for generation of the focused image with the blurred image or intermediate image to generate a new image and/or combines one or more blurred images selected by the operation portion or the intermediate image with the focused image to generate a new image.
According to the present invention, there are the advantages that the optical system can be simplified, the costs can be reduced, and in addition an image blurred only in the desired region or a restored image (that is, a focused image) having little influence of noise and further a combined image of those can be obtained by a single imaging operation.
100, 100A . . . imaging apparatuses, 110 . . . optical system, 110A . . . optical unit, 120 . . . imaging element, 130 . . . analog front end portion (AFE), 140 . . . image processing device, 150 . . . camera signal processing portion, 180 . . . operation portion, 190 . . . exposure control device, 200 . . . system control device, 201 . . . optical system switch control portion, 111 . . . object side lens, 112 . . . focus lens, 113 . . . wavefront forming optical element, 113a . . . phase plate (optical wavefront modulation element), 142 . . . convolution processor, 143 . . . kernel data ROM, and 144 . . . convolution control portion.
Below, embodiments of the present invention will be explained with reference to the accompanying drawings.
An imaging apparatus 100 according to the present embodiment has an optical system 110, imaging element 120, analog front end portion (AFE) 130, image processing device 140, camera signal processing portion 150, image display memory 160, image monitoring device 170, operation portion 160, and exposure control device 190.
The optical system 110 supplies an image obtained by capturing an image of an object OBJ to the imaging element 120.
The optical system 110 of the present embodiment includes an optical wavefront modulation element as will be explained in detail later.
The imaging element 120 is formed by a CCD or CMOS sensor at which the image captured at the optical system 110 including the optical wavefront modulation element is focused and which outputs focused first-order image information as a first-order image signal FIM of an electric signal to the image processing device 140 via the analog front end portion 130.
In
The analog front end portion (hereinafter referred to as an “AFE”) 130 has a timing generator 131 and an analog/digital (A/D) converter 132.
The timing generator 131 generates a drive timing of the CCD of the imaging element 120, while the A/D converter 132 converts an analog signal input from the CCD to a digital signal and outputs the same to the image processing device 140.
The image processing device (two-dimensional convolution means) 140 forming a portion of the signal processing portion receives as input the digital signal of the captured image coming from the AFE 130 in a front stage, applies two-dimensional convolution processing to this, and transfers the same to the camera signal processing portion (DSP) 150 in a latter stage.
The image processing device 140 performs filter processing on the optical transfer function (OTF) in accordance with the exposure information of the exposure control device 190.
The image processing device 140 has a function of generating a diffusion-free image signal from a diffused image signal of the object from the imaging element 120. Further, the signal processing portion has a function of applying noise reduction filtering in the first step.
The processing of the image processing device 140 will be explained in further detail later.
The camera signal processing portion (DSP) 150 performs color interpolation, white balancing, YCbCr conversion processing, compression, filtering, and other processing and performs storage of data into the memory 160, an image display in the image monitoring device 170, and so on.
The exposure control device 190 performs the exposure control and, at the same time, waits for operation inputs of the operation portion 180 etc., determines the operation of the system as a whole in accordance with those inputs, controls the AFE 130, image processing device 140, DSP 150, etc., and conducts mediation control of the system as a whole.
The imaging apparatus 100 of the present embodiment has a plurality of imaging modes, for example, a macro imaging mode (proximate) and a distant view imaging mode (infinitely distant) other than the portrait mode and is configured so that these imaging modes can be selected and input by the operation portion 180.
The operation portion 180 is, for example as shown in
Note that the portrait mode is one of imaging modes set in accordance with the object at the time of the normal imaging and is an imaging mode suitable for capturing the image of a person. It makes the image of the background a blurred image by focusing on a person at the center. As other settable modes, there are a sports mode, sunset mode, night view mode, black-and-white mode, sepia mode, and so on.
Each mode can be selected and set by the MENU button 1801 and cross key 1803. In the present embodiment, the apparatus is configured so that a horizontal imaging use portrait and vertical imaging use portrait can be selected as the portrait mode. Note that the modes may be switched by a touch panel method on the liquid crystal screen 1701.
The imaging apparatus 100 in the present embodiment has the following function for making the portrait imaging easier.
Namely, the signal processing portion formed by the image processing device 140, DSP 150, and exposure control device 190 has a generation function of performing predetermined signal processing with respect to the diffused image signal, for example, generation of a diffusion-free image signal from a diffused image signal of the object from the imaging element 120, and combining the image before the processing of this signal processing portion and the image after the processing to form a new image.
This generation function generates a plurality of images by blurred image processing in the background region and combines a focused image of an object region including a main object after the processing to generate a new image.
Further, provision is made of a recording function of recording images before the signal processing in the signal processing portions (image processing device, DSP) 140, 150, etc., restored images after the processing, and combined new images in for example a not shown memory buffer or image display memory 160.
Since this recording function is provided, the present imaging apparatus 100 has the effect that portrait imaging by providing a generation function can be easily carried out. In addition, it can give the following effect.
Namely, by recording the image before the signal processing and the image after the signal processing, it is possible to select the position and size of an area desired to be made clear (conversely, an area desired to be made blurred) after the imaging and recording and prepare a new image.
For this reason, a portrait captured image can be prepared from an image which was captured and recorded in a mode other than the portrait mode at the time of imaging.
The signal processing portion of the imaging apparatus 100 having such a function extracts a focused image of the object region including the main object from the image after the image restoration processing and extracts an unfocused image of the background region contacting the object region from the image before the image restoration processing. It combines these extracted focused image of the object region and unfocused image of the background region to thereby generate a new image. Then, it records the generated image.
Further, in the present embodiment, the operation portion 180 functions as a designation portion for making the user designate the object region as well.
Below, first to third examples of processing for preparation of a portrait image according to the present embodiment will be explained with reference to
Further,
The analog signal obtained by the imaging element 120 is digitalized at the AFE 130, is digitally processed in the image processing portion 140, becomes the Y, Cb, and Cr signals at the DSP 150, and is displayed as a through image in the image monitoring device 170 serving as the display portion.
When the operation portion 180 is used to select the vertical imaging portrait mode or horizontal imaging portrait mode, as shown in
Then, as shown in
Note that whether to set vertical imaging or horizontal imaging may be automatically detected by using an angular velocity sensor.
If explaining this processing operation with reference to
Then, when the user depresses a shutter key during the preview image display (ST2), the image is recorded in the RAM of the buffer (ST3), the image is restored for only the center region set in advance (ST4), and the recording processing is carried out (ST5).
In this case, after completion of imaging, in the preview image, as shown in
If explaining this processing operation with reference to
Then, when the user depresses the shutter key during the display of the preview image (ST12), the image is recorded in the RAM of the buffer (ST13), the preview image is displayed, and the user selects the object by the operation portion 180 (ST14).
Then, processing is performed for restoring the image of the selected region portion (ST15), and processing is preformed for recording the restored image (ST16).
For the purpose of blurring the background image more, as shown in
As the extent of blurring at this time, the operation portion 180 of the first example is operated to select the vertical imaging portrait mode or horizontal imaging portrait mode. At this time, the user is made select this extent of blurring and perform more filter processing for blurring (image processing) the farther from the frame at the center portion.
In this example, the filters are formed so that the degree of blurring becomes stronger in the filter FLT2 than the filter FLT3.
Note that, as the blurring filters FLT2 and FLT3, general smoothing filters may be used as well.
In this way, the imaging apparatus 100 of the present embodiment can easily perform portrait imaging. By recording the image before the signal processing and the image after the signal processing, it is possible to select the position and size of an area desired to be made clear (conversely, an area desired to be blurred) after the imaging and recording to prepare a new image. For this reason, this apparatus has the advantage that a portrait captured image can be prepared from an image captured and recorded in a mode other than the portrait mode at the time of imaging.
Here, a specific example of the feature of the present invention, that is, generating a focused image in the set range or out of the set range on the liquid crystal screen 1701 by the operation portion 180, combining this with a blurred image, and thereby generating a new image will be explained.
The left side of
The present invention is characterized in that the size and position of the range of the focused image can be freely changed by the operation portion 180. The left side of
Further, the left side of
Here, a concrete method for designating the range will be explained with reference to the drawings.
For example, a cursor (cross mark) on the liquid crystal screen 1701 may be moved by the cross key 1803 to designate the center and radius to determine a circular shape, three points may be designated to determine a circular shape, or the center and two radii may be designated to determine an elliptical shape as shown in
Further, the corners of the shape may be designated to determine a polygonal shape. Further, in a case of dividing the screen into two, it is possible to designate two points to determine division by 2. For example, as shown in
As shown in
When designation by points is selected, as shown in
Note that, in the present embodiment, a case where the range (position and size) of the blurred image or focused image was designated and the blurred image and focused image were combined to generate and record a new image was explained. As another embodiment, an extent of blurring of the blurred image may be made selectable by the selection of the kernel data explained later by the operation portion 180 or the selection of any of a plurality of filters shown in
Further, in the present embodiment, as another embodiment, it is possible to suitably blur everything other than the blurred image and focused image to generate a suitably focused intermediate image. In the present invention, the intermediate image means an image which is not more focused than the focused image, but not more blurred than the blurred image. This can be generated by performing processing which is the same as the processing for generating the focused image, but does not generate a perfect focused image, for example processing by a coefficient different from the coefficient for obtaining the focused image. This other embodiment of the present invention is characterized by combining the intermediate image after the signal processing and the focused image to form a new image.
According to this generation function, it becomes possible to generate an intermediate image in the background region, generate a focused image in the object region including the main object, and combine these images to generate a new image. When combining a blurred image and focused image, a big difference occurs in the image quality in the vicinity of the combined portions and there is a possibility of unnatural blurriness. However, as explained above, by employing an intermediate image in place of the blurred image, the difference of image quality in the vicinity of the combined portions is reduced and it becomes possible to exhibit a more natural blurriness. Due to this, even in a case where a blurred image is replaced by an intermediate image in the present embodiment, the effects of the present invention can be obtained.
Further, in the present embodiment, the case of single signal processing portions 140, 150, and 190 was explained. However, when an intermediate image is generated, two of each of the signal processing portions 140, 150, 190, etc. may be provided as well. By providing two, one can be used as the signal processing portion for generating the focused image, and the other can be used as the signal processing portion for generating the intermediate image. The processing speed can be raised since the generation of the focused image and the generation of the intermediate image can be performed simultaneously.
The imaging apparatus 100 of the present embodiment has characterizing configurations in the optical system and image processing device as will be explained below so that a person etc. can be made more distinct without being influenced by camera shake etc.
Below, the configurations and functions of the optical system and image processing device of the present embodiment will be explained concretely.
Further,
Further,
The zoom optical system 110 of
Note that, in the present embodiment, a case where a phase plate was used was explained, but the optical wavefront modulation elements of the present invention may include any elements so far as they deform the wavefront. They may include optical elements changing in thickness (for example, the above-explained third-order phase plate), optical elements changing in refractive index (for example, a refractive index distribution type wavefront modulation lens), optical elements changing in thickness and refractive index by the coding on the lens surface (for example, a wavefront coding hybrid lens), liquid crystal elements able to modulate the phase distribution of the light (for example, liquid crystal spatial phase modulation elements), and other optical wavefront modulation elements.
The zoom optical system 110 of
The phase plate 113a shown in the figure is an optical lens regularly diffusing the light beams converged by the optical system. By inserting this phase plate, an image not focused anywhere on the imaging element 120 is realized.
In other words, the phase plate 113a forms light beams having a deep depth (playing a central role in the image formation) and flare (blurred portion).
A means for restoring this regularly diffused image to a focused image by digital processing will be referred to as a wavefront aberration control optical system. This processing is carried out in the image processing device 140.
Here, the basic principle of the wavefront aberration control optical system will be explained.
As shown in
This is represented by the following equation.
g=H*f (Equation 1)
Note that, * represents convolution.
In order to find the object from the generated image, the next processing is required.
f=H
−1
*g (Equation 2)
Here, the kernel size and operational coefficients concerning H will be explained.
Assume that the zoom positions are Zpn, Zpn−1, . . . . Further, assume that the individual H functions are Hn, Hn−1, . . . .
The spots are different, therefore the H functions become as follows.
The difference of the number of rows and/or the number of columns of this matrix is referred to as the “kernel size”. The numbers are the operational coefficients.
Here, each H function may be stored in the memory.
By using the PSF as a function of the object distance, using the object distance for calculation, and calculating the H function, it is also possible to set the system so as to create the optimum filter for any object distance. Further, it is also possible to use the H function as a function of the object distance and directly find the H function by the object distance.
In the present embodiment, as shown in
Note that, in the present embodiment, “diffusion” means the phenomenon where as explained above, inserting the phase plate 113a causes the formation of an image not focused anywhere on the imaging element 120 and the formation of light beams having a deep depth (playing a central role in the image formation) and flare (blurred portion) by the phase plate 113a and includes the same meaning as aberration because of the behavior of the image being diffused and forming a blurred portion. Accordingly, in the present embodiment, there also exists a case where diffusion is explained as aberration.
Next, the configuration and processing of the image processing device 140 will be explained.
The image processing device 140, as shown in
The convolution control portion 144 turns the convolution processing ON/OFF, controls the screen size, replaces kernel data, etc. and is controlled by the exposure control device 190.
Further, the kernel data storage ROM 143, as shown in
In the example of
First, the exposure information (RP) is detected and supplied to the convolution control portion 144 (ST21).
In the convolution control portion 144, the kernel size and numerical value operational coefficients are set in a register from the exposure information RP (ST22).
Then, the convolution operation is carried out on the image data captured at the imaging element 120 and input via the AFE 130 to the two-dimensional convolution processing portion 142 based on the data stored in the register. The processed and converted data is transferred to the camera signal processing portion 150 (ST23).
Below, a more specific example of the signal processing portion and kernel data storage ROM of the image processing device 140 will be explained.
The example of
The exposure information determined at the time of setting the exposure is acquired, and the kernel data is selected and controlled through the convolution control portion 144. In the two-dimensional convolution operation portion 142, the convolution processing is applied by using the kernel data.
The example of
The exposure information determined at the time of setting the exposure is acquired, and the kernel data is selected and controlled through the convolution control portion 144.
In the two-dimensional convolution operation portion 142, after applying the noise reduction filter ST31, the color space is converted by color conversion processing ST32, then convolution processing ST33 is applied by using the kernel data after that.
The noise processing ST34 is carried out again, and the color space is returned to the original one by the color conversion processing ST35. As the color conversion processing, for example YCbCr conversion can be mentioned, but another conversion may be employed.
Note that, it is also possible to omit the second noise processing ST34.
The example of
The exposure information determined at the time of setting the exposure is acquired, and the kernel data is selected and controlled through the convolution control portion 144.
The two-dimensional convolution operation portion 142 applies convolution processing ST43 by using the OTF restoration filter after noise reduction processing ST41 and color conversion processing ST42.
The noise processing ST44 is carried out again, and the color space is returned to the original one by the color conversion processing ST45. As the color conversion processing, for example YCbCr conversion can be mentioned, but other conversion may also be employed.
Note that, either the noise recording processing ST41 or ST44 may be carried out as well.
The example of
Note that it is also possible to omit the second noise processing ST5.
The exposure information determined at the time of setting the exposure is acquired, and the kernel data is selected and controlled through the convolution control portion 144.
In the two-dimensional convolution operation portion 142, after applying the noise reduction filter ST51, the color space is converted by color conversion processing ST52, then convolution processing STS3 is applied by using the kernel data after that.
The noise processing ST54 in accordance with the exposure information is carried out again, and the color space is returned to the original one by the color conversion processing ST55. As the color conversion processing, for example, the YCbCr conversion can be mentioned, but other conversion may be employed.
Note that, it is also possible to omit the noise reduction processing ST51.
An explanation was given above of the example of performing the filter processing in the two-dimensional convolution operation portion 143 in accordance with only the exposure information. However, it becomes possible to perform extraction or operation of the suitable operational coefficient by combining for example the object distance information, zoom information, or imaging mode information with the exposure information.
Note that the imaging system in
The image processing device 300, as shown in
In this image processing device 300, the image processing computation processor 303 obtaining information concerning the approximate distance of the object distance of the object read out from the object approximate distance information detection device 400 and the exposure information stores the kernel size and its operational coefficients used in suitable operation with respect to the object distance position in the kernel and/or numerical value operational coefficient storage register 302 and performs suitable operation at the convolution device 301 by using those values for operation to restore the image.
As explained above, in the case of an imaging apparatus provided with a phase plate (wavefront coding optical element) as an optical wavefront modulation element, if within a predetermined focal distance range, a suitable aberration-free image signal can be generated by image processing concerning that range, but if out of the predetermined focal length range, there is a limit to the correction of the image processing, therefore only an object out of the above range ends up becoming an image signal with aberration.
Further, on the other hand, by applying image processing not causing aberration within a predetermined narrow range, it also becomes possible to give blurriness to an image out of the predetermined narrow range.
The present example is configured so as to detect the distance up to the main object by the object approximate distance information detection device 400 including the distance detection sensor and perform processing for image correction different in accordance with the detected distance.
The above image processing is carried out by a convolution operation. In order to accomplish this, for example, it is possible to employ a configuration commonly storing one type of operational coefficient of the convolution operation, storing in advance a correction coefficient in accordance with the focal length, correcting the operational coefficient by using this correction coefficient, and performing suitable convolution operation by the corrected operational coefficient.
Other than this configuration, it is possible to employ the following configurations.
It is possible to employ a configuration storing in advance the kernel size and the operational coefficient itself of the convolution in accordance with the focal length and performing convolution operation by these stored kernel size and operational coefficient, a configuration storing in advance the operational coefficient in accordance with a focal length as a function, finding the operational coefficient by this function according to the focal length, and performing the convolution operation by the calculated operational coefficient, and so on.
When linked with the configuration of
At least two conversion coefficients corresponding to the aberration due to at least the phase plate 113a are stored in advance in the register 302 serving as the conversion coefficient storing means in accordance with the object distance. The image processing computation processor 303 functions as the coefficient selecting means for selecting a conversion coefficient in accordance with the distance up to the object from the register 302 based on the information generated by the object approximate distance information detection device 400 as the object distance information generating means.
Then, the convolution device 301 serving as the converting means converts the image signal according to the conversion coefficient selected at the image processing computation processor 303 as the coefficient selecting means.
Alternatively, as explained above, the image processing computation processor 303 as the conversion coefficient processing means computes the conversion coefficient based on the information generated by the object approximate distance information detection device 400 as the object distance information generating means and stores the same in the register 302.
Then, the convolution device 301 serving as the converting means converts the image signal according to the conversion coefficient obtained by the image processing computation processor 303 serving as the conversion coefficient operation means and stored in the register 302.
Alternatively, at least one correction value in accordance with the zoom position or zoom amount of the zoom optical system 210 is stored in advance in the register 302 serving as the correction value storing means. This correction value includes the kernel size of the object aberration image.
The register 302, functioning also as the second conversion coefficient storing means, stores in advance the conversion coefficient corresponding to the aberration due to the phase plate 113a.
Then, based on the distance information generated by the object approximate distance information detection device 400 serving as the object distance information generating means, the image processing computation processor 303 serving as the correction value selecting means selects the correction value in accordance with the distance up to the object from the register 302 serving as the correction value storing means.
The convolution device 301 serving as the converting means converts the image signal based on the conversion coefficient obtained from the register 302 serving as the second conversion coefficient storing means and the correction value selected by the image processing computation processor 303 serving as the correction value selecting means.
The image processing device 300A, in the same way as
In this image processing device 300A, the image processing computation processor 303 obtaining information concerning the zoom position or zoom amount read out from the zoom information detection device 500 and exposure information stores the kernel size and its operational coefficients used in suitable operation with respect to the exposure information and its zoom position in the kernel and/or numerical value operational coefficient storage register 302 and performs suitable operation at the convolution device 301 by using those values for operation to restore the image.
As explained above, when applying a phase plate serving as the optical wavefront modulation element to an imaging apparatus provided in a zoom optical system, the generated spot image differs according to the zoom position of the zoom optical system. For this reason, when performing the convolution operation of a focal point deviated image (spot image) obtained by the phase plate in a later DSP etc., in order to obtain the suitable focused image, convolution operation differing in accordance with the zoom position becomes necessary.
Therefore, the present embodiment is configured provided with the zoom information detection device 500, performing a suitable convolution operation in accordance with the zoom position, and obtaining a suitable focused image without regard as to the zoom position.
For suitable convolution operation in the image processing device 300A, it is possible to employ a configuration commonly storing one type of operational coefficient of convolution in the register 302.
Other than this configuration, it is also possible to employ the following configurations.
It is possible to employ a configuration storing in advance a correction coefficient in the register 302 in accordance with each zoom position, correcting the operational coefficient by using this correction coefficient, and performing a suitable convolution operation by the corrected operational coefficient, a configuration storing in advance the kernel size and the operational coefficient per se of the convolution in the register 302 in accordance with each zoom position and performing the convolution operation by these stored kernel size and operational coefficient, a configuration storing in advance the operational coefficient in accordance with the zoom position as a function in the register 302, finding the operational coefficient by this function according to the zoom position, and performing the convolution operation by the computed operational coefficient, and so on.
When linking this with the configuration of
At least two conversion coefficients corresponding to aberrations caused by the phase plate 113a in accordance with the zoom position or zoom amount of the zoom optical system 210 are stored in advance in the register 302 serving as the conversion coefficient storing means. The image processing computation processor 303 functions as the coefficient selecting means for selecting the conversion coefficient in accordance with the zoom position or zoom amount of the zoom optical system 210 from the register 302 based on the information generated by the zoom information detection device 400 serving as the zoom information generating means.
Further, the convolution device 301 serving as the converting means converts the image signal according to the conversion coefficient selected at the image processing computation processor 303 serving as the coefficient selecting means.
Alternatively, as explained before, the image processing computation processor 303 serving as the conversion coefficient operation means processes the conversion coefficient based on the information generated by the zoom information detection device 500 serving as the zoom information generating means and stores the same in the register 302.
Then, the convolution device 301 serving as the converting means converts the image signal according to the conversion coefficient obtained in the image processing computation processor 303 serving as the conversion coefficient operation means and stored in the register 302.
Alternatively, at least one correction value in accordance with the zoom position or zoom amount of the zoom optical system 210 is stored in advance in the register 302 serving as the correction value storing means. This correction value includes the kernel size of the object aberration image.
The register 302 functioning also as the second conversion coefficient storing means stores in advance a conversion coefficient corresponding to the aberration due to the phase plate 113a.
Then, based on the zoom information generated by the zoom information detection device 500 serving as the zoom information generating means, the image processing computation processor 303 serving as the correction value selecting means selects the correction value in accordance with the zoom position or zoom amount from the register 302 serving as the correction value storing means.
The convolution device 301 serving as the converting means converts the image signal based on the conversion coefficient obtained from the register 302 serving as the second conversion coefficient storing means and the correction value selected by the image processing computation processor 303 serving as the correction value selecting means.
In this example, two-dimensional information is formed by the object distance information and zoom information, and the exposure information forms information like the depth.
The image processing device 300B, in the same way as
In this image processing device 300B, the image processing computation processor 303 obtaining information concerning the approximate distance of the object distance of the object read out from the object approximate distance information detection device 400 and the exposure information stores the kernel size and its operational coefficients used in suitable operation with respect to the object distance position in the kernel and/or numerical value operational coefficient storage register 302 and performs suitable operation at the convolution device 301 by using those values for operation to restore the image.
Also in this case, as explained above, in the case of an imaging apparatus provided with a phase plate (wavefront coding optical element) serving as an optical wavefront modulation element, if within a predetermined focal distance range, a suitable aberration-free image signal can be generated by image processing concerning that range, but if out of the predetermined focal length range, there is a limit to the correction of the image processing, therefore only an object out of the above range ends up becoming an image signal with aberration.
Further, on the other hand, by applying image processing not causing aberration within a predetermined narrow range, it also becomes possible to give blurriness to an image out of the predetermined narrow range.
The present example is configured so as to detect the distance up to the main object by the object approximate distance information detection device 400 including the distance detection sensor and perform processing for image correction different in accordance with the detected distance.
The above image processing is carried out by a convolution operation. In order to accomplish this, it is possible to employ a configuration commonly storing one type of operational coefficient of a convolution operation, storing in advance a correction coefficient in accordance with the object distance, correcting the operational coefficient by using this correction coefficient, and performing the suitable convolution operation with the corrected operational coefficient, a configuration storing in advance an operational coefficient in accordance with the object distance as a function, finding the operational coefficient by this function according to the focal length, and performing the convolution operation with the computed operational coefficient, and a configuration storing in advance the kernel size and the operational coefficient per se of convolution and performing the convolution operation by these stored kernel size and operational coefficient in accordance with the focal length, and so on.
In the present embodiment, as explained above, the image processing is changed in accordance with the mode setting of the DSC (portrait, infinitely distant (scene), and macro).
When linking this with the configuration of
As explained before, a conversion coefficient differing in accordance with each imaging mode set by the imaging mode setup portion 700 of the operation portion 180 through the image processing computation processor 303 serving as the conversion coefficient processing means is stored in the register 302 serving as the conversion coefficient storing means.
The image processing computation processor 303 extracts the conversion coefficient from the register 302 serving as the conversion coefficient storing means based on the information generated by the object approximate distance information detection device 400 serving as the object distance information generating means in accordance with the imaging mode set by the operation switches 701 of the imaging mode setting portion 700. At this time, for example the image processing computation processor 303 functions as a conversion coefficient extracting means.
Further, the convolution device 301 serving as the converting means performs the conversion processing in accordance with the imaging mode of the image signal according to the conversion coefficient stored in the register 302.
Note that, the optical system in
Further, the kernel data storage ROM of
By employing three dimensions and further four or more dimensions as shown in
Note that, as explained above, in the case of an imaging apparatus provided with a phase plate (wavefront coding optical element) serving as an optical wavefront modulation element, if within a predetermined focal distance range, a suitable aberration-free image signal can be generated by image processing concerning that range, but if out of the predetermined focal length range, there is a limit to the correction of the image processing, therefore only an object out of the above range ends up becoming an image signal with aberration.
Further, on the other hand, by applying image processing not causing aberration within a predetermined narrow range, it also becomes possible to give blurriness to an image out of the predetermined narrow range.
In the present embodiment, the wavefront aberration control optical system is employed so it is possible to obtain a high definition image quality. In addition, the optical system can be simplified, and the cost can be reduced.
Below, these characteristic features will be explained.
As seen also from
In this way, the first-order image FIM formed in the imaging apparatus 100 of the present embodiment is given light beam conditions of extremely deep depth.
In the present embodiment, the high definition final image is left to the correction processing of the latter stage image processing device 140 configured by, for example, a digital signal processor. Therefore, as shown in
The image processing device 140, as explained above, receives the first-order image FIM by the imaging element 120, applies predetermined correction processing etc. for boosting the MTF at the spatial frequency of the first-order image, and forms a high definition final image FNLIM.
The MTF correction processing of the image processing device 140 performs correction so that, for example as indicated by a curve A of
The characteristic indicated by the curve B in
Note that all corrections in the present embodiment are according to the parameter of the spatial frequency.
In the present embodiment, as shown in
For example, in the case of the MTF characteristic of
Namely, by performing the correction by weakening the edge enhancement on the low frequency side and high frequency side within a predetermined bandwidth of the spatial frequency and strengthening the edge enhancement in an intermediate frequency zone, the desired MTF characteristic curve B is virtually realized.
In this way, the imaging apparatus 100 according to the embodiment is an image forming system basically configured by the optical system 110 and imaging element 120 forming the first-order image and by the image processing device 140 forming the first-order image to the high definition final image, wherein the optical system is newly provided with a wavefront forming optical element or is provided with a glass, plastic, or other optical element with a surface shaped for wavefront forming use so as to deform (modulate) the wavefront of the image formed, such a wavefront is focused onto the imaging surface (light receiving surface) of the imaging element 120 formed by a CCD or CMOS sensor, and the focused first-order image is passed through the image processing device 140 to obtain the high definition image.
In the present embodiment, the first-order image from the imaging element 120 is given light beam conditions with very deep depth. For this reason, the MTF of the first-order image inherently becomes a low value, and the MTF thereof is corrected by the image processing device 140.
Here, the process of image formation in the imaging apparatus 100 of the present embodiment will be considered in terms of wave optics.
A spherical wave scattered from one point of an object point becomes a converged wave after passing through the imaging optical system. At that time, when the imaging optical system is not an ideal optical system, aberration occurs. The wavefront becomes not spherical, but a complex shape. Geometric optics and wave optics are bridged by wavefront optics. This is convenient in the case where a wavefront phenomenon is handled.
When handling a wave optical MTF on an imaging plane, the wavefront information at an exit pupil position of the imaging optical system becomes important.
The MTF is calculated by a Fourier transform of the wave optical intensity distribution at the imaging point. The wave optical intensity distribution is obtained by squaring the wave optical amplitude distribution. That wave optical amplitude distribution is found from a Fourier transform of a pupil function at the exit pupil.
Further, the pupil function is the wavefront information (wavefront aberration) at the exit pupil position, therefore if the wavefront aberration can be strictly calculated as a numerical value through the optical system 110, the MTF can be calculated.
Accordingly, if modifying the wavefront information at the exit pupil position by a predetermined technique, the MTF value on the imaging plane can be freely changed.
In the present embodiment as well, the shape of the wavefront is mainly changed by a wavefront forming optical element. It is truly the phase (length of light path along the beams) that is adjusted to form the desired wavefront.
Then, when forming the target wavefront, the light beams from the exit pupil are formed by a dense beam portion and a sparse beam portion as seen from the geometric optical spot images shown in
The MTF of this state of light beams exhibits a low value at a position where the spatial frequency is low and somehow maintains the resolution up to the position where the spatial frequency is high.
Namely, if this low MTF value (or, geometric optically, the state of the spot image), the phenomenon of aliasing will not be caused.
That is, a low pass filter is not necessary.
Further, the flare-like image causing a drop in the MTF value may be eliminated by the image processing device 140 configured by the later stage DSP etc. Due to this, the MTF value is remarkably improved.
Next, responses of MTF of the present embodiment and conventional optical system will be considered.
Further,
As seen from the figures as well, in the case of the optical system having the optical wavefront modulation element, even in the case where the object is out of the focal point position, the change of the response of MTF becomes smaller than that of the optical system without an optical wavefront modulation element inserted.
By the processing by the convolution filter of the image formed by this optical system, the response of MTF is improved.
As explained above, according to the present embodiment, the signal processing portion configured by the image processing device 140, DSP 150, and exposure control device 190 performs the predetermined signal processing with respect to the diffused image signal, for example, generation of a diffusion-free image signal from the diffused image of the object from the imaging element 120. This has a generation function of combining the image before the processing of this signal processing portion and the image after the processing to combine a new image. This generation function generates a plurality of images in the background region by blurred image processing, combines a focused image of the object region including the main object after the above processing to generate a new image, and provides a recording function of recording the image before the signal processing, the restored image after the processing, and combined new image in for example a not shown memory buffer or image display memory 160. Therefore the following effects can be obtained.
There are the advantages that portrait imaging can be easily carried out, it is possible to record the image before the signal processing and the image after the signal processing and thereby select the position or size of an area desired to be made clear (conversely, an area desired to be blurred) after the imaging and recording to prepare a new image, and it is possible to prepare a portrait captured image from an image captured in a mode other than the portrait mode at the time of imaging.
Further, it includes the optical system 110 and imaging element 120 forming the first-order image and the image processing device 140 forming the first-order images into the high definition final image. The image processing device 140 performs filter processing with respect to the optical transfer function (OTF) in accordance with the exposure information from the exposure control device 190. Therefore, there are the advantages that the optical system can be simplified, the costs can be reduced, and in addition restored images with little influence of noise can be obtained.
Further, by making the kernel size used at the time of the convolution operation and coefficients used for its numerical value operation variable, and linking the kernel size learned by inputs of the operation portion 180 etc. and found suitable and the coefficients explained above, there are the advantages that the lens design can be carried out without worrying about the magnification and defocus range, and high precision image restoration by convolution becomes possible.
Further, there are the advantages that an optical lens of a high degree of difficulty, expense, and large size is not needed. Further, a so-called natural image where the object to be captured is focused, but the background is blurred can be obtained without driving the lens.
Further, the imaging apparatus 100 according to the present embodiment can be used for wavefront aberration control optical systems of zoom lenses of digital cameras, camcorders, and other consumer apparatuses for which smaller size, lighter weight, and lower costs have to be considered.
Further, in the present embodiment, since the apparatus has an imaging lens system having a wavefront forming optical element for deforming the wavefront of an image formed on the light receiving surface of the imaging element 120 by the imaging lens 112 and the image processing device 140 for receiving the first-order image FIM by the imaging element 120 and applying predetermined correction processing etc. to boost the MTF at the spatial frequency of the first-order image and form the high definition final image FNLIM, there is the advantage that the acquisition of a high definition image quality becomes possible.
Further, the configuration of the optical system 110 can be simplified, production becomes easier, and the cost can be reduced.
The fact that when using a CCD or CMOS sensor as the imaging element, there is a resolution limit determined from the pixel pitch and, when the resolution of the optical system is over that limit resolution power, the phenomenon of aliasing occurs and exerts an adverse influence upon the final image is known.
For the improvement of the image quality, desirably the contrast is raised as much as possible, but this requires a high performance lens system.
However, as explained above, when using a CCD or CMOS sensor as the imaging element, aliasing occurs.
At present, in order to avoid the occurrence of aliasing, the imaging lens system jointly uses a low pass filter made of a monoaxial crystalline system to thereby avoid the phenomenon of aliasing.
The joint usage of the low pass filter in this way is correct in terms of principle, but the low pass filter per se is made of crystal, therefore is expensive and hard to manage. Further, there is the disadvantage that the optical system is more complicated due to the use in the optical system.
As described above, a higher definition image quality is demanded as a trend of the times. In order to form a high definition image, the optical system in a general imaging lens device must be made more complicated. If it is complicated, production becomes difficult. Also, the utilization of expensive low pass filters leads to an increase in the cost.
However, according to the present embodiment, the occurrence of the phenomenon of aliasing can be avoided without using a low pass filter, and it becomes possible to obtain a high definition image quality
Note that, in the present embodiment, the example of arranging the wavefront forming optical element of the optical system on the object side from the stop was shown, but functional effects the same as those described above can be obtained even by arranging the wavefront forming optical element at a position the same as the position of the stop or on the focus lens side from the stop.
Further, the optical system in
Further, the kernel data storage ROM of
Further, the above embodiment was explained by taking as an example the case where there was one optical system, but the present invention can also be applied with respect to an imaging apparatus having a plurality of optical systems.
The difference between the present imaging apparatus 100A and the imaging apparatus 100 of
The optical unit 110A has a plurality of (two in the present embodiment) optical systems 110-1 and 110-2 and sequentially supplies images obtained by capturing an image of the object OBJ to the imaging element 120 in response to the switch processing of the optical system switch control portion 201.
The optical systems 110-1 and 110-2 have different optical magnifications and optically fetch the image of the captured target object (object) OBJ.
The system control device 200 basically has the same function as that of the exposure control device, waits for the operation inputs of the operation portion 180 etc., determines the operation of the overall system in response to those inputs, controls the optical system switch control portion 201, AFE 130, image processing device 140, DSP 150, etc. and conducts the mediation control of the whole system.
The rest of the configuration is the same as
First, the optical system is confirmed (ST61), then the kernel data is set (ST62).
Then, when the switching instruction of the optical systems is given by the operation of the operation portion 180 (ST63), the output of the optical system of the optical unit 110A is switched by the optical system switch control portion 210, and the processing of step ST61 is carried out (ST64).
According to the embodiment of
Namely, the imaging apparatus of
Further, there are the advantages that an optical lens of a high degree of difficulty, expense, and large size is not needed. Further, a so-called natural image where the object to be captured is focused, but the background is blurred can be obtained without driving the lens.
Further, the imaging apparatus 100 according to the present embodiment can be used for wavefront aberration control optical systems of zoom lenses of digital cameras, camcorders, and other consumer apparatuses for which smaller size, lighter weight, and lower costs have to be considered.
According to the imaging apparatus and image processing method of the present invention, the optical system can be simplified, the costs can be reduced, and in addition it is possible to obtain restored images with little influence of noise. Therefore, they can be applied to a digital still camera, a camera mounted in a mobile phone, a camera mounted in a digital personal assistant, an image inspection system, an industrial camera for automatic control, and so on.
Number | Date | Country | Kind |
---|---|---|---|
2005-303131 | Oct 2005 | JP | national |
2006-043657 | Feb 2006 | JP | national |
2006-173507 | Jun 2006 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2006/318388 | 9/15/2006 | WO | 00 | 10/8/2009 |