The present invention is directed to optical systems and, more particularly, to optical systems that are located at a fixed position within an imaging system and that provide a customized depth of field.
In a conventional digital imaging system, light received by an optical system consisting of one or more lenses is projected onto an image plane at which an image sensor is located. The received light may be, for example, light reflected off of or light emitted by one or more objects located at various distances from the lens. The image sensor detects the image directed onto the image plane and generates raw data that is processed by an image processor which produces a processed image that is available for storage or for viewing. In such conventional systems, however, only objects located within a small range of distances are focused onto the image plane for a given lens position, namely, for a given distance between the lens or lenses and the image plane. Objects located outside this range of object distances, which is known as the depth of field, may be directed onto the image plane but are not focused and appear blurred in the processed image. As a result, some objects in the processed image may appear in focus whereas other objects in the processed image appear out of focus depending on the distance of each object from the lens.
To focus objects located at other image distances using such conventional digital imaging systems, the lens must be moved within the imaging system to change the distance between the lens and the image plane. However, such movement also causes objects that previously appeared in focus to now appear out-of focus. Thus, only a portion of the objects in an image will appear in focus regardless of the lens position.
Recently, hand-held cellular telephones and various other devices have been introduced which incorporate a digital imaging system. Such devices typically require the optical surfaces to remain at a fixed position within the imaging system. It is not practical to include moving parts because of size and cost constraints, and thus their imaging systems have a fixed focal range.
To enable the imaging systems in such devices to provide acceptable processed images over a wider range of object distances, a known approach is to modify the imaging system by blurring the image directed onto the image plane. Because the optical properties of the blurring are known, the blurred image can be digitally processed to obtain an in focus processed image. The blurring and subsequent digital processing of the image allows for a wider range of object distances at which the processed image appears acceptably sharp, thereby extending the depth of field of the imaging system.
Such blurring is attained by incorporating a single blurring optical surface into an existing optical system that was originally intended to provide images having optimum image quality. Namely, the blurring optical surface is added to an optical system that was initially designed to provide the best possible focusing, i.e., provide optimal aberration control. For example, over a small range of distances, a diffraction limited optical system is such a system. Such optical systems that provide high quality focusing and aberration control are typically expensive to design and manufacture, an unnecessary expense given that the image that is directed onto the image plane is blurred. Moreover, such optical systems have an increased length.
According to an aspect of the invention, an optical system has a plurality of optical surfaces configured to provide blurred images of objects located within a selected range of object distances. At least two of the plurality of optical surfaces are configured to contribute to the blurring.
Preferably, the optical surfaces are configured so that no subset of the plurality of optical surfaces may provide a focused image of the object. Stated another way, the optical surfaces desirably are configured so that simply removing one or more of the optical surfaces would not leave a focusing optical system. Desirably, no subset of the plurality of optical surfaces forms a diffraction limited optical system. At object distances located near a hyperfocal distance of the optical system, the blurring may widen a point spread function (PSF) of the optical system. At object distances located near each end of the selected range of object distances, the blurring may narrow a point spread function (PSF) of the optical system. Within the selected range of object, distances, a modulation transfer function (MTF) of the optical system may be greater than a predetermined value.
Each one of the plurality of optical surfaces may contribute to broadening a through-focus modulation transfer function curve of the optical system. A peak modulation transfer function (MTF) of the optical system may be reduced when any one of the plurality of optical surfaces is removed. At least one of the optical surfaces may be configured to contribute differently to the blurring in different regions of that optical surface, and may be configured so that its contribution to the blurring changes discontinuously across a boundary between any two of these regions.
According to another aspect of the invention, a blurring optical system has a plurality of optical surfaces configured to provide images of objects located within a selected range of object distances. The blurring optical system has a particular f-number, field of view, number of optical surfaces, and optical track length. The blurring optical system according to this aspect of the invention desirably is arranged so that the peak modulation transfer function (MTF) of the blurring optical system is at least 50% of the peak MTF of a conventional non-blurring optical system having the same f-number, field of view, number of surfaces, and optical track length. The f-number, field of view, number of surfaces, and optical track length of the blurring optical system typically are such that the peak MTF of the conventional system having the same f-number, field of view, number of surfaces, and optical track length is less than about 70% of an MTF of a diffraction limited optical system having that f-number and field of view.
In accordance with this aspect of the invention, the number of optical surfaces may be less than that of a diffraction limited optical system, and may be less than that of a conventional optical system having a peak MTF of at least 80% of that of the diffraction limited system. The selected range of object distances may define a depth of field of the optical system. Each one of the plurality of optical surfaces may contribute to broadening the MTF of the optical system. The peak MTF of the optical system may be reduced when any one of the number of optical surfaces is removed.
A further aspect of the invention provides an imaging system which includes a blurring optical system and an image sensor operable to capture light directed by said optical system and to generate raw data representing the directed light. The imaging system according to this aspect of the invention desirably includes an image processor operable to process a first set of the raw data representing at least a portion of a field of view of the optical system by applying a plurality of different first deblurring functions to the first set of the raw data to yield a plurality of first processed image portions. The image processor desirably is also operable to select one of the plurality of first processed image portions having the best image quality. Systems according to this aspect of the invention can provide an auto-focusing capability. In one arrangement, the first set of raw data includes all of the raw data to be incorporated into a finished image, and the image processor simply supplies the selected first processed image portion as the finished image.
In other arrangements, the first set of raw data represents a first portion of the field of view of the optical system. The image processor may be arranged to select the first deblurring function which yielded the selected first processed image portion and apply the selected first deblurring function to additional sets of the raw data representing additional portions of the field of view to yield additional processed image portions.
In another variant, each of the plurality of different first deblurring functions is associated with an object distance. The image processor is operable to select an object distance associated with the first deblurring function which yielded the selected first processed image portion, and to select a set of additional deblurring functions associated with the selected object distance from among a plurality of sets of additional deblurring functions. The image processor is arranged to apply the deblurring functions in the selected set of deblurring functions to additional sets of the raw data representing additional portions of the field of view to yield additional processed image portions. For at least some object distances, the PSF of the optical system may differ for different portions of the field of view, and at least one of the sets of deblurring functions may include a plurality of different deblurring functions, each associated with a different portion of the field of view.
In yet another variant, the first set of raw data again represents a first portion of the field of view of the optical system. In this variant, the image processor is operable to process one or more additional sets of the raw data representing one or more additional portions of the field of view by applying a plurality of different additional deblurring functions to each additional set of the raw data to yield a plurality of processed image portions for such additional set of the raw data. The image processor is arranged to select one of the plurality of processed image portions having the best image quality for such additional set of the raw data, the selection step for each set of the raw data being performed independently of the selection step for other sets of raw data.
Yet another aspect of the present invention provides an imaging system which includes a blurring optical system having a point spread function (PSF) which differs for different portions of a field of view. The system according to this aspect of the invention includes an image sensor operable to capture light directed by said optical system and to generate raw data representing the directed light; and further includes an image processor operable to process at least part of the raw data to yield a processed image. Desirably, the image processor according to this aspect of the invention is operable to apply different deblurring functions to different portions of the raw data representing different portions of the field of view.
Further aspects of the invention include image processing methods incorporating generation of raw data and processing of the raw data as discussed above with reference to the imaging systems.
The foregoing aspects, features and advantages of the present invention will be further appreciated when considered with reference to the following detailed description and accompanying drawings.
The image sensor 110 receives light directed by the optical system 101 at the image surface 112, converts the captured light into raw data representing the captured light, and delivers the raw data to the image processor 120. The image sensor may be, for example, a charge coupled device (CCD) or a CMOS digital image sensor. The raw data typically is provided to the processor 120 in digital form.
The image processor 120 processes the raw data received from the image sensor 112 and generates a processed image that may be outputted to, for example, a memory device or a display device. As discussed further below, the image processor 120 may apply a mathematical function, referred to herein as a “deblurring” function, which sharpens the processed image.
If only focusing optical surfaces are present in the optical system 101 and the object O is located at a given object distance OD froth the optical system, a focused image I is present at a distance ID from the optical system. When the image surface 112 of the image sensor 110 is also located at distance ID from the optical system, the sharpest possible image is directed onto the image sensor 110. If the object O is moved nearer to or further from the optical system but the position of the image surface does not change (or, conversely, if the image surface is moved but the object O remains at the same distance from the optical system), the image quality of the projected image decreases. For a sufficiently small movement of the object O (or the image surface 112), the image quality of the image remains within an acceptable tolerance, namely, the image directed onto the image plane is of sufficient image quality for the image processor to provide an acceptable processed image. The range of object distances within which the image quality remains within this tolerance is referred to as the depth of field. Similarly, the range of image distances within which the image surface 112 may be moved while the image quality remains within tolerance is referred to as the depth of focus.
For a given object distance, the TF-MTF curve is essentially bell shaped and peaks at a particular image distance; this image distance is the point denoted as zero focus shift. The MTF at zero focus shift is referred to herein as the “peak MTF.” The object distance used in measuring the TF-MTF curve typically is the hyperfocal distance F, namely, the nearest distance at which the optical system can be focused while the projected image of an object located at a substantially infinite distance remains acceptably sharp. The TF-MTF curve is shown for a given wavelength of light and will shift with a change in wavelength. For systems such as cameras intended to image visible light, the TF-MTF curve typically is taken at a wavelength of 589 nm, in the yellow region of the spectrum. Also, the MTF depends upon the spatial frequency at which the image is sampled. The MTF typically is calculated or measured at a sampling frequency of one-half of the Nyquist frequency to one quarter of the Nyquist frequency. Also, the MTF may vary with position within the image plane; the MTF typically referred to is the MTF on the optical axis. Though the TF-MTF curve is shown in
When the blurring optical surface is also included, such as by incorporating a blurring phase element, the range of image distances at, which the TF-MTF is above an acceptable value is increased, as shown by curve 204 in
By increasing the depth of focus, the corresponding depth of field is also widened. Stated another way, for objects located at object distances within this depth of field, the images that are directed onto the image surface 112 of the sensor are out of focus within an acceptable range, and have MTF within the acceptable range above 0.1. Such images are then converted into raw data by the image sensor 110 and delivered to the image processor 120. Because the blurring function is known, the image processor is able to generate processed images that are in focus using the raw data representing the blurred images.
When the blurring optical surface is added, the images I1, I2, I3 are each stretched in the direction along the optical axis 99 by varying the focal depth of the various rays emanating from each of the objects O1, O2, O3. The paths taken by the various rays emanating from object O2, for example, are changed by the blurring optical surface such that some rays are now focused in front of the image plane and some rays are now focused behind the image plane, rather than all the rays being focused at the image plane. The paths of the rays emanating from objects O1 and are each similarly changed by the blurring optical surface so that some of the rays emanating from each of these objects are now focused at the image plane. Namely, the addition of the blurring surface widens the point spread function (PSF) of the optical system at a nominal focus i.e., the PSF at object distance O2 corresponding to zero focus shift, and reduces the PSF for objects at the edges of a desired range of object distances. The objects O1, O2, and O3 are each directed onto the image plane 112 as blurred images which, after the corresponding raw data is processed by the image processor, are restored to the images I1, I2, and I3 shown in
The known optical system 101 shown in
At least two of the optical surfaces in the optical system 302 contribute to the blurring, and no grouping of optical surfaces 304 through 309 selected from the optical system 302 forms a diffraction limited optical system or conventional non-blurring optical system. Rather, the two or more optical surfaces that contribute to the blurring each contribute to broadening the modulation transfer function (MTF) of the optical system, and when any one of the optical surfaces 304, . . . , 309 is removed, the directed image deteriorates because the peak MTF of the optical system is reduced. It should be noted that the shapes of the optical surfaces 304, . . . , 309 are for illustrative purposes only and merely indicate that the configurations of some or all of these optical surfaces differ from those shown in
By dividing the blurring PSF among the optical elements and by designing the optical system from the start as an optical system that is to provide a blurred image, fewer optical elements are required when compared to the known optical systems which combine a conventional optical system designed to provide the best possible focusing with an additional blurring surface. As a result, the cost of the optical system can be reduced. Also, a much greater choice of configurations is available for each optical surface.
The imaging system 300 also includes an image sensor 310 having an image plane 312 that may function in a manner similar to that of the image sensor 110 of
The distribution of the blurring and focusing functions in optical system 302 can be described in terms of a function f1 which defines a PSF resulting from the focusing and function f2 which defines a PSF resulting from the blurring. The overall point spread function (PSF) of the optical system is a convolution of functions f1 and f2. Moreover, if function fA defines an optical transmittance function of a particular one of the optical surfaces and function fB defines an optical transmittance function of another one of the optical surfaces, the functions fA and fB each contribute to the function f2. Also, no subset including less than all of the optical surfaces provides f1.
According to another embodiment of the invention, the contribution by a given optical surface to the blurring of an image is not uniform over the entire surface. Instead, the optical surface includes a plurality of regions each of which contributes differently to the blurring.
Another aspect of the invention relates to the processing of the raw data a digital imaging system having a blurring optical system. The digital imaging system incorporates an optical system 302, sensor 310 and processor 320 generally as discussed above with reference to
The image processor 320 applies a function, referred to herein as a “deblurring function,” to the raw data from the sensor 310. As used in this disclosure, the term “deblurring function” refers to a function which at least partially reverses the effects of blurring in the raw data. The deblurring function thus produces a processed image which is appreciably sharper than the image represented by the raw data. Stated another way, the deblurring function at least partially reverses the effect of the point spread function of the optical system. Deblurring functions per se are known in the art. The deblurring function which yields the best image depends in part on the point spread function of the optical system.
In one embodiment, the optical system is arranged such that the PSF is substantially constant across the field of view of the system but the PSF changes as a function of image distance or, equivalently, object distance. This is represented diagrammatically in
Processor 320 has access to stored deblurring functions associated with the various ranges of object distances. The deblurring functions may be stored in a conventional digital memory (not shown) which is incorporated in the processor itself or connected to the processor. The deblurring functions may be stored in any format. For example, each deblurring function may be stored in the form of a set of algorithmic instructions, coefficients or other information which can be used directly in processing of data from the sensor. Alternatively, each deblurring function may be stored as information from which information useful in processing the raw data can be derived. For example, a deblurring function associated with a particular range of object distances can be stored by storing the PSF associated with that range of object distances along with instructions for deriving the deblurring function from the PSF.
The stored deblurring function associated with a particular range of object distances can be calculated by calculating the PSF for an object distance within the range based on the design of the optical system, and determining the deblurring function based on the calculated PSF. Alternatively, the deblurring function for a particular range of object distances can be derived by measuring the PSF of an actual system at an image distance corresponding to an object distance within the range, and determining the deblurring function based on the measured PSF. In yet another arrangement, the deblurring function for a particular range of object distances can be derived by applying a variety of deblurring functions to raw data from the sensor of an actual system imaging an object at an object distance within the range, measuring one or more aspects of image quality such as image sharpness achieved by each deblurring function, and storing the particular deblurring function which yields the best image quality. Using measurements of PSF or image quality to derive each deblurring function compensates for manufacturing tolerances. In a mass-production process, such measurements can be performed for every system, or for a representative sample of the systems. Where systems are produced in batches, the measurements can be performed on one or a few samples in each batch, and the results applied to the remaining systems in the batch. For example, where the sensors are formed on semiconductor wafers and the optical elements are assembled to the sensors in a wafer-level process, the assemblies formed from each wafer may constitute a batch.
In operation, the when the image sensor of the imaging system captures an image directed by the optical system, the image processor 320 processes the raw data generated by the image sensor using each of the stored deblurring functions. The processor applies each of the stored deblurring functions separately to the raw data to obtain a plurality of processed images. At this stage, each processed image is in the form of digital data defining an image. The image processor tests each of these images for image quality. The step of testing for image quality may include testing for image sharpness. The processed image having the best image quality is selected. The selected processed image typically is stored or displayed, whereas the other processed images may be discarded. The processing carried out by the image processor allows the imaging system to provide the same effect as an auto-focus operation in a conventional camera without requiring movement of the optical surfaces. The imaging system does not require the moving parts that are otherwise needed to move the optical surfaces in known auto-focus systems. Moreover, the optical system need not provide a substantially constant PSF over a wide range of object distances. Stated another way, the ability of the processor to select from among a plurality of deblurring functions removes a constraint on the design of the optical system.
In another embodiment, the image processor 320 applies the various deblurring functions to a first set of the raw data which represents a first region of the field of view which is smaller than the entire field of view, such as region R1 (
In this embodiment, the deblurring function which is used to reconstruct the image is selected based on the object distance of the objects represented in the first region. For example, if the objects represented in the first region are foreground objects disposed in the first range of object distances (10 cm to 50 cm), the selected deblurring function will be appropriate for PSF 601. This deblurring function typically will not provide optimal deblurring for objects in other ranges of object distance. Thus, background objects disposed at large object distances may not be sharp in the processed image. This effect can provide aesthetically desirable results, similar to the effect of manually or automatically focusing a conventional lens having limited depth of field.
Regions R1-R4, though depicted as rectangular in
In a further variant, the first region may be a user-selectable region of the image. For example, in a system with an optical viewfinder or an electronic viewfinder which displays a crude image based on the raw data, the system may be provided with user controls which allow the user to move a cursor in the viewfinder and thus select a particular region of the image. This allows the user to select a region of the image depicting particularly important objects, and assures that the deblurring function will be selected to maximize the image quality of those objects.
In another variant, the image processor respectively selects a deblurring function for each region of the field of view independently of the selection for other fields of view. For the example shown in
In another embodiment, a digital imaging system includes an optical system that is designed to provide a PSF that differs in different regions of the field of view, as described above, but which does not change substantially with object distance. For example,
In a still further embodiment of the invention, a digital imaging system includes an optical system that provides a PSF that differs in different regions of the field of view and also varies as a function of image distance. For example,
The image processor has access to stored deblurring functions appropriate for each of the PSF curves. A first set of deblurring functions is associated with the first range of image distances. The first set includes a first deblurring function which will compensate for PSF 801 associated with first region R1; a second deblurring function which will compensate for PSF 802 associated with second region R2, a third deblurring function appropriate to PSF 803 associated with third region R3, and a fourth deblurring function appropriate to PSF 804 for region R4. Likewise, a second set of deblurring functions, associated with the second range of object distances, includes first through fourth deblurring functions associated with regions R1-R4, respectively, the deblurring functions being appropriate to compensate for PSF functions 805, 806, 807 and 808, respectively. A third set of deblurring functions includes first through fourth deblurring functions appropriate to PSF functions 809-812.
When an image is captured by the image sensor of the imaging system, the image processor processes the first set of raw data from region R1 with each of the first deblurring functions associated with region R1, i.e., with the deblurring functions appropriate to PSFs 801, 805 and 809, to obtain three first processed image portions. Here again, the image processor tests the image quality of each first processed image portions and selects the first processed image portion having the best image quality. This selection implicitly selects an object distance and a set of deblurring functions. For example, if the first processed image portion obtained using a deblurring function appropriate to PSF 805 has the best image quality, the system has implicitly selected the second range of object distances (
The ability of the system to provide satisfactory image quality despite variation in PSF across the field of view and variation in PSF with object distance and image distance removes constraints from the design of the optical system. Here again, the stored deblurring functions can be derived by calculation from the design of the optical system, from measurements of PSF, or from measurements of image quality using different deblurring functions during manufacture of the system. In this embodiment, the calculations or measurements are performed separate for each region and for each range of object distance or image distance. Here again, where measurements are used to determine the stored deblurring functions, the processor will compensate for variations in manufacture of the system.
In another variant, the image processor selects a deblurring function for each region independently of the selection made for other portions. For example, the raw data of region R1 is processed with each of the deblurring functions associated with the first region R1 (the deblurring functions appropriate for PSF curves 801, 805, and 809) to provide a plurality of first processed image portions, and the first processed image portion having the most preferred image quality is selected for region R1. Similarly, the raw data for region R2 is processed with each of the deblurring functions associated with region R2 (the deblurring functions appropriate for PSF curves 802, 806, and 810) to provide a plurality of second processed image portions. The second processed image portion having the most preferred image quality is selected for region R2. The raw data for the third and fourth regions are treated similarly, using the deblurring functions associated with those regions.
The selection of the image portion having the most preferred image quality in each respective region of the field allows the image processor to correct for differences in the distance of objects in the different regions and thus extends the depth of field of the imaging system.
Although the features described above can be applied in a wide range of applications, they are particularly useful in systems of the type used in small, inexpensive digital cameras of the type found in cellular telephones, PDAs and other portable electronic devices. For example, the features described above can be implemented with an optical system having a track length of 1 cm or less, and more preferably 5 mm or less, and having 3 optical elements (6 optical surfaces) or fewer, such as those having only 2 optical elements or only 1 optical element. Also, the image processor typically is located within the same electronic device as the image sensor. In some applications, the optical sensor, image processor and image sensor may be provided as a module which can be mounted to a circuit panel or otherwise installed in a larger device. The image processor may perform functions other than deblurring as, for example, correction of geometric distortion.
Although the invention herein has been described with reference to particular embodiments, it is to be understood that these embodiments are merely illustrative of the principles and applications of the present invention. It is therefore to be understood that numerous modifications may be made to the illustrative embodiments and that other arrangements may be devised without departing from the spirit and scope of the present invention as defined by the appended claims.
This application claims the benefit of the filing date of U.S. Provisional Application No. 61/001,988 filed Nov. 6, 2007, the disclosure of which is incorporated herein by reference.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/US2008/012528 | 11/4/2008 | WO | 00 | 8/10/2010 |
Number | Date | Country | |
---|---|---|---|
61001988 | Nov 2007 | US |