A technique has been developed for producing a zoom camera image by processing and combining the images from two lenses with two different fixed focal lengths or fields of view (see International patent application PCT/US2009/069804, filed Dec. 30, 2009). The image from the longer focal length (e.g., narrow field) lens may produce the central part of the final image, while the shorter focal length (e.g., wide field) lens may produce the remainder of the final image. Digital processing may adjust these two parts to produce a single image equivalent to that from a lens with an intermediate focal length. While this process may enable two fixed lenses to emulate the effect of a zoom lens, the line of demarcation between the two portions of the final image may be visible and distracting.
Some embodiments of the invention may be better understood by referring to the following description and accompanying drawings that are used to illustrate embodiments of the invention. In the drawings:
In the following description, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In other instances, well-known circuits, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
References to “one embodiment”, “an embodiment”, “example embodiment”, “various embodiments”, etc., indicate that the embodiment(s) of the invention so described may include particular features, structures, or characteristics, but not every embodiment necessarily includes the particular features, structures, or characteristics. Further, some embodiments may have some, all, or none of the features described for other embodiments.
In the following description and claims, the terms “coupled” and “connected,” along with their derivatives, may be used. It should be understood that these terms are not intended as synonyms for each other. Rather, in particular embodiments, “connected” is used to indicate that two or more elements are in direct physical or electrical contact with each other. “Coupled” is used to indicate that two or more elements co-operate or interact with each other, but they may or may not be in direct physical or electrical contact.
As used in the claims, unless otherwise specified the use of the ordinal adjectives “first”, “second”, “third”, etc., to describe a common element, merely indicate that different instances of like elements are being referred to, and are not intended to imply that the elements so described must be in a given sequence, either temporally, spatially, in ranking, or in any other manner.
Various embodiments of the invention may be implemented in one or any combination of hardware, firmware, and software. The invention may also be implemented as instructions contained in or on a computer-readable medium, which may be read and executed by one or more processors to enable performance of the operations described herein. A computer-readable medium may include any mechanism for storing information in a form readable by one or more computers. For example, a computer-readable medium may include a tangible storage medium, such as but not limited to read only memory (ROM); random access memory (RAM); magnetic disk storage media; optical storage media; a flash memory device, etc.
Various embodiments of the invention pertain to a blending technique used on an image created from a first digitized image from a fixed lens with a narrow field of view (referred to herein as a ‘narrow field lens’) and a second digitized image from a fixed lens with a wide field of view (referred to herein as a ‘wide field lens’). In this document, the terms ‘narrow’ and ‘wide’ are meant to be relative to each other, not to any external reference or industry standard. Within this document, an ‘image’ is a collection of pixel values that represent a visual picture. The pixels are typically thought of as being arranged in a rectangular array to achieve an easily understood correspondence between the image and the picture, but other embodiments may use other arrangements of pixels. Even if the image is not being displayed, processing the pixels may be described as if the image were being displayed, with terms such as ‘inner’, ‘outer’, ‘zoom’, ‘reduced’, ‘enlarged’, etc., describing how processing this data would effects the visual picture if it were displayed.
Once images have been obtained from both lenses, with all or at least a portion of the scene depicted by the narrow field lens being a subset of the scene depicted by the wide field lens, a composite image may be formed by using pixels from the narrow field image to form an inner portion (e.g., a central portion) of the composite, and using pixels from the wide field image to form an outer portion of the composite. The inner and outer portions may overlap to form an intermediate portion. Pixels within this intermediate portion may be derived by processing pixels from the narrow field image with the associated pixels from the wide field image, to gradually transition from the inner portion to the outer portion in a way that reduces visual discontinuities between the inner and outer portions.
As shown in
The sizes of the intermediate zone, the inner zone, and the outer zone, relative to each other, may depend on various factors, and in some embodiments may be dynamically variable. In other embodiments, these relative sizes may be fixed. The intermediate zone is shown as having a hollow rectangular shape, but may have any other feasible shape, such as but not limited to an annular ring. In some embodiments, each pixel in the intermediate zone may be processed individually, while in other embodiments, multi-pixel groups may be processed together. In some embodiments that contain multi-element pixels (e.g., color pixels consisting of red, blue, and green elements or yellow, magenta, and cyan elements), each element may be processed separately from the other elements in that pixel. Within this document, including the claims, any processing that is described as being performed on a pixel may be performed separately on individual elements within a pixel, and that element-by-element process shall be encompassed by the description and/or claim.
In one embodiment, each pixel in the intermediate zone that is close to the inner zone may be processed so as to result in a value nearly identical to the value it would have if it were in the inner zone (i.e., derived solely from the narrow field image). In a similar manner, each pixel in the intermediate zone that is close to the outer zone may be processed so as to result in a value nearly identical to the value it would have if it were in the outer zone (i.e., derived solely from the wide field image). As each pixel's location is farther from the inner zone and closer to the outer zone, it may be processed in a way that is influenced less by the pixel derived from the narrow field image and more by the associated pixel derived from the wide field image.
Pf=(X*Pw)+(1−X)*Pn
where Pf is the final pixel value,
Pw is the associated pixel value derived from the wide field image,
Pn is the associated pixel value derived from the narrow field image, and
X is a value between 0 and 1 that is related to the relative spatial position of the pixel between the inner zone and outer zone. In one embodiment, X may vary linearly across the distance from the inner zone to the outer zone (i.e., represent the fractional distance), while in other embodiments it may vary non-linearly (e.g., change more slowly or quickly near the borders of the intermediate zone than in the middle portions of that zone).
In this example, X=0 at the border between the inner and intermediate zones, while X=1 at the border between the outer and intermediate zones.
In some embodiments (e.g., where the intermediate zone has a hollow rectangular shape as in
At 420 the scale of the two images may be adjusted so that they both reflect the same scale. For example, the previously described method of image registration, through cropping and resampling, may be used so that a given portion of the scene from one image is represented by approximately the same number of pixels as it is in the other image. In some instances, only the wide field image may be cropped/upsampled in this manner. In other instances, the narrow field image may also be cropped and/or downsampled. To decide how much to crop and resample, in some embodiments it may be necessary to first determine the field of view and pixel dimensions of the final image. In other embodiments this may be predetermined.
At 430 a composite image may be created by combining the outer portion of the modified wide field image with the (modified or unmodified) narrow field image. These two portions may be defined such that they overlap to form an intermediate zone containing corresponding pixels from both. In some embodiments the size and location of this intermediate zone may be fixed and predetermined. In other embodiments the size and/or location of this intermediate zone may be variable, and determined either through an automatic process or by the user.
At 440 an algorithm may be determined for blending the pixels in the intermediate zone. In some embodiments there will only be one algorithm, and this step may be skipped. In other embodiments, there may be multiple algorithms to select from, either automatically or by the user. In some embodiments, multiple algorithms may be used during the same processing, either in parallel or sequentially.
At 450 the algorithm(s) may be used to process the pixels in the intermediate zone. In combination with the pixels in the inner and outer zones, these pixels may then produce the final image at 460. At 470, this final image may then be converted to a picture for display on a screen (e.g., for viewing by the person taking the picture), but the final image may alternately sent to a printer, or simply saved for use at a later time. In some embodiments, the user may examine the final image on the device's display and decide if the image needs further processing, using either the same algorithm(s) or different algorithm(s).
In some situations, the blending process described here may not produce a satisfactory improvement in the final image, and if that determination can be predicted, a decision may be made (either automatically or by a user) not to use a blending process. In some situations, merging the wide field image and the narrow field image (with or without blending) may not produce a satisfactory improvement in the final image, and a decision may be made (either automatically or by a user) not to combine those two initial images. If either of these situations is true, then one of the initial images may be used as is, one of the initial images may be modified in some way, or neither image may be used.
The foregoing description is intended to be illustrative and not limiting. Variations will occur to those of skill in the art. Those variations are intended to be included in the various embodiments of the invention, which are limited only by the scope of the following claims.