Luminance source selection in a multi-lens camera

Information

  • Patent Grant
  • 9413984
  • Patent Number
    9,413,984
  • Date Filed
    Monday, October 24, 2011
    13 years ago
  • Date Issued
    Tuesday, August 9, 2016
    8 years ago
Abstract
The luminance information of an image captured by a multi-lens camera system can be improved by selecting a luminance information source for each portion of the captured image. Each lens of the camera system can capture an initial image. For each portion of a final image, a corresponding initial image portion can be selected as the luminance information source. The portions of the final image and initial images can be pixels, groups of pixels, or other image portions. The luminance information from the selected initial image portions is combined to form final image luminance information. Chrominance information can also be selected from the initial images to form final image chrominance information, and the final image chrominance information and the final image luminance information can be combined to form a final image.
Description
BACKGROUND

1. Technical Field


This disclosure relates to a camera system, and more specifically, to a multi-lens camera system configured to select between sources of luminance information in generating an image.


2. Description of the Related Art


An imaging system typically consists of an imaging lens and an image sensor. An imaging lens collects light emitted or reflected from objects in a scene and directs collected light upon the image sensor. An image sensor is a photosensitive device that converts light incident upon the image sensor during an image capture to an electronic signal representative of the captured light. To obtain color image data, a color filter array (such as a Bayer filter) is used in conjunction with the image sensor to separate between different spectral regions of the total light spectrum of the image being captured. Color filter arrays separate captured light into (for instance) green image planes, red image planes, and blue image planes.


Given an image sensor's active area dimension (the image sensor's “format”) and the desired field of view, the focal length of a lens can be calculated. The size of the aperture of the lens can be set according to image sensor's photo sensitivity, exposure time, and noise level tolerance. The focal length divided by the aperture's size is called the “F-number,” and indicates the ability of the lens to collect light. Lower F-Numbers are associated with more light being collected by the lens and directed upon the image sensor.


A phenomena caused by the use of color filter arrays is the appearance of color artifacts also caused by the spatial disposition of the different colors. For example, in a captured image of a white line 1 pixel deep on a black background, the white line will appear in various colors depending on the position of the light from the line incident upon the image sensor. Multi-lens systems can be implemented to reduce such artifacts, but can be accompanied by issues of increased system cost and complexity.





BRIEF DESCRIPTION OF THE DRAWINGS

The disclosed embodiments have other advantages and features which will be more readily apparent from the following detailed description of the invention and the appended claims, when taken in conjunction with the accompanying drawings, in which:


FIG. (or “FIG.”) 1 illustrates a side view of a single lens camera, according to one example embodiment.



FIG. 2 illustrates a color filter array having multiple pixels, according to one example embodiment.



FIG. 3 illustrates a side view of a three lens camera having one image sensor and three lenses, according to one example embodiment.



FIG. 4 illustrates an example of a scene as projected on to an image sensor, according to one example embodiment.



FIG. 5 illustrates a front view of a three lens camera using one rectangular image sensor divided in to three regions, according to one example embodiment.



FIG. 6 illustrates a front view of a three lens camera having one image sensor, one large lens and two smaller lenses, according to one example embodiment.



FIG. 7 illustrates a front view of a four lens camera having a one image sensor and four lenses, according to one example embodiment.



FIG. 8 illustrates a 16 lens camera having four regions, each containing four lenses as illustrated in FIG. 7, according to one example embodiment.



FIG. 9 illustrates a light source incident upon two image sensor regions, according to one example embodiment.





DETAILED DESCRIPTION

The figures and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.


Reference will now be made in detail to several embodiments, examples of which are illustrated in the accompanying figures. It is noted that wherever practicable similar or like reference numbers may be used in the figures and may indicate similar or like functionality. The figures depict embodiments of the disclosed system (or method) for purposes of illustration only. One skilled in the art will readily recognize from the following description that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles described herein.


A multi-lens camera system can improve image luminance by, for each image portion of a field of view, selecting one of the lenses as a source for luminance information for the portion of the field of view. One or more image sensors can be divided into a plurality of image sensor regions, each associated with a different lens. Each lens collects light from the field of view and directs the collected light onto an image sensor region. Each image sensor region captures an initial image representing the light incident upon the image sensor region from an associated lens, and the camera system can combine the initial images to create a final image.


The field of view can be divided into a plurality of portions (such as pixels), and each field of view portion can be associated with an image sensor region portion of each image sensor region. For example, the field of view can be divided into 100 by 100 pixels, and each image sensor region can capture an initial image measuring 100 by 100 pixels, with a 1-to-1 correspondence between each field of view pixel and an associated initial image pixel for each captured initial image. The final image generated by the camera system can include a plurality of portions corresponding to the plurality of field of view portions and the associated plurality of image sensor region portions. The source of luminance information for each final image portion can be selected from among the plurality of associated initial image portions. Continuing with the previous example, if the final image is 100 by 100 pixels, the luminance information for each final image pixel can be selected from among the plurality of associated initial image pixels. By selecting the source of luminance information for each final image portion, the final image can have higher dynamic range and better low light performance than any of the initial images captured by the camera system.


Luminance Source Selection Overview


The multi-lens camera system described herein can improve the effective resolution and spatial resolution of a captured image while improving the low light performance and increasing the dynamic range of the camera system. Each lens in the multi-lens camera system is associated with one or more optical surfaces having a non-zero optical power. For a spatial resolution off, each lens can resolve image details of the size 1/f.


As noted above, each lens of the multi-lens camera system collects light from a field of view and directs the light onto an image sensor region. The image sensor region captures light incident upon the image sensor region from an associated lens to form an initial image. The multi-lens camera system can generate a final image based at least in part on the combination of image information from one or more of the initial images. The final image can be generated by selecting, for each final image portion, luminance information from one of the plurality of associated initial image portions.


The multi-lens camera system can include one or more color filters, polarized filters, chromatic filters, and neutral density filters integrated within the system configured to filter collected light prior to capture by the image sensor. Each initial image can have a different light intensity from other initial images. The camera system can include an algorithm for adding initial images to form a final image having higher dynamic range than the initial images. Each lens in the multi-lens camera system can have a different F-Number than the other lenses, and can be focused to a different distance than the other lenses.


In one embodiment, selecting the source of luminance information for each final image portion includes selecting the sharpest initial image portion associated with the final image portion from among the initial portions. As noted above, the final image portions and initial image portions can be single pixels. Thus, selecting the luminance information for a final image portion can include selecting an associated initial image pixel from among the initial images for a final image pixel. The source of luminance information for any two final image pixels can be different initial images or the same initial image. In addition to selecting the source of luminance information at the pixel level, the source of luminance information can be selected for pixel groups, or for any image portion according to the principles described herein.


Selecting luminance information for the final image at the pixel level can beneficially allow for maximizing the signal to noise ratio for dark image portions and avoiding saturation at bright image portions, particularly at image portions including borders between bright areas and dark area. Selecting luminance information at the pixel group level can beneficially reduce computation time in generating the final image. Computation time can be further reduced by performing edge detection on the initial images, and selecting pixel groups based on the detected edges.


The lenses of the multi-camera system can be focused at the same distance or at different distances, and can have the same field of view or different fields of view. Further, each lens can include various different filters, and the image sensor regions associated with the lenses can be exposed at different times or the same time, or for different intervals of time or the same interval of time. By adjusting the parameters of each lens and the image sensor region exposures, the dynamic range of the final image can be increased, the white balance of the final image can be better controlled, and the low light performance of the camera system can be increased. In one embodiment, luminance information is chosen for dark areas of the final image from an initial image associated with a lens with a broader spectral filter, and for light areas of the final image from other image sensor regions. This can allow for choosing the source of luminance from initial images that are not saturated, further increasing the dynamic range of the final image.


The multi-lens camera system described herein can perform a method for selecting a source of luminance information for each pixel in a final image, including the steps:

    • 1. for each pixel in a final image, comparing the luminance information of a set of corresponding pixels from a plurality of initial images,
    • 2. for each pixel in the final image, selecting a pixel from the set of corresponding pixels as a source of luminance information for the pixel of the final image based on the comparisons,
    • 3. combining the luminance information from the selected pixels to form final image luminance information,
    • 4. selecting chrominance information from each of the initial images to form final image chrominance information, and
    • 5. combining the final image luminance information and the final image chrominance information to form the final image.


      System Overview


The system and method described herein provide high quality imaging while considerably reducing the length of the camera as compared to other systems and methods.



FIG. 1 illustrates a side view of a single lens camera having a single lens (102) that can include one or more elements and a single sensor (101). FIG. 2 illustrates a sensor array (201) having multiple pixels where the position of the green filter, red filter and blue filter are marked by (202), (203) and (204) respectively. The image that will be taken using this configuration needs to be processed in order to separate the green, red and blue images.



FIG. 3 illustrates a side view of a three lens camera having one sensor (310) and three lenses (301), (302) and (303). Each one of the said lens will project the image of the same scene on to segments of the sensor marked by (311), (312), and (313) respectively. Each one of the three lenses will have different color filters integrated within the lens, in front of it or between the lens and sensor (310). Using the described configuration the image acquired by the sensor will be composed of two or more smaller images, each imaging information from the scene at different spectrums.



FIG. 4 illustrates an example of a scene as projected on to the sensor (401), in each region of the sensor (402), (403) and (404) the same scene is projected but each region will contain information for light at different wavelengths representing different colors according to the filters integrated within the lens that forms the image on each region.


The described configuration does not require the use of a color filter array and therefore the maximal spatial frequency that can be resolved by the sensor can be higher. On the other hand, using smaller lens and smaller active area per channel can result in a smaller focal length of the lens. Therefore, the spatial resolution of objects can be decreased, and the maximal resolvable resolution for each color can remain the same.


The image acquired by the sensor is composed of two or more smaller images, each containing information of the same scene but in different colors. The complete image is then processed and separated in to 3 or more smaller images and combined together to one large color image.


The described method of imaging has many advantages:

    • 1. Shorter lens track (height): Each one of the lenses used can be smaller in size than the single lens covering the same field of view. The total track (height) of each lens can be smaller, allowing the camera to be smaller in height, an important factor for mobile phone cameras, notebook cameras and other applications requiring short optical track.
    • 2. Reduced Color artifacts: Since each color is captured separately, artifacts originating from spatial dependency of each color in a color filter array can be reduced.
    • 3. Lens requirements: Each lens does not have to be optimal for all spectrums used, simplifying the lens design and possibly decreasing the amount of elements used in each lens as no color correction may be needed.
    • 4. Larger Depth of Focus: The depth of focus of a system depends on its focal length. Since smaller lenses are used with smaller focal lengths, the depth of focus is increased by the scale factor, squared.
    • 5. Elimination of focus mechanism: Focus mechanisms can change the distance between the lens and the sensor to compensate for the change in object distance and to assure that the desired distance is in focus during the exposure time. Such a mechanism can be costly and can have many other disadvantages such as increased size, increased power consumption, shutter lag, decreased reliability, and increased price.


Using a fourth lens in addition to the three used for each color red, green and blue (or other colors) with a broad spectral transmission can allow extension of the sensor's dynamic range and can improve the signal-to-noise performance of the camera in low light conditions.


All configurations described above using a fourth lens element can be applied to configurations having two or more lenses.


Another configuration uses two or more lenses with one sensor having a color filter array integrated or on top of the sensor such as a Bayer filter array. In such a configuration no color filter will be integrated in to each lens channel and all lenses will create a color image on the sensor region corresponding to the specific lens. The resulting image will be processed to form one large image combining the two or more color images that are projected on to the sensor.


Dividing the sensor's active area in to 3 areas, one for each of red, green, and blue (for example), can be achieved by placing 3 lenses as illustrated in the figures. The resulting image will include 3 small images containing information of the same scene in a different color.



FIG. 5 illustrates a front view of a three lens camera using one rectangular sensor (500) divided in to three regions (501), (502) and (503). The three lenses (511), (512) and (513) each having different color filters integrated within the lens, in front of the lens or between the lens and the sensor are used to form an image of the same scene but in different colors. In this example each region of the sensor (501), (502) and (503) are rectangular having the longer dimension of the rectangle perpendicular to the long dimension of the complete sensor.


Other three lens configuration can be used, such as using a larger green filtered lens and two smaller lenses for blue and red, such a configuration will results in higher spatial resolution in the green channel since more pixels are being used.



FIG. 6 illustrates a front view of a three lens camera having one sensor (600), one large lens (613) and two smaller lenses (611) and (612). The large lens (613) is used to form an image on the sensor segment marked (603) while the two smaller lenses form an image on the sensor's segments marked with (601) and (602) respectively. The larger lens (613) can use a green color filter while the two smaller lenses (611) and (612) can use a blue and red filter respectively. Other color filters could be used for each lens.


A four lens camera includes 4 lenses each having a different color filter integrated within the lens, in front of the lens, or between the lens and the sensor region corresponding to the lens. The color filter used for two lenses can be the same, resulting in a particular color filter appearing twice among the 4 lenses.



FIG. 7 illustrates a front view of a four lens camera having one sensor (700) and four lenses (711), (712),(713) and (714). Each lens forms an image on the corresponding sensor region marked with (701), (702),(703) and (704) respectively. Each one of the lenses will be integrated with a color filter within the lens, in front of the lens, or between the lens and the sensor. All four lenses can be integrated with different color filters, or two of the four lenses can be integrated with the same color filter. For example, two green filters, one blue filter, and one red filter can allow more light collection in the green spectrum.



FIG. 8 illustrates a 16 lens camera having 4 regions (801), (802), (803) and (804), each containing four lenses as illustrated in FIG. 7.



FIG. 9 illustrates a light source 901 incident upon an image sensor 900 including two image sensor regions 910 and 920. The light source 901 is primarily incident upon two pixels of the image sensor region 910, and upon one pixel of the image sensor region 920. Accordingly, the sharpness of the light source in the initial images captured by the image sensor regions 910 and 920 is greater in the pixel of the initial image captured by the image sensor region 920 than in either of the two pixels of the initial image captured by the image sensor region 910. Accordingly, the camera system can select luminance information for a pixel of a final image corresponding to the light source from the initial image captured by the image sensor region 920 over the initial image captured by the image sensor region 910. The camera system can similarly select luminance information for each pixel of the final image, and can combined selected luminance information with chrominance information retrieved from the initial images to form the final image.


Additional Configuration Considerations


Throughout this specification, some embodiments have used the expression “coupled” along with its derivatives. The term “coupled” as used herein is not necessarily limited to two or more elements being in direct physical or electrical contact. Rather, the term “coupled” may also encompass two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other, or are structured to provide a thermal conduction path between the elements.


Likewise, as used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.


In addition, use of the “a” or “an” are employed to describe elements and components of the embodiments herein. This is done merely for convenience and to give a general sense of the invention. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.


Finally, as used herein any reference to “one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.


Upon reading this disclosure, those of skill in the art will appreciate still additional alternative structural and functional designs from the principles herein. Thus, while particular embodiments and applications have been illustrated and described, it is to be understood that the disclosed embodiments are not limited to the precise construction and components disclosed herein. Various modifications, changes and variations, which will be apparent to those skilled in the art, may be made in the arrangement, operation and details of the method and apparatus disclosed herein without departing from the spirit and scope defined in the appended claims.

Claims
  • 1. A multi-lens camera system, comprising: a first lens configured to collect light from a field of view and direct the light onto a first image sensor region, the first image sensor region being configured to capture the light incident upon the first image sensor region and to generate a first initial image having a plurality of first portions;a second lens configured to collect light from the field of view and direct the light onto a second image sensor region, the second image sensor region being configured to capture the light incident upon the second image sensor region and to generate a second initial image with a plurality of second portions; anda controller configured to generate luminance information for a plurality of final image portions of a final image based on luminance information of the first initial image and the second initial image;wherein the controller is configured to, for each final image portion, compare luminance information of each first portion with luminance information of each corresponding second portion, and to select luminance information from either each of the first or the corresponding second portions for each of the final image portions based on the comparison and to select chrominance information by combining each of the first and the second portions for each of the final image portions.
  • 2. The multi-lens camera system of claim 1, wherein each final image portion comprises one pixel of the final image, and wherein the corresponding portions of the first initial image and the second initial image comprise one corresponding pixel in each of the first initial image and the second initial image.
  • 3. The multi-lens camera system of claim 1, wherein each final image portion comprises a set of pixels of the final image, and wherein the corresponding portions of the first initial image and the second initial image comprise a set of corresponding pixels in each of the first initial image and the second initial image.
  • 4. The multi-lens camera system of claim 3, wherein a set of pixels of the final image comprises a set of pixels determined to represent an image edge.
  • 5. The multi-lens camera system of claim 4, wherein the set of pixels is determined to represent the image edge by an edge detection algorithm.
  • 6. The multi-lens camera system of claim 1, wherein the first lens and the second lens are configured to be focused at the same distance.
  • 7. The multi-lens camera system of claim 1, wherein the first lens and the second lens are configured to be focused at different distances.
  • 8. The multi-lens camera system of claim 1, wherein at least one of the first lens and the second lens includes one or more of: a neutral density filter, a chromatic filter, and a polarizing filter.
  • 9. The multi-lens camera system of claim 1, wherein the first image sensor region and the second image sensor region are exposed for different amounts of time.
  • 10. The multi-lens camera system of claim 1, wherein the first lens and the second lens have a substantially similar field of view.
  • 11. The multi-lens camera system of claim 1, wherein each of the final image portions maximizes the signal-to-noise ratio and avoids saturation for each final image portion based on the comparison.
  • 12. A method for capturing an image by a multi-lens camera system, comprising: generating, by a first image sensor region, a first initial image having a plurality of first portions, the first initial image representative of light collected by a first lens and incident upon the first image sensor region during a first exposure period;generating, by a second image region, a second initial image having a plurality of second portions, the second initial image representative of light collected by a second lens and incident upon the second image sensor region during a second exposure period;for each portion of a final image:comparing luminance information of each first portion with luminance information of each corresponding second portion; andselecting luminance information from either the each of the first portions or the corresponding second portions based on the comparison; andcombining the selected luminance information for the plurality of corresponding final image portions to form final image luminance information;selecting chrominance information for each of the final image portions by combining each of the first and the corresponding second portions for each of the final image portions.
  • 13. The method of claim 12, wherein each portion of the final image comprises one pixel of the final image, and wherein the corresponding portions of the first initial image and the second initial image comprise one corresponding pixel in each of the first initial image and the second initial image.
  • 14. The method of claim 12, wherein each portion of the final image comprises a set of pixels of the final image, and wherein the corresponding portions of the first initial image and the second initial image comprise a set of corresponding pixels in each of the first initial image and the second initial image.
  • 15. The method of claim 14, wherein a set of pixels of the final image comprises a set of pixels determined to represent an image edge.
  • 16. The method of claim 15, wherein the set of pixels is determined to represent the image edge by an edge detection algorithm.
  • 17. The method of claim 12, wherein the first lens and the second lens are configured to be focused at the same distance.
  • 18. The method of claim 12, wherein the first lens and the second lens are configured to be focused at different distances.
  • 19. The method of claim 12, wherein at least one of the first lens and the second lens includes one or more of: a neutral density filter, a chromatic filter, and a polarizing filter.
  • 20. The method of claim 12, wherein the first image sensor region and the second image sensor region are exposed for different amounts of time.
  • 21. The method of claim 12, wherein the first lens and the second lens have a substantially similar field of view.
  • 22. The method of claim 12, further comprising maximizing the signal-to-noise ratio and avoiding saturation for each final image portion based on the comparison.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a national phase application of PCT/NL2011/050725, entitled “System and Method for Imaging Using Multi Aperture Camera,” filed on Oct. 24, 2011, which in turn claims priority to U.S. Provisional Application No. 61/406,148, filed on Oct. 24, 2010, the contents of which are all hereby incorporated by reference in their entirety.

PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/NL2011/050725 10/24/2011 WO 00 7/8/2013
Publishing Document Publishing Date Country Kind
WO2012/057622 5/3/2012 WO A
US Referenced Citations (107)
Number Name Date Kind
4890905 Van Rosmalen et al. Jan 1990 A
5347340 Tsukada Sep 1994 A
6115065 Yadid-Pecht et al. Sep 2000 A
6765617 Tangen Jul 2004 B1
6773638 Kloosterboer et al. Aug 2004 B2
6809766 Krymski et al. Oct 2004 B1
6898331 Tiana May 2005 B2
6980248 Suda Dec 2005 B1
7151259 Koch et al. Dec 2006 B2
7199348 Olsen Apr 2007 B2
7224384 Iddan May 2007 B1
7233359 Suda Jun 2007 B2
7564019 Olsen Jul 2009 B2
7620309 Georgiev Nov 2009 B2
7718940 Hirasawa May 2010 B2
7916181 Nilehn Mar 2011 B2
7932941 Hayasaka Apr 2011 B2
8023016 Iijima Sep 2011 B2
8049806 Feldman Nov 2011 B2
8228417 Georgiev et al. Jul 2012 B1
8259212 Brady et al. Sep 2012 B2
8290358 Georgiev Oct 2012 B1
8315476 Georgiev et al. Nov 2012 B1
8345144 Georgiev et al. Jan 2013 B1
8436909 Farina May 2013 B2
8514491 Duparre Aug 2013 B2
8629390 Olsen Jan 2014 B2
8749694 Georgiev et al. Jun 2014 B2
8817015 Georgiev Aug 2014 B2
8885059 Venkataraman Nov 2014 B1
20010026322 Takahashi Oct 2001 A1
20020067416 Yoneda et al. Jun 2002 A1
20020089596 Suda Jul 2002 A1
20020122124 Suda Sep 2002 A1
20030108240 Gutta Jun 2003 A1
20030234907 Kawai Dec 2003 A1
20040047518 Tiana Mar 2004 A1
20040080661 Afsenius et al. Apr 2004 A1
20040109004 Bastos et al. Jun 2004 A1
20040201748 Goldstein Oct 2004 A1
20050052751 Liu Mar 2005 A1
20050128323 Choi Jun 2005 A1
20050128335 Kolehmainen Jun 2005 A1
20050128509 Tokkonen Jun 2005 A1
20050134699 Nagashima et al. Jun 2005 A1
20050160112 Makela Jul 2005 A1
20050225654 Feldman et al. Oct 2005 A1
20050259169 Ito Nov 2005 A1
20050270395 Yoneda et al. Dec 2005 A1
20060003328 Grossberg Jan 2006 A1
20060055811 Frtiz Mar 2006 A1
20060108505 Gruhlke et al. May 2006 A1
20060125936 Gruhike Jun 2006 A1
20060193509 Criminisi et al. Aug 2006 A1
20070075218 Gates Apr 2007 A1
20070091197 Okayama Apr 2007 A1
20070153086 Usui et al. Jul 2007 A1
20070177004 Kolehmainen et al. Aug 2007 A1
20070189748 Drimbarean et al. Aug 2007 A1
20070211164 Olsen et al. Sep 2007 A1
20070252074 Ng Nov 2007 A1
20070252908 Kolehmainen Nov 2007 A1
20070258006 Olsen et al. Nov 2007 A1
20080068452 Nakao Mar 2008 A1
20080079839 Sung et al. Apr 2008 A1
20080084486 Enge Apr 2008 A1
20080218611 Parulski et al. Sep 2008 A1
20080219493 Tadmor Sep 2008 A1
20080240508 Nakao et al. Oct 2008 A1
20080278610 Boettinger Nov 2008 A1
20090021612 Hamilton Jan 2009 A1
20090103792 Rahn et al. Apr 2009 A1
20090140131 Utagawa Jun 2009 A1
20090262987 Ioffe et al. Oct 2009 A1
20090321861 Oliver et al. Dec 2009 A1
20100020078 Shpunt Jan 2010 A1
20100021064 Lee et al. Jan 2010 A1
20100039713 Lusinchi et al. Feb 2010 A1
20100052192 Hasegawa et al. Mar 2010 A1
20100085468 Park Apr 2010 A1
20100097491 Farina et al. Apr 2010 A1
20100117176 Uekawa May 2010 A1
20100127157 Tamaki et al. May 2010 A1
20100128137 Guidash May 2010 A1
20100171866 Brady Jul 2010 A1
20100253833 Deever Oct 2010 A1
20100259607 Kennedy et al. Oct 2010 A1
20100283837 Oohchida Nov 2010 A1
20110019048 Raynor Jan 2011 A1
20110069189 Venkataraman et al. Mar 2011 A1
20110080487 Venkataraman et al. Apr 2011 A1
20110122308 Duparre May 2011 A1
20110134282 Morita et al. Jun 2011 A1
20110157387 Han et al. Jun 2011 A1
20110242356 Aleksic Oct 2011 A1
20120007942 Michrowski et al. Jan 2012 A1
20120140100 Shibazaki Jun 2012 A1
20130002928 Imai Jan 2013 A1
20130121615 Intwala et al. May 2013 A1
20130128087 Georgiev et al. May 2013 A1
20130293744 Attar et al. Nov 2013 A1
20130308197 Duparre Nov 2013 A1
20130321674 Cote et al. Dec 2013 A1
20130335598 Gustavsson Dec 2013 A1
20150234151 Venkataraman Aug 2015 A1
20150235371 Venkataraman Aug 2015 A1
20150296193 Cote Oct 2015 A1
Foreign Referenced Citations (25)
Number Date Country
1206126 May 2002 EP
2336816 Jun 2011 EP
2000 152281 May 2000 JP
2001078212 Mar 2001 JP
2002135795 May 2002 JP
2002171430 Jun 2002 JP
2005 109622 Apr 2005 JP
2005 303694 Oct 2005 JP
2006246193 Sep 2006 JP
2007 158825 Jun 2007 JP
2008099329 Apr 2008 JP
2010004090 Jan 2010 JP
2011 109484 Jun 2011 JP
WO 03049035 Jun 2003 WO
WO 2004021264 Mar 2004 WO
WO 2004027880 Apr 2004 WO
2006039486 Apr 2006 WO
WO 2007005714 Jan 2007 WO
WO 2008085679 Jul 2008 WO
WO 2008087652 Jul 2008 WO
WO 2009123278 Oct 2009 WO
WO 2009151903 Dec 2009 WO
WO 0022566 Apr 2010 WO
WO 2010059182 May 2010 WO
WO 2010078563 Jul 2010 WO
Non-Patent Literature Citations (13)
Entry
PCT International Search Report, PCT Application No. PCT/NL2011/050722, Apr. 16, 2012, 7 pages.
PCT Written Opinion, PCT Application No. PCT/NL2011/050722, Apr. 16, 2012, 14 pages.
PCT International Search Report, PCT Application No. PCT/NL2011/050723, Apr. 20, 2012, 6 pages.
PCT Written Opinion, PCT Application No. PCT/NL2011/050723, Apr. 20, 2012, 12 pages.
PCT International Search Report, PCT Application No. PCT/NL2011/050724, Feb. 6, 2012, 4 pages.
PCT Written Opinion, PCT Application No. PCT/NL2011/050724, Feb. 6, 2012, 5 pages.
PCT International Search Report, PCT Application No. PCT/NL2011/050725, Feb. 1, 2012, 4 pages.
PCT Written Opinion, PCT Application No. PCT/NL2011/050725, Feb. 1, 2012, 10 pages.
PCT International Search Report, PCT Application No. PCT/US2011/050726, Mar. 29, 2012, 6 pages.
PCT Written Opinion, PCT Application No. PCT/US2011/050726, Mar. 29, 2012, 10 pages.
Mirotznik, M. et al., “A Practical Enhanced-Resolution Integrated Optical-Digital Imaging Camera,” Proceedings of SPIE, Modeling and Simulation for Military Operations IV, Trevisani, D.A. (ed.), Jan. 2009, pp. 743806-1-743806-9, vol. 7348.
Horstmeyer, R. et al., “Flexible Multimodal Camera Using a Light Field Architecture,” 2009 IEEE International Conference on Computational Photography (ICCP2009), IEEE, Apr. 16, 2009, pp. 1-8.
Horisaki, R. et al., “A Compound-Eye Imaging System with Irregular Lens-Array Arrangement,” Proceedings of SPIE, Optics and Photonics for Information Processing II, Awwal, A.A.S. et al. (eds.), Aug. 13-14, 2008, pp. 70720G1-70720G-9, vol. 7072.
Related Publications (1)
Number Date Country
20130293744 A1 Nov 2013 US
Provisional Applications (1)
Number Date Country
61406148 Oct 2010 US