Many digital cameras have autofocus capability. Autofocus may be fully automatic such that the camera identifies objects in the scene and focuses on the objects. In some cases, the camera may even decide which objects are more important than other objects and subsequently focus on the more important objects. Alternatively, autofocus may utilize user input specifying which portion or portions of the scene are of interest. Based thereupon, the autofocus function identifies objects within the portion(s) of the scene, specified by the user, and focuses the camera on such objects.
To achieve market adoption, the autofocus function must be reliable and fast such that every time a user captures an image, the camera quickly brings the desired portion, or portions, of the scene into focus. Preferably, the autofocus function is sufficiently fast that the user does not notice any delay between pressing the trigger button and image capture. The autofocus is particularly important for cameras having no means for manual focus, such as compact digital cameras and camera phones.
Many digital cameras use contrast autofocus, wherein the autofocus function adjusts the imaging objective to maximize contrast in at least a portion of the scene, thus bringing that portion of the scene into focus. More recently, phase-detection autofocus has gained popularity because it is faster than contrast autofocus. Phase-detection autofocus directly measures the degree of misfocus by comparing light passing through one portion of the imaging objective, e.g., the left portion, with light passing through another portion of the imaging objective, e.g., the right portion. Some digital single-lens reflex cameras include a dedicated phase-detection sensor in addition to the image sensor that captures images.
However, this solution is not feasible for more compact and/or less expensive cameras. Therefore, camera manufacturers are developing image sensors with on-chip phase detection, i.e., image sensors with integrated phase detection capability via the inclusion of phase-detection auto-focus (PDAF) pixels in the image sensor's pixel array.
Image sensor 101 has a pixel array 200A that includes at least one PDAF pixel detector 200.
In
In
In
One indicator of the accuracy of phase-detection auto-focusing by image sensor 101, hereinafter “PDAF accuracy,” is how well the magnitude of Δx indicates the magnitude of misfocus Δz. Specifically, with reference to
In a first embodiment, an image sensor with an asymmetric-microlens PDAF detector is disclosed. The asymmetric-microlens PDAF detector includes a plurality of pixels and a microlens. The plurality of pixels forms a sub-array having at least two rows and two columns. The microlens is located above each of the plurality of pixels, and is rotationally asymmetric about an axis perpendicular to the sub-array. The axis intersects a local extremum of a top surface of the microlens.
In a second embodiment, PDAF imaging system is disclosed. The PDAF imaging system includes an image sensor and an image data processing unit. The image sensor has an asymmetric-microlens PDAF detector that includes: (a) a plurality of pixels forming a sub-array having at least two rows and two columns, and (b) a microlens located above each of the plurality of pixels and being rotationally asymmetric about an axis perpendicular to the sub-array. The axis intersects a local extremum of a top surface of the microlens. The image data processing unit is capable of receiving electrical signals from each of the plurality of pixels and generating a PDAF signal from the received electrical signals.
In a third embodiment, a method for forming a gull-wing microlens is disclosed. The method includes forming, on a substrate, a plate having a hole therein. The method also includes reflowing the plate.
Applicant has determined that PDAF accuracy depends on angular sensitivity of dual-diode PDAF pixels 200.
PDAF detector 500 is at a distance rp from the center of PDAF pixel array 500A, where rp is measured from the pixel array center to a location related to PDAF detector 500, such as optical axis 231 or interface 531. Distance rp is similar to distance 352R,
hereinafter referred to as Equation (1). For example, Eq. (1) applies at least for a singlet lens.
Design CRA χp may be defined without reference to an imaging lens. For example, PDAF detector 500 may include an opaque structure 525 that has an aperture 525A therethrough. Aperture 525A has a center axis 525A′. Design CRA χp may correspond to the propagation angle of a chief ray transmitted by symmetric microlens 230 that passes through a specific position within aperture 525A, such as through center axis 525A′. Alternatively, design CRA χp may be an angle formed by optical axis 231 and a line connecting focus 232P and a point on center axis 525A′.
Alternatively, design CRA χp may be defined with reference to edges of pixels 211 and 212. Pixel 211 has a left edge 211L. Pixel 212 has a right edge 212R. Design CRA χp may be the propagation angle of a chief ray transmitted by symmetric microlens 230 that passes through a mid-point between edges 211L and 212R. Alternatively, design CRA χp may be an angle formed by optical axis 231 and a line connecting focus 232P and a mid-point between edges 211L and 212R.
In Eq. (1), distance dpa is a characteristic distance between pixel array 200A and lens 310 along the z-axis of coordinate system 298. Herein, distances from lens 310 are referenced to a principal plane of lens 310, unless noted otherwise. Distances 311A-C of
In
Pixel responses 611 and 612 have respective peak regions 611P and 612P that are symmetric about crossing angle θx. Pixel responses 611 and 612 also have respective “valley” regions 611V and 612V. The difference in pixel response at peak regions and valley regions is indicative of the pixel's angular sensitivity. In PDAF detector 500, pixels 211-212 have respective peak-to-valley values 611PV and 612PV.
Applicant has determined that PDAF accuracy decreases as crossing angle θx deviates from zero degrees. For PDAF detector 500, crossing angle θx increases with radial distance rp (e.g., distance 352R,
Decreasing crossing angle θx may be achieved by modifying symmetric microlens 230 to minimize distance 532, such that pixel responses 611 and 612 are shifted in plot 600 to the right (e.g., by crossing angle θx) while maintaining their respective shapes. The shapes of pixel responses 611 and 612 result in part from symmetric microlens 230, which due to its being a focusing lens, imparts a quadratic phase shift (as a function of x and/or y in coordinate system 298) on light transmitted therethrough. The positions of pixel responses 611 and 612, e.g., with respect to θr=0, is determined at least partially by any linear phase shift (as a function of x and/or y in coordinate system 298) imparted by symmetric microlens 230 on incident light. As a symmetric lens, symmetric microlens 230 does not impart any such linear phase shift. The simplest optical element that imparts a linear phase is a prism, which is asymmetric and imparts only a linear phase shift on light transmitted therethrough.
Image sensor 701 has a plurality of rows 701R. A line 702 between detector-center 700C and pixel-array center 701C forms an angle 733A with respect to a line parallel to rows 701R and including pixel-array center 701C. Detector-center 700C is located at a distance 700D from pixel-array center 701C.
Color filters 721 and 722 each transmit a specified range or ranges of visible electromagnetic radiation to its associated underlying pixel. For example, visible color filters based on primary colors have pass bands corresponding to the red, green, or blue (RGB) region of the electromagnetic spectrum, and are referred to as red filters, green filters, and blue filters respectively. Visible color filters based on secondary colors have pass bands corresponding to combinations of primary colors, resulting in filters that transmit either cyan, magenta, or yellow (CMY) light, and are referred to as cyan filters, magenta filters, and yellow filters, respectively. A panchromatic color filter (Cl) transmits all colors of visible light equally. Since the transmission spectrum of a pixel's color filter distinguishes it from its neighboring pixels, a pixel is referred to by its filter type, for example, a “red pixel” includes a red filter. Herein, the transmission of a pixel refers to the transmission spectrum of its color filter.
Symmetry planes 731 and 732 may be perpendicular to each other and intersect each other at a detector-center 700C. Color filters 721 and 722 have reflection symmetry with respect to both symmetry planes 731 and 732. Symmetric multi-pixel phase-difference detector 700 also has two-fold rotational symmetry. Table 1 shows fourteen exemplary color filter configurations of symmetric multi-pixel phase-difference detectors 700, where R, G, B, C, M, Y, and Cl denote red, green, blue, cyan, magenta, yellow, and panchromatic color filters respectively. In any of the fourteen configurations, the two color filters may be switched without departing from the scope hereof. For example, in configuration (c), color filter 721 is a green filter and color filter 722 is a red filter.
While asymmetric microlens 730 is shown to not completely cover pixels 711-714 in the plan view of
Cross-sectional plane 7A-7A′ is orthogonal to the x-y plane of coordinate system 298, includes detector-center 700C, and forms angle 733A with pixel-array center 701C. Asymmetric microlens 830 extends a distance 838 past pixel 712 toward pixel-array center 701C. Distance 838 may equal zero without departing from the scope hereof.
Apex 839 is located at a distance 831 from detector center 700C. Microlens 830 is rotationally asymmetric about an axis 839A that intersects apex 839 and is perpendicular to a top surface 720T of color filter array 720.
Distance 831(i) is one indicator of the asymmetry of microlens 830. In an embodiment, image sensor 701 has a plurality of asymmetric-microlens PDAF detectors 800(i), where i=1, 2, . . . with respective detector centers 700C(i) located at a respective distances 700D(i) from pixel-array center 701C. Each PDAF detector 800(i) has a respective microlens 830(i) and surface 830S(i) having an apex 839(i) located at a respective distance 831(i) from pixel-array center 701C. Distance 831(i) is for example a monotonically increasing function of distance 700D(i). The shape of each surface 830S(i) is for example designed such that each microlens 830(i) focuses a chief-ray incident thereon on detector center 700C(i).
A second indicator of the asymmetry of microlens 830 is its tilt in cross-sectional plane 7A-7A′, as indicated by the height difference of microlens 830 at opposite sides of PDAF detector 800. PDAF detector 800 includes an inner side 802 and an outer side 804. Microlens 830 has an inner height 832 at inner edge 802 and an outer height 834 equal to zero (and hence not shown) at outer edge 804. Microlens 830 has a cross-sectional width 800W in cross-sectional plane 7A-7A. Heights 832 and 834 and width 800W determine a microlens tilt angle α. In an embodiment, image sensor 701 has a plurality of asymmetric-microlens PDAF detectors 800(i), where i=1, 2, . . . with respective detector centers 700C(i) located at a respective distances 700D(i) from pixel-array center 701C. Each PDAF detector 800(i) also has a respective microlens tilt angle α(i) that is a monotonically increasing function of distance 700D(i).
Pixel responses 1011 and 1013 have respective peak regions 1011P and 1013P and respective “valley” regions 1011V and 1013V. The difference in pixel response at peak regions and valley regions is indicative of the pixel's angular sensitivity. In PDAF detector 800, pixels 711 and 713 have respective peak-to-valley values 1011PV and 1013PV. Peak-to-valley values 1011PV and 1013PV are less than peak-to-valley values 611PV and 612PV respectively, which indicates that decreasing crossing angle θx, while beneficial, results in decreased angular sensitivity of pixels in PDAF detector.
Such decreased angular sensitivity may be overcome by adding degrees of freedom to the microlens of an asymmetric-microlens PDAF detector. For example,
Asymmetric gull-wing microlens 1130 is asymmetric about detector-center 700C between pixels 711 and 713. Asymmetric gull-wing microlens 1130 extends distance 838 past pixel 711 toward pixel-array center 701C.
Asymmetric gull-wing microlens 1130 has a planar bottom surface 1130B and a non-planar top surface 1130S. Non-planar surface 1130S includes a local minimum 1133 intersected by axis 1133B about which asymmetric gull-wing microlens 1130 is rotationally asymmetric. Asymmetric gull-wing microlens 1130 is also rotationally asymmetric about an axis perpendicular to top surface 720T and through either local maxima 1139(0) and 1139(15) of non-planar top surface 1130S.
Axis 1133B denotes a boundary at a concave region of asymmetric gull-wing microlens 1130 between two convex regions of asymmetric gull-wing microlens 1130: microlens region 1130(1) and 1130(2). Microlens region 1130(1) and 1130(2) have respective surface regions 1130S(1) and surface region 1130S(2) of top surface 1130S. Axis 1133B may be located such that microlens regions 1130(1) and 1130(2) have equal widths in cross-section 7A-7A′. Surface regions 1130S(1) and 1130S(2) have respective local maxima 1139(15) and 1139(0) located above phase-detection pixels 712 and 713 respectively, as denoted in both
Asymmetric gull-wing microlens 1130 may be symmetric about cross-sectional plane 7A-7A′, which intersects local maxima 1139(15) and 1139(0) as shown in
Local maximum 1139(15) and local minimum 1133 are located at respective distances 1131 and 1134 from detector center 700C. In an embodiment, image sensor 701 has a plurality of asymmetric-microlens PDAF detectors 1100(i), where i=1, 2, . . . with respective detector centers 700C(i) located at a respective distances 700D(i) from pixel-array center 701C. Each PDAF detector 1100(i) has a respective asymmetric gull-wing microlens 1130(i) with a local minimum 1139(i) located at a respective distance 1131(i) from pixel-array center 701C. Each asymmetric gull-wing microlens 1130(i) also has a respective local minimum 1133(i) located at a respective distance 1134(i) from pixel-array center 701C. Distances 1131(i) and 1134(i) are for example a monotonically increasing function of distance 700D(i).
Distance 1134(i) may equal zero, for example, in an asymmetric gull-wing microlens 1130 on a PDAF detectors 1100 close to pixel-array center 701C, such that asymmetric gull-wing microlens 1130 has at least one of (a) Δh=0, and (b) a surface similar to a top surface of a ring torus, a horn torus, or a spindle torus. When both (a) and (b) apply, such an asymmetric gull-wing microlens 1130 is rotationally symmetric about its axis 1133B, and hence is only nominally “asymmetric.” PDAF detectors 1100 close to pixel-array center 701C are for example those that include only pixels of pixel array 700A closer to pixel-array center 701C than ninety percent of all pixels of pixel array 700A.
While curve h(θ) is symmetric about θ=360°, it may be asymmetric about θ=360° without departing from the scope hereof.
Whereas microlens 830, in cross-section, may be viewed as a single hypothetical plano-convex lens on top of a prism, microlens 1130, in cross-section, may be viewed as two plano-convex lenses on top of a prism.
Surface regions 1130S(1,2) have respective best-fit radii of curvature R1 and R2 such that microlens regions 1130(1,2) may have different respective focal lengths f1 and f2 determined by R1 and R2, respectively, and the refractive index of asymmetric gull-wing microlens 1130. Local minimum 1133 and adjacent surface regions 1130S(1, 2) provide asymmetric gull-wing microlens 1130 with additional degrees of freedom, compared to microlens 830, for optimizing pixel responses of pixels 711 and 712 as a function of angle θr. For example, while asymmetric gull-wing microlens 1130 may be optimized to impose an appropriate linear phase shift on light transmitted therethrough to decrease crossing angle θx to zero while minimizing a loss in angular sensitivity, for example, such that peak-to-valley values 1011PV and 1013PV are closer to 611PV and 612PV, respectively.
In step 1410, method 1400 forms, on a substrate, a plate having a hole therein. The plate may be formed via a photolithography. In an example of step 1410, a plate 1520 is formed on a top surface 1510T of a substrate 1510, as shown in
Plate 1520 has a top surface 1520T and a hole 1521 therein. Hole 1521 has a perimeter 1521P, the average height of which is located at a height 1521H above substrate top surface 1510T. Hole 1521 extends to a depth 1521D toward substrate 1510. Depth 1521D may extend to top surface 1510T. Hole 1521 is a through hole that exposes a portion of top surface 1510T. Plate top surface 1520T may be planar and parallel to substrate top surface 1510T, for example, to form an asymmetric gull-wing microlens 1130 with Δh=0. Alternatively, at least part of top plate surface 1520T may be a nonparallel to substrate top surface 1510T, for example, to form an asymmetric gull-wing microlens 1130 with Δh≠0, as illustrated in
Substrate 1510 may be above an image sensor pixel array. For example, substrate 1510 is color-filter array 720 or a layer thereon, plate 1520 covers pixels 711-714 of
In step 1420, method 1400 reflows the plate. In an example of step 1420, plate 1520 is reflowed to yield asymmetric gull-wing microlens 1130. Step 1420 may include reflowing the plate with a spatially-varying reflow temperature beneath plate top surface 1520T, and result in asymmetric microlens 1130 with spatially-varying heights 1135(i). Plate 1520 is for example formed of a positive photoresist having a glass transition temperature Tg between the aforementioned Tmin and Tmax, which enables reflow behavior that is sufficiently stable to form asymmetric gull-wing microlens 1130. Accordingly, the reflow of step 1420 may include step 1422, in which the plate is heated to a temperature between Tmin and Tmax.
In an exemplary mode of operation, imaging objective 310 form an image of a scene 1691 on image sensor 701. Pixel array 700A and imaging objective are separated by distance 312Z, illustrated in
Combinations of Features
Features described above as well as those claimed below may be combined in various ways without departing from the scope hereof. The following examples illustrate some possible, non-limiting combinations:
(A1) An image sensor includes an asymmetric-microlens PDAF detector. The asymmetric-microlens PDAF detector includes a plurality of pixels and a microlens. The plurality of pixels forms a sub-array having at least two rows and two columns. The microlens is located above each of the plurality of pixels, and is rotationally asymmetric about an axis perpendicular to the sub-array. The axis intersects a local extremum of a top surface of the microlens.
(A2) The image sensor denoted by (A1) is capable of capturing an image formed thereon, and may further include an image data processing unit capable of (i) receiving electrical signals from each of the plurality of pixels, and (ii) outputting data associated with misfocus of the image.
(A3) In an image sensor denoted by one of (A1) and (A2), the first microlens may have, in a cross-sectional plane intersecting the first microlens and perpendicular to a top surface of the first sub-array, a height profile with more than one local maximum
(A4) In the image sensor denoted by (A3), the first sub-array may be a two-by-two planar array, the height profile having two local maxima.
(A5) An image sensor denoted by one of (A1) through (A4) may further include a second asymmetric-microlens PDAF detector that has a plurality of second pixels and a second microlens. The second plurality of pixels forms a second sub-array having at least two rows and two columns. The second microlens is located above each of the second plurality of pixels and is rotationally asymmetric about a second axis perpendicular to the second sub-array and intersecting a second local extremum of a top surface of the second microlens. The first and second plurality of pixels are part of a pixel array having a pixel-array top surface and a pixel-array center. The second asymmetric-microlens PDAF detector is further from the pixel-array center than the first asymmetric-microlens PDAF detector. A location on the pixel-array top surface directly beneath the first local extremum is at a first distance from a center of the first sub-array. A location on the pixel-array top surface directly beneath the second local extremum is at a second distance from a center of the second sub-array, the second distance exceeding the first distance.
(A6) In an image sensor denoted by one of (A1) through (A5), in which the first plurality of pixels each have a respective color filter thereon, each color filter having a transmission spectrum, the sub-array, by virtue of the transmission spectrum of each color filter, may have reflection symmetry with respect to a center of the pixel sub-array.
(B1) A PDAF imaging system includes an image sensor and an image data processing unit. The image sensor has an asymmetric-microlens PDAF detector that includes: (a) a plurality of pixels forming a sub-array having at least two rows and two columns, and (b) a microlens located above each of the plurality of pixels and being rotationally asymmetric about an axis perpendicular to the sub-array. The axis intersects a local extremum of a top surface of the microlens. The image data processing unit is capable of receiving electrical signals from each of the plurality of pixels and generating a PDAF signal from the received electrical signals.
(B2) The PDAF imaging system denoted by (B1) may further include an autofocus module capable of receiving the PDAF signal and generating a misfocus signal indicative of a degree of misfocus between the image sensor and an imaging lens that has an optical axis intersecting the image sensor.
(B3) In a PDAF imaging system denoted by one of (B1) and (B2), the first microlens may have, in a cross-sectional plane intersecting the first microlens and perpendicular to a top surface of the first sub-array, a height profile having more than one local maximum.
(B4) In the PDAF imaging system denoted by (B3), the first sub-array may be a two-by-two planar array, and the height profile may have two local maxima.
(B5) A PDAF imaging system denoted by one of (B1) through (B4) may further include the second asymmetric-microlens PDAF detector of the image sensor denoted by (A5).
(B6) In an image sensor denoted by one of (B1) through (B5), in which the first plurality of pixels each have a respective color filter thereon, each color filter having a transmission spectrum, the sub-array, by virtue of the transmission spectrum of each color filter, may have reflection symmetry with respect to a center of the pixel sub-array.
(C1) A method for forming a gull-wing microlens includes forming, on a substrate, a plate having a hole therein. The method also includes reflowing the plate.
(C2) In the method denoted by (C1), the step of reflowing may include heating the plate to a temperature between 140° C. and 180° C.
(C3) In a method denoted by one of (C1) and (C2), the hole may be a through hole
(C4) In a method denoted by one of (C1) through (C3), (a) the substrate may be above a pixel array of an image sensor, (b) the plate may cover a two-by-two array of pixels of the pixel array, and (c) the pixel-array center, the hole, and the center of the two-by-two array of pixels may be collinear.
(C5) In a method denoted by one of (C1) through (C4), in which the plate is formed on a surface of the substrate, the plate may have a top surface that is not parallel to the surface of the substrate.
Changes may be made in the above methods and systems without departing from the scope hereof. It should thus be noted that the matter contained in the above description or shown in the accompanying drawings should be interpreted as illustrative and not in a limiting sense. The following claims are intended to cover all generic and specific features described herein, as well as all statements of the scope of the present method and system, which, as a matter of language, might be said to fall therebetween.
Number | Name | Date | Kind |
---|---|---|---|
20060170810 | Kim | Aug 2006 | A1 |
20060273239 | McGrath | Dec 2006 | A1 |
20070014019 | Mouli | Jan 2007 | A1 |
20070181923 | Tanaka | Aug 2007 | A1 |
20080090182 | Suzuki | Apr 2008 | A1 |
20080151378 | Kim | Jun 2008 | A1 |
20090174945 | Lin | Jul 2009 | A1 |
20100045849 | Yamasaki | Feb 2010 | A1 |
20130341493 | Ando | Dec 2013 | A1 |
20140055661 | Imamura | Feb 2014 | A1 |
20160071896 | Kawabata | Mar 2016 | A1 |
20160073077 | Kurahashi | Mar 2016 | A1 |
20160178913 | Endoh | Jun 2016 | A1 |
20160181309 | Uehira | Jun 2016 | A1 |
20160233259 | Sekine | Aug 2016 | A1 |
20160334621 | Kawabata | Nov 2016 | A1 |
20170041525 | Liu | Feb 2017 | A1 |
20170077164 | Kawabata | Mar 2017 | A1 |
20170104020 | Lee | Apr 2017 | A1 |
20170338258 | Kim | Nov 2017 | A1 |
20170347042 | Borthakur | Nov 2017 | A1 |
Number | Date | Country |
---|---|---|
2015-153975 | Aug 2015 | JP |
Entry |
---|
Cox C. et al., Microlens formation using heavily dyed photoresist in a single step, Copyright 2006 Society of Photo-Optical Instrumentation Engineers, 7 pages. |
Krivoshapko, S.N., et al. Classification of Cyclic Surfaces and Geometrical Research of Canal Surfaces, IJRRAS 12 (3) Sep. 2012, pp. 360-374. |
Wei et al., Real-time Observation for the Formation of Microlens Arrays Fabricated Using Thermal Reflow Process, Tamkang Journal of Science and Engineering, vol. 7, No. 2, pp. 81-86 (2004), pp. 81-86. |
Number | Date | Country | |
---|---|---|---|
20180084185 A1 | Mar 2018 | US |