This relates generally to imaging systems and, more particularly, to imaging systems with phase detection capabilities.
Modern electronic devices such as cellular telephones, cameras, and computers often use digital image sensors. Imager sensors (sometimes referred to as imagers) may be formed from a two-dimensional array of image sensing pixels. Each pixel receives incident photons (light) and converts the photons into electrical signals. Image sensors are sometimes designed to provide images to electronic devices using a Joint Photographic Experts Group (JPEG) format.
Some applications such as automatic focusing and three-dimensional (3D) imaging may require electronic devices to provide stereo and/or depth sensing capabilities. For example, to bring an object of interest into focus for an image capture, an electronic device may need to identify the distances between the electronic device and object of interest. To identify distances, conventional electronic devices use complex arrangements. Some arrangements require the addition of lenticular arrays that focus incident light on sub-regions of a two-dimensional pixel array. However, these arrangements may lead to reduced spatial resolution, reduced color fidelity, increased cost, and increased complexity.
It would therefore be desirable to be able to provide improved imaging systems with depth sensing capabilities.
Embodiments of the present invention relate to image sensors with phase detection capabilities. An electronic device with a digital camera module is shown in
Still and video image data from image sensor(s) 14 may be provided to image processing and data formatting circuitry 16 via path 26. Image processing and data formatting circuitry 16 may be used to perform image processing functions such as automatic focusing functions, depth sensing, data formatting, adjusting white balance and exposure, implementing video image stabilization, face detection, etc. For example, during automatic focusing operations, image processing and data formatting circuitry 16 may process data gathered by phase detection pixels in image sensor 14 to determine the magnitude and direction of lens movement (e.g., movement of lens 28) needed to bring an object of interest into focus.
Image processing and data formatting circuitry 16 may also be used to compress raw camera image files if desired (e.g., to Joint Photographic Experts Group or JPEG format). In a typical arrangement, which is sometimes referred to as a system on chip (SOC) arrangement, camera sensor 14 and image processing and data formatting circuitry 16 are implemented on a common integrated circuit. The use of a single integrated circuit to implement camera sensor 14 and image processing and data formatting circuitry 16 can help to reduce costs. This is, however, merely illustrative. If desired, camera sensor 14 and image processing and data formatting circuitry 16 may be implemented using separate integrated circuits. If desired, camera sensor 14 and image processing circuitry 16 may be formed on separate semiconductor substrates. For example, camera sensor 14 and image processing circuitry 16 may be formed on separate substrates that have been stacked.
Camera module 12 may convey acquired image data to host subsystems 20 over path 18 (e.g., image processing and data formatting circuitry 16 may convey image data to subsystems 20). Electronic device 10 typically provides a user with numerous high-level functions. In a computer or advanced cellular telephone, for example, a user may be provided with the ability to run user applications. To implement these functions, host subsystem 20 of electronic device 10 may include storage and processing circuitry 24 and input-output devices 22 such as keypads, input-output ports, joysticks, and displays. In certain embodiments, input-output devices 22 may include an infrared light source such as an infrared LED. Storage and processing circuitry 24 may include volatile and nonvolatile memory (e.g., random-access memory, flash memory, hard drives, solid-state drives, etc.). Storage and processing circuitry 24 may also include microprocessors, microcontrollers, digital signal processors, application specific integrated circuits, or other processing circuits.
It may be desirable to provide image sensors with depth sensing capabilities (e.g., to use in automatic focusing applications, 3D imaging applications such as machine vision applications, etc.). To provide depth sensing capabilities, image sensor(s) 14 may include phase detection pixel groups. Image sensor(s) 14 may include phase detection pixels such as phase detection pixel group 100 shown in
Color filters such as color filter elements 104 may be interposed between microlens 102 and substrate 108. Color filter elements 104 may filter incident light by only allowing predetermined wavelengths to pass through color filter elements 104 (e.g., color filter 104 may only be transparent to the wavelengths corresponding to a green color, a red color, a blue color, a yellow color, a cyan color, a magenta color, visible light, infrared light, etc.). Color filter 104 may be a broadband color filter. Examples of broadband color filters include yellow color filters (e.g., yellow color filter material that passes red and green light) and clear color filters (e.g., transparent material that passes red, blue, and green light). In general, broadband filter elements may pass two or more colors of light. If desired, no color filter element may be provided and the photodiodes may receive unfiltered light. Photodiodes PD1 and PD2 may serve to absorb incident light focused by microlens 102 and produce pixel signals that correspond to the amount of incident light absorbed.
Photodiodes PD1 and PD2 may each cover approximately half of the substrate area under microlens 102 (as an example). By only covering half of the substrate area, each photosensitive region may be provided with an asymmetric angular response (e.g., photodiode PD1 may produce different image signals based on the angle at which incident light reaches pixel pair 100). The angle at which incident light reaches pixel pair 100 relative to a normal axis 116 (i.e., the angle at which incident light strikes microlens 102 relative to the optical axis 116 of lens 102) may be herein referred to as the incident angle or angle of incidence.
An image sensor can be formed using front side illumination imager arrangements (e.g., when circuitry such as metal interconnect circuitry is interposed between the microlens and photosensitive regions) or backside illumination imager arrangements (e.g., when photosensitive regions are interposed between the microlens and the metal interconnect circuitry). The example of
In the example of
The positions of photodiodes PD1 and PD2 may sometimes be referred to as asymmetric or displaced positions because the center of each photosensitive area 110 is offset from (i.e., not aligned with) optical axis 116 of microlens 102. Due to the asymmetric formation of individual photodiodes PD1 and PD2 in substrate 108, each photosensitive area 110 may have an asymmetric angular response (e.g., the signal output produced by each photodiode 110 in response to incident light with a given intensity may vary based on an angle of incidence). It should be noted that the example of
Line 160 may represent the output image signal for photodiode PD2 whereas line 162 may represent the output image signal for photodiode PD1. For negative angles of incidence, the output image signal for photodiode PD2 may increase (e.g., because incident light is focused onto photodiode PD2) and the output image signal for photodiode PD1 may decrease (e.g., because incident light is focused away from photodiode PD1). For positive angles of incidence, the output image signal for photodiode PD2 may be relatively small and the output image signal for photodiode PD1 may be relatively large.
The size and location of photodiodes PD1 and PD2 of pixel pair 100 of
Output signals from pixel pairs such as pixel pair 100 may be used to adjust the optics (e.g., one or more lenses such as lenses 28 of
For example, by creating pairs of pixels that are sensitive to light from one side of the lens or the other, a phase difference can be determined. This phase difference may be used to determine both how far and in which direction the image sensor optics should be adjusted to bring the object of interest into focus.
When an object is in focus, light from both sides of the image sensor optics converges to create a focused image. When an object is out of focus, the images projected by two sides of the optics do not overlap because they are out of phase with one another. By creating pairs of pixels where each pixel is sensitive to light from one side of the lens or the other, a phase difference can be determined. This phase difference can be used to determine the direction and magnitude of optics movement needed to bring the images into phase and thereby focus the object of interest. Pixel blocks that are used to determine phase difference information such as pixel pair 100 are sometimes referred to herein as phase detection pixels or depth-sensing pixels.
A phase difference signal may be calculated by comparing the output pixel signal of PD1 with that of PD2. For example, a phase difference signal for pixel pair 100 may be determined by subtracting the pixel signal output of PD1 from the pixel signal output of PD2 (e.g., by subtracting line 162 from line 160). For an object at a distance that is less than the focused object distance, the phase difference signal may be negative. For an object at a distance that is greater than the focused object distance, the phase difference signal may be positive. This information may be used to automatically adjust the image sensor optics to bring the object of interest into focus (e.g., by bringing the pixel signals into phase with one another).
In order to improve phase detection pixel group 100, phase detection pixel group 100 may include pedestal 105, as shown in
In certain embodiments, an imaging system may include more than one image sensor, only one of which includes phase detection pixels. An example of an embodiment of this type is shown in
At least one of the image sensors may include phase detection pixels. As shown in
The phase detection auto focus algorithm may be calibrated during assembly to account for the differences between lens modules 28-1 and 28-2. This way, both image sensors 14-1 and 14-2 can be focused using the phase detection data.
Importantly, it should be noted that image sensor 14-2 may not include any phase detection pixels. Although image sensor 14-2 does not include phase detection pixels, the phase detection data from image sensor 14-1 may be used to generate focus feedback that adjusts lens module 28-2. This concept of using phase detection data from a first image sensor to help focus the lens module of a second image sensor can be used to implement imaging systems that focus quickly.
Image sensor 14-1 may be a monochrome sensor, while image sensor 14-2 may be a color sensor. The monochrome sensor may include pixels of only one color. For example, image sensor 14-1 may include pixels with no color filtration. There may be no color filter material included in image sensor 14-1 at all, or image sensor 14-1 may include exclusively clear or white color filter elements. Alternatively, image sensor 14-1 may include color filter material that is configured to filter a certain color of visible light (e.g., red, green, blue, etc.), infrared light, or ultraviolet light. Image sensor 14-2, on the other hand, may include color filter elements of different colors. Image sensor 14-2 may include, for example, blue, red, and green color filter elements that are arranged according to a Bayer color filter pattern. Other colors or color filter patterns may be used in image sensor 14-2 if desired.
Including phase detection pixels in a color sensor typically necessitates using color correction algorithms in order to account for the unique structure of the phase detection pixels. By including the phase detection pixels on only the monochrome sensor, the camera module shown in
Another advantage of a monochrome sensor with phase detection pixels is that the monochrome sensor allows for maximum light input, which results in optimal low-light focusing. Additionally, the color sensor can take advantage of the phase detection data from the monochrome sensor to have a similarly high responsivity in all light conditions.
In addition to being used for phase detection applications, monochrome sensor 14-1 may be used for imaging applications. For example, monochrome sensor 14-1 may have some phase detection pixel groups and some imaging pixels. In addition to using phase detection data from the phase detection pixel groups for focusing purposes, the imaging pixels may be used for imaging purposes. One example of this type of application is when a camera module of the type shown in
In embodiments where image sensor 14-1 is a monochrome infrared or near-infrared sensor, the imaging system may also include infrared or near-infrared light sources that are configured to emit infrared or near-infrared light. The light source may be an infrared LED, for example. In another embodiment, sensor 14-1 may be a monochrome ultraviolet sensor and an ultraviolet light source may be included in the imaging system.
As shown in
Signals from phase detection pixels may also be binned if desired. For example, if it is desired to use monochrome sensor 14-1 for imaging purposes, the signals from each pixel in a 2×2 group of pixels may be binned or summed. If each 2×2 group of pixels is binned, the data can be used for imaging purposes. In general, the data from the phase detection pixels or imaging pixels can be binned in any desired manner for any desired purpose.
Including phase detection pixel groups across the entire array (as shown in
Examples of numerous phase detection pixel groups have been described above.
Additionally, phase detection pixel groups may be included where more than one microlens covers the pixels. For example, three adjacent pixels in a 1×3 group may make up a phase detection pixel group. Instead of a single microlens covering all three pixels, two microlenses may each cover approximately 1.5 pixels. In yet another embodiment, phase detection pixels may have various sub-pixels, such as an inner sub-pixel that is nested within an outer sub-pixel. Microlenses of any shape can be used in phase detection pixel groups (e.g., circular, elliptical, toroidal, etc.). In general, image sensor 14-1 may include any pixel group capable of generating phase detection data. The phase detection data may then be used to focus lens modules 28-1 and 28-2.
In various embodiments of the present invention, an imaging system may include a first image sensor that includes phase detection pixels, a first lens module that is configured to focus light on the first image sensor, a second image sensor that does not include phase detection pixels, a second lens module that is configured to focus light on the second image sensor, and processing circuitry that is configured to adjust the second lens module based on phase detection data from the phase detection pixels.
The first image sensor may be a monochrome image sensor, and the second image sensor may be a color image sensor. The monochrome image sensor may be configured to detect white light. The monochrome image sensor may include a color filter material that covers all of the pixels in the monochrome image sensor. The color filter material may be configured to pass light of a given type, wherein the given type is selected from the group consisting of: visible light, infrared light, near-infrared light, and ultraviolet light. The phase detection pixels may be organized in phase detection pixel groups, and each phase detection pixel group may include adjacent pixels covered by a single microlens. The single microlens may be formed on a pedestal. At least one phase detection pixel group may include a shielding element that covers portions of underlying pixels. The phase detection pixels may be organized in phase detection pixel groups, and each phase detection pixel group may include a group of pixels covered by a single microlens, and each group of pixels may be selected from the group consisting of: a 1×2 group, a 1×3 group, a 2×2 group, a 2×4 group, a 3×3 group, and a 4×4 group.
A method of operating an imaging system that includes at least one monochrome sensor with phase detection pixels and at least one color sensor may include generating phase detection pixel data with the phase detection pixels, and adjusting a first lens based on the phase detection pixel data. The first lens may be positioned above the at least one color sensor. The method may also include adjusting a second lens based on the phase detection pixel data. The second lens may be positioned above the at least one monochrome sensor. The at least one color sensor may not include any phase detection pixels. The method may also include generating image data using the at least one monochrome sensor.
An imaging system may include a monochrome image sensor, a first lens module that is configured to focus light on the monochrome image sensor, a color image sensor, and a second lens module that is configured to focus light on the color image sensor. The monochrome image sensor may include phase detection pixels, the color image sensor may include imaging pixels, the color image sensor may not include phase detection pixels, the color image sensor may include a color filter array with a plurality of color filter elements, and the plurality of color filter elements may include at least color filter elements of a first color and color filter elements of a second color that is different than the first color.
The imaging system may also include processing circuitry that is configured to receive data from the monochrome image sensor and color image sensor. The processing circuitry may be configured to adjust the second lens module based on phase detection data from the phase detection pixels. The processing circuitry may be configured to adjust the first lens module based on the phase detection data from the phase detection pixels. The phase detection pixels may include at least first and second pixels covered by a single microlens. The plurality of color filter elements may be arranged according to a Bayer color filter pattern.
The foregoing is merely illustrative of the principles of this invention and various modifications can be made by those skilled in the art without departing from the scope and spirit of the invention.