The present disclosure relates to an imaging device incorporating an interference enhanced diffractive grating to enable the identification of incident light states.
Digital image sensors are commonly used in a variety of electronic devices, such as handheld cameras, medical devices, security systems, telephones, computers, and tablets, to capture images. In a typical arrangement, light sensitive areas or pixels are arranged in a two-dimensional array having multiple rows and columns of pixels. Each pixel generates an electrical charge in response to receiving photons as a result of being exposed to incident light. For example, each pixel can include a photodiode that generates charge in an amount that is generally proportional to the amount of light (i.e. the number of photons) incident on the pixel during an exposure period. The charge can then be read out from each of the pixels, for example through peripheral circuitry.
In conventional color image sensors, absorptive color filters are used to enable the image sensor to detect the color of incident light. The color filters are typically disposed in sets (e.g. of red, green, and blue (RGB); cyan, magenta, and yellow (CMY); or red, green, blue, and infrared (RGBIR)). Such arrangements have about 3-4 times lower sensitivity and signal to noise ratio (SNR) at low light conditions, color crosstalk, color shading at high chief ray angles (CRA), and lower spatial resolution due to color filter patterning, resulting in lower spatial frequency as compared to monochrome sensors without color filters. However, the image information provided by a monochrome sensor does not include information about the color of the imaged object.
In addition, conventional color and monochrome image sensors incorporate non-complementary metal-oxide semiconductor (CMOS), polymer-based materials, for example to form filters and micro lenses for each of the pixels, that result in image sensor fabrication processes that are more time-consuming and expensive than processes that only require CMOS materials. Moreover, the resulting devices suffer from compromised reliability and operational life, as the included color filters and micro lenses are subject to weathering and performance that degrades at a much faster rate than inorganic CMOS materials. In addition, the processing required to interpolate between pixels of different colors in order to produce a continuous image is significant.
Image sensors have been developed that utilize uniform, non-focusing metal gratings or grids, that attenuate light outside of a selected polarization direction, before that light is absorbed in a silicon substrate. Such an approach enables the polarization characteristics of incident light to be determined, without requiring the use of absorptive filters. However, the non-focusing diffractive grating results in light loss before the light reaches the substrate. As a result, the quantum efficiency of such sensors in very low. Such an approach also requires an adjustment or shift in the microlens and the grating position and structures across the image plane to accommodate high chief ray angles (CRAs).
Accordingly, it would be desirable to provide an improved image sensor that was capable of sensing characteristics of incident light, including the color and polarization state of the light.
Embodiments of the present disclosure provide image sensors, image sensing methods, and methods for in-pixel interferometric recording of an image of a light field using diffractive grating and scattering elements, referred to herein as diffractive elements. The provided interferometric imaging enables information about the light field (light state) incident at the pixels to be determined. Characteristics of the light that can be decoded by embodiments of the present disclosure include color or spectral information; polarization; phase; divergence; and super-resolution information. Such information can be obtained in connection with conventional imaging operations, or in connection with imaging complex light fields and their characteristics.
An image sensor in accordance with embodiments of the present disclosure includes a plurality of pixels disposed across an imaging surface. Each of the pixels includes a plurality of sub-pixels. A diffraction layer is disposed adjacent a light incident surface side of the image sensor. The diffraction layer includes a set of electrically conductive or semi-conductive diffraction elements or features for each pixel in the plurality of pixels. The diffraction features scatter incident light onto the sub-pixels. The diffraction pattern produced across the area of the pixel by the diffraction features is dependent on the light state of the incident light. For example, the light state measured or determined by embodiments of the present disclosure can include the wavelength, polarization state, phase and amplitude of incident light. In accordance with the least some embodiments of the present disclosure, the light state of the light incident on a pixel can be determined from ratios of relative signal intensities at each of the sub-pixels within the pixel. Accordingly, embodiments of the present disclosure provide an image sensor that does not require color filters in order to identify the wavelength of incident light. In addition, embodiments of the present disclosure do not require micro lenses or infrared filters in order to provide high resolution images and high resolution color identification. Furthermore, embodiments of the present disclosure provide an image sensor that can determine a polarization state of incident light from ratios of relative signal intensities at each of the sub-pixels within the pixel. Because embodiments of the present disclosure allow the polarization state to be determined from the scattering of light achieved by the included scattering elements, without requiring the use of polarization grids, sensitivity and resolution can be improved as compared to conventional polarization sensing devices. The resulting light state image sensor has high sensitivity, high spatial resolution, high color resolution, wide spectral range, the ability to detect a number of linear and circular polarization states, the ability to detect the phase of incident light, a low stack height, and can be manufactured using conventional CMOS processes.
An imaging device or apparatus in accordance with embodiments of the present disclosure incorporates an image sensor having a diffraction layer on a light incident side of a sensor substrate. The sensor substrate includes an array of pixels, each of which includes a plurality of light sensitive areas or sub-pixels. The diffraction layer includes a set of electrically conductive or semi-conductive diffraction features for each pixel. The diffraction features can include linear elements particularly configured for the detection of different light polarizations. In accordance with the least some embodiments of the present disclosure, the linear elements are disposed to extend radially from a centerline of the associated pixel, and increase in length from one line to the next in one of a clockwise or anticlockwise direction. The diffraction features can additionally include scattering elements. In accordance with the least some embodiments of the present disclosure, the diffraction features are sized such that an area of each sub-pixel covered by the diffraction features is the same or about the same. Accordingly, the diffraction features operate as diffractive pixel micro lenses, which create asymmetric diffractive light patterns that are strongly dependent on the color and polarization of incident light. Because the diffraction layer is relatively thin (e.g. about 500 nm or less), it provides a very high coherence degree for the incident light, which facilitates the formation of stable interference patterns.
The relative distribution of the incident light amongst the sub-pixels of a pixel is determined by comparing the signal ratios. For example, in a configuration in which each pixel includes a 2×2 array of sub-pixels, there are 6 possible combinations of sub-pixel signal ratios that can be used to identify the color of light incident at the pixel with very high accuracy. In particular, because the interference pattern produced by the diffraction elements strongly correlates with the color and polarization of the incident light, the incident light color and polarization can be identified with very high accuracy (e.g. within 25 nm or less). The identification or assignment of the color and polarization of the incident light from the ratios of signals produced by the sub-pixels can be determined by comparing those ratios to pre-calibrated sub-pixel photodiode signal ratios (attributes) for different light states, such as color and polarization, of incident light.
An imaging device or apparatus incorporating an image sensor in accordance with embodiments of the present disclosure can include an imaging lens that focuses collected light onto an image sensor. The light from the lens is focused and diffracted onto pixels included in the image sensor by electrically conductive or semi-conductive diffraction features. More particularly, each pixel includes a plurality of sub-pixels, and is associated with a set of diffraction features. The diffraction features function to create an asymmetrical diffraction pattern across the sub-pixels. Differences in the strength of the signals at each of the sub-pixels within a pixel can be applied to determine a light state of the light incident on the pixel.
Imaging sensing methods in accordance with embodiments of the present disclosure include focusing light collected from within a scene onto an image sensor having a plurality of pixels disposed in an array. The light incident on each pixel is focused and diffracted by a set of diffraction features onto a plurality of included sub-pixels. The diffraction pattern produced by the diffraction features depends on the light state of the incident light. Accordingly, the amplitude of the signal generated by the incident light at each of the sub-pixels in each pixel can be read to determine the light state of that incident light. In accordance with embodiments of the present disclosure, the assignment of a light state to light incident on a pixel includes determining ratios of signal strengths produced by sub-pixels within the pixel, and comparing those ratios to values stored in a lookup table for light state assignment. An image sensor produced in accordance with embodiments of the present disclosure therefore does not require micro lenses for each pixel or color filters, and provides high light state sensitivity over a range of wavelengths that can be coincident with the full wavelength sensitivity of the image sensor pixels.
Additional features and advantages of embodiments of the present disclosure will become more readily apparent from the following description, particularly when considered together with the accompanying drawings.
The control circuit 132 can receive data for instructing an input clock, an operation mode, and the like, and can output data such as internal information related to the image sensor 100. Accordingly, the control circuit 132 can generate a clock signal that provides a standard for operation of the vertical drive circuit 116, the column signal processing circuit 120, and the horizontal drive circuit 124, and control signals based on a vertical synchronization signal, a horizontal synchronization signal, and a master clock. The control circuit 132 outputs the generated clock signal in the control signals to the various other circuits and components.
The vertical drive circuit 116 can, for example, be configured with a shift register, can operate to select a pixel drive wiring 136, and can supply pulses for driving sub-pixels of a pixel 104 through the selected drive wiring 136 in units of a row. The vertical drive circuit 116 can also selectively and sequentially scan elements of the array 108 in units of a row in a vertical direction, and supply the signals generated within the pixels 104 according to an amount of light they have received to the column signal processing circuit 120 through a vertical signal line 140.
The column signal processing circuit 120 can operate to perform signal processing, such as noise removal, on the signal output from the pixels 104. For example, the column signal processing circuit 120 can perform signal processing such as a correlated double sampling (CDS) for removing a specific fixed patterned noise of a selected pixel 104 and an analog to digital (A/D) conversion of the signal.
The horizontal drive circuit 124 can include a shift register. The horizontal drive circuit 124 can select each column signal processing circuit 120 in order by sequentially outputting horizontal scanning pulses, causing each column signal processing circuit 122 to output a pixel signal to a horizontal signal line 144.
The output circuit 128 can perform predetermined signal processing with respect to the signals sequentially supplied from each column signal processing circuit 120 through the horizontal signal line 144. For example, the output circuit 128 can perform a buffering, black level adjustment, column variation correction, various digital signal processing, and other signal processing procedures. An input and output terminal 148 exchanges signals between the image sensor 100 and external components or systems.
Accordingly, a light state image sensor 100 in accordance with at least some embodiments of the present disclosure can be configured as a CMOS image sensor of a column A/D type in which column signal processing is performed.
With reference now to
With reference now to
Each pixel 104 in a light state image sensor 100 in accordance with embodiments of the present disclosure includes a plurality of sub-pixels 604. The sub-pixels 604 within a pixel 104 generally include adjacent photoelectric conversion elements or areas within the image sensor substrate 112. In operation, each sub-pixel 604 generates a signal in proportion to an amount of light incident thereon. As an example, each sub-pixel 604 is a photodiode. As represented in
As previously discussed, each pixel 104 is associated with a set of diffraction features or elements 524. The diffraction elements 528 within a set 524 can be centered on a line or axis that is perpendicular to the light incident surface of the image sensor 100, and that is coincident with or parallel to a centerline 612 extending from the geometric center of the light incident area of the pixel 104. In accordance with further embodiments of the present disclosure, the diffraction features 528 within a set 524 can be centered along a line that extends from the center of the pixel 104 area, and that follows or is parallel to a chief ray angle associated with the relevant pixel 104. In accordance with the least some embodiments of the present disclosure, the set of diffraction features 524 associated with a pixel 104 include a central feature 704 that is centered on or adjacent the centerline 612 of the pixel 104. As an example, but without limitation, the central feature 704 can be configured as a cylindrical element. Moreover, the centerline 612 associated with the pixel 104 area and the central feature 704 can be disposed on or adjacent a line intersecting a point that is centered between the light incident surfaces 608 of the sub-pixels 604.
In accordance with embodiments of the present disclosure, the set of diffraction features 524 includes a number of elongated or linear elements 708. As examples, but without limitation, the linear elements 708 are configured in a rectangular shape. In accordance with embodiments of the present disclosure, the linear elements 708 are disposed along or parallel to lines extending radially from the centerline 612 of the pixel 104 when the pixel 104 is seen in a plan view. In accordance with embodiments of the present disclosure, the radially extending linear elements 708 can be disposed at equal intervals about the centerline 612. For example, but without limitation, and as illustrated in
In accordance with still further embodiments of the present disclosure, the sets of diffraction elements 524 can also include a plurality of scattering elements 712. The scattering elements 712 can be configured similarly to the central element 704. Accordingly, as an example, each scattering element 712 can be configured as a cylinder, with the ends of the cylinder parallel to the light incident surface of the pixel 104. The scattering elements 712 can be disposed in areas generally between adjacent linear elements 708. For instance, and as illustrated in
With reference now to
As depicted in
The electrically conductive or semiconductive diffraction elements 528 enable the modulation of light amplitude and phase incident on a pixel 104 with a relatively small amount of attenuation. In particular, the attenuation of incident light is much less than in a sensor using a conventional polarization grid. The linear diffraction elements 708 in accordance with embodiments of the present disclosure have their longitudinal axes disposed along lines radiating from the centerline 612 of the pixel 104 that are spaced apart from one another by 45°. As can be appreciated by one of skill in the art after consideration of the present disclosure, light polarized such that its oscillating electric field is parallel to an electrically conductive or semi-conductive material of a linear diffraction element 708 is strongly attenuated. Accordingly, incident light having different polarization states will produce different interference patterns across the sub-pixels 604 of the pixel 104. In addition, the inclusion of scattering elements 712 enables areas of coverage of the diffraction elements 528 with respect to the sub-pixels 604 to be balanced. For example, in accordance with the least some embodiments of the present disclosure, the areas of coverage of the diffraction elements 528 with respect to any one sub-pixel 604 is the same as any other sub-pixel 604 within the pixel 104, ±20%. In accordance with further embodiments of the present disclosure the areas of coverage can be within about ±10% of one another. The combination of a length of linear diffraction elements 708 that changes with radial position about the centerline 612 of the pixel 104, and scattering elements 712 allow additional light states, such as non-linearly polarized light states, to be detected. For example, left and right circularly polarized light will produce different interference patterns, which can be recognized from the different sub-pixel 604 relative signal combinations that will be produced.
The different speckle patterns produced by light interference interactions with the diffraction elements 528 strongly correlates with the illumination conditions, such as light convergence or shape of the incident light wavefront, and can be used to identify the phase state of the light incident on the pixel 104. The decoded phase state of the light can provide possibility of super resolution since it relies on light amplitude interactions recorded as interference patterns rather than light intensity summation (known as Abbe limit for resolution). In addition, the interference pattern will be different for overfocus (light diverging), in focus (light front parallel), or under focus (light converging) states. A representation of the different diffraction patterns or light distribution 904 produced across the different sub-pixels 604 of a pixel 104 by the associated set of diffraction elements 528 is illustrated in
More particularly, the differences in the amount of light incident on the different sub-pixels 604 results in the generation of different signal amounts by those sub-pixels 604. This is illustrated in
In order to enable the detection of a state of incident light, the pixels 104 included in the image sensor 100 can be calibrated. Calibration can include exposing a pixel 104 to light having a known state, and recording the resulting output from each of sub-pixel 604. For example, where each pixel 104 has the same diffraction element set 524 configuration, the output of the sub-pixels 604 from a single representative pixel 104, the output of the sub-pixels 604 from representative pixels 104 selected from different areas of the image sensor 100, the output of the sub-pixels 604 from randomly selected pixels 104, or the output of the sub-pixels 604 from each pixel 104 within the image sensor 100 can be recorded. As a further example, where different regions or areas of the pixel array 108 have different diffraction element set 524 configurations, an output from the sub-pixels 604 of one or more representative pixels 104 from each area can be recorded. This process can be repeated for each light state of interest.
In accordance with embodiments of the present disclosure, ratios of the signal strength or amplitude of the signals produced by pairs of sub-pixels 604 within a pixel 104 are determined. For example, where a pixel 104 includes four sub-pixels 604, six unique sets of sub-pixel 604 amplitudes can be created. In accordance with embodiments of the present disclosure, the values of the ratios between different sub-pixel 604 pairs obtained through the calibration process can be stored in a calibration table 1004, for example as shown in
Accordingly, by identifying a light state in a table of different light states associated with sub-pixel 604 signal strength ratios that most closely matches the ratios of signal strengths observed (i.e. measured) for a sample of light of an unknown light state, the light state of the light sample can be accurately assigned. Moreover, identification of the light state of the incident light is possible across a wide range of states. For example, the identification of any wavelength to which the sub-pixels 604 are sensitive is possible. Moreover, a polarization state, including a linear or circular polarization direction, can be detected. For instance, it is possible to provide light state information for light having a wavelength within a range of from 400 nm to 1000 nm, a linear polarization, or a circular polarization. In accordance with embodiments of the present disclosure, an identification of the state of the light incident on a pixel 104 is performed using a simple analytical expression:
Light State Vector=PD1/PD2i−PD1/PD3*j+PD1/PD4*k−PD2/PD*i+PD2/PD4*m−PD3/PD4*n
More particularly, and as shown in the example calibration table 1004, the calibrated light state identification can be stored in a column of color (wavelength) and polarization identification values. Additional columns are provided for tabulating the ratios of the amplitude or intensity of unique pairs of the sub-pixels 604 within the pixel 104 for each light state. As can be appreciated by one of skill in the art after consideration of the present disclosure, the values within the illustrated table 1004 are provided for illustration purposes, and actual values will depend on the particular configuration of the diffraction features 524 and other characteristics of the pixel 104 and associated components of the image sensor 100 as implemented.
This calculating and storing of difference values in the difference matrix 1304 is repeated for each ratio represented in the calibration table for the selected wavelength.
At step 1124, a determination is made as to whether all of the lines (i.e. all of the light states) represented within the calibration table 1204 have been considered. If lines remain to be considered, the next line of ratios (corresponding to the next light state) is selected from the calibration table 1004 (step 1128). The ratios measured for the unknown case are then compared to the ratios stored in the table for the next selected line of the calibration table 1004, corresponding to a next calibrated light state, to determine a next set of differences, which are then saved to the difference matrix 1304 (see
Once all of the lines (light states) of values within the calibration table 1004 have been compared to the measured signal ratios, the line (light state) in the difference matrix 1304 with the smallest row difference is identified, and the associated light state is assigned as the light state of the light incident on the subject pixel 104 (step 1132). Where the difference for one line is zero, the wavelength associated with that line is identified as the light state of the light incident on the subject pixel 104. Where no line has a difference of zero, a row difference for each line is calculated. The row with the smallest calculated row difference value is then selected as identifying the light state of the light incident on the subject pixel. In accordance with embodiments of the present disclosure, the row difference can be calculated as follows:
After the light state of the incident light has been determined, that information can be stored, output to an application, or otherwise utilized. (step 1136). As can be appreciated by one of skill in the art after consideration of the present disclose, this process can be performed for each pixel 104 in the image sensor 100. As can further be appreciated by one of skill in the art after consideration of the present disclosure, where different calibration tables 1004 have been generated for different pixels 104, the process of light state determination is performed in connection with the calibration table 1004 that is applicable to the subject pixel 104.
The optical system 504 includes an objective lens of the camera 1400. The optical system 504 collects light from within a field of view of the camera 1400, which can encompass a scene containing an object. As can be appreciated by one of skill in the art after consideration of the present disclosure, the field of view is determined by various parameters, including a focal length of the lens, the size of the effective area of the image sensor 100, and the distance of the image sensor 100 from the lens. In addition to a lens, the optical system 504 can include other components, such as a variable aperture and a mechanical shutter. The optical system 504 directs the collected light to the image sensor 100 to form an image of the object on a light incident surface of the image sensor 100.
As discussed elsewhere herein, the image sensor 100 includes a plurality of pixels 104 disposed in an array 108. Moreover, the image sensor 100 can include a semiconductor element or substrate 112 in which the pixels 104 each include a number of sub-pixels 604 that are formed as photosensitive areas or photodiodes within the substrate 112. In addition, as also described elsewhere herein, each pixel 104 is associated with a set of diffraction features 524 formed in a diffraction layer 520 positioned between the optical system 504 and the sub-pixels 604. The photosensitive sites or sub-pixels 604 generate analog signals that are proportional to an amount of light incident thereon. These analog signals can be converted into digital signals in a circuit, such as a column signal processing circuit 120, included as part of the image sensor 100, or in a separate circuit or processor. As discussed herein the distribution of light amongst the sub-pixels 604 of a pixel 104 is dependent on the light state of the incident light. The digital signals can then be output.
The imaging control unit 1403 controls imaging operations of the image sensor 100 by generating and outputting control signals to the image sensor 100. Further, the imaging control unit 1403 can perform autofocus in the camera 1400 on the basis of image signals output from the image sensor 100. Here, “autofocus” is a system that detects the focus position of the optical system 504 and automatically adjusts the focus position. As this autofocus, a method in which an image plane phase difference is detected by phase difference pixels arranged in the image sensor 100 to detect a focus position (image plane phase difference autofocus) can be used. Further, a method in which a position at which the contrast of an image is highest is detected as a focus position (contrast autofocus) can also be applied. The imaging control unit 1403 adjusts the position of the lens 1001 through the lens driving unit 1404 on the basis of the detected focus position, to thereby perform autofocus. Note that, the imaging control unit 1403 can include, for example, a DSP (Digital Signal Processor) equipped with firmware.
The lens driving unit 1404 drives the optical system 504 on the basis of control of the imaging control unit 1403. The lens driving unit 1404 can drive the optical system 504 by changing the position of included lens elements using a built-in motor.
The image processing unit 1405 processes image signals generated by the image sensor 100. This processing includes, for example, assigning a light state to light incident on a pixel 104 by determining ratios of signal strength between pairs of sub-pixels 604 included in the pixel 104, and determining an amplitude of the pixel 104 signal from the individual sub-pixel 604 signal intensities, as discussed elsewhere herein. In addition, this processing includes determining a light state of light incident on a pixel 104 by comparing the observed ratios of signal strengths from pairs of sub-pixels 604 to calibrated ratios for those pairs stored in a calibration table 1004. The image processing unit 1405 can include, for example, a microcomputer equipped with firmware, and/or a processor that executes application programming, to implement processes for identifying color information in collected image information as described herein.
The operation input unit 1406 receives operation inputs from a user of the camera 1400. As the operation input unit 1406, for example, a push button or a touch panel can be used. An operation input received by the operation input unit 1406 is transmitted to the imaging control unit 1403 and the image processing unit 1405. After that, processing corresponding to the operation input, for example, the collection and processing of imaging an object or the like, is started.
The frame memory 1407 is a memory configured to store frames that are image signals for one screen or frame of image data. The frame memory 1407 is controlled by the image processing unit 1405 and holds frames in the course of image processing.
The display unit 1408 can display information processed by the image processing unit 1405. For example, a liquid crystal panel can be used as the display unit 1408.
The recording unit 1409 records image data processed by the image processing unit 1405. As the recording unit 1409, for example, a memory card or a hard disk can be used.
An example of a camera 1400 to which embodiments of the present disclosure can be applied has been described above. The image sensor 100 of the camera 1400 can be configured as described herein. Specifically, the image sensor 100 can diffract incident light across different light sensitive areas or sub-pixels 604 of a pixel 104, and can compare ratios of signals from pairs of the sub-pixels 604 to corresponding stored ratios for a number of different light states, to identify a closest match, and thus the light state of the incident light. Moreover, the color identification capabilities of the image sensor 100 can be described as hyperspectral, as wavelength identification is possible across the full range of wavelengths to which the sub-pixels are sensitive. In accordance with embodiments of the present disclosure, the light states that can be detected include different combinations of wavelength and polarization of incident light. Moreover, polarization states sensitivity can be combined with super resolution, complex field, and wavefront sensitivity.
Note that, although a camera has been described as an example of an electronic apparatus, an image sensor 100 and other components, such as processors and memory for executing programming or instructions and for storing calibration information as described herein, can incorporated into other types of devices. Such devices include, but are not limited to, surveillance systems, automotive sensors, scientific instruments, medical instruments, etc. As further examples, embodiments of the present disclosure can detect the rotation to the polarization of state of light caused by biologically active molecules. In addition, the ability to detect polarization state provided by embodiments of the present disclosure facilitates the encryption of light using polarization state. In accordance with still other embodiments, a system 100 as disclosed herein be implemented in connection with a communication system, in which information is encoded or is distinguished from other units of information using the color and polarization state of light. Still other applications of embodiments of the present disclosure include quantum computing.
As can be appreciated by one of skill in the art after consideration of the present disclosure, an image sensor 100 as disclosed herein utilizes interference effects to obtain light state information, such as wavelength and polarization information, over a wide spectral range. In addition, an image sensor 100 as disclosed herein can be produced using CMOS processes entirely. Implementations of an image sensor 100 or devices incorporating an image sensor 100 as disclosed herein can utilize calibration tables 1004 developed for each pixel 104 of the image sensor 100. Alternatively, calibration tables 1004 can be developed for each different pattern of diffraction features 524. In addition to providing calibration tables 1004 that are specific to particular pixels 104 and/or particular patterns of diffraction features 524, calibration tables 1004 can be developed for use in selected regions of the array 108.
Methods for producing an image sensor in accordance with embodiments of the present disclosure include applying conventional CMOS production processes to produce an array of pixels in an image sensor substrate in which each pixel includes a plurality of sub-pixels or photodiodes. As an example, the material of the sensor substrate is silicon (Si), and each sub-pixel is a photodiode formed therein. A thin layer of material is disposed on or adjacent a light incident side of the image sensor substrate. Moreover, the thin layer of material can be disposed on a back surface side of the image sensor substrate. As an example, the thin layer of material is silicon oxide (SiO2), and has a thickness of 500 nm or less. In accordance with the least some embodiments of the present disclosure, an anti-reflection layer can be disposed between the light incident surface of the image sensor substrate and the thin layer of material. A light focusing, electrically conductive or semiconductive diffractive grating pattern is formed in the thin layer of material. In particular, a set 524 of diffraction features 528 is disposed adjacent each of the pixels. The diffraction features can be formed as metal or semiconductor material features embedded in trenches formed in the thin layer of material. For example, the diffraction features can be formed from silicon nitride (SiN) or aluminum. Different diffraction features within a set 524 of diffraction features 528 can be formed from different materials. Moreover, the diffraction features 528 can be relatively thin (i.e. from about 100 to about 200 nm), and the pattern can include a plurality of linear diffraction elements 708 of various lengths disposed along lines that extend radially from a central circular feature 704 and scattering elements 712 interspersed between the linear diffraction elements 708. Production of an image sensor in accordance with embodiments of the present disclosure can be accomplished using only CMOS processes. Moreover, an image sensor produced in accordance with embodiments of the present disclosure does not require micro lenses or color filters for each pixel.
The foregoing has been presented for purposes of illustration and description. Further, the description is not intended to limit the disclosed systems and methods to the forms disclosed herein. Consequently, variations and modifications commensurate with the above teachings, within the skill or knowledge of the relevant art, are within the scope of the present disclosure. The embodiments described hereinabove are further intended to explain the best mode presently known of practicing the disclosed systems and methods, and to enable others skilled in the art to utilize the disclosed systems and methods in such or in other embodiments and with various modifications required by the particular application or use. It is intended that the appended claims be construed to include alternative embodiments to the extent permitted by the prior art.