The disclosure relates generally to image sensors, and more specifically to pixel cell structure including interfacing circuitries for determining light intensity for image generation.
A typical pixel in an image sensor includes a photodiode to sense incident light by converting photons into charge (e.g., electrons or holes). The incident light can include components of different wavelength ranges for different applications, such as 2D and 3D sensing. Moreover, to reduce image distortion, a global shutter operation can be performed in which each photodiode of the array of photodiodes senses the incident light simultaneously in a global exposure period to generate the charge. The charge can be converted by a charge sensing unit (e.g., a floating diffusion) to convert to a voltage. The array of pixel cells can measure different components of the incident light based on the voltages converted by the charge sensing unit and provide the measurement results for generation of 2D and 3D images of a scene.
The present disclosure relates to image sensors. More specifically, and without limitation, this disclosure relates to a pixel cell. This disclosure also relates to operating the circuitries of pixel cells to generate a digital representation of the intensity of incident light.
In one example, an apparatus comprises: a plurality of photodiodes, each photodiode being configured to convert a component of incident light of a wavelength range to charge; one or more charge sensing units; one or more analog-to-digital converters (ADCs); a memory; and a controller configured to: enable the each photodiode to generate charge in response to a different component of the incident light; transfer the charge from the plurality of photodiodes to the one or more charge sensing units to convert to voltages; receive a selection of one or more quantization processes of a plurality of quantization processes corresponding to a plurality of intensity ranges; based on the selection, control the one or more ADCs to perform the selected one or more quantization processes to quantize the voltages from the one or more charge sensing units to digital values representing components of a pixel of different wavelength ranges; store at least some of the digital values in the memory; and generate a pixel value based on the at least some of the digital values stored in the memory.
In some aspects, the each photodiode is configured to, within an integration period, accumulate at least a part of the charge as residual charge until the each photodiode saturates, and to transfer the remaining charge as overflow charge to the one or more charge sensing unit after the photodiode saturates. The one or more charge sensing unit comprises a charge storage device having a configurable capacitance. The plurality of quantization operations comprise: a first quantization operation to generate a first digital value representing a quantity of the overflow charge received by the charge storage device configured at a maximum capacitance, the first quantization operating being associated with a first intensity range; and a second quantization operation to, after the residual charge is transferred to the charge storage device configured at a minimum capacitance, generate a second digital value representing a quantity of the residual charge stored at the charge storage device, the second quantization operation being associated with a second intensity range lower than the first intensity range.
In some aspects, the plurality of quantization operations comprises a third quantization operation to generate a third digital value representing a time-of-saturation of the charge storage device caused by the overflow charge.
In some aspects, the apparatus further comprises a light receiving surface through which the plurality of the photodiodes receives the incident light. The plurality of photodiodes forms a stack structure with respect to the light receiving surface such that the each photodiode is separated from the light receiving surface by a different distance. The component converted by the each photodiode is based on the respective distance between the each photodiode and the light receiving surface.
In some aspects, the apparatus further comprises a filter array on a first side of the light receiving surface, the filter array having filter elements positioned at a plurality of locations on the first side of the light receiving surface to set a component of the incident light that enters the light receiving surface at the respective location. The plurality of photodiodes correspond to a plurality of sub-pixels and are positioned at the plurality of locations on a second side of the light receiving surface to receive the respective components of the incident light.
In some aspects, the apparatus further comprises a single microlens over a plurality of filter arrays including the filter array and configured to project the incident light received from one spot of a scene towards the plurality of locations on the first side of the light receiving surface.
In some aspects, the apparatus further comprises a plurality of microlenses including a first microlens, the first microlens covering the filter array and configured to project the incident light received from one spot of a scene towards the plurality of locations on the first side of the light receiving surface.
In some aspects, the one or more sensing unit comprises a single charge sensing unit. The one or more ADCs comprises a single ADC coupled with an output of the single charge sensing unit. The apparatus further comprises a plurality of switches, each switch coupled between a photodiode of the plurality of photodiodes and an input of the single charge sensing unit. The controller is configured to: control the plurality of switches to transfer the charge generated by the each photodiode to the single charge sensing unit to convert to voltages; and control the single ADC to quantize the voltages generated by the single charge sensing unit.
In some aspects, the controller is configured to: control a first switch of the plurality of switches to transfer a first overflow charge from a first photodiode of the plurality of photodiodes to the single charge sensing unit to convert to a first voltage; based on the selection, control the single ADC to perform at least one of the first or third quantization operations of the first voltage to generate a first digital value; control the first switch to transfer a first residual charge from the first photodiode to the single charge sensing unit to convert to a second voltage; based on the selection, control the single ADC to perform the second quantization operation of the second voltage to generate a second digital value; control a second switch of the plurality of switches to transfer a second residual charge from a second photodiode of the plurality of photodiodes to the single charge sensing unit to convert to a third voltage; based on the selection, control the single ADC to perform the second quantization operation of the third voltage to generate a third digital value; and output, from the memory, generate the pixel value based on one of first digital value and the second digital value.
In some aspects, each of the photodiodes has a different full well capacity for storing the residual charge. The controller is configured to: control the plurality of switches to transfer overflow charge from each photodiode of the plurality of photodiodes to the single charge storage unit simultaneously to generate a first voltage; and based on the selection, control the single ADC to quantize the first voltage using at least one of the first or third quantization operations to generate a first digital value; control the plurality of switches to transfer residual charge from the each photodiode to the single charge storage unit at different times to generate second voltages each corresponding the respective residual charge from the each photodiode; based on the selection, control the single ADC to quantize the second voltages using the second quantization operation to generate second digital values; and generate the pixel value based on the first digital value and the second digital values.
In some aspects, the controller is configured to, within a first time period: control the plurality of switches to transfer overflow charge from each photodiode of the plurality of photodiodes to the single charge sensing unit at different times to generate first voltages each corresponding the respective overflow charge from the each photodiode; and based on the selection, control the single ADC to quantize the first voltages using at least one of the first or third quantization operations to generate first digital values. The controller is further configured to, within a second time period: control the plurality of switches to transfer residual charge from the each photodiode to the single charge storage unit at different times to generate second voltages each corresponding to the respective residual charge from the each photodiode; and based on the selection, control the single ADC to quantize the second voltages using the second quantization operation to generate second digital values; and generate the pixel value based on at least some of the first digital values and the second digital values.
In some aspects, the one or more sensing unit comprises a plurality of charge sensing units corresponding to the plurality of photodiodes. The apparatus further comprises a plurality of switches each coupled between each charge sensing unit of the plurality of charge sensing units and a corresponding photodiode of the plurality of photodiodes.
In some aspects, the controller is configured to: enable a first photodiode of the plurality of the photodiodes to transfer a first charge to a first charge sensing unit of the plurality of charge sensing units to generate a first voltage; enable a second photodiode of the plurality of the photodiodes to transfer a second charge to a second charge sensing unit of the plurality of charge sensing units to generate a second voltage; based on the selection, control the one or more ADCs to perform the first quantization operation of the first voltage to generate a first digital value, followed by the second or third quantization operations of the second voltage to generate a second digital value; and generate the pixel value based on the first digital value and the second digital value.
In some aspects, the controller is configured to, within a first time period: enable a first photodiode of the plurality of the photodiodes to generate a first charge in response to the incident light; enable the first photodiode to transfer a first overflow charge of the first charge to a first charge sensing unit of the plurality of charge sensing units to generate a first voltage; and based on the selection, control the one or more ADCs to perform the third quantization operation of the first voltage to generate a first digital value representing a first time-to-saturation. The controller is further configured to, within a second time period: enable a second photodiode of the plurality of the photodiodes to generate a second charge in response to the incident light; enable the second photodiode to transfer a second overflow charge of the second charge to a second charge sensing unit of the plurality of charge sensing units to generate a second voltage; and based on the selection, control the one or more ADCs to perform the third quantization operation of the second voltage to generate a second digital value representing a second time-to-saturation. The controller is further configured to generate the pixel value based on the first digital value and the second digital value.
In some aspects, the plurality of charge sensing units comprises a first charge sensing unit, a second charge sensing unit, a third charge sensing unit, and a fourth charge sensing unit. The one or more ADCs comprise a first ADC and a second ADC. The controller is configured to: control the first ADC to quantize a first voltage from the first charge sensing unit and a second voltage from the second charge sensing unit; and control the second ADC to quantize a third voltage from the third charge sensing unit and a third voltage from the second charge sensing unit.
In some aspects, the controller is configured to store each of the digital values in the memory.
In some aspects, the controller is configured to: control the one or more ADC to generate a first digital value based on quantizing a first voltage corresponding to charge generated by a first photodiode of the plurality of photodiodes; store the first digital value in the memory; read the first digital value to compute the pixel value; control the one or more ADC to generate a second digital value based on quantizing a second voltage corresponding to charge generated by a second photodiode of the plurality of photodiodes; overwrite the first digital value with a second digital value in the memory; and read the second digital value to compute the pixel value.
In one example, a method is provided. The method comprises: enabling each photodiode of a plurality of photodiodes of a pixel cell to generate charge in response to a different component of incident light received by the pixel cell; transferring the charge from the plurality of photodiodes to the one or more charge sensing units to convert to voltages; receiving, for each photodiode of the plurality of photodiodes, a selection of one or more quantization processes of a plurality of quantization processes corresponding to a plurality of intensity ranges; based on the selection, controlling the one or more ADCs to perform the selected one or more quantization processes to quantize the voltages from the one or more charge sensing units to digital values representing components of a pixel of different wavelength ranges; storing at least some of the digital values in a memory; and generating a pixel value based on the at least some of the digital values stored in the memory.
In some aspects, the plurality of quantization processes comprises a first quantization process to measure a quantity of residual charge accumulated at a first photodiode of the plurality of photodiodes before the first photodiode saturates, a second quantization process to measure a quantity of overflow charge transferred by the first photodiode after the first photodiode saturates, and a third quantization process to measure a time-to-saturation of the one or more charge sensing units caused by the overflow charge from the first photodiode.
Illustrative examples are described with reference to the following figures.
The figures depict examples of the present disclosure for purposes of illustration only. One skilled in the art will readily recognize from the following description that alternative examples of the structures and methods illustrated may be employed without departing from the principles, or benefits touted, of this disclosure.
In the appended figures, similar components and/or features may have the same reference label. Further, various components of the same type may be distinguished by following the reference label by a dash and a second label that distinguishes among the similar components. If only the first reference label is used in the specification, the description is applicable to any one of the similar components having the same first reference label irrespective of the second reference label.
In the following description, for the purposes of explanation, specific details are set forth in order to provide a thorough understanding of certain inventive examples. However, it will be apparent that various examples may be practiced without these specific details. The figures and description are not intended to be restrictive.
A typical image sensor includes an array of pixel cells. Each pixel cell includes a photodiode to measure the intensity incident light by converting photons into charge (e.g., electrons or holes). The charge generated by the photodiode can be converted to a voltage by a charge sensing unit, which can include a floating drain node. The voltage can be quantized by an analog-to-digital converter (ADC) into a digital value. The digital value can represent an intensity of light received by the pixel cell and can form a pixel, which can correspond to light received from a spot of a scene. An image comprising an array of pixels can be derived from the digital outputs of the array of pixel cells.
An image sensor can be used to perform different modes of imaging, such as 2D and 3D sensing. The 2D and 3D sensing can be performed based on light of different wavelength ranges. For example, visible light can be used for 2D sensing, whereas invisible light (e.g., infra-red light) can be used for 3D sensing. An image sensor may include an optical filter array to allow visible light of different optical wavelength ranges and colors (e.g., red, green, blue, monochrome, etc.) to a first set of pixel cells assigned for 2D sensing, and invisible light to a second set of pixel cells assigned for 3D sensing.
To perform 2D sensing, a photodiode at a pixel cell can generate charge at a rate that is proportional to an intensity of visible light component (e.g., red, green, blue, monochrome, etc.) incident upon the pixel cell, and the quantity of charge accumulated in an exposure period can be used to represent the intensity of visible light (or a certain color component of the visible light). The charge can be stored temporarily at the photodiode and then transferred to a capacitor (e.g., a floating diffusion) to develop a voltage. The voltage can be sampled and quantized by an analog-to-digital converter (ADC) to generate an output corresponding to the intensity of visible light. An image pixel value can be generated based on the outputs from multiple pixel cells configured to sense different color components of the visible light (e.g., red, green, and blue colors).
Moreover, to perform 3D sensing, light of a different wavelength range (e.g., infra-red light) can be projected onto an object, and the reflected light can be detected by the pixel cells. The light can include structured light, light pulses, etc. The pixel cells outputs can be used to perform depth sensing operations based on, for example, detecting patterns of the reflected structured light, measuring a time-of-flight of the light pulse, etc. To detect patterns of the reflected structured light, a distribution of quantities of charge generated by the pixel cells during the exposure time can be determined, and pixel values can be generated based on the voltages corresponding to the quantities of charge. For time-of-flight measurement, the timing of generation of the charge at the photodiodes of the pixel cells can be determined to represent the times when the reflected light pulses are received at the pixel cells. Time differences between when the light pulses are projected to the object and when the reflected light pulses are received at the pixel cells can be used to provide the time-of-flight measurement.
A pixel cell array can be used to generate information of a scene. In some examples, a subset (e.g., a first set) of the pixel cells within the array can detect visible components of light to perform 2D sensing of the scene, and another subset (e.g., a second set) of the pixel cells within the array can detect an infra-red component of the light to perform 3D sensing of the scene. The fusion of 2D and 3D imaging data are useful for many applications that provide virtual-reality (VR), augmented-reality (AR) and/or mixed reality (MR) experiences. For example, a wearable VR/AR/MR system may perform a scene reconstruction of an environment in which the user of the system is located. Based on the reconstructed scene, the VR/AR/MR can generate display effects to provide an interactive experience. To reconstruct a scene, a subset of pixel cells within a pixel cell array can perform 3D sensing to, for example, identify a set of physical objects in the environment and determine the distances between the physical objects and the user. Another subset of pixel cells within the pixel cell array can perform 2D sensing to, for example, capture visual attributes including textures, colors, and reflectivity of these physical objects. The 2D and 3D image data of the scene can then be merged to create, for example, a 3D model of the scene including the visual attributes of the objects. As another example, a wearable VR/AR/MR system can also perform a head tracking operation based on a fusion of 2D and 3D image data. For example, based on the 2D image data, the VR/AR/AR system can extract certain image features to identify an object. Based on the 3D image data, the VR/AR/AR system can track a location of the identified object relative to the wearable device worn by the user. The VR/AR/AR system can track the head movement based on, for example, tracking the change in the location of the identified object relative to the wearable device as the user's head moves.
Using different sets of pixel cells for sensing different incident light components (e.g., different visible light components, visible light versus infra-red light, etc.) can pose a number of challenges. First, because only a subset of the pixel cells of the array is used to measure a particular incident light component, the spatial resolution of the imaging can be reduced. For example, in a case where different subsets of the pixel cells are used for 2D imaging or 3D imaging, the spatial resolutions of both of the 2D image and 3D image are lower than the maximum spatial resolution available at the pixel cell array. Likewise, the spatial resolution for imaging of a particular visible light component is also lower than the maximum spatial resolution available at the pixel cell array. Although the resolutions can be improved by including more pixel cells, such an approach can lead to increases in the form-factor of the image sensor as well as power consumption, both of which are undesirable especially for a wearable device.
Moreover, since pixel cells assigned to measure light of different wavelength ranges are not collocated, different pixel cells may capture information of different spots of a scene, which can complicate the mapping between images of different incident light components. For example, a pixel cell that receives a certain color component of visible light (for 2D imaging) and a pixel cell that receives invisible light (for 3D imaging) may also capture information of different spots of the scene. The output of these pixel cells cannot be simply merged to generate the 2D and 3D images. The lack of correspondence between the output of the pixel cells due to their different locations can be worsened when the pixel cell array is capturing 2D and 3D images of a moving object. While there are processing techniques available to correlate different pixel cell outputs to generate pixels for a 2D image, and to correlate between 2D and 3D images (e.g., interpolation), these techniques are typically computation-intensive and can also increase power consumption. Similar problems can occur in the mapping of pixel cells associated with different visible light components for reconstruction of a 2D image.
The present disclosure relates to an image sensor having an array of pixel cells and can provide collocated imaging of different components of incident light from a spot of a scene, and to provide a global shutter operation. Specifically, each pixel cell can include a plurality of photodiodes, one or more charge sensing units, one or more analog-to-digital converters (ADCs), a memory, and a controller. Each photodiode of the plurality of photodiodes is configured to convert a component of incident light to charge. The controller can transfer the charge from the plurality of photodiodes to the one or more charge sensing units to convert to voltages. The controller can also receive a selection of one or more quantization processes of a plurality of quantization processes corresponding to a plurality of intensity ranges and, based on the selection, control the ADCs to perform the selected one or more quantization processes to quantize the voltages from the one or more charge sensing units to digital values representing components of a pixel, which can correspond to the spot of the scene. The controller can also store at least some of the digital values in the memory, and generate a pixel value based on the at least some of the digital values stored in the memory. The pixel values from the array of the pixel cells can represent the incident light received by each pixel cell within a global exposure period to support the global shutter operation.
Various techniques are proposed to enable the plurality of photodiodes to convert different components of incident light from a spot of a scene. the pixel cell includes a light receiving surface through which the plurality of the photodiodes receives the incident light. In some examples, the plurality of photodiodes forms a stack structure with respect to the light receiving surface such that the each photodiode is separated from the light receiving surface by a different distance. As the incident light propagates through the stack structure, different components can be absorbed and converted by the each photodiode based on the respective distance between the each photodiode and the light receiving surface. In some examples, each photodiode can be formed in a semiconductor substrate. The pixel cell can be formed by stacking a plurality of semiconductor substrates comprising the plurality of photodiodes with the light receiving surface being formed on the top or the bottom semiconductor substrate in the stack.
In some examples, the plurality of photodiodes can be arranged laterally and having the same distance from the receiving surface, with each photodiode corresponding to a sub-pixel and configured to receive incident light from the same spot of the scene. A filter array can be positioned on a first side of the light receiving surface of the pixel cell. The filter array may include filter elements positioned at a plurality of locations on the first side of the light receiving surface to set a component of the incident light that enters the light receiving surface at the respective location. The plurality of photodiodes can be in a single semiconductor substrate and positioned at the plurality of locations on a second side of the light receiving surface to receive the respective components of the incident light. In some examples, one or more microlens can be positioned over the filter array to project the incident light received from that same spot of the scene towards the plurality of locations on the first side of the light receiving surface, such that each photodiode, as a sub-pixel, can receive incident light from that same spot.
The one or more charge sensing units may include a configurable charge storage device and a buffer to convert the charge generated by the plurality of photodiodes to voltages. The configurable charge storage device may include a floating drain to accumulate the charge from the plurality of photodiodes to generate the voltages. An auxiliary capacitor (e.g., a metal capacitor, a metal-oxide-semiconductor (MOS) capacitor, etc.) can be connected or disconnected from the floating drain to expand or reduce the capacitance of the charge storage device. A buffer can buffer the voltage at the charge storage device to increase its driving strength. In one example, the one or more charge sensing units may include a single charge sensing unit which can be shared among the plurality of photodiodes to perform charge-to-voltage conversion. In one example, the one or more charge sensing units may include a plurality of charge sensing units each corresponding to a photodiode of the plurality of photodiodes and configured to convert the charge generated by the respective photodiode to a voltage. In both examples, the pixel cell further includes a plurality of switches, with each switch coupled between a photodiode and the shared/corresponding charge sensing unit, which can be controlled by the controller to control the transfer of the charge from the photodiode to the charge sensing unit. In a case where the one or more ADCs are shared among the plurality of charge sensing units, each charge sensing unit can also include a switch coupled between the buffer and the one or more ADCs which can be controlled by the controller to select which of the charge sensing unit to provide an output voltage to the one or more ADCs.
The controller can control the one or more ADCs to quantize the output voltages output by the one or more charge sensing units. The one or more ADCs can be shared between the charge sensing units (in a case where the pixel cell includes multiple charge sensing units), or may include a single ADC coupled with a single charge sensing unit of the pixel cell. The one or more ADCs can quantize the voltages based on different quantization operations associated with different intensity ranges, which can increase the dynamic range of the light intensity measurement operation. Specifically, each photodiode can generate a quantity of charge within an exposure period, with the quantity of charge representing the incident light intensity. Each photodiode also has a quantum well to store at least some of the charge as residual charge. For a low light intensity range, the photodiode can store the entirety of the charge as residual charge in the quantum well. For a medium light intensity range, the quantum well can be saturated by the residual charge, and the photodiode can transfer the remaining charge as overflow charge to the one or more charge sensing units. The quantum well capacity can be set based on a bias voltage on the switch between the photodiode and the charge sensing unit. For a high light intensity range, the charge storage device in the one or more charge sensing unit can be saturated by the overflow charge. The controller can control the one or more ADCs to perform different quantization operations for the high intensity range, the medium intensity range, and for the low intensity range based on measuring the overflow charge and residual charge.
Specifically, for measurement of the high light intensity range, the controller can control the one or more ADCs to perform a time-to-saturation (TTS) measurement operation by quantizing the time it takes for the charge storage device to become saturated by the overflow charge. The TTS measurement can be performed within the exposure time, and when the photodiode is coupled with the charge sensing unit and the output of the charge sensing unit is coupled with the one or more ADCs. The capacity of the charge storage device can be maximized by connecting the floating drain with the auxiliary capacitor for the TTS operation. In the TTS operation, a counter can start at the beginning of the exposure time. The one or more ADCs can compare the output voltage of the charge sensing unit with a static threshold representing the saturation limit of the charge storage device to generate a decision. When the decision indicates that the output of the charge sensing unit reaches the threshold, a count value from the counter can be stored in the memory to represent a time-to-saturation. The TTS operation can be performed when the light intensity is so high that the charge storage device becomes saturated and the output of the charge sensing unit no longer correlates with the quantity of charge generated (and the light intensity), which can extend the upper limit of the dynamic range.
For measurement of the medium intensity range in which the charge storage device is not saturated by the overflow charge, the controller can control the one or more ADCs to perform a FD ADC measurement operation to measure a quantity of the overflow charge stored in the charge storage device. The FD ADC measurement can be performed within the exposure period, and when the photodiode is coupled with the charge sensing unit which can temporarily store the overflow charge and convert the stored charge to a first voltage. The one or more ADCs can compare the first voltage with a first ramping voltage to generate a decision. The counter can start counting at the starting point of the ramping voltage, and the memory can store the count value when the first ramping voltage reaches the first voltage. The stored count value can represent the quantity of the overflow charge. In a case where multiple charge sensing units share an ADC, the ADC can compare the first voltages output by each charge sensing unit sequentially with the ramping voltage to generate the count values.
For measurement of the low intensity range in which the photodiode's quantum well is not saturated by the residual charge, the controller can control the one or more ADCs to perform a PD ADC measurement operation to measure a quantity of the residual charge stored in photodiode. To perform the PD ADC measurement operation, the controller can control the switch between the photodiode and the charge sensing unit to transfer the residual charge out of the photodiode into the charge storage device to convert to a second voltage. The capacitance of the charge storage device can also be reduced (e.g., by disconnecting the auxiliary capacitor) to maximize the charge-to-voltage conversion gain, which can reduce the quantization error. The one or more ADCs can compare the second voltage with a second ramping voltage to generate a decision. The counter can start counting at the starting point of the second ramping voltage, and the memory can store the count value when the second ramping voltage reaches the second voltage. The stored count value can represent the quantity of the residual charge. The PD ADC operation (and the transfer of the residual charge) can be performed after the exposure period. In a case where multiple photodiodes share a charge sensing unit, the controller can the perform the PD ADC operation for each photodiode sequentially, in which the controller can control the switches to transfer the residual charge from one photodiode to the charge sensing unit to generate the second voltage, and control the ADC to quantize the second voltage, and then repeat the operation for other photodiodes. As the residual charge is typically much less susceptible to dark current in the photodiode, the noise floor of the low light intensity measurement can be lowered, which can further extend the lower limit of the dynamic range.
The controller can control the ADCs to perform one or more of the aforementioned quantization operations for each photodiode based on the sharing arrangements of the one or more charge sensing units and the one or more ADCs among the plurality of the photodiodes, as well as the selection of the quantization operations. As described above, a single charge sensing unit can be shared among the plurality of photodiodes. In some examples, the controller can control the switches to allow each photodiode to take turn to transfer overflow charge to the single charge sensing unit to generate the first voltage, which can then be quantized by the ADC in a TTS operation and/or a FD ADC operation based on the selection. The charge sensing unit can be reset between the start of quantization operations for a photodiode. Followed by the completion of the quantization of the overflow charge, the controller can control the switches to allow each photodiode to take turn in transferring residual charge to the charge sensing unit, followed by a PD ADC operation to measure the residual charge. While such arrangements allow each photodiode to have the same access to the charge sensing unit and the same set of quantization operations can be performed on the outputs of each photodiode, each photodiode may have different effective exposure periods for accumulating overflow charge, which can degrade the global shutter operation.
In some examples, the controller can also allow, based on the selection, a first photodiode to transfer overflow charge to the charge sensing unit to perform the TTS and FD ADC operations within the exposure period. After the exposure period ends, the controller can control the switches to allow all of the photodiodes to take turn to transfer residual charge to the charge sensing unit and to perform the PD ADC operations. Such arrangements can be used when, for example, the intensity of a particular component is very high compared with other components (e.g., in a dark environment with strong infra-red illumination for 3D sensing). The same exposure period can be provided for each photodiode to either accumulate charge for the strong infra-red component or for the other much weaker visible light components. The TTS and FD ADC operations can be performed on the output of the photodiode that detects the strong infra-red component, while PD ADC operations can be performed on the outputs of other photodiodes, which can improve the dynamic range of both the low intensity (for visible light) and the high intensity (for infra-red light) measurement operations.
In some examples, the controller can also some or all of the photodiodes to transfer overflow charge to the charge sensing unit simultaneously. For example, photodiodes configured to detect the same wavelength range can transfer overflow charge to the charge sensing unit simultaneously. Moreover, for photodiodes that detect different wavelength ranges, the controller can set the biases of their switches to set different quantum well capacities for these photodiodes. For example, the controller can lower the quantum well capacity for a photodiode associated with a particular wavelength range which is expected to the be strongest among other wavelength ranges, such that that photodiode is more likely to transfer overflow charge to the charge sensing unit than other photodiodes. Such arrangements not only provide same exposure period for each photodiode, as in the example described above, but also enhance flexibility in the light measurement operation. Specifically, while the TTS/FD ADC operation output is more likely to represent the output of the expected strongest component of the incident light, when the operation condition changes and the intensities of other components also increase, the TTS/FD ADC operation output can reflect the other high intensity components of the incident light as well.
In a case where each photodiode has full access to a charge sensing unit, but the charge sensing units share one or more ADCs, the controller can also adapt the quantization operations based on the sharing of the ADCs. For example, based on the selection, the controller can control the switches in the charge sensing units to connect one of the photodiode (and its charge sensing unit) to the ADC to perform the TTS operation within the integration period. After the TTS operation completes, the controller can control the ADC to perform a FD ADC operation and/or a PD ADC operation (e.g., based on the selection) for each of the photodiodes and the corresponding charge sensing unit. In some examples, the controller can also control the switches in the charge sensing units to take turn in connecting with the ADC to perform the TTS operation. After the TTS operations for all of the photodiodes complete, the controller can control the ADC to perform a FD ADC operation and/or a PD ADC operation for each of the photodiodes sequentially.
The memory can be configured to store the quantization results for each of the photodiodes. In some examples, the memory can be configured to store the quantization results for all of the photodiodes simultaneously. In some examples, the quantization results of each photodiode can be stored in the memory sequentially. The quantization results of one photodiode of the pixel cell can be stored in the memory, read out for generation of the pixel value, and then overwritten by the quantization results of another photodiode of the pixel cell.
With examples of the present disclosure, a pixel cell can perform collocated imaging for different components of incident light, which can support collocated 2D and 3D imaging operations. Having the same set of pixel cells to perform sensing of different components can facilitate the correspondence between images of different components generated by the pixel cells. Moreover, given that every pixel cell of a pixel cell array can be used to generate the image, the full spatial resolution of the pixel cell array can be utilized. As a result, the spatial resolutions of the images can also be improved. Further, typically the ADCs consume a lot of power and occupies a lot of space. By sharing the ADCs among the photodiodes of each pixel cell, the form factor and power consumption of the image sensor can also be reduced, while the ADCs of each pixel cell of a pixel cell array can generate a pixel value based on incident light received by each pixel cell within the same global exposure period to support a global shutter operation to reduce image distortion.
The disclosed techniques may include or be implemented in conjunction with an artificial reality system. Artificial reality is a form of reality that has been adjusted in some manner before presentation to a user, which may include, e.g., a virtual reality (VR), an augmented reality (AR), a mixed reality (MR), a hybrid reality, or some combination and/or derivatives thereof. Artificial reality content may include completely generated content or generated content combined with captured (e.g., real-world) content. The artificial reality content may include video, audio, haptic feedback, or some combination thereof, any of which may be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to the viewer). Additionally, in some examples, artificial reality may also be associated with applications, products, accessories, services, or some combination thereof, that are used to, e.g., create content in an artificial reality and/or are otherwise used in (e.g., perform activities in) an artificial reality. The artificial reality system that provides the artificial reality content may be implemented on various platforms, including a head-mounted display (HMD) connected to a host computer system, a standalone HMD, a mobile device or computing system, or any other hardware platform capable of providing artificial reality content to one or more viewers.
Near-eye display 100 includes a frame 105 and a display 110. Frame 105 is coupled to one or more optical elements. Display 110 is configured for the user to see content presented by near-eye display 100. In some examples, display 110 comprises a waveguide display assembly for directing light from one or more images to an eye of the user.
Near-eye display 100 further includes image sensors 120a, 120b, 120c, and 120d. Each of image sensors 120a, 120b, 120c, and 120d may include a pixel cell array configured to generate image data representing different fields of views along different directions. For example, sensors 120a and 120b may be configured to provide image data representing two fields of view towards a direction A along the Z axis, whereas sensor 120c may be configured to provide image data representing a field of view towards a direction B along the X axis, and sensor 120d may be configured to provide image data representing a field of view towards a direction C along the X axis.
In some examples, sensors 120a-120d can be configured as input devices to control or influence the display content of the near-eye display 100, to provide an interactive VR/AR/MR experience to a user who wears near-eye display 100. For example, sensors 120a-120d can generate physical image data of a physical environment in which the user is located. The physical image data can be provided to a location tracking system to track a location and/or a path of movement of the user in the physical environment. A system can then update the image data provided to display 110 based on, for example, the location and orientation of the user, to provide the interactive experience. In some examples, the location tracking system may operate a SLAM algorithm to track a set of objects in the physical environment and within a view of field of the user as the user moves within the physical environment. The location tracking system can construct and update a map of the physical environment based on the set of objects, and track the location of the user within the map. By providing image data corresponding to multiple fields of views, sensors 120a-120d can provide the location tracking system a more holistic view of the physical environment, which can lead to more objects to be included in the construction and updating of the map. With such an arrangement, the accuracy and robustness of tracking a location of the user within the physical environment can be improved.
In some examples, near-eye display 100 may further include one or more active illuminators 130 to project light into the physical environment. The light projected can be associated with different frequency spectrums (e.g., visible light, infra-red light, ultra-violet light, etc.), and can serve various purposes. For example, illuminator 130 may project light in a dark environment (or in an environment with low intensity of infra-red light, ultra-violet light, etc.) to assist sensors 120a-120d in capturing images of different objects within the dark environment to, for example, enable location tracking of the user. Illuminator 130 may project certain markers onto the objects within the environment, to assist the location tracking system in identifying the objects for map construction/updating.
In some examples, illuminator 130 may also enable stereoscopic imaging. For example, one or more of sensors 120a or 120b can include both a first pixel cell array for visible light sensing and a second pixel cell array for infra-red (IR) light sensing. The first pixel cell array can be overlaid with a color filter (e.g., a Bayer filter), with each pixel of the first pixel cell array being configured to measure intensity of light associated with a particular color (e.g., one of red, green or blue colors). The second pixel cell array (for IR light sensing) can also be overlaid with a filter that allows only IR light through, with each pixel of the second pixel cell array being configured to measure intensity of IR lights. The pixel cell arrays can generate an RGB image and an IR image of an object, with each pixel of the IR image being mapped to each pixel of the RGB image. Illuminator 130 may project a set of IR markers on the object, the images of which can be captured by the IR pixel cell array. Based on a distribution of the IR markers of the object as shown in the image, the system can estimate a distance of different parts of the object from the IR pixel cell array, and generate a stereoscopic image of the object based on the distances. Based on the stereoscopic image of the object, the system can determine, for example, a relative position of the object with respect to the user, and can update the image data provided to display 100 based on the relative position information to provide the interactive experience.
As discussed above, near-eye display 100 may be operated in environments associated with a very wide range of light intensities. For example, near-eye display 100 may be operated in an indoor environment or in an outdoor environment, and/or at different times of the day. Near-eye display 100 may also operate with or without active illuminator 130 being turned on. As a result, image sensors 120a-120d may need to have a wide dynamic range to be able to operate properly (e.g., to generate an output that correlates with the intensity of incident light) across a very wide range of light intensities associated with different operating environments for near-eye display 100.
As discussed above, to avoid damaging the eyeballs of the user, illuminators 140a, 140b, 140c, 140d, 140e, and 140f are typically configured to output lights of very low intensities. In a case where image sensors 150a and 150b comprise the same sensor devices as image sensors 120a-120d of
Moreover, the image sensors 120a-120d may need to be able to generate an output at a high speed to track the movements of the eyeballs. For example, a user's eyeball can perform a very rapid movement (e.g., a saccade movement) in which there can be a quick jump from one eyeball position to another. To track the rapid movement of the user's eyeball, image sensors 120a-120d need to generate images of the eyeball at high speed. For example, the rate at which the image sensors generate an image frame (the frame rate) needs to at least match the speed of movement of the eyeball. The high frame rate requires short total exposure time for all of the pixel cells involved in generating the image frame, as well as high speed for converting the sensor outputs into digital values for image generation. Moreover, as discussed above, the image sensors also need to be able to operate at an environment with low light intensity.
Waveguide display assembly 210 is configured to direct image light to an eyebox located at exit pupil 230 and to eyeball 220. Waveguide display assembly 210 may be composed of one or more materials (e.g., plastic, glass, etc.) with one or more refractive indices. In some examples, near-eye display 100 includes one or more optical elements between waveguide display assembly 210 and eyeball 220.
In some examples, waveguide display assembly 210 includes a stack of one or more waveguide displays including, but not restricted to, a stacked waveguide display, a varifocal waveguide display, etc. The stacked waveguide display is a polychromatic display (e.g., a red-green-blue (RGB) display) created by stacking waveguide displays whose respective monochromatic sources are of different colors. The stacked waveguide display is also a polychromatic display that can be projected on multiple planes (e.g., multi-planar colored display). In some configurations, the stacked waveguide display is a monochromatic display that can be projected on multiple planes (e.g., multi-planar monochromatic display). The varifocal waveguide display is a display that can adjust a focal position of image light emitted from the waveguide display. In alternate examples, waveguide display assembly 210 may include the stacked waveguide display and the varifocal waveguide display.
Waveguide display 300 includes a source assembly 310, an output waveguide 320, and a controller 330. For purposes of illustration,
Source assembly 310 generates image light 355. Source assembly 310 generates and outputs image light 355 to a coupling element 350 located on a first side 370-1 of output waveguide 320. Output waveguide 320 is an optical waveguide that outputs expanded image light 340 to an eyeball 220 of a user. Output waveguide 320 receives image light 355 at one or more coupling elements 350 located on the first side 370-1 and guides received input image light 355 to a directing element 360. In some examples, coupling element 350 couples the image light 355 from source assembly 310 into output waveguide 320. Coupling element 350 may be, e.g., a diffraction grating, a holographic grating, one or more cascaded reflectors, one or more prismatic surface elements, and/or an array of holographic reflectors.
Directing element 360 redirects the received input image light 355 to decoupling element 365 such that the received input image light 355 is decoupled out of output waveguide 320 via decoupling element 365. Directing element 360 is part of, or affixed to, first side 370-1 of output waveguide 320. Decoupling element 365 is part of, or affixed to, second side 370-2 of output waveguide 320, such that directing element 360 is opposed to the decoupling element 365. Directing element 360 and/or decoupling element 365 may be, e.g., a diffraction grating, a holographic grating, one or more cascaded reflectors, one or more prismatic surface elements, and/or an array of holographic reflectors.
Second side 370-2 represents a plane along an x-dimension and a y-dimension. Output waveguide 320 may be composed of one or more materials that facilitate total internal reflection of image light 355. Output waveguide 320 may be composed of e.g., silicon, plastic, glass, and/or polymers. Output waveguide 320 has a relatively small form factor. For example, output waveguide 320 may be approximately 50 mm wide along x-dimension, 30 mm long along y-dimension and 0.5-1 mm thick along a z-dimension.
Controller 330 controls scanning operations of source assembly 310. The controller 330 determines scanning instructions for the source assembly 310. In some examples, the output waveguide 320 outputs expanded image light 340 to the user's eyeball 220 with a large field of view (FOV). For example, the expanded image light 340 is provided to the user's eyeball 220 with a diagonal FOV (in x and y) of 60 degrees and/or greater and/or 150 degrees and/or less. The output waveguide 320 is configured to provide an eyebox with a length of 20 mm or greater and/or equal to or less than 50 mm; and/or a width of 10 mm or greater and/or equal to or less than 50 mm.
Moreover, controller 330 also controls image light 355 generated by source assembly 310, based on image data provided by image sensor 370. Image sensor 370 may be located on first side 370-1 and may include, for example, image sensors 120a-120d of
After receiving instructions from the remote console, mechanical shutter 404 can open and expose the set of pixel cells 402 in an exposure period. During the exposure period, image sensor 370 can obtain samples of lights incident on the set of pixel cells 402, and generate image data based on an intensity distribution of the incident light samples detected by the set of pixel cells 402. Image sensor 370 can then provide the image data to the remote console, which determines the display content, and provide the display content information to controller 330. Controller 330 can then determine image light 355 based on the display content information.
Source assembly 310 generates image light 355 in accordance with instructions from the controller 330. Source assembly 310 includes a source 410 and an optics system 415. Source 410 is a light source that generates coherent or partially coherent light. Source 410 may be, e.g., a laser diode, a vertical cavity surface emitting laser, and/or a light emitting diode.
Optics system 415 includes one or more optical components that condition the light from source 410. Conditioning light from source 410 may include, e.g., expanding, collimating, and/or adjusting orientation in accordance with instructions from controller 330. The one or more optical components may include one or more lenses, liquid lenses, mirrors, apertures, and/or gratings. In some examples, optics system 415 includes a liquid lens with a plurality of electrodes that allows scanning of a beam of light with a threshold value of scanning angle to shift the beam of light to a region outside the liquid lens. Light emitted from the optics system 415 (and also source assembly 310) is referred to as image light 355.
Output waveguide 320 receives image light 355. Coupling element 350 couples image light 355 from source assembly 310 into output waveguide 320. In examples where coupling element 350 is diffraction grating, a pitch of the diffraction grating is chosen such that total internal reflection occurs in output waveguide 320, and image light 355 propagates internally in output waveguide 320 (e.g., by total internal reflection), toward decoupling element 365.
Directing element 360 redirects image light 355 toward decoupling element 365 for decoupling from output waveguide 320. In examples where directing element 360 is a diffraction grating, the pitch of the diffraction grating is chosen to cause incident image light 355 to exit output waveguide 320 at angle(s) of inclination relative to a surface of decoupling element 365.
In some examples, directing element 360 and/or decoupling element 365 are structurally similar. Expanded image light 340 exiting output waveguide 320 is expanded along one or more dimensions (e.g., may be elongated along x-dimension). In some examples, waveguide display 300 includes a plurality of source assemblies 310 and a plurality of output waveguides 320. Each of source assemblies 310 emits a monochromatic image light of a specific band of wavelength corresponding to a primary color (e.g., red, green, or blue). Each of output waveguides 320 may be stacked together with a distance of separation to output an expanded image light 340 that is multi-colored.
Near-eye display 100 is a display that presents media to a user. Examples of media presented by the near-eye display 100 include one or more images, video, and/or audio. In some examples, audio is presented via an external device (e.g., speakers and/or headphones) that receives audio information from near-eye display 100 and/or control circuitries 510 and presents audio data based on the audio information to a user. In some examples, near-eye display 100 may also act as an AR eyewear glass. In some examples, near-eye display 100 augments views of a physical, real-world environment, with computer-generated elements (e.g., images, video, sound, etc.).
Near-eye display 100 includes waveguide display assembly 210, one or more position sensors 525, and/or an inertial measurement unit (IMU) 530. Waveguide display assembly 210 includes source assembly 310, output waveguide 320, and controller 330.
IMU 530 is an electronic device that generates fast calibration data indicating an estimated position of near-eye display 100 relative to an initial position of near-eye display 100 based on measurement signals received from one or more of position sensors 525.
Imaging device 535 may generate image data for various applications. For example, imaging device 535 may generate image data to provide slow calibration data in accordance with calibration parameters received from control circuitries 510. Imaging device 535 may include, for example, image sensors 120a-120d of
The input/output interface 540 is a device that allows a user to send action requests to the control circuitries 510. An action request is a request to perform a particular action. For example, an action request may be to start or end an application or to perform a particular action within the application.
Control circuitries 510 provide media to near-eye display 100 for presentation to the user in accordance with information received from one or more of: imaging device 535, near-eye display 100, and input/output interface 540. In some examples, control circuitries 510 can be housed within system 500 configured as a head-mounted device. In some examples, control circuitries 510 can be a standalone console device communicatively coupled with other components of system 500. In the example shown in
The application store 545 stores one or more applications for execution by the control circuitries 510. An application is a group of instructions, that, when executed by a processor, generates content for presentation to the user. Examples of applications include: gaming applications, conferencing applications, video playback applications, or other suitable applications.
Tracking module 550 calibrates system 500 using one or more calibration parameters and may adjust one or more calibration parameters to reduce error in determination of the position of the near-eye display 100.
Tracking module 550 tracks movements of near-eye display 100 using slow calibration information from the imaging device 535. Tracking module 550 also determines positions of a reference point of near-eye display 100 using position information from the fast calibration information.
Engine 555 executes applications within system 500 and receives position information, acceleration information, velocity information, and/or predicted future positions of near-eye display 100 from tracking module 550. In some examples, information received by engine 555 may be used for producing a signal (e.g., display instructions) to waveguide display assembly 210 that determines a type of content presented to the user. For example, to provide an interactive experience, engine 555 may determine the content to be presented to the user based on a location of the user (e.g., provided by tracking module 550), or a gaze point of the user (e.g., based on image data provided by imaging device 535), a distance between an object and user (e.g., based on image data provided by imaging device 535).
In addition, image sensor 600 also includes an illuminator 622, an optical filter 624, an imaging module 628, and a sensing controller 630. Illuminator 622 may be an infra-red illuminator, such as a laser, a light emitting diode (LED), etc., that can project infra-red light for 3D sensing. The projected light may include, for example, structured light, light pulses, etc. Optical filter 624 may include an array of filter elements overlaid on the plurality of photodiodes 612a-612d of each pixel cell including pixel cell 606a. Each filter element can set a wavelength range of incident light received by each photodiode of pixel cell 606a. For example, a filter element over photodiode 612a may transmit the visible blue light component while blocking other components, a filter element over photodiode 612b may transmit the visible green light component, a filter element over photodiode 612c may transmit the visible red light component, whereas a filter element over photodiode 612d may transmit the infra-red light component.
Image sensor 600 further includes an imaging module 628. Imaging module 628 may further include a 2D imaging module 632 to perform 2D imaging operations and a 3D imaging module 634 to perform 3D imaging operations. The operations can be based on digital values provided by ADCs 616. For example, based on the digital values from each of photodiodes 612a-612c, 2D imaging module 632 can generate an array of pixel values representing an intensity of an incident light component for each visible color channel, and generate an image frame for each visible color channel. Moreover, 3D imaging module 634 can generate a 3D image based on the digital values from photodiode 612d. In some examples, based on the digital values, 3D imaging module 634 can detect a pattern of structured light reflected by a surface of an object, and compare the detected pattern with the pattern of structured light projected by illuminator 622 to determine the depths of different points of the surface with respect to the pixel cells array. For detection of the pattern of reflected light, 3D imaging module 634 can generate pixel values based on intensities of infra-red light received at the pixel cells. As another example, 3D imaging module 634 can generate pixel values based on time-of-flight of the infra-red light transmitted by illuminator 622 and reflected by the object.
Image sensor 600 further includes a sensing controller 640 to control different components of image sensor 600 to perform 2D and 3D imaging of an object. Reference is now made to
Furthermore, image sensor 600 can also perform 3D imaging of object 704. Referring to
Each the photodiodes 612a, 612b, 612c, and 612d can be in a separate semiconductor substrate, which can be stacked to form image sensor 600. For example, photodiode 612a can be in a semiconductor substrate 840, photodiode 612b can be in a semiconductor substrate 842, photodiode 612c can be in a semiconductor substrate 844, whereas photodiode 612d can be in a semiconductor substrate 846. Each semiconductor substrate can include other photodiodes of other pixel cells, such as pixel cells 602b to receive light from spot 804b. Image sensor 600 can include another semiconductor substrate 848 which can include pixel cell processing circuits 849 which can include, for example, charge sensing units 614, ADCs 616, etc. Each semiconductor substrate can be connected to a metal interconnect, such as metal interconnects 850, 852, 854, and 856 to transfer the charge generated at each photodiode to processing circuit 849.
The arrangements of
Reference is now made to
Specifically, shutter switch M0 can be controlled by an AB signal provided by controller 920 to start an exposure period, in which the photodiode PD can generate and accumulate charge in response to incident light. Transfer switch M1 can be controlled by a TG signal provided by controller 920 to transfer some of the charge to charge storage device 902. In one quantization operation, transfer switch M1 can be biased at a partially-on state to set a quantum well capacity of photodiode PD, which also sets a quantity of residual charge stored at the photodiode PD. After the photodiode PD is saturated by the residual charge, overflow charge can flow through transfer switch M1 to charge storage device 902. In another quantization operation, transfer switch M1 can be fully turned on to transfer the residual charge from the photodiode PD to charge storage device for measurement.
Charge storage device 902 has a configurable capacity and can convert the charge transferred from switch M1 to a voltage at the OF node. Charge storage device 902 includes a CFD capacitor (e.g., a floating drain) and a CEXT capacitor (e.g., an MOS capacitor) connected by a M6 switch. M6 switch can be enabled by a LG signal to expand the capacity of charge storage device 902 by connecting CFD and CEXT capacitors in parallel, or to reduce the capacity by disconnecting the capacitors from each other. The capacity of charge storage device 902 can be reduced for measurement of residual charge to increase the charge-to-voltage gain and to reduce the quantization error. Moreover, the capacity of charge storage device 902 can also be increased for measurement of overflow charge to reduce the likelihood of saturation and to improve non-linearity. As to be described below, the capacity of charge storage device 902 can be adjusted for measurement of different light intensity ranges. Charge storage device 902 is also coupled with a reset switch M2 which can be controlled by a reset signal RST, provided by controller 920, to reset CFD and CEXT capacitors between different quantization operations.
Switchable buffer 904 can be include a switch M3 configured as a source follower to buffer the voltage at the OF node to improve its driving strength. The buffered voltage can be at the input node PIXEL_OUT of ADC 616. The M4 transistor provides a current source for switchable buffer 904 and can be biased by a VB signal. Switchable buffer 904 also includes a switch M5 which be enabled or disabled by a SEL signal. When switch M5 is disabled, source follower M3 can be disconnected from the PIXEL_OUT node. As to be described below, pixel cell 602a may include multiple charge sensing units 614 each including a switchable buffer 904, and one of the charge sensing units can be coupled with PIXEL_OUT (and ADC 616) at one time based on the SEL signal.
As described above, charge generated by photodiode PD within an exposure period can be temporarily stored in charge storage device 902 and converted to a voltage. The voltage can be quantized to represent an intensity of the incident light based on a pre-determined relationship between the charge and the incident light intensity. Reference is now made to
The definitions of low light intensity range 1006 and medium light intensity range 1008, as well as thresholds 1002 and 1004, can be based on the full well capacity of photodiode PD and the capacity of charge storage device 902. For example, low light intensity range 706 can be defined such that the total quantity of residual charge stored in photodiode PD, at the end of the exposure period, is below or equal to the storage capacity of the photodiode, and threshold 1002 can be based on the full well capacity of photodiode PD. Moreover, medium light intensity range 1008 can be defined such that the total quantity of charge stored in charge storage device 902, at the end of the exposure period, is below or equal to the storage capacity of the measurement capacitor, and threshold 1004 can be based on the storage capacity of charge storage device 902. Typically threshold 1004 is can be based on a scaled storage capacity of charge storage device 902 to ensure that when the quantity of charge stored in charge storage device 902 is measured for intensity determination, the measurement capacitor does not saturate, and the measured quantity also relates to the incident light intensity. As to be described below, thresholds 1002 and 1004 can be used to detect whether photodiode PD and charge storage device 902 saturate, which can determine the intensity range of the incident light.
In addition, in a case where the incident light intensity is within high light intensity range 1010, the total overflow charge accumulated at charge storage device 902 may exceed threshold 1004 before the exposure period ends. As additional charge is accumulated, charge storage device 902 may reach full capacity before the end of the exposure period, and charge leakage may occur. To avoid measurement error caused due to charge storage device 902 reaching full capacity, a time-to-saturation measurement can be performed to measure the time duration it takes for the total overflow charge accumulated at charge storage device 902 to reach threshold 1004. A rate of charge accumulation at charge storage device 902 can be determined based on a ratio between threshold 1004 and the time-to-saturation, and a hypothetical quantity of charge (Q3) that could have been accumulated at charge storage device 902 at the end of the exposure period (if the capacitor had limitless capacity) can be determined by extrapolation according to the rate of charge accumulation. The hypothetical quantity of charge (Q3) can provide a reasonably accurate representation of the incident light intensity within high light intensity range 1010.
Referring back to
Moreover, to measure low light intensity range 1006, transfer switch M1 can be controlled in a fully turned-on state to transfer the residual charge stored in photodiode PD to charge storage device 902. The transfer can occur after the quantization operation of the overflow charge stored at charge storage device 902 completes and after charge storage device 902 is reset. Moreover, the capacitance of charge storage device 902 can be reduced. As described above, the reduction in the capacitance of charge storage device 902 can increase the charge-to-voltage conversion ratio at charge storage device 902, such that a higher voltage can be developed for a certain quantity of stored charge. The higher charge-to-voltage conversion ratio can reduce the effect of measurement errors (e.g., quantization error, comparator offset, etc.) introduced by subsequent quantization operation on the accuracy of low light intensity determination. The measurement error can set a limit on a minimum voltage difference that can be detected and/or differentiated by the quantization operation. By increasing the charge-to-voltage conversion ratio, the quantity of charge corresponding to the minimum voltage difference can be reduced, which in turn reduces the lower limit of a measurable light intensity by pixel cell 602a and extends the dynamic range.
The charge (residual charge and/or overflow charge) accumulated at charge storage device 902 can develop an analog voltage at the OF node, which can be buffered by switchable buffer 904 at PIXEL_OUT and quantized by ADC 616. As shown in
Comparator 906 can compare an analog voltage COMP_IN, which is derived from PIXEL_OUT by the CC capacitor, against a threshold VREF, and generate a decision VOUT based on the comparison result. The CC capacitor can be used in a noise/offset compensation scheme to store the reset noise and comparator offset information in a VCC voltage, which can be added to the PIXEL_OUT voltage to generate the COMP_IN voltage, to cancel the reset noise component in the PIXEL_OUT voltage. The offset component remains in the COMP_IN voltage and can be cancelled out by the offset of comparator 906 when comparator 906 compares the COMP_IN voltage against threshold VREF to generate the decision VOUT. Comparator 906 can generate a logical one for VOUT if the COMP_IN voltage equals or exceeds VREF. Comparator 906 can also generate a logical zero for VOUT if the COMP_IN voltage falls below VREF. VOUT can control a latch signal which controls memory 912 to store a count value from counter 914.
As discussed above, ADC 616 can introduce quantization errors when there is a mismatch between a quantity of charge represented by the quantity level output by ADC 616 (e.g., represented by the total number of quantization steps) and the actual input quantity of charge that is mapped to the quantity level by ADC 808. The quantization error can be reduced by using a smaller quantization step size. In the example of
Although quantization error can be reduced by using smaller quantization step sizes, area and performance speed may limit how far the quantization step can be reduced. With smaller quantization step size, the total number of quantization steps needed to represent a particular range of charge quantities (and light intensity) may increase. A larger number of data bits may be needed to represent the increased number of quantization steps (e.g., 8 bits to represent 255 steps, 7 bits to represent 127 steps, etc.). The larger number of data bits may require additional buses to be added to pixel output buses 816, which may not be feasible if pixel cell 601 is used on a head-mounted device or other wearable devices with very limited spaces. Moreover, with a larger number of quantization step size, ADC 808 may need to cycle through a larger number of quantization steps before finding the quantity level that matches (with one quantization step), which leads to increased processing power consumption and time, and reduced rate of generating image data. The reduced rate may not be acceptable for some applications that require a high frame rate (e.g., an application that tracks the movement of the eyeball).
One way to reduce quantization error is by employing a non-uniform quantization scheme, in which the quantization steps are not uniform across the input range.
One advantage of employing a non-uniform quantization scheme is that the quantization steps for quantizing low input charge quantities can be reduced, which in turn reduces the quantization errors for quantizing the low input charge quantities, and the minimum input charge quantities that can be differentiated by ADC 616 can be reduced. Therefore, the reduced quantization errors can push down the lower limit of the measureable light intensity of the image sensor, and the dynamic range can be increased. Moreover, although the quantization errors are increased for the high input charge quantities, the quantization errors may remain small compared with high input charge quantities. Therefore, the overall quantization errors introduced to the measurement of the charge can be reduced. On the other hand, the total number of quantization steps covering the entire range of input charge quantities may remain the same (or even reduced), and the aforementioned potential problems associated with increasing the number of quantization steps (e.g., increase in area, reduction in processing speed, etc.) can be avoided.
Referring back to
Reference is now made to
Vcc(T1)=(Vref_high+Vcomp_offset)−(Vpixel_out_rst+VσKTC) (Equation 1)
At time T1, the RST signal, the AB signal, and the COMP_RST signal are released, which starts an exposure period (labelled Texposure) in which photodiode PD can accumulate and transfer charge. Exposure period Texposure can end at time T2. Between times T1 and T3, TG signal can set transfer switch M1 in a partially turned-on state to allow PD to accumulate residual charge before photodiode PD saturates. If the light intensity in the medium or high intensity ranges of
Vcomp_in(Tx)=Vpixel_out_sig1−Vpixel_out_rstVref_high+Vcomp_offset (Equation 2)
In Equation 2, the difference between Vpixel_out_sig1−Vpixel_out_rst represents the quantity of overflow charge stored in charge storage device 902. The comparator offset in the COMP_IN voltage can also cancel out the comparator offset introduced by comparator 906 when performing the comparison.
Between times T1 and T3, two phases of measurement of the COMP_IN voltage can be performed, including a time-to-saturation (TTS) measurement phase for high light intensity range 1010 and an FD ADC phase for measurement of overflow charge for medium light intensity 1008. Between times T1 and T2 (Texposure) the TTS measurement can be performed by comparing COMP_IN voltage with a static Vref_low representing a saturation level of charge storage device 902 by comparator 906. When PIXEL_OUT voltage reaches the static VREF, the output of comparator 906 (VOUT) can trip, and a count value from counter 914 at the time when VOUT trips can be stored into memory 912. At time T2, controller 920 can determine the state of VOUT of comparator 906 at the end of the TTS phase, and can assert FLAG_1 signal if VOUT is asserted. The assertion of the FLAG_1 signal can indicate that charge storage device 902 saturates and can prevent subsequent measurement phases (FD ADC and PD ADC) from overwriting the count value stored in memory 912. The count value from TTS can then be provided to represent the intensity of light received by the photodiode PD during the integration period.
Between times T2 and T3 (labelled TFDADC), the FD ADC operation can be performed by comparing COMP_IN voltage with a ramping VREF voltage that ramps from Vref_low to Vref_high, which represents the saturation level of photodiode PD (e.g., threshold 1002), as described in
Between times T3 and T4 (labelled TPDADC-transfer) can be the second reset phase, in which both RST and COMP_RST signals are asserted to reset charge storage device 902 (comprising the parallel combination of CFD capacitor and CEXT capacitor) and comparator 906 to prepare for the subsequent PD ADC operation. The VCC voltage can be set according to Equation 1.
After RST and COMP_RST are released, LG is turned off to disconnect CEXT from CFD to increase the charge-to-voltage conversion rate for the PD ADC operation. TG is set at a level to fully turn on the M1 transfer switch to transfer the residual charge stored in the photodiode PD to CFD. The residual charge develops a new PIXEL_OUT voltage, Vpixel_out_sig2. The CC capacitor can AC couple the new PIXEL_OUT voltage Vpixel_out_sig2 into COMP_IN voltage by adding the VCC voltage. Between times T3 and T4, the photodiode PD remains capable of generating additional charge in addition to the charge generated between times T1 to T3, and transferring the additional charge to charge storage device 902. The Vpixel_out_sig2 also represents the additional charge transferred between times T3 and T4. At time T4, the COMP_IN voltage can be as follows:
Vcomp_in(T4)=Vpixel_out_sig2−Vpixel_out_rst+Vref_high+Vcomp_offset (Equation 3)
In Equation 3, the difference between Vpixel_out_sig2-Vpixel_out_rst represents the quantity of charge transferred by the photodiode to charge storage device 902 between times T3 and T4. The comparator offset in the COMP_IN voltage can also cancel out the comparator offset introduced by comparator 906 when performing the comparison.
At time T4, the AB signal is asserted to prevent the photodiode PD from accumulating and transferring additional charge. Moreover, VREF can be set a static level Vref_low_margin. Comparator 906 can compare the COMP_IN voltage with Vref_low_margin to determine whether the photodiode PD saturates. Vref_low_margin is slightly higher than Vref_low, which represents the saturation level of photodiode PD (e.g., threshold 1002), to prevent false tripping of comparator 906 when the quantity of residual charge is close to but does not exceed the saturation level. Controller 920 can determine the state of VOUT of comparator 906 and can assert FLAG_2 if VOUT is asserted to indicate that photodiode PD saturates. If the FLAG_2 is asserted, memory 912 can be locked to preserve the count value stored in memory 912 (from FD ADC) and prevents memory 912 from be overwritten by the subsequent PD ADC operation.
Between times T4 and T5, controller 920 can perform the PD ADC operation by comparing the COMP_IN voltage with a VREF ramp that starts from Vref_low_margin to Vref_high. In PD ADC phase, Vref_high can represent the minimum detectable quantity of residual charge stored in photodiode PD, whereas Vref_low_margin can represent the saturation threshold of photodiode PD with margin to account for dark current, as described above. If neither FLAG_1 nor FLAG_2 is asserted prior to PD ADC, the count value obtained when comparator 906 trips during PD ADC can be stored into memory 912, and the count value from PD ADC can be provided to represent the intensity of light.
Although
Reference is now made to
After ADC 616 completes the quantization operations of the overflow charge from each photodiode, controller 920 can allow each photodiode to take turn in transferring residual charge to charge sensing unit 614 followed by a PD ADC operation. For example, between times T7 and T8 controller 920 can configure charge storage device 902 at minimum capacity, and bias transfer switch M0a at the fully-on state to transfer the residual charge from PD0 to charge sensing unit 614, followed by a PD ADC operation to quantize the residual charge between times T8 and T9. Controller 920 can reset charge storage device 902 between times T9 and T10 to remove the residual charge, and then repeat these operations for other photodiodes (e.g., PD1, PDn) after time T10.
While the arrangements of operation 1302 allow each photodiode to have the same access to the charge sensing unit and the same set of quantization operations can be performed on the outputs of each photodiode, each photodiode may have different integration periods for the overflow charge, which can degrade the global shutter operation. For example, the integration period of overflow charge for photodiode PD0 is between times T0 and T2, whereas the integration period of overflow charge for photodiode PD1 is between times T3 and T5. The different integration periods can cause motion blur as the different photodiodes may capture light from different spots when the image sensor is moving at a high speed, which can distort the images and degrade the global shutter operation.
In operation 1304, on the other hand, controller 920 can control shutter switches M0a-M0n and transfer switches M1a-M1n to allow only one photodiode to transfer overflow charge to charge sensing unit 614 to perform the TTS and FD ADC operations within the exposure period. For example, controller 920 can set shutter switch M0a at a partially-on state to allow photodiode PD0 to transfer overflow charge between times T0 and T2, in which ADC 616 can perform TTS and FD ADC operations. Controller 920 can reset charge storage device 902 between times T2 and T3. After time T3, controller 920 can control transfer switches M1a-M1n to allow all of the photodiodes to take turn to transfer residual charge to the charge sensing unit and to perform the PD ADC operations. For example, between times T3 and T5, photodiode PD0 can transfer residual charge, followed by a PD ADC operation. After charge storage device 902 is reset between times T5 and T6, another photodiode can transfer residual charge, followed by another PD ADC operation. Such arrangements can be used when, for example, the intensity of a particular component is very high compared with other components (e.g., in a dark environment with strong infra-red illumination for 3D sensing). The same exposure period can be provided for each photodiode to either accumulate charge for the strong infra-red component or for the other much weaker visible light components. The TTS and FD ADC operations can be performed on the output of the photodiode that detects the strong infra-red component (e.g., photodiode PD0), while PD ADC operations can be performed on the outputs of other photodiodes, which can improve the dynamic range of both the low intensity (for visible light) and the high intensity (for infra-red light) measurement operations.
In some examples, controller 920 can also allow some or all the photodiodes to transfer overflow charge or residual charge to the charge sensing unit simultaneously. Such arrangements can be part of a binning operation, in which the transfer switches of photodiodes configured detect the same component of incident light (e.g., photodiodes 612a and 612b of
In both operations 1302 and 1304, the quantization operations performed for each photodiode can be based on selection 922. For example, in operation 1302, it can be that only TTS operation is performed for PD0, only FD operation is performed for PD1, etc., based on selection 922. Moreover, the selection of PD0 to perform TTS and/or FD ADC operation in operation 1304 can also be based on selection 922.
Reference is now made to
In operation 1404, controller 920 can also allow each photodiode to take turn in performing the TTS operations during the exposure period, which starts at time T0. For example, between times T0 and T1 controller 920 can assert control signal SEL0 to connect charge sensing unit 614a to ADC 616 to perform a TTS operation for PD0. Also, between times T1 and T2 controller 920 can assert control signal SEL1 to connect charge sensing unit 614b to ADC 616 to perform a TTS operation for PD1. Controller 920 can repeat the TTS operations for other photodiodes. Starting at time T3, controller 920 can control the timing of the SEL0 and SEL1 signals to allow each charge sensing unit to take turn in connecting with ADC 616 to perform FD ADC operation and PD ADC operations. For example, a FD ADC operation is performed for PD0 between times T3 and T4, a PD ADC operation is performed for PD0 between times T5 and T6, followed by a FD ADC operation performed for PD1 between times T6 and T7, and a PD ADC operation is performed for PD1 between times T7 and T8.
In both operations 1402 and 1404, the quantization operations performed for each photodiode can be based on selection 922. For example, in operation 1402, it can be that only TTS operation is performed for PD0, only FD operation is performed for PD1, etc., based on selection 922. Moreover, the selection of PD0 to perform TTS and/or FD ADC operation in operation 1402 can also be based on selection 922.
In
Although
Method 1500 starts with step 1502, in which controller 920 can enable each photodiode of a plurality of photodiodes (e.g., PD0, PD1, PDn, etc.) of pixel cell 602a to generate charge in response to a different component of the incident light. As described above, pixel cell 602a may include a plurality of photodiodes. In some examples, as shown in
In step 1504, the controller can transfer the charge generated by the plurality of photodiodes to one or more charge sensing units 614 to convert to voltages. The charge may include residual charge accumulated at each photodiode, as well as overflow charge transferred by the each photodiode after the photodiode saturates. In some examples, as shown in
In step 1506, the controller can receive, for each photodiode of the plurality of photodiodes, a selection of one or more quantization processes of a plurality of quantization processes corresponding to a plurality of intensity ranges. The plurality of quantization processes may include, for example, a time-to-saturation (TTS) measurement operation to measure a time for the charge sensing unit to become saturated by the overflow charge, a FD ADC measurement operation to measure a quantity of overflow charge, and a PD ADC measurement operation to measure a quantity of residual charge. The TTS operation can be for a high light intensity range, the FD ADC measurement operation can be for a medium light intensity range, whereas the PD ADC measurement operation can be for a low light intensity range.
The selection of the quantization operations may be specific to, for example, the sharing of the charge sensing units among the photodiodes. For example, referring to
In step 1508, the controller can control one or more ADCs 616 to perform the selected quantization operations to quantize the voltages from the one or more charge sensing units to digital values representing components of a pixel of different wavelength ranges. For example, a digital value can be generated for one of the photodiodes based on the TTS, FD ADC, and PD ADC operations, whereas a digital value can be generated for other photodiodes based on the FD ADC and PD ADC operations. The quantization can be performed by a comparator (e.g., comparator 906) comparing the voltages with a static threshold (for TTS) or a ramping threshold (for FD ADC and PD ADC) to control when memory 912 stores a count value from counter 914. The count value can be the digital values.
In step 1510, the controller can store at least some of the digital values in memory 912. In some examples, memory 912 may have sufficient capacity to store a digital value for each photodiode representing a component of the pixel. In some examples, memory 912 may store the digital value of one photodiode, output the digital value to an imaging module (e.g., an imaging module 628), and then store the digital value of another photodiode. The imaging module can acquire the digital values of the pixel and construct the pixel based on the digital values, in step 1512.
Some portions of this description describe the examples of the disclosure in terms of algorithms and symbolic representations of operations on information. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, and/or hardware.
Steps, operations, or processes described may be performed or implemented with one or more hardware or software modules, alone or in combination with other devices. In some examples, a software module is implemented with a computer program product comprising a computer-readable medium containing computer program code, which can be executed by a computer processor for performing any or all of the steps, operations, or processes described.
Examples of the disclosure may also relate to an apparatus for performing the operations described. The apparatus may be specially constructed for the required purposes, and/or it may comprise a general-purpose computing device selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a non-transitory, tangible computer readable storage medium, or any type of media suitable for storing electronic instructions, which may be coupled to a computer system bus. Furthermore, any computing systems referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
Examples of the disclosure may also relate to a product that is produced by a computing process described herein. Such a product may comprise information resulting from a computing process, where the information is stored on a non-transitory, tangible computer readable storage medium and may include any example of a computer program product or other data combination described herein.
The language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the disclosure be limited not by this detailed description, but rather by any claims that issue on an application based hereon. Accordingly, the disclosure of the examples is intended to be illustrative, but not limiting, of the scope of the disclosure, which is set forth in the following claims.
This patent application claims priority to U.S. Provisional Patent Application Ser. No. 62/690,571, filed Jun. 27, 2018, entitled “Digital Pixel Sensor with Multiple Photodiodes,” and which is/are assigned to the assignee hereof and is incorporated herein by reference in its entirety for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
4596977 | Bauman et al. | Jun 1986 | A |
5053771 | McDermott | Oct 1991 | A |
5844512 | Gorin et al. | Dec 1998 | A |
5963369 | Steinthal et al. | Oct 1999 | A |
6384905 | Barrows | May 2002 | B1 |
6522395 | Bamji et al. | Feb 2003 | B1 |
6529241 | Clark | Mar 2003 | B1 |
6864817 | Salvi et al. | Mar 2005 | B1 |
6963369 | Olding | Nov 2005 | B1 |
7326903 | Ackland | Feb 2008 | B2 |
7362365 | Reyneri et al. | Apr 2008 | B1 |
7659772 | Nomura et al. | Feb 2010 | B2 |
7659925 | Krymski | Feb 2010 | B2 |
7719589 | Turchetta et al. | May 2010 | B2 |
7956914 | Xu | Jun 2011 | B2 |
8134623 | Purcell et al. | Mar 2012 | B2 |
8144227 | Kobayashi | Mar 2012 | B2 |
8369458 | Wong et al. | Feb 2013 | B2 |
8426793 | Barrows | Apr 2013 | B1 |
8754798 | Lin | Jun 2014 | B2 |
8773562 | Fan | Jul 2014 | B1 |
8779346 | Fowler et al. | Jul 2014 | B2 |
8946610 | Iwabuchi et al. | Feb 2015 | B2 |
9001251 | Smith et al. | Apr 2015 | B2 |
9094629 | Ishibashi | Jul 2015 | B2 |
9185273 | Beck et al. | Nov 2015 | B2 |
9274151 | Lee et al. | Mar 2016 | B2 |
9282264 | Park et al. | Mar 2016 | B2 |
9332200 | Hseih et al. | May 2016 | B1 |
9343497 | Cho | May 2016 | B2 |
9363454 | Ito et al. | Jun 2016 | B2 |
9478579 | Dai et al. | Oct 2016 | B2 |
9497396 | Choi | Nov 2016 | B2 |
9531990 | Wilkins et al. | Dec 2016 | B1 |
9800260 | Banerjee | Oct 2017 | B1 |
9819885 | Furukawa et al. | Nov 2017 | B2 |
9832370 | Cho et al. | Nov 2017 | B2 |
9909922 | Schweickert et al. | Mar 2018 | B2 |
9935618 | Fenigstein | Apr 2018 | B1 |
9948316 | Yun et al. | Apr 2018 | B1 |
9955091 | Dai et al. | Apr 2018 | B1 |
9967496 | Ayers et al. | May 2018 | B2 |
10003759 | Fan | Jun 2018 | B2 |
10015416 | Borthakur et al. | Jul 2018 | B2 |
10090342 | Gambino et al. | Oct 2018 | B1 |
10096631 | Ishizu | Oct 2018 | B2 |
10154221 | Ogino et al. | Dec 2018 | B2 |
10157951 | Kim et al. | Dec 2018 | B2 |
10321081 | Watanabe et al. | Jun 2019 | B2 |
10345447 | Hicks | Jul 2019 | B1 |
10419701 | Liu | Sep 2019 | B2 |
10574925 | Otaka | Feb 2020 | B2 |
10594974 | Ivarsson et al. | Mar 2020 | B2 |
10598546 | Liu | Mar 2020 | B2 |
10608101 | Liu | Mar 2020 | B2 |
10686996 | Liu | Jun 2020 | B2 |
10726627 | Liu | Jul 2020 | B2 |
10750097 | Liu | Aug 2020 | B2 |
10764526 | Liu et al. | Sep 2020 | B1 |
10804926 | Gao et al. | Oct 2020 | B2 |
10812742 | Chen et al. | Oct 2020 | B2 |
10825854 | Liu | Nov 2020 | B2 |
10834344 | Chen et al. | Nov 2020 | B2 |
10897586 | Liu et al. | Jan 2021 | B2 |
10903260 | Chen et al. | Jan 2021 | B2 |
10917589 | Liu | Feb 2021 | B2 |
10923523 | Liu et al. | Feb 2021 | B2 |
10931884 | Liu et al. | Feb 2021 | B2 |
10951849 | Liu | Mar 2021 | B2 |
10969273 | Berkovich et al. | Apr 2021 | B2 |
11004881 | Liu et al. | May 2021 | B2 |
11057581 | Liu | Jul 2021 | B2 |
11089210 | Berkovich et al. | Aug 2021 | B2 |
11089241 | Chen et al. | Aug 2021 | B2 |
20020067303 | Lee et al. | Jun 2002 | A1 |
20020113886 | Hynecek | Aug 2002 | A1 |
20030001080 | Kummaraguntla et al. | Jan 2003 | A1 |
20030020100 | Guidash | Jan 2003 | A1 |
20030049925 | Layman et al. | Mar 2003 | A1 |
20040095495 | Inokuma et al. | May 2004 | A1 |
20040118994 | Mizuno | Jun 2004 | A1 |
20040251483 | Ko et al. | Dec 2004 | A1 |
20050046715 | Lim et al. | Mar 2005 | A1 |
20050057389 | Krymsk | Mar 2005 | A1 |
20050104983 | Raynor | May 2005 | A1 |
20050206414 | Cottin et al. | Sep 2005 | A1 |
20050237380 | Kaku et al. | Oct 2005 | A1 |
20050280727 | Sato et al. | Dec 2005 | A1 |
20060023109 | Mabuchi et al. | Feb 2006 | A1 |
20060146159 | Farrier | Jul 2006 | A1 |
20060158541 | Ichikawa | Jul 2006 | A1 |
20070013983 | Kitamura et al. | Jan 2007 | A1 |
20070076109 | Krymski | Apr 2007 | A1 |
20070076481 | Tennant | Apr 2007 | A1 |
20070092244 | Pertsel | Apr 2007 | A1 |
20070102740 | Ellis-Monaghan et al. | May 2007 | A1 |
20070131991 | Sugawa | Jun 2007 | A1 |
20070208526 | Staudt et al. | Sep 2007 | A1 |
20070222881 | Mentzer | Sep 2007 | A1 |
20080001065 | Ackland | Jan 2008 | A1 |
20080007731 | Botchway et al. | Jan 2008 | A1 |
20080042046 | Mabuchi | Feb 2008 | A1 |
20080042888 | Danesh | Feb 2008 | A1 |
20080068478 | Watanabe | Mar 2008 | A1 |
20080088014 | Adkisson et al. | Apr 2008 | A1 |
20080191791 | Nomura et al. | Aug 2008 | A1 |
20080226170 | Sonoda | Sep 2008 | A1 |
20080226183 | Lei et al. | Sep 2008 | A1 |
20090002528 | Manabe et al. | Jan 2009 | A1 |
20090033588 | Kajita et al. | Feb 2009 | A1 |
20090040364 | Rubner | Feb 2009 | A1 |
20090066820 | Jiang et al. | Mar 2009 | A1 |
20090091645 | Trimeche et al. | Apr 2009 | A1 |
20090128640 | Yumiki | May 2009 | A1 |
20090224139 | Buettgen et al. | Sep 2009 | A1 |
20090244328 | Yamashita | Oct 2009 | A1 |
20090244346 | Funaki | Oct 2009 | A1 |
20090245637 | Barman et al. | Oct 2009 | A1 |
20090261235 | Lahav et al. | Oct 2009 | A1 |
20090303371 | Watanabe et al. | Dec 2009 | A1 |
20090321615 | Sugiyama et al. | Dec 2009 | A1 |
20100013969 | Ui | Jan 2010 | A1 |
20100140732 | Eminoglu et al. | Jun 2010 | A1 |
20100194956 | Yuan et al. | Aug 2010 | A1 |
20100232227 | Lee | Sep 2010 | A1 |
20100276572 | Iwabuchi et al. | Nov 2010 | A1 |
20110049589 | Chuang et al. | Mar 2011 | A1 |
20110122304 | Sedelnikov | May 2011 | A1 |
20110149116 | Kim | Jun 2011 | A1 |
20110155892 | Neter et al. | Jun 2011 | A1 |
20110254986 | Nishimura et al. | Oct 2011 | A1 |
20110298074 | Funao | Dec 2011 | A1 |
20120016817 | Smith et al. | Jan 2012 | A1 |
20120039548 | Wang et al. | Feb 2012 | A1 |
20120068051 | Ahn | Mar 2012 | A1 |
20120075511 | Tay | Mar 2012 | A1 |
20120092677 | Suehira et al. | Apr 2012 | A1 |
20120105475 | Tseng | May 2012 | A1 |
20120105668 | Velarde et al. | May 2012 | A1 |
20120113119 | Massie | May 2012 | A1 |
20120127284 | Bar-Zeev et al. | May 2012 | A1 |
20120133807 | Wu et al. | May 2012 | A1 |
20120138775 | Cheon et al. | Jun 2012 | A1 |
20120153123 | Mao et al. | Jun 2012 | A1 |
20120188420 | Black et al. | Jul 2012 | A1 |
20120200499 | Osterhout et al. | Aug 2012 | A1 |
20120212465 | White et al. | Aug 2012 | A1 |
20120241591 | Wan et al. | Sep 2012 | A1 |
20120262616 | Sa et al. | Oct 2012 | A1 |
20120267511 | Kozlowski | Oct 2012 | A1 |
20120273654 | Hynecek et al. | Nov 2012 | A1 |
20120273906 | Mackey et al. | Nov 2012 | A1 |
20120305751 | Kusuda | Dec 2012 | A1 |
20120327279 | Hashimoto et al. | Dec 2012 | A1 |
20130020466 | Ayers et al. | Jan 2013 | A1 |
20130056809 | Mao et al. | Mar 2013 | A1 |
20130057742 | Nakamura et al. | Mar 2013 | A1 |
20130068929 | Solhusvik et al. | Mar 2013 | A1 |
20130069787 | Petrou | Mar 2013 | A1 |
20130082313 | Manabe | Apr 2013 | A1 |
20130113969 | Manabe et al. | May 2013 | A1 |
20130120615 | Hirooka et al. | May 2013 | A1 |
20130120625 | Ishii et al. | May 2013 | A1 |
20130126710 | Kondo | May 2013 | A1 |
20130141619 | Lim et al. | Jun 2013 | A1 |
20130187027 | Qiao et al. | Jul 2013 | A1 |
20130207219 | Ahn | Aug 2013 | A1 |
20130214371 | Asatsuma et al. | Aug 2013 | A1 |
20130218728 | Hashop et al. | Aug 2013 | A1 |
20130221194 | Manabe | Aug 2013 | A1 |
20130229543 | Hashimoto et al. | Sep 2013 | A1 |
20130229560 | Kondo | Sep 2013 | A1 |
20130234029 | Bikumandla | Sep 2013 | A1 |
20130248685 | Ahn | Sep 2013 | A1 |
20130293752 | Peng et al. | Nov 2013 | A1 |
20130299674 | Fowler et al. | Nov 2013 | A1 |
20130300906 | Yan | Nov 2013 | A1 |
20140021574 | Egawa | Jan 2014 | A1 |
20140042299 | Wan et al. | Feb 2014 | A1 |
20140042582 | Kondo | Feb 2014 | A1 |
20140078336 | Beck et al. | Mar 2014 | A1 |
20140085523 | Hynecek | Mar 2014 | A1 |
20140176770 | Kondo | Jun 2014 | A1 |
20140211052 | Choi | Jul 2014 | A1 |
20140232890 | Yoo et al. | Aug 2014 | A1 |
20140247382 | Moldovan et al. | Sep 2014 | A1 |
20140306276 | Yamaguchi | Oct 2014 | A1 |
20140313387 | Vogelsang et al. | Oct 2014 | A1 |
20140368687 | Yu et al. | Dec 2014 | A1 |
20150048427 | Hu et al. | Feb 2015 | A1 |
20150083895 | Hashimoto et al. | Mar 2015 | A1 |
20150085134 | Novotny et al. | Mar 2015 | A1 |
20150090863 | Mansoorian et al. | Apr 2015 | A1 |
20150172574 | Honda et al. | Jun 2015 | A1 |
20150189209 | Yang et al. | Jul 2015 | A1 |
20150201142 | Smith et al. | Jul 2015 | A1 |
20150208009 | Oh et al. | Jul 2015 | A1 |
20150229859 | Guidash et al. | Aug 2015 | A1 |
20150237274 | Yang et al. | Aug 2015 | A1 |
20150279884 | Kusumoto | Oct 2015 | A1 |
20150287766 | Kim et al. | Oct 2015 | A1 |
20150309311 | Cho | Oct 2015 | A1 |
20150309316 | Osterhout et al. | Oct 2015 | A1 |
20150312461 | Kim et al. | Oct 2015 | A1 |
20150312502 | Borremans | Oct 2015 | A1 |
20150312557 | Kim | Oct 2015 | A1 |
20150350582 | Korobov et al. | Dec 2015 | A1 |
20150358569 | Egawa | Dec 2015 | A1 |
20150358571 | Dominguez Castro et al. | Dec 2015 | A1 |
20150358593 | Sato | Dec 2015 | A1 |
20150381907 | Boettiger et al. | Dec 2015 | A1 |
20150381911 | Shen et al. | Dec 2015 | A1 |
20160011422 | Thurber et al. | Jan 2016 | A1 |
20160018645 | Haddick et al. | Jan 2016 | A1 |
20160021302 | Cho et al. | Jan 2016 | A1 |
20160028974 | Guidash et al. | Jan 2016 | A1 |
20160028980 | Kameyama et al. | Jan 2016 | A1 |
20160035770 | Ahn et al. | Feb 2016 | A1 |
20160037111 | Dai et al. | Feb 2016 | A1 |
20160078614 | Ryu et al. | Mar 2016 | A1 |
20160088253 | Tezuka | Mar 2016 | A1 |
20160093659 | Nakamura et al. | Mar 2016 | A1 |
20160100115 | Kusano | Apr 2016 | A1 |
20160111457 | Sekine | Apr 2016 | A1 |
20160112626 | Shimada | Apr 2016 | A1 |
20160118992 | Milkov | Apr 2016 | A1 |
20160165160 | Hseih et al. | Jun 2016 | A1 |
20160197117 | Nakata et al. | Jul 2016 | A1 |
20160204150 | Oh et al. | Jul 2016 | A1 |
20160210785 | Balachandreswaran et al. | Jul 2016 | A1 |
20160225813 | Liao et al. | Aug 2016 | A1 |
20160240570 | Barna et al. | Aug 2016 | A1 |
20160249004 | Saeki et al. | Aug 2016 | A1 |
20160255293 | Gesset | Sep 2016 | A1 |
20160276394 | Chou et al. | Sep 2016 | A1 |
20160307945 | Madurawe | Oct 2016 | A1 |
20160337605 | Ito | Nov 2016 | A1 |
20160353045 | Kawahito et al. | Dec 2016 | A1 |
20160360127 | Dierickx et al. | Dec 2016 | A1 |
20170013215 | McCarten | Jan 2017 | A1 |
20170039906 | Jepsen | Feb 2017 | A1 |
20170041571 | Tyrrell et al. | Feb 2017 | A1 |
20170053962 | Oh et al. | Feb 2017 | A1 |
20170059399 | Suh et al. | Mar 2017 | A1 |
20170062501 | Velichko et al. | Mar 2017 | A1 |
20170069363 | Baker | Mar 2017 | A1 |
20170070691 | Nishikido | Mar 2017 | A1 |
20170099422 | Goma et al. | Apr 2017 | A1 |
20170099446 | Cremers et al. | Apr 2017 | A1 |
20170104021 | Park et al. | Apr 2017 | A1 |
20170104946 | Hong | Apr 2017 | A1 |
20170111600 | Wang et al. | Apr 2017 | A1 |
20170141147 | Raynor | May 2017 | A1 |
20170154909 | Ishizu | Jun 2017 | A1 |
20170170223 | Hynecek et al. | Jun 2017 | A1 |
20170201693 | Sugizaki et al. | Jul 2017 | A1 |
20170207268 | Kurokawa | Jul 2017 | A1 |
20170228345 | Gupta et al. | Aug 2017 | A1 |
20170270664 | Hoogi et al. | Sep 2017 | A1 |
20170272667 | Hynecek | Sep 2017 | A1 |
20170272768 | Tall et al. | Sep 2017 | A1 |
20170280031 | Price et al. | Sep 2017 | A1 |
20170293799 | Skogo et al. | Oct 2017 | A1 |
20170310910 | Smith et al. | Oct 2017 | A1 |
20170324917 | Mlinar et al. | Nov 2017 | A1 |
20170338262 | Hirata | Nov 2017 | A1 |
20170339327 | Koshkin et al. | Nov 2017 | A1 |
20170346579 | Barghi | Nov 2017 | A1 |
20170350755 | Geurts | Dec 2017 | A1 |
20170359497 | Mandelli et al. | Dec 2017 | A1 |
20170366766 | Geurts et al. | Dec 2017 | A1 |
20180019269 | Klipstein | Jan 2018 | A1 |
20180077368 | Suzuki | Mar 2018 | A1 |
20180084164 | Hynecek et al. | Mar 2018 | A1 |
20180115725 | Zhang et al. | Apr 2018 | A1 |
20180136471 | Miller et al. | May 2018 | A1 |
20180143701 | Suh et al. | May 2018 | A1 |
20180152650 | Sakakibara et al. | May 2018 | A1 |
20180167575 | Watanabe et al. | Jun 2018 | A1 |
20180176545 | Aflaki Beni | Jun 2018 | A1 |
20180204867 | Kim et al. | Jul 2018 | A1 |
20180213205 | Oh | Jul 2018 | A1 |
20180220093 | Murao et al. | Aug 2018 | A1 |
20180224658 | Teller | Aug 2018 | A1 |
20180227516 | Mo et al. | Aug 2018 | A1 |
20180241953 | Johnson | Aug 2018 | A1 |
20180270436 | Ivarsson et al. | Sep 2018 | A1 |
20180276841 | Krishnaswamy et al. | Sep 2018 | A1 |
20180286896 | Kim et al. | Oct 2018 | A1 |
20180376046 | Liu | Dec 2018 | A1 |
20180376090 | Liu | Dec 2018 | A1 |
20190035154 | Liu | Jan 2019 | A1 |
20190046044 | Tzvieli et al. | Feb 2019 | A1 |
20190052788 | Liu | Feb 2019 | A1 |
20190056264 | Liu | Feb 2019 | A1 |
20190057995 | Liu | Feb 2019 | A1 |
20190058058 | Liu | Feb 2019 | A1 |
20190098232 | Mori et al. | Mar 2019 | A1 |
20190104263 | Ochiai et al. | Apr 2019 | A1 |
20190104265 | Totsuka et al. | Apr 2019 | A1 |
20190110039 | Linde et al. | Apr 2019 | A1 |
20190123088 | Kwon | Apr 2019 | A1 |
20190124285 | Otaka | Apr 2019 | A1 |
20190141270 | Otaka et al. | May 2019 | A1 |
20190149751 | Wise | May 2019 | A1 |
20190157330 | Sato et al. | May 2019 | A1 |
20190172227 | Kasahara | Jun 2019 | A1 |
20190172868 | Chen et al. | Jun 2019 | A1 |
20190191116 | Madura | Jun 2019 | A1 |
20190246036 | Wu et al. | Aug 2019 | A1 |
20190253650 | Kim | Aug 2019 | A1 |
20190327439 | Chen et al. | Oct 2019 | A1 |
20190331914 | Lee et al. | Oct 2019 | A1 |
20190335151 | Rivard et al. | Oct 2019 | A1 |
20190348460 | Chen et al. | Nov 2019 | A1 |
20190355782 | Do et al. | Nov 2019 | A1 |
20190363118 | Berkovich et al. | Nov 2019 | A1 |
20190371845 | Chen et al. | Dec 2019 | A1 |
20190379827 | Berkovich et al. | Dec 2019 | A1 |
20190379846 | Chen et al. | Dec 2019 | A1 |
20200007800 | Berkovich et al. | Jan 2020 | A1 |
20200053299 | Zhang et al. | Feb 2020 | A1 |
20200059589 | Liu et al. | Feb 2020 | A1 |
20200068189 | Chen et al. | Feb 2020 | A1 |
20200186731 | Chen et al. | Jun 2020 | A1 |
20200195875 | Berkovich et al. | Jun 2020 | A1 |
20200217714 | Liu | Jul 2020 | A1 |
20200228745 | Otaka | Jul 2020 | A1 |
20200374475 | Fukuoka et al. | Nov 2020 | A1 |
20210026796 | Graif et al. | Jan 2021 | A1 |
20210099659 | Miyauchi et al. | Apr 2021 | A1 |
20210185264 | Wong et al. | Jun 2021 | A1 |
20210227159 | Sambonsugi | Jul 2021 | A1 |
20210368124 | Berkovich et al. | Nov 2021 | A1 |
Number | Date | Country |
---|---|---|
1490878 | Apr 2004 | CN |
1728397 | Feb 2006 | CN |
1812506 | Aug 2006 | CN |
103207716 | Jul 2013 | CN |
103730455 | Apr 2014 | CN |
104125418 | Oct 2014 | CN |
104204904 | Dec 2014 | CN |
104469195 | Mar 2015 | CN |
104704812 | Jun 2015 | CN |
104733485 | Jun 2015 | CN |
104754255 | Jul 2015 | CN |
105706439 | Jun 2016 | CN |
106255978 | Dec 2016 | CN |
106791504 | May 2017 | CN |
109298528 | Feb 2019 | CN |
202016105510 | Oct 2016 | DE |
0675345 | Oct 1995 | EP |
1681856 | Jul 2006 | EP |
1732134 | Dec 2006 | EP |
1746820 | Jan 2007 | EP |
1788802 | May 2007 | EP |
2037505 | Mar 2009 | EP |
2063630 | May 2009 | EP |
2538664 | Dec 2012 | EP |
2804074 | Nov 2014 | EP |
2833619 | Feb 2015 | EP |
3032822 | Jun 2016 | EP |
3229457 | Oct 2017 | EP |
3258683 | Dec 2017 | EP |
3425352 | Jan 2019 | EP |
3439039 | Feb 2019 | EP |
3744085 | Dec 2020 | EP |
2002199292 | Jul 2002 | JP |
2003319262 | Nov 2003 | JP |
2005328493 | Nov 2005 | JP |
2006203736 | Aug 2006 | JP |
2007074447 | Mar 2007 | JP |
2011216966 | Oct 2011 | JP |
2012095349 | May 2012 | JP |
2013009087 | Jan 2013 | JP |
2013172203 | Sep 2013 | JP |
2014165733 | Sep 2014 | JP |
2016092661 | May 2016 | JP |
100574959 | Apr 2006 | KR |
20110050351 | May 2011 | KR |
20120058337 | Jun 2012 | KR |
20150095841 | Aug 2015 | KR |
20160008267 | Jan 2016 | KR |
20160008287 | Jan 2016 | KR |
2006124592 | Nov 2006 | WO |
WO-2014055391 | Apr 2014 | WO |
WO-2016095057 | Jun 2016 | WO |
WO-2017003477 | Jan 2017 | WO |
WO-2017013806 | Jan 2017 | WO |
WO-2017047010 | Mar 2017 | WO |
2017058488 | Apr 2017 | WO |
2017069706 | Apr 2017 | WO |
2017169882 | Oct 2017 | WO |
WO-2017169446 | Oct 2017 | WO |
WO-2019018084 | Jan 2019 | WO |
WO-2019111528 | Jun 2019 | WO |
WO-2019145578 | Aug 2019 | WO |
2019168929 | Sep 2019 | WO |
Entry |
---|
U.S. Appl. No. 16/431,693, “Non-Final Office Action”, dated Jan. 30, 2020, 6 pages. |
U.S. Appl. No. 16/436,049, “Non-Final Office Action”, dated Mar. 4, 2020, 9 pages. |
U.S. Appl. No. 15/668,241 , Advisory Action, dated Oct. 23, 2019, 5 pages. |
U.S. Appl. No. 15/668,241 , Final Office Action, dated Jun. 17, 2019, 19 pages. |
U.S. Appl. No. 15/668,241 , Non-Final Office Action, dated Dec. 21, 2018, 3 pages. |
U.S. Appl. No. 15/668,241 , Notice of Allowance, dated Jun. 29, 2020, 8 pages. |
U.S. Appl. No. 15/668,241 , Notice of Allowance, dated Mar. 5, 2020, 8 pages. |
U.S. Appl. No. 15/668,241 , “Supplemental Notice of Allowability”, dated Apr. 29, 2020, 5 pages. |
U.S. Appl. No. 15/719,345 , Final Office Action, dated Apr. 29, 2020, 14 pages. |
U.S. Appl. No. 15/719,345 , Non-Final Office Action, dated Nov. 25, 2019, 14 pages. |
U.S. Appl. No. 15/719,345 , Notice of Allowance, dated Aug. 12, 2020, 11 pages. |
U.S. Appl. No. 15/801,216 , Advisory Action, dated Apr. 7, 2020, 3 pages. |
U.S. Appl. No. 15/801,216 , Final Office Action, dated Dec. 26, 2019, 5 pages. |
U.S. Appl. No. 15/801,216 , Non-Final Office Action, dated Jun. 27, 2019, 13 pages. |
U.S. Appl. No. 15/801,216 , Notice of Allowance, dated Jun. 23, 2020, 5 pages. |
U.S. Appl. No. 15/847,517 , Non-Final Office Action, dated Nov. 23, 2018, 21 pages. |
U.S. Appl. No. 15/847,517 , Notice of Allowance, dated May 1, 2019, 11 pages. |
U.S. Appl. No. 15/861,588 , Non-Final Office Action, dated Jul. 10, 2019, 11 pages. |
U.S. Appl. No. 15/861,588 , Notice of Allowance, dated Nov. 26, 2019, 9 pages. |
U.S. Appl. No. 15/876,061 , “Corrected Notice of Allowability”, dated Apr. 28, 2020, 3 pages. |
U.S. Appl. No. 15/876,061 , Non-Final Office Action, dated Sep. 18, 2019, 23 pages. |
U.S. Appl. No. 15/876,061 , “Notice of Allowability”, dated May 6, 2020, 2 pages. |
U.S. Appl. No. 15/876,061 , Notice of Allowance, dated Feb. 4, 2020, 13 pages. |
U.S. Appl. No. 15/927,896 , Non-Final Office Action, dated May 1, 2019, 10 pages. |
U.S. Appl. No. 15/983,379 , Notice of Allowance, dated Oct. 18, 2019, 9 pages. |
U.S. Appl. No. 15/983,391 , Non-Final Office Action, dated Aug. 29, 2019, 12 pages. |
U.S. Appl. No. 15/983,391 , Notice of Allowance, dated Apr. 8, 2020, 8 pages. |
U.S. Appl. No. 16/177,971 , Final Office Action, dated Feb. 27, 2020, 9 pages. |
U.S. Appl. No. 16/177,971 , Non-Final Office Action, dated Sep. 25, 2019, 9 pages. |
U.S. Appl. No. 16/177,971 , Notice of Allowance, dated Apr. 24, 2020, 6 pages. |
U.S. Appl. No. 16/210,748 , Final Office Action, dated Jul. 7, 2020, 11 pages. |
U.S. Appl. No. 16/210,748 , Non-Final Office Action, dated Jan. 31, 2020, 11 pages. |
U.S. Appl. No. 16/249,420 , Non-Final Office Action, dated Jul. 22, 2020, 9 pages. |
U.S. Appl. No. 16/286,355 , Non-Final Office Action, dated Oct. 1, 2019, 6 pages. |
U.S. Appl. No. 16/286,355 , Notice of Allowance, dated Feb. 12, 2020, 7 pages. |
U.S. Appl. No. 16/286,355 , Notice of Allowance, dated Jun. 4, 2020, 7 pages. |
U.S. Appl. No. 16/369,763 , Non-Final Office Action, dated Jul. 22, 2020, 15 pages. |
U.S. Appl. No. 16/382,015 , Notice of Allowance, dated Jun. 11, 2020, 11 pages. |
U.S. Appl. No. 16/384,720 , Non-Final Office Action, dated May 1, 2020, 6 pages. |
U.S. Appl. No. 16/431,693 , Notice of Allowance, dated Jun. 24, 2020, 7 pages. |
U.S. Appl. No. 16/435,449 , Notice of Allowance, dated Jul. 27, 2020, 8 pages. |
U.S. Appl. No. 16/436,049 , Non-Final Office Action, dated Jun. 30, 2020, 11 pages. |
U.S. Appl. No. 16/454,787 , Notice of Allowance, dated Jul. 9, 2020, 9 pages. |
U.S. Appl. No. 16/566,583 , Final Office Action, dated Apr. 15, 2020, 24 pages. |
U.S. Appl. No. 16/566,583 , Non-Final Office Action, dated Oct. 1, 2019, 10 pages. |
U.S. Appl. No. 16/566,583 , Non-Final Office Action, dated Jul. 27, 2020, 11 pages. |
Application No. EP18179838.0 , Extended European Search Report, dated May 24, 2019, 17 pages. |
EP18179838.0, “Partial European Search Report”, dated Dec. 5, 2018, 14 pages. |
Application No. EP18179846.3 , Extended European Search Report, dated Dec. 7, 2018, 10 pages. |
Application No. EP18179851.3 , Extended European Search Report, dated Dec. 7, 2018, 8 pages. |
Application No. EP18188684.7 , Extended European Search Report, dated Jan. 16, 2019, 10 pages. |
Application No. EP18188684.7 , Office Action, dated Nov. 26, 2019, 9 pages. |
Application No. EP18188962.7 , Extended European Search Report, dated Oct. 23, 2018, 8 pages. |
Application No. EP18188962.7 , Office Action, dated Aug. 28, 2019, 6 pages. |
Application No. EP18188968.4 , Extended European Search Report, dated Oct. 23, 2018, 8 pages. |
Application No. EP18188968.4 , Office Action, dated Aug. 14, 2019, 5 pages. |
Application No. EP18189100.3 , Extended European Search Report, dated Oct. 9, 2018, 8 pages. |
Application No. PCT/US2018/039350 , International Preliminary Report on Patentability, dated Jan. 9, 2020, 10 pages. |
Application No. PCT/US2018/039350 , International Search Report and Written Opinion, dated Nov. 15, 2018, 13 pages. |
Application No. PCT/US2018/039352 , International Search Report and Written Opinion, dated Oct. 26, 2018, 10 pages. |
Application No. PCT/US2018/039431 , International Search Report and Written Opinion, dated Nov. 7, 2018, 14 pages. |
Application No. PCT/US2018/045661 , International Search Report and Written Opinion, dated Nov. 30, 2018, 11 Pages. |
Application No. PCT/US2018/045666 , International Preliminary Report on Patentability, dated Feb. 27, 2020, 11 pages. |
Application No. PCT/US2018/045666 , International Search Report and Written Opinion, dated Dec. 3, 2018, 13 pages. |
Application No. PCT/US2018/045673 , International Search Report and Written Opinion, dated Dec. 4, 2018, 13 pages. |
Application No. PCT/US2018/046131 , International Search Report and Written Opinion, dated Dec. 3, 2018, 10 pages. |
Application No. PCT/US2018/064181 , International Preliminary Report on Patentability, dated Jun. 18, 2020, 9 pages. |
Application No. PCT/US2018/064181 , International Search Report and Written Opinion, dated Mar. 29, 2019, 12 pages. |
Application No. PCT/US2019/014044 , International Search Report and Written Opinion, dated May 8, 2019, 11 pages. |
Application No. PCT/US2019/019756 , International Search Report and Written Opinion, dated Jun. 13, 2019, 11 pages. |
Application No. PCT/US2019/025170 , International Search Report and Written Opinion, dated Jul. 9, 2019, 11 pages. |
Application No. PCT/US2019/027727 , International Search Report and Written Opinion, dated Jun. 27, 2019, 11 pages. |
Application No. PCT/US2019/027729, International Search Report and Written Opinion, dated Jun. 27, 2019, 10 pages. |
Application No. PCT/US2019/031521 , International Search Report and Written Opinion, dated Jul. 11, 2019, 11 pages. |
Application No. PCT/US2019/036575, International Search Report and Written Opinion, dated Sep. 30, 2019, 16 pages. |
Application No. PCT/US2019/049756 , International Search Report and Written Opinion, dated Dec. 16, 2019, 8 pages. |
Application No. PCT/US2019/059754 , International Search Report and Written Opinion, dated Mar. 24, 2020, 15 pages. |
Application No. PCT/US2019/065430 , International Search Report and Written Opinion, dated Mar. 6, 2020, 15 pages. |
Tanner et al., “Low-Power Digital Image Sensor for Still Picture Image Acquisition”, Visual Communications and Image Processing, vol. 4306, Jan. 22, 2001, 8 pages. |
Xu et al., “A New Digital-Pixel Architecture for CMOS Image Sensor With Pixel-Level ADC and Pulse Width Modulation using A 0.18 Mu M CMOS Technology”, Institute of Electrical and Electronics Engineers Conference on Electron Devices and Solid-State Circuits, Dec. 16-18, 2003, pp. 265-268. |
U.S. Appl. No. 16/249,420, “Notice of Allowance”, dated Nov. 18, 2020, 8 pages. |
U.S. Appl. No. 15/719,345, “Notice of Allowance”, dated Sep. 3, 2020, 12 pages. |
U.S. Appl. No. 16/435,449, “Notice of Allowance”, dated Sep. 16, 2020, 7 pages. |
U.S. Appl. No. 16/454,787, “Notice of Allowance”, dated Sep. 9, 2020, 9 pages. |
U.S. Appl. No. 16/707,988, “Non-Final Office Action”, dated Sep. 22, 2020, 15 pages. |
U.S. Appl. No. 16/454,787, Notice of Allowance dated Apr. 22, 2020, 10 pages. |
Cho et al., Low Power Dual CDS for a Column-Parallel CMOS Image Sensor, JSTS: Journal of Semiconductor Technology and Science, vol. 12, No. 4, Dec. 30, 2012, pp. 388-396. |
Kavusi et al., Quantitative Study of High-Dynamic-Range Image Sensor Architectures, Proceedings of SPIE, The International Society for Optical Engineering, vol. 5301, Jun. 7, 2004, pp. 264-275. |
International Application No. PCT/US2019/035724, International Search Report and Written Opinion dated Sep. 10, 2019, 12 pages. |
International Application No. PCT/US2019/036484, International Search Report and Written Opinion dated Sep. 19, 2019, 10 pages. |
International Application No. PCT/US2019/036492, International Search Report and Written Opinion dated Sep. 25, 2019, 9 pages. |
International Application No. PCT/US2019/036536, International Search Report and Written Opinion dated Sep. 26, 2019, 14 pages. |
International Application No. PCT/US2019/039410, International Search Report and Written Opinion dated Sep. 30, 2019, 11 pages. |
International Application No. PCT/US2019/039758, International Search Report and Written Opinion dated Oct. 11, 2019, 13 pages. |
International Application No. PCT/US2019/047156, International Search Report and Written Opinion dated Oct. 23, 2019, 9 pages. |
International Application No. PCT/US2019/048241, International Search Report and Written Opinion dated Jan. 28, 2020, 16 pages. |
Snoeij, A Low Power Column-Parallel 12-Bit ADC for CMOS Imagers, Proceedings IEEE Workshop on Charge-Coupled Devices Andadvanced Image Sensors (CCDS & AIS), IEEE, Jun. 1, 2005, pp. 169-172. |
U.S. Appl. No. 16/829,249, “Non-Final Office Action”, dated Apr. 27, 2021, 9 pages. |
Advisory Action dated Oct. 8, 2020 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 4 Pages. |
Amir M.F., et al., “3-D Stacked Image Sensor With Deep Neural Network Computation,” IEEE Sensors Journal, IEEE Service Center, New York, NY, US, May 15, 2018, vol. 18 (10), pp. 4187-4199, XP011681876. |
Chuxi L., et al., “A Memristor-Based Processing-in-Memory Architechture for Deep Convolutional Neural Networks Approximate Computation,” Journal of Computer Research and Development, Jun. 30, 2017, vol. 54 (6), pp. 1367-1380. |
Communication Pursuant Article 94(3) dated Dec. 23, 2021 for European Application No. 19744961.4, filed Jun. 28, 2019, 8 pages. |
Communication Pursuant Article 94(3) dated Jan. 5, 2022 for European Application No. 19740456.9, filed Jun. 27, 2019, 12 pages. |
Corrected Notice of Allowability dated Feb. 3, 2021 for U.S. Appl. No. 16/566,583, filed Sep. 10, 2019, 2 Pages. |
Corrected Notice of Allowability dated Apr. 9, 2021 for U.S. Appl. No. 16/255,528, filed Jan. 23, 2019, 5 Pages. |
Corrected Notice of Allowability dated Dec. 11, 2020 for U.S. Appl. No. 16/566,583, filed Sep. 10, 2019, 2 Pages. |
Corrected Notice of Allowability dated Jul. 26, 2021 for U.S. Appl. No. 16/707,988, filed Dec. 9, 2019, 2 Pages. |
Extended European Search Report for European Application No. 19743908.6, dated Sep. 30, 2020, 9 Pages. |
Final Office Action dated Dec. 3, 2021 for U.S. Appl. No. 17/072,840, filed Oct. 16, 2020, 23 pages. |
Final Office Action dated Jul. 12, 2021 for U.S. Appl. No. 16/435,451, filed Jun. 7, 2019, 13 Pages. |
Final Office Action dated Oct. 18, 2021 for U.S. Appl. No. 16/716,050, filed Dec. 16, 2019, 18 Pages. |
Final Office Action dated Oct. 21, 2021 for U.S. Appl. No. 16/421,441, filed May 23, 2019, 23 Pages. |
Final Office Action dated Jan. 27, 2021 for U.S. Appl. No. 16/255,528, filed Jan. 23, 2019, 31 Pages. |
Final Office Action dated Jul. 28, 2021 for U.S. Appl. No. 17/083,920, filed Oct. 29, 2020, 19 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2019/014904, dated Aug. 5, 2019, 7 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2019/019765, dated Jun. 14, 2019, 9 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2019/034007, dated Oct. 28, 2019, 18 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2019/066805, dated Mar. 6, 2020, 9 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2019/066831, dated Feb. 27, 2020, 11 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2020/044807, dated Sep. 30, 2020, 12 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2020/058097, dated Feb. 12, 2021, 09 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2020/059636, dated Feb. 11, 2021, 18 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2021/031201, dated Aug. 2, 2021, 13 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2021/033321, dated Sep. 6, 2021, 11 pages. |
International Search Report and Written Opinion for International Application No. PCT/US2021/041775, dated Nov. 29, 2021, 14 pages. |
Millet L., et al., “A 5500-Frames/s 85-GOPS/W3-D Stacked BSI Vision Chip Based on Parallel In-Focal-Plane Acquisition and Processing,” IEEE Journal of Solid-State Circuits, USA, Apr. 1, 2019, vol. 54(4), pp. 1096-1105, XP011716786. |
Non-Final Office Action dated Feb. 1, 2021 for U.S. Appl. No. 16/435,451, filed Jun. 7, 2019, 14 Pages. |
Non-Final Office Action dated Jan. 1, 2021 for U.S. Appl. No. 16/715,792, filed Dec. 16, 2019, 15 Pages. |
Non-Final Office Action dated Jul. 2, 2021 for U.S. Appl. No. 16/820,594, filed Mar. 16, 2020, 8 Pages. |
Non-Final Office Action dated Sep. 2, 2021 for U.S. Appl. No. 16/910,844, filed Jun. 24, 2020, 7 Pages. |
Non-Final Office Action dated Dec. 4, 2020 for U.S. Appl. No. 16/436,137, filed Jun. 10, 2019, 12 Pages. |
Non-Final Office Action dated May 7, 2021 for U.S. Appl. No. 16/421,441, filed May 23, 2019, 17 Pages. |
Non-Final Office Action dated Jun. 8, 2021 for U.S. Appl. No. 17/072,840, filed Oct. 16, 2020, 7 Pages. |
Non-Final Office Action dated Jul. 10, 2020 for U.S. Appl. No. 16/255,528, filed Jan. 23, 2019, 27 Pages. |
Non-Final Office Action dated May 14, 2021 for U.S. Appl. No. 16/716,050, filed Dec. 16, 2019, 16 Pages. |
Non-Final Office Action dated Mar. 15, 2021 for U.S. Appl. No. 16/896,130, filed Jun. 8, 2020, 16 Pages. |
Non-Final Office Action dated Apr. 21, 2021 for U.S. Appl. No. 17/083,920, filed Oct. 29, 2020, 17 Pages. |
Non-Final Office Action dated Oct. 21, 2021 for U.S. Appl. No. 17/083,920, filed Oct. 29, 2020, 19 Pages. |
Non-Final Office Action dated Jul. 25, 2019 for U.S. Appl. No. 15/909,162, filed Mar. 1, 2018, 20 Pages. |
Notice of Allowance dated Apr. 1, 2021 for U.S. Appl. No. 16/255,528, filed Jan. 23, 2019, 7 Pages. |
Notice of Allowance dated Nov. 3, 2020 for U.S. Appl. No. 16/566,583, filed Sep. 10, 2019, 11 Pages. |
Notice of Allowance dated May 5, 2021 for U.S. Appl. No. 16/707,988, filed Dec. 9, 2019, 14 Pages. |
Notice of Allowance dated Jan. 7, 2022 for U.S. Appl. No. 16/899,908, filed Jun. 12, 2020, 10 pages. |
Notice of Allowance dated Dec. 8, 2021 for U.S. Appl. No. 16/829,249, filed Mar. 25, 2020, 6 pages. |
Notice of Allowance dated Jul. 8, 2021 for U.S. Appl. No. 17/150,925, filed Jan. 15, 2021, 10 Pages. |
Notice of Allowance dated Jul. 13, 2021 for U.S. Appl. No. 16/896,130, filed Jun. 8, 2020, 8 Pages. |
Notice of Allowance dated Oct. 14, 2020 for U.S. Appl. No. 16/384,720, filed Apr. 15, 2019, 8 Pages. |
Notice of Allowance dated Oct. 15, 2020 for U.S. Appl. No. 16/544,136, filed Aug. 19, 2019, 11 Pages. |
Notice of Allowance dated Apr. 16, 2021 for U.S. Appl. No. 16/715,792, filed Dec. 16, 2019, 10 Pages. |
Notice of Allowance dated Nov. 17, 2021 for U.S. Appl. No. 16/899,908, filed Jun. 12, 2020, 7 Pages. |
Notice of Allowance dated Sep. 17, 2021 for U.S. Appl. No. 16/899,908, filed Jun. 12, 2020, 11 Pages. |
Notice of Allowance dated Mar. 18, 2020 for U.S. Appl. No. 15/909,162, filed Mar. 1, 2018, 9 Pages. |
Notice of Allowance dated Dec. 21, 2021 for U.S. Appl. No. 16/550,851, filed Aug. 26, 2019, 10 pages. |
Notice of Allowance dated Oct. 21, 2020 for U.S. Appl. No. 16/436,049, filed Jun. 10, 2019, 8 Pages. |
Notice of Allowance dated Dec. 22, 2021 for U.S. Appl. No. 16/910,844, filed Jun. 24, 2020, 7 pages. |
Notice of Allowance dated Jan. 22, 2021 for U.S. Appl. No. 16/369,763, filed Mar. 29, 2019, 8 Pages. |
Notice of Allowance dated Nov. 22, 2021 for U.S. Appl. No. 16/820,594, filed Mar. 16, 2020, 18 pages. |
Notice of Allowance dated Nov. 22, 2021 for U.S. Appl. No. 16/820,594, filed Mar. 16, 2020, 8 pages. |
Notice of Allowance dated Nov. 24, 2021 for U.S. Appl. No. 16/910,844, filed Jun. 24, 2020, 8 pages. |
Notice of Allowance dated Aug. 25, 2021 for U.S. Appl. No. 16/715,792, filed Dec. 16, 2019, 9 Pages. |
Notice of Allowance dated Oct. 25, 2021 for U.S. Appl. No. 16/435,451, filed Jun. 7, 2019, 8 Pages. |
Notice of Allowance dated Aug. 26, 2020 for U.S. Appl. No. 16/384,720, filed Apr. 15, 2019, 8 Pages. |
Notice of Allowance dated Oct. 26, 2021 for U.S. Appl. No. 16/896,130, filed Jun. 8, 2020, 8 Pages. |
Notice of Allowance dated Aug. 30, 2021 for U.S. Appl. No. 16/829,249, filed Mar. 25, 2020, 8 pages. |
Notice of Reason for Rejection dated Nov. 16, 2021 for Japanese Application No. 2019-571699, filed Jun. 25, 2018, 13 pages. |
Office Action dated Jul. 3, 2020 for Chinese Application No. 201810821296, filed Jul. 24, 2018, 17 Pages. |
Office Action dated Jul. 7, 2021 for European Application No. 19723902.3, filed Apr. 1, 2019, 3 Pages. |
Office Action dated Jul. 7, 2021 for European Application No. 19737299.8, filed Jun. 11, 2019, 5 Pages. |
Office Action dated Mar. 9, 2021 for Chinese Application No. 201810821296, filed Jul. 24, 2018, 10 Pages. |
Office Action dated Dec. 14, 2021 for Japanese Application No. 2019571598, filed Jun. 26, 2018, 12 pages. |
Office Action dated Jun. 28, 2020 for Chinese Application No. 201810821296, filed Jul. 24, 2018, 2 Pages. |
Office Action dated Sep. 30, 2021 forTaiwan Application No. 107124385, 17 Pages. |
Partial International Search Report and Provisional Opinion for International Application No. PCT/US2021/041775, dated Oct. 8, 2021, 12 pages. |
Restriction Requirement dated Feb. 2, 2021 for U.S. Appl. No. 16/716,050, filed Dec. 16, 2019, 7 Pages. |
Sebastian A., et al., “Memory Devices and Applications for In-memory Computing,” Nature Nanotechnology, Nature Publication Group, Inc, London, Mar. 30, 2020, vol. 15 (7), pp. 529-544, XP037194929. |
Shi C., et al., “A 100fps Vision Chip Based on a Dynamically Reconfigurable Hybrid Architecture Comprising a PE Array and Self-Organizing Map Neural Network,” International Solid-State Circuits Conference, Session 7, Image Sensors, Feb. 10, 2014, pp. 128-130, XP055826878. |
Snoeij M.F., et al., “A low Power col. Parallel 12-bit ADC for CMOS Imagers,” XP007908033, Jun. 1, 2005, pp. 169-172. |
Advisory Action dated Oct. 1, 2021 for U.S. Appl. No. 17/083,920, filed Oct. 29, 2020, 4 pages. |
Corrected Notice of Allowability dated Dec. 1, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 3 pages. |
Corrected Notice of Allowability dated Dec. 7, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 3 pages. |
Corrected Notice of Allowability dated Jan. 25, 2021 for U.S. Appl. No. 16/384,720, filed Apr. 15, 2019, 4 Pages. |
Corrected Notice of Allowability dated Jan. 29, 2021 for U.S. Appl. No. 16/544,136, filed Aug. 19, 2019, 2 Pages. |
Corrected Notice of Allowance dated Mar. 7, 2022 for U.S. Appl. No. 17/150,925, filed Jan. 15, 2021, 2 Pages. |
Extended European Search Report for European Application No. 18886564.6, dated Jan. 26, 2021, 6 Pages. |
Final Office Action dated Nov. 3, 2021 for U.S. Appl. No. 16/560,665, filed Sep. 4, 2019, 19 Pages. |
Final Office Action dated Jul. 8, 2021 for U.S. Appl. No. 16/672,427, filed Nov. 1, 2019, 10 Pages. |
International Preliminary Report on Patentability for International Application No. PCT/US2019/027727, dated Oct. 29, 2020, 8 Pages. |
International Search Report and Written Opinion for International Application No. PCT/US2021/054327, dated Feb. 14, 2022, 8 pages. |
International Search Report and Written Opinion for International Application No. PCT/US2021/057966, dated Feb. 22, 2022, 15 pages. |
Non-Final Office Action dated Feb. 2, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 8 Pages. |
Non-Final Office Action dated Mar. 2, 2022 for U.S. Appl. No. 17/127,670, filed Dec. 18, 2020, 18 pages. |
Non-Final Office Action dated Dec. 7, 2020 for U.S. Appl. No. 16/672,427, filed Nov. 1, 2019, 9 Pages. |
Non-Final Office Action dated Jan. 8, 2020 for U.S. Appl. No. 16/285,873, filed Feb. 26, 2019, 14 Pages. |
Non-Final Office Action dated Feb. 11, 2022 for U.S. Appl. No. 16/672,427, filed Nov. 1, 2019, 9 pages. |
Non-Final Office Action dated Dec. 24, 2020 for U.S. Appl. No. 16/407,072, filed May 8, 2019, 15 Pages. |
Non-Final Office Action dated Apr. 29, 2021 for U.S. Appl. No. 16/560,665, filed Sep. 4, 2019, 17 Pages. |
Notice of Allowance dated Apr. 1, 2021 for U.S. Appl. No. 16/436,049, filed Jun. 10, 2019, 8 Pages. |
Notice of Allowance dated Dec. 1, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 11 pages. |
Notice of Allowance dated Jun. 1, 2021 for U.S. Appl. No. 16/407,072, filed May 8, 2019, 11 Pages. |
Notice of Allowance dated Dec. 7, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 11 pages. |
Notice of Allowance dated Mar. 7, 2022 for U.S. Appl. No. 16/421,441, filed May 23, 2019, 18 pages. |
Notice of Allowance dated Jul. 9, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 6 Pages. |
Notice of Allowance dated Nov. 10, 2021 for U.S. Appl. No. 16/672,427, filed Nov. 1, 2019, 6 Pages. |
Notice of Allowance dated Mar. 11, 2022 for U.S. Appl. No. 16/716,050, filed Dec. 16, 2019, 13 pages. |
Notice of Allowance dated Sep. 13, 2021 for U.S. Appl. No. 16/407,072, filed May 8, 2019, 10 Pages. |
Notice of Allowance dated Feb. 14, 2022 for U.S. Appl. No. 16/715,792, filed Dec. 16, 2019, 9 pages. |
Notice of Allowance dated Dec. 16, 2021 for U.S. Appl. No. 16/407,072, filed May 8, 2019, 2 pages. |
Notice of Allowance dated Jun. 18, 2020 for U.S. Appl. No. 16/285,873, filed Feb. 26, 2019, 10 Pages. |
Notice of Allowance dated Apr. 19, 2022 for U.S. Appl. No. 16/550,851, filed Aug. 26, 2019, 08 pages. |
Notice of Allowance dated Feb. 22, 2022 for U.S. Appl. No. 17/083,920, filed Oct. 29, 2020, 10 pages. |
Notice of Allowance dated Sep. 24, 2020 for U.S. Appl. No. 15/668,241, filed Aug. 3, 2017, 13 Pages. |
Notice of Allowance dated Oct. 27, 2021 for U.S. Appl. No. 16/210,748, filed Dec. 5, 2018, 6 Pages. |
Notice of Allowance dated Oct. 29, 2021 for U.S. Appl. No. 16/672,427, filed Nov. 1, 2019, 9 Pages. |
Notification of the First Office Action dated Oct. 28, 2021 for Chinese Application No. 2019800218483, filed Jan. 24, 2019, 17 pages. |
Supplemental Notice of Allowability dated Jul. 8, 2021 for U.S. Appl. No. 16/436,049, filed Jun. 10, 2019, 2 Pages. |
Office Action dated May 10, 2022 for Taiwan Application No. 108122610, 19 pages. |
Office Action dated Jul. 6, 2022 for Chinese Application No. 201980024435.0, filed Apr. 1, 2019, 16 pages. |
Number | Date | Country | |
---|---|---|---|
20200007800 A1 | Jan 2020 | US |
Number | Date | Country | |
---|---|---|---|
62690571 | Jun 2018 | US |