As is known in the art, some known ranging systems can include laser radar (ladar), light-detection and ranging (lidar), and range-finding systems, to measure the distance to objects in a scene. A laser ranging and imaging system emits pulses toward a particular location and measures the return echoes to extract ranges to objects at the location, from which a three-dimensional representation of the objects can be computed.
Time-of-flight laser ranging systems generally work by emitting a laser pulse and recording the time it takes for the laser pulse to travel to a target, reflect, and return to a photoreceiver. The laser ranging instrument records the time of the outgoing pulse and records the time that a laser pulse returns. The difference between these two times is the time of flight to and from the target. Using the speed of light, the round-trip time of the pulses is used to calculate the distance to the target.
Photodetector arrays, such as a focal plane array (FPA), can be used to detect signal returns. Multi-pixel time-of-flight (ToF) receivers having linear mode avalanche photodiode (APD) sensing elements in the FPA often suffer from gain mismatch across the photodiode array despite all elements operating under identical reverse bias. This mismatch may result from APD manufacturing process variations which cannot easily be reduced below the level of practical impact. As a result of APD gain mismatch, conventional APD-based multi-pixel ToF receivers may suffer from variation in optical sensitivity across the pixels of the array.
Example embodiments of the disclosure provide methods and apparatus for non-uniformity correction (NUC) that can be used to equalize the gain of the APD elements across an array for improving uniformity of sensitivity across the receiver. Example embodiments of a ToF receiver having NUC may include a secondary mode of operation in which the receiver can operate in a passive mode where pixels of the receiver are not measuring the timing and amplitude of optical return pulses. Instead, the receiver measures the direct current (DC) photocurrent in the individual APD elements of the array.
A passive module may implement a passive mode of operation in a variety of ways. In one embodiment, passive mode can include integration of APD photocurrent over a defined time period. In another embodiment, passive mode can include measuring the amplified APD current using a current/voltage (I/V) converter, such as a transimpedance amplifier (TIA) for example. In a further embodiment, passive mode can include passive measurement circuits to produce a potential (or digitized potential) that is representative of the APD DC current and/or a time measurement that is representative of the APD DC current.
Sensor embodiments may include biasing for avalanche photodiodes (APDs) based on a direct injection (DI) structure that supports one or more features. In embodiments, a sensor includes individual pixelated control of the APD anode for controlling photodiode reverse biasing using a DI transistor, e.g., PFET, a digital-to-analog converter (DAC), and/or SRAM. High output impedance of the DI PFET device at high frequencies results in good charge injection to the active detection path for transient signals and desired attenuation of low frequency signals. An integrated passive mode can be used for passive imaging and/or for gain non-uniformity correction (NUC) of the APD. In embodiments, passive mode operation is based on a direct injection (DI) integration. An integrated voltage may be passed to a discriminator and time-to-digital converter (TDC) to generate time-to-threshold timing information. In some embodiments, instead of DI based biasing, buffered direct injection (BDI) structure is used that can improve regulation of APD anode potential. In other embodiments, the passive mode is accomplished with an instantaneous measurement of the background current generated in the APD.
In one aspect, a method comprises: for a sensor having an avalanche photodiode (APD) array and an integrated circuit having a bias control module and a passive/active mode module; measuring currents from the APD array; individually controlling reverse bias voltages applied across each element in the APD array to reduce response nonuniformity among the elements of the APD array; and selecting a nonuniformity bias correction for each element based on the measured currents from the APD array.
A method can further include one or more of the following features in any combination: integrating the currents from the APD array for one or more defined amounts of time, converting currents from the APD array to voltage levels, the APD array has a common cathode configuration, the APD array has a common anode configuration, illuminating the APD array with a first light level L1 and a second light level L2, illuminating the APD array at the first and second light levels L1 and L2 with a continuous wave (CW) light source, one of first and second light levels L1 or L2 is the non-illuminated dark condition, performing a current measurement I(V1,L1) at a first APD bias condition V1 and the first light level L1, for each element in the APD array, determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, performing a current measurement I(V1,L2) at a first APD bias condition V1 and the second light level L2, for each element in the APD array, determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, performing a current measurement I(V2,L1) at a second APD bias condition V2 and the first light level L1, for each element in the APD array, determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, performing a current measurement I(V2,L2) at a second APD bias condition V2 and the second light level L2, for each element in the APD array, determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, determining a relative gain of each element in the APD array in the two bias conditions, V1 and V2, from the four current measurements I(V1,L1), I(V1,L2), I(V2,L1), and I(V2,L2), determining the gain of each element of the APD array at bias condition V2 relative to its gain at bias condition V1 as: Abs[I(V2,L1)−I(V2,L2)]/Abs[I(V1,L1)−I(V1,L2)], where Abs[x] represents the absolute value of quantity x, and using the relative gains computed for each element of the APD array to control the reverse bias voltages applied across each element to reduce the response nonuniformity among the elements of the APD array, the bias control module comprises a memory, a digital-to-analog converter (DAC), and a transistor, which is coupled to a first one of the elements in the APD array, the passive/active mode module comprises an active/passive switch and an integration capacitor, wherein the first one of the elements in the APD array, the transistor, the active/passive switch, and the integration capacitor provide a direct injection integrator mode for the sensor in active mode, the sensor comprises a time-to-digital converter (TDC) for generating time-stamp when a voltage level on the integration capacitor exceeds a threshold, performing a plurality of total current measurements I(V1,L1), I(V2,L1), I(V3,L1), etc., at a plurality of APD bias conditions and at the first light level L1, for each element in the APD array, and determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, performing a plurality of total current measurements I(V1,L2), I(V2,L2), I(V3,L2), at a plurality of APD bias conditions V1, V2, V3, and at the second light level L2, for each element in the APD array, and determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, determining the gains of each element of the APD array at the plurality of bias conditions V2, V3 relative to its gain at bias condition V1, as: Abs[I(V,L1)−I(V,L2)]/Abs[I(V1,L1)−I(V1,L2)], where V represents one of the plurality of APD bias conditions V2, V3 and Abs[x] represents the absolute value of quantity x, and using the relative gains computed for each element of the APD array to control the reverse bias voltages applied across each element of the APD array to reduce the response nonuniformity among the elements of the APD array, the nonuniformity correction of the reverse bias voltages applied across each element of the APD array are found by selecting from the plurality of bias conditions tested, V2, V3, the bias condition for each element resulting in the relative gain closest to some specified value, the nonuniformity correction of the reverse bias voltages applied across each element of the APD array are found by selecting from the plurality of bias conditions tested, V2, V3, the two bias conditions for each element resulting in relative gains closest to some specified value, and estimation of the reverse bias that will result in the specified value of relative gain by linear interpolation, the nonuniformity correction of the reverse bias voltages applied across each element of the APD array are found by fitting the plurality of relative gain values calculated for the plurality of bias conditions tested, V2, V3, to a mathematical function that models the gain-vs-reverse bias characteristic of the APD elements, and applying the mathematical function as fit to the relative gain data of each element to estimate the reverse bias that will result in a specified value of the relative gain for each element, the mathematical function used to model the gain-vs-reverse bias characteristic of the APD elements is
where V is the reverse bias voltage, VB is the best-fit APD breakdown voltage, and nA and nB are fit parameters, including illuminating the APD array with a first light level L1, chosen such that the resulting photocurrent of each element of the APD array greatly exceeds that element's dark current in magnitude, performing a plurality of total current measurements I(V1,L1), I(V2,L1), I(V3,L1), at a plurality of APD bias conditions V1, V2, V3, and at the first light level L1, for each element in the APD array, determining each element's total current, which comprises the sum of each element's photocurrent and dark current in that condition, and using the total current measurements for each element of the APD array to control the reverse bias voltages applied across each element of the APD array to reduce response nonuniformity among the elements of the APD array, the nonuniformity correction of the reverse bias voltages applied across each element of the APD array are found by selecting from the plurality of bias conditions tested, V2, V3, the bias conditions for each element resulting in total current closest to some specified value, and/or total current measurements at a plurality of APD bias conditions and reverse bias voltage adjustments are performed iteratively to equalize total current among the elements of the APD array, selecting APD bias conditions in successive iterations based on the current measurements of the previous iteration.
The foregoing features of this disclosure, as well as the disclosure itself, may be more fully understood from the following description of the drawings in which:
Prior to describing example embodiments of the disclosure some information is provided. Laser ranging systems can include laser radar (ladar), light-detection and ranging (lidar), and rangefinding systems, which are generic terms for the same class of instrument that uses light to measure the distance to objects in a scene. This concept is similar to radar, except optical signals are used instead of radio waves. Similar to radar, a laser ranging and imaging system emits a pulse toward a particular location and measures the return echoes to extract the range.
Laser ranging systems generally work by emitting a laser pulse and recording the time it takes for the laser pulse to travel to a target, reflect, and return to a photoreceiver. The laser ranging instrument records the time of the outgoing pulse—either from a trigger or from calculations that use measurements of the scatter from the outgoing laser light—and then records the time that a laser pulse returns. The difference between these two times is the time of flight to and from the target. Using the speed of light, the round-trip time of the pulses is used to calculate the distance to the target.
Lidar systems may scan the beam across a target area to measure the distance to multiple points across the field of view, producing a full three-dimensional range profile of the surroundings. More advanced flash lidar cameras, for example, contain an array of detector elements, each able to record the time of flight to objects in their field of view.
When using light pulses to create images, the emitted pulse may intercept multiple objects, at different orientations, as the pulse traverses a 3D volume of space. The echoed laser-pulse waveform contains a temporal and amplitude imprint of the scene. By sampling the light echoes, a record of the interactions of the emitted pulse is extracted with the intercepted objects of the scene, allowing an accurate multi-dimensional image to be created. To simplify signal processing and reduce data storage, laser ranging and imaging can be dedicated to discrete-return systems, which record only the time of flight (TOF) of the first, or a few, individual target returns to obtain angle-angle-range images. In a discrete-return system, each recorded return corresponds, in principle, to an individual laser reflection (i.e., an echo from one particular reflecting surface, for example, a tree, pole or building). By recording just a few individual ranges, discrete-return systems simplify signal processing and reduce data storage, but they do so at the expense of lost target and scene reflectivity data. Because laser-pulse energy has significant associated costs and drives system size and weight, recording the TOF and pulse amplitude of more than one laser pulse return per transmitted pulse, to obtain angle-angle-range-intensity images, increases the amount of captured information per unit of pulse energy. All other things equal, capturing the full pulse return waveform offers significant advantages, such that the maximum data is extracted from the investment in average laser power. In full-waveform systems, each backscattered laser pulse received by the system is digitized at a high sampling rate (e.g., 500 MHz to 1.5 GHz). This process generates digitized waveforms (amplitude versus time) that may be processed to achieve higher-fidelity 3D images.
Of the various laser ranging instruments available, those with single-element photoreceivers generally obtain range data along a single range vector, at a fixed pointing angle. This type of instrument—which is, for example, commonly used by golfers and hunters—either obtains the range (R) to one or more targets along a single pointing angle or obtains the range and reflected pulse intensity (I) of one or more objects along a single pointing angle, resulting in the collection of pulse range-intensity data, (R,I)i, where i indicates the number of pulse returns captured for each outgoing laser pulse.
More generally, laser ranging instruments can collect ranging data over a portion of the solid angle of a sphere, defined by two angular coordinates (e.g., azimuth and elevation), which can be calibrated to three-dimensional (3D) rectilinear cartesian coordinate grids; these systems are generally referred to as 3D lidar and ladar instruments. The terms “lidar” and “ladar” are often used synonymously and, for the purposes of this discussion, the terms “3D lidar,” “scanned lidar,” or “lidar” are used to refer to these systems without loss of generality. 3D lidar instruments obtain three-dimensional (e.g., angle, angle, range) data sets. Conceptually, this would be equivalent to using a rangefinder and scanning it across a scene, capturing the range of objects in the scene to create a multi-dimensional image. When only the range is captured from the return laser pulses, these instruments obtain a 3D data set (e.g., angle, angle, range), where the index n is used to reflect that a series of range-resolved laser pulse returns can be collected, not just the first reflection.
Some 3D lidar instruments are also capable of collecting the intensity of the reflected pulse returns generated by the objects located at the resolved (angle, angle, range) objects in the scene. When both the range and intensity are recorded, a multi-dimensional data set [e.g., angle, angle, (range-intensity)n] is obtained. This is analogous to a video camera in which, for each instantaneous field of view (FOV), each effective camera pixel captures both the color and intensity of the scene observed through the lens. However, 3D lidar systems, instead capture the range to the object and the reflected pulse intensity.
Lidar systems can include different types of lasers, including those operating at different wavelengths, including those that are not visible (e.g., those operating at a wavelength of 840 nm or 905 nm), and in the near-infrared (e.g., those operating at a wavelength of 1064 nm or 1550 nm), and the thermal infrared including those operating at wavelengths known as the “eyesafe” spectral region (i.e., generally those operating at a wavelength beyond 1300-nm, which is blocked by the cornea), where ocular damage is less likely to occur. Lidar transmitters are generally invisible to the human eye. However, when the wavelength of the laser is close to the range of sensitivity of the human eye—roughly 350 nm to 730 nm—the light may pass through the cornea and be focused onto the retina, such that the energy of the laser pulse and/or the average power of the laser must be lowered to prevent ocular damage. Thus, a laser operating at, for example, 1550 nm, can—without causing ocular damage—generally have 200 times to 1 million times more laser pulse energy than a laser operating at 840 nm or 905 nm.
One challenge for a lidar system is detecting poorly reflective objects at long distance, which requires transmitting a laser pulse with enough energy that the return signal—reflected from the distant target—is of sufficient magnitude to be detected. To determine the minimum required laser transmission power, several factors must be considered. For instance, the magnitude of the pulse returns scattering from the diffuse objects in a scene is proportional to their range and the intensity of the return pulses generally scales with distance according to 1/R4 for small objects and 1/R2 for larger objects; yet, for highly-specularly reflecting objects (i.e., those reflective objects that are not diffusively-scattering objects), the collimated laser beams can be directly reflected back, largely unattenuated. This means that—if the laser pulse is transmitted, then reflected from a target 1 meter away—it is possible that the full energy (J) from the laser pulse will be reflected into the photoreceiver; but—if the laser pulse is transmitted, then reflected from a target 333 meters away—it is possible that the return will have a pulse with energy approximately 1012 weaker than the transmitted energy. To provide an indication of the magnitude of this scale, the 12 orders of magnitude (1012) is roughly the equivalent of: the number of inches from the earth to the sun, 10× the number of seconds that have elapsed since Cleopatra was born, or the ratio of the luminous output from a phosphorescent watch dial, one hour in the dark, to the luminous output of the solar disk at noon.
In many cases of lidar systems highly-sensitive photoreceivers are used to increase the system sensitivity to reduce the amount of laser pulse energy that is needed to reach poorly reflective targets at the longest distances required, and to maintain eyesafe operation. Some variants of these detectors include those that incorporate photodiodes, and/or offer gain, such as avalanche photodiodes (APDs) or single-photon avalanche detectors (SPADs). These variants can be configured as single-element detectors-segmented-detectors, linear detector arrays, or area detector arrays. Using highly sensitive detectors such as APDs or SPADs reduces the amount of laser pulse energy required for long-distance ranging to poorly reflective targets. The technological challenge of these photodetectors is that they must also be able to accommodate the incredibly large dynamic range of signal amplitudes.
As dictated by the properties of the optics, the focus of a laser return changes as a function of range; as a result, near objects are often out of focus. Furthermore, also as dictated by the properties of the optics, the location and size of the “blur”—i.e., the spatial extent of the optical signal—changes as a function of range, much like in a standard camera. These challenges are commonly addressed by using large detectors, segmented detectors, or multi-element detectors to capture all of the light or just a portion of the light over the full-distance range of objects. It is generally advisable to design the optics such that reflections from close objects are blurred, so that a portion of the optical energy does not reach the detector or is spread between multiple detectors. This design strategy reduces the dynamic range requirements of the detector and prevents the detector from damage.
Acquisition of the lidar imagery can include, for example, a 3D lidar system embedded in the front of car, where the 3D lidar system, includes a laser transmitter with any necessary optics, a single-element photoreceiver with any necessary dedicated or shared optics, and an optical scanner used to scan (“paint”) the laser over the scene. Generating a full-frame 3D lidar range image—where the field of view is 20 degrees by 60 degrees and the angular resolution is 0.1 degrees (10 samples per degree)—requires emitting 120,000 pulses [(20*10*60*10)=120,000)]. When update rates of 30 frames per second are required, such as is required for automotive lidar, roughly 3.6 million pulses per second must be generated and their returns captured.
There are many ways to combine and configure the elements of the lidar system—including considerations for the laser pulse energy, beam divergence, detector array size and array format (single element, linear, 2D array), and scanner to obtain a 3D image. If higher power lasers are deployed, pixelated detector arrays can be used, in which case the divergence of the laser would be mapped to a wider field of view relative to that of the detector array, and the laser pulse energy would need to be increased to match the proportionally larger field of view. For example— compared to the 3D lidar above—to obtain same-resolution 3D lidar images 30 times per second, a 120,000-element detector array (e.g., 200×600 elements) could be used with a laser that has pulse energy that is 120,000 times greater. The advantage of this “flash lidar” system is that it does not require an optical scanner; the disadvantages are that the larger laser results in a larger, heavier system that consumes more power, and that it is possible that the required higher pulse energy of the laser will be capable of causing ocular damage. The maximum average laser power and maximum pulse energy are limited by the requirement for the system to be eyesafe.
As noted above, while many lidar systems operate by recording only the laser time of flight and using that data to obtain the distance to the first target return (closest) target, some lidar systems are capable of capturing both the range and intensity of one or multiple target returns created from each laser pulse. For example, for a lidar system that is capable of recording multiple laser pulse returns, the system can detect and record the range and intensity of multiple returns from a single transmitted pulse. In such a multi-pulse lidar system, the range and intensity of a return pulse from a closer—by object can be recorded, as well as the range and intensity of later reflection(s) of that pulse—one(s) that moved past the closer—by object and later reflected off of more-distant object(s). Similarly, if glint from the sun reflecting from dust in the air or another laser pulse is detected and mistakenly recorded, a multi-pulse lidar system allows for the return from the actual targets in the field of view to still be obtained.
The amplitude of the pulse return is primarily dependent on the specular and diffuse reflectivity of the target, the size of the target, and the orientation of the target. Laser returns from close, highly-reflective objects, are many orders of magnitude greater in intensity than the intensity of returns from distant targets. Many lidar systems require highly sensitive photodetectors, for example APDs, which along with their CMOS amplification circuits may be damaged by very intense laser pulse returns.
For example, if an automobile equipped with a front-end lidar system were to pull up behind another car at a stoplight, the reflection off of the license plate may be significant—perhaps 1012 higher than the pulse returns from targets at the distance limits of the lidar system. When a bright laser pulse is incident on the photoreceiver, the large current flow through the photodetector can damage the detector, or the large currents from the photodetector can cause the voltage to exceed the rated limits of the CMOS electronic amplification circuits, causing damage. For this reason, it is generally advisable to design the optics such that the reflections from close objects are blurred, so that a portion of the optical energy does not reach the detector or is spread between multiple detectors.
However, capturing the intensity of pulses over a larger dynamic range associated with laser ranging may be challenging because the signals are too large to capture directly. One can infer the intensity by using a recording of a bit-modulated output obtained using serial-bit encoding obtained from one or more voltage threshold levels. This technique is often referred to as time-over-threshold (TOT) recording or, when multiple-thresholds are used, multiple time-over-threshold (MTOT) recording.
A data processing and calibration circuit 213 may be inserted between the memories 212 and the readout 214 which may perform any number of data correction or mapping functions. For example, the circuit may compare timing return information to timing reference information and convert timing return information into specific range information. Additionally, the circuit may correct for static or dynamic errors using calibration and correction algorithms. Other possible functions include noise reduction based on multi-return data or spatial correlation or objection detection. A possible mapping function may be to reshape the data into point-cloud data or to include additional probability data of correct measurement values based on additionally collected information from the sensor.
where V is the reverse bias, VB is the APD breakdown voltage, and nA and nB are fit parameters. Best-fit values of the free parameters VB, nA, and nB may be determined from a set of empirical measurements of avalanche gain (M) at particular values of the reverse bias (V) using curve-fitting software which implements standard algorithms known to those skilled in the art, such as the method of least squares. This equation can be used to approximate the gain-voltage characteristics of APDs manufactured in a given lot by curve-fitting data taken from a sample of APDs produced in the lot. Once fit to a given manufacturing lot, the equation can then be used to estimate required non-uniformity corrections to the anode bias of a given APD from that lot, based on a measurement of that APD's uncorrected gain at some non-adjusted reverse bias. Alternatively, this equation can be fit to gain-voltage data collected from an individual pixel, allowing highly accurate and individualized estimation of the needed nonuniformity correction to that individual APD's anode bias.
In embodiments, a sensor, such as a ToF sensor, includes a mode of operation in which the receiver can operate in a “passive” mode in which the pixels of the receiver are not measuring the timing and amplitude of optical return pulses. In the passive mode, the receiver is measuring the direct current (DC) photocurrent in the individual APD elements. The passive mode of operation can include one or more features for measuring APD direct current. In some embodiments, APD photocurrent is integrated over a defined time period. In embodiments, measurement of the amplified APD current is performed by a current/voltage (I/V) converter. In general, passive measurement circuits can produce a potential (or digitized potential) that is representative of the APD DC current or a time measurement that is representative of the APD DC current.
In addition to the passive operational mode, in embodiments the sensor should have the integrated ability to control the pixelated anodes of the APD array, such as a common cathode based APD array, thus allowing of control of the reverse bias for the individual elements within the APD array. Pixelated control of the anode potential can be such as by integrating a user programmable digital-to-analog converter (DAC) and local memory cell in each pixel to set the APD anode potential. The performance of such a pixelated DAC may be limited, such as to a 1-2 V range with 4-6 bits of resolution.
While example embodiments of the disclosure show a particular diode polarity, in other embodiments the diode polarity is switched. For example, instead of anode bias control, as shown and described above, a cathode bias control module can bias the cathodes of diodes sharing a common anode.
As noted above, the example embodiment of
Additional calibration measurements may be taken for different operating conditions such as multiple APD gain (reverse bias) settings or operating temperatures. These multiple calibration measurements may be used to extrapolate calibration configurations for any number or combinations of conditions over which calibration of APD gain is desired.
It is understood that APD data can be collected in a variety of ways to meet the needs of a particular application for NUC. For example, in some embodiments, NUC data collection can include collecting coarse, uncorrected gain-voltage data from each pixel in a given array in the vicinity of breakdown. For instance, if based on earlier die screening it is known that none of the pixels in a particular array die break down below 49 V, one might make gain measurements at 45 V, 46, V, 47 V, 47.5 V, 48 V, 48.25 V, 48.5 V, and 48.75 V. This individualized gain-voltage data can be fitted to Equation (1), for example, for computing a more accurate correction to the APD anode bias to achieve a desired gain. In this arrangement, multiple light and dark frames (light_2,3,4,5,6 & dark_2,3,4,5,6 etc., as shown and described in
In another embodiment, NUC data collection can include collecting fine, uncorrected gain-voltage data from each pixel in a given array, in the vicinity of breakdown, and either: sorting the data points to find the one closest to the desired gain and using that to find an associated pixel anode adjustment; and/or extrapolating between the two data points closest to the desired gain to find an associated pixel anode adjustment. In general, this approach only works well with relatively dense data point information due to the nonlinear dependence of gain on voltage near breakdown. In contrast to the above embodiment, each curve is not gain fitted. Rather, reverse bias correction is determined directly, or with interpolation between measured data points.
The calibration measurement data may also be used as reference information to help identify APD failures during operation or over sensor lifetime. One embodiment method of this concept is to compare the dark passive frame measurement at calibration to dark passive measurements periodically during operation or at sensor power-on. This embodiment requires a method for optically shielding the sensor from any photo input, however, this may be a common feature in system applications to protect the sensitive detectors and amplification circuitry from excessive optical inputs. Further, these fault detection methods may be utilized by automotive safety systems to achieve a specific Automotive Safety Integration Level (ASIL) as required by ISO 26262 “Road Vehicles—Functional Safety.”
Advantages of passive APD NUC in accordance with example embodiments of the disclosure include that the optical illumination of the receiver during the NUC process is CW, as opposed to a pulsed optical source. The ability to produce uniform CW illumination across the entire APD array is simpler than producing a uniform pulsed illumination. In addition, the use of uniform pulsed illumination of the receiver can also result in crosstalk artifacts in the receiver output data due to simultaneous activation of all the pixels within the receiver at the same time.
It is understood that a circuit of equivalent function can be configured to adjust the bias of an APD from the cathode side in common-anode embodiments.”
In the illustrated embodiment, the example sensor 800 includes photodetectors 802, such as photodiodes, in a common cathode configuration. The anode 804 of each photodetector 802 is coupled to a transistor 806, such as a PFET, having a conduction state coupled to a digital to analog converter (DAC) 808, which is coupled to memory 810, such as SRAM. The memory 810 can store NUC information for the bias level for each APD element. The PFET 806, DAC 808, and memory 810 can comprise an anode bias control module 812, which is an example implementation of the anode bias control module 510 in
The photodiode 802 is coupled to a front end module, which may include an amplifier 814, and discriminator 816, such as a voltage discriminator. A time-to-digital converter (TDC) 818 can receive the output of the discriminator 816 for providing a digital. representation of the time of a received pulse. in embodiments, an optical pulse received by the photodiode 802 results in a current pulse from the photodiode that can be converted by the front end module 814 into a. voltage signal used for comparison with a threshold by the discriminator 816. When the threshold is exceeded, the TDC 818 can record the start time, end time, and/or duration of the exceedance.
In the illustrated embodiment, a multiplexer 820 can select the AFE 814 output or a passive mode signal 822. The PFET 806 may be coupled to an integration capacitor 824 in the passive mode and to ground in the non-passive mode. As described above, the integration capacitor 824 can integrate current from the photodiode 802 in the passive mode as part of NUC processing.
In other embodiments, as noted above, the polarity of the diode can be reversed.
In another embodiment, the passive mode operation captures an instantaneous photo-current from the photo-diode/APD and the individual APD bias is modulated by common mode circuit control which is similarly programmable for each individual pixel/APD with a memory element and a DAC. Other embodiments must generally include a method for measuring passive photo-current from the APD and must have a method for individually controlling the APD reverse bias.
For the circuit implementation of
During active mode operation 1000, APD current 1006 includes a pulse 1020 generated in response to a receive optical pulse. The pulse 1020, if of sufficient magnitude, causes the TDC to generate a time stamp 1022 corresponding to time-of-arrival for the pulse 1020.
Processing may be implemented in hardware, software, or a combination of the two. Processing may be implemented in computer programs executed on programmable computers/machines that each includes a processor, a storage medium or other article of manufacture that is readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device, and one or more output devices. Program code may be applied to data entered using an input device to perform processing and to generate output information.
The system can perform processing, at least in part, via a computer program product, (e.g., in a machine-readable storage device), for execution by, or to control the operation of, data processing apparatus (e.g., a programmable processor, a computer, or multiple computers). Each such program may be implemented in a high-level procedural or object-oriented programming language to communicate with a computer system. However, the programs may be implemented in assembly or machine language. The language may be a compiled or an interpreted language and it may be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network. A computer program may be stored on a storage medium or device (e.g., CD-ROM, hard disk, or magnetic diskette) that is readable by a general or special purpose programmable computer for configuring and operating the computer when the storage medium or device is read by the computer.
Processing may also be implemented as a machine-readable storage medium, configured with a computer program, where upon execution, instructions in the computer program cause the computer to operate.
Processing may be performed by one or more programmable embedded processors executing one or more computer programs to perform the functions of the system. All or part of the system may be implemented as, special purpose logic circuitry (e.g., an FPGA (field programmable gate array) and/or an ASIC (application-specific integrated circuit)).
All or portions of this processing may also be accomplished in the sensor. Advantages of this include limiting data output to control or applications systems or increasing the speed of corrected, remapped, filtered, or calibrated data.
Having described exemplary embodiments of the disclosure, it will now become apparent to one of ordinary skill in the art that other embodiments incorporating their concepts may also be used. The embodiments contained herein should not be limited to disclosed embodiments but rather should be limited only by the spirit and scope of the appended claims. All publications and references cited herein are expressly incorporated herein by reference in their entirety.
Elements of different embodiments described herein may be combined to form other embodiments not specifically set forth above. Various elements, which are described in the context of a single embodiment, may also be provided separately or in any suitable subcombination. Other embodiments not specifically described herein are also within the scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
6760145 | Taylor et al. | Jul 2004 | B1 |
6778728 | Taylor et al. | Aug 2004 | B2 |
6894823 | Taylor et al. | May 2005 | B2 |
6989921 | Bernstein et al. | Jan 2006 | B2 |
7015780 | Bernstein et al. | Mar 2006 | B2 |
7160753 | Williams, Jr. | Jan 2007 | B2 |
7432537 | Huntington | Oct 2008 | B1 |
7504053 | Alekel | Mar 2009 | B1 |
7764719 | Munroe et al. | Jul 2010 | B2 |
7782911 | Munroe et al. | Aug 2010 | B2 |
7852549 | Alekel et al. | Dec 2010 | B2 |
7885298 | Munroe | Feb 2011 | B2 |
7994421 | Williams et al. | Aug 2011 | B2 |
8145060 | Hufstedler | Mar 2012 | B2 |
8207484 | Williams | Jun 2012 | B1 |
8319307 | Williams | Nov 2012 | B1 |
8570372 | Russell | Oct 2013 | B2 |
8597544 | Alekel | Dec 2013 | B2 |
8630036 | Munroe | Jan 2014 | B2 |
8630320 | Munroe et al. | Jan 2014 | B2 |
8730564 | Alekel | May 2014 | B2 |
8743453 | Alekel et al. | Jun 2014 | B2 |
8760499 | Russell | Jun 2014 | B2 |
8766682 | Williams | Jul 2014 | B2 |
8853639 | Williams, Jr. | Oct 2014 | B2 |
9121762 | Williams et al. | Sep 2015 | B2 |
9197233 | Gaalema et al. | Nov 2015 | B2 |
9269845 | Williams et al. | Feb 2016 | B2 |
9368933 | Nijjar et al. | Jun 2016 | B1 |
9397469 | Nijjar et al. | Jul 2016 | B1 |
9447299 | Schut et al. | Sep 2016 | B2 |
9451554 | Singh et al. | Sep 2016 | B1 |
9466745 | Williams et al. | Oct 2016 | B2 |
9553216 | Williams et al. | Jan 2017 | B2 |
9591238 | Lee et al. | Mar 2017 | B2 |
9693035 | Williams et al. | Jun 2017 | B2 |
9759602 | Williams | Sep 2017 | B2 |
9804264 | Villeneuve et al. | Oct 2017 | B2 |
9810775 | Welford et al. | Nov 2017 | B1 |
9810777 | Williams et al. | Nov 2017 | B2 |
9810786 | Welford et al. | Nov 2017 | B1 |
9812838 | Villeneuve et al. | Nov 2017 | B2 |
9823353 | Eichenholz et al. | Nov 2017 | B2 |
9835490 | Williams et al. | Dec 2017 | B2 |
9841495 | Campbell et al. | Dec 2017 | B2 |
9843157 | Williams | Dec 2017 | B2 |
9847441 | Huntington | Dec 2017 | B2 |
9857468 | Eichenholz et al. | Jan 2018 | B1 |
9869754 | Campbell et al. | Jan 2018 | B1 |
9874635 | Eichenholz et al. | Jan 2018 | B1 |
9897687 | Campbell et al. | Feb 2018 | B1 |
9905992 | Welford et al. | Feb 2018 | B1 |
9923331 | Williams | Mar 2018 | B2 |
9941433 | Williams et al. | Apr 2018 | B2 |
9958545 | Eichenholz et al. | May 2018 | B2 |
9989629 | LaChapelle | Jun 2018 | B1 |
9995622 | Williams | Jun 2018 | B2 |
10003168 | Villeneuve | Jun 2018 | B1 |
10007001 | LaChapelle et al. | Jun 2018 | B1 |
10012732 | Eichenholz et al. | Jul 2018 | B2 |
10061019 | Campbell et al. | Aug 2018 | B1 |
10088559 | Weed et al. | Oct 2018 | B1 |
10094925 | LaChapelle | Oct 2018 | B1 |
10114111 | Russell et al. | Oct 2018 | B2 |
10121813 | Eichenholz et al. | Nov 2018 | B2 |
10132879 | Latham et al. | Nov 2018 | B2 |
10139478 | Gaalema et al. | Nov 2018 | B2 |
10169678 | Sachdeva et al. | Jan 2019 | B1 |
10169680 | Sachdeva et al. | Jan 2019 | B1 |
10175345 | Rhee et al. | Jan 2019 | B2 |
10175697 | Sachdeva et al. | Jan 2019 | B1 |
10191155 | Curatu | Jan 2019 | B2 |
10209359 | Russell et al. | Feb 2019 | B2 |
10211592 | Villeneuve et al. | Feb 2019 | B1 |
10211593 | Lingvay et al. | Feb 2019 | B1 |
10217889 | Dhulla et al. | Feb 2019 | B2 |
10218144 | Munroe et al. | Feb 2019 | B2 |
10241198 | LaChapelle et al. | Mar 2019 | B2 |
10254388 | LaChapelle et al. | Apr 2019 | B2 |
10254762 | McWhirter et al. | Apr 2019 | B2 |
10267898 | Campbell et al. | Apr 2019 | B2 |
10267899 | Weed et al. | Apr 2019 | B2 |
10267918 | LaChapelle et al. | Apr 2019 | B2 |
10275689 | Sachdeva et al. | Apr 2019 | B1 |
10295668 | LaChapelle et al. | May 2019 | B2 |
10310058 | Campbell et al. | Jun 2019 | B1 |
10324170 | Engberg, Jr. et al. | Jun 2019 | B1 |
10324185 | McWhirter et al. | Jun 2019 | B2 |
10338199 | McWhirter et al. | Jul 2019 | B1 |
10338223 | Englard et al. | Jul 2019 | B1 |
10340651 | Drummer et al. | Jul 2019 | B1 |
10345437 | Russell et al. | Jul 2019 | B1 |
10345447 | Hicks | Jul 2019 | B1 |
10348051 | Shah et al. | Jul 2019 | B1 |
10386489 | Albelo et al. | Aug 2019 | B2 |
10394243 | Ramezani et al. | Aug 2019 | B1 |
10401480 | Gaalema et al. | Sep 2019 | B1 |
10401481 | Campbell et al. | Sep 2019 | B2 |
10418776 | Welford et al. | Sep 2019 | B2 |
10445599 | Hicks | Oct 2019 | B1 |
10451716 | Hughes et al. | Oct 2019 | B2 |
10473788 | Englard et al. | Nov 2019 | B2 |
10481605 | Maila et al. | Nov 2019 | B1 |
10488496 | Campbell et al. | Nov 2019 | B2 |
10491885 | Hicks | Nov 2019 | B1 |
10502831 | Eichenholz | Dec 2019 | B2 |
10503172 | Englard et al. | Dec 2019 | B2 |
10509127 | Englard et al. | Dec 2019 | B2 |
10514462 | Englard et al. | Dec 2019 | B2 |
10520602 | Villeneuve et al. | Dec 2019 | B2 |
10523884 | Lee et al. | Dec 2019 | B2 |
10535191 | Sachdeva et al. | Jan 2020 | B2 |
10539665 | Danziger et al. | Jan 2020 | B1 |
10545240 | Campbell et al. | Jan 2020 | B2 |
10551485 | Maheshwari et al. | Feb 2020 | B1 |
10551501 | LaChapelle | Feb 2020 | B1 |
10557939 | Campbell et al. | Feb 2020 | B2 |
10557940 | Eichenholz et al. | Feb 2020 | B2 |
10571567 | Campbell et al. | Feb 2020 | B2 |
10571570 | Paulsen et al. | Feb 2020 | B1 |
10578720 | Hughes et al. | Mar 2020 | B2 |
10591600 | Villeneuve et al. | Mar 2020 | B2 |
10591601 | Hicks et al. | Mar 2020 | B2 |
10606270 | Englard et al. | Mar 2020 | B2 |
10627495 | Gaalema et al. | Apr 2020 | B2 |
10627512 | Hicks | Apr 2020 | B1 |
10627516 | Eichenholz | Apr 2020 | B2 |
10627521 | Englard et al. | Apr 2020 | B2 |
10636285 | Haas et al. | Apr 2020 | B2 |
10641874 | Campbell et al. | May 2020 | B2 |
10663564 | LaChapelle | May 2020 | B2 |
10663585 | McWhirter | May 2020 | B2 |
10677897 | LaChapelle et al. | Jun 2020 | B2 |
10677900 | Russell et al. | Jun 2020 | B2 |
10684360 | Campbell | Jun 2020 | B2 |
10908232 | Latham et al. | Feb 2021 | B2 |
20180069367 | Villeneuve et al. | Mar 2018 | A1 |
20180284239 | LaChapelle et al. | Oct 2018 | A1 |
20180284240 | LaChapelle et al. | Oct 2018 | A1 |
20180284275 | LaChapelle | Oct 2018 | A1 |
20180284280 | Eichenholz et al. | Oct 2018 | A1 |
20190310368 | LaChapelle | Oct 2019 | A1 |
Entry |
---|
U.S. Appl. No. 17/352,829, filed Jun. 21, 2021, Huntington et al. |
U.S. Appl. No. 17/352,937, filed Jun. 21, 2021, Cadugan et al. |
U.S. Appl. No. 17/376,607, filed Jul. 15, 2021, Stewart et al. |
U.S. Appl. No. 17/400,300, filed Aug. 12, 2021, Myers et al. |
U.S. Appl. No. 17/197,314, filed Mar. 10, 2021, Taylor et al. |
U.S. Appl. No. 17/197,328, filed Mar. 30, 2021, Taylor et al. |
U.S. Appl. No. 17/230,253, filed Apr. 14, 2021, Judkins, III et al. |
U.S. Appl. No. 17/230,276, filed Apr. 14, 2021, Cadugan. |
U.S. Appl. No. 17/230,277, filed Apr. 14, 2021, Judkins, III et al. |
TW201422772 (w/Machine English Translation from Espacenet.com) 70 Pages, Jun. 16, 2014, Schut et al. |