The present disclosure is directed to a novel sensor device for 2D (two-dimensional) imaging and for 3D (three-dimensional) imaging. The present disclosure is further directed to a method for 2D imaging and for 3D imaging. The present disclosure is moreover directed to a device for scanning depth and to a method for scanning depth.
3D imaging has been around for decades. Increased demand for imaging solutions and technological evolution have resulted in a broad spectrum of applications.
3D imaging is useful in different fields. For example, depth sensing has gained importance in the automotive industry in application areas, such as prevention of car crash, monitoring of the physical conditions of the car driver, or autonomous driving. Such applications require substantially real-time responses to be effective.
In medical diagnosis and surgery, delicate and rare operations have been made feasible thanks to the ability to reproduce the structure and locations of internal organs based on 3D imaging, which enables surgeons to train operations in advance. For this kind of application, slow and precise scanning systems to generate the 3D images can be used since there is no need for real time operation.
In the areas of security and data protection, face recognition in combination with fingerprinting can provide a method to protect people from theft of their personal data. Here, again, relatively slow systems can be used for 3D imaging since the main issue is to create a detailed 3D image of the object, e.g., of the face and/or the fingerprint, and real time response is not a concern. Structured light systems and multi camera systems are widely used for this purpose.
Some imaging techniques build on depth information. To gain depth information, various solutions for depth detection have been developed. Depending on the application in question, some of these solutions may be more suitable than others.
Stereovision, which is a known 3D imaging technique, obtains 3D geometric information of an object from multiple images based on the principle of human binocular vision. Two images of the object are acquired simultaneously or consecutively from different viewing angles and these two images are used to derive the object's 3D geometric information based on trigonometry. The object's 3D profile and location can thereby be calculated. The stereovision method is a well-developed 3D imaging technique yielding satisfactory results. However, the stereovision method is computationally intensive and can perform poorly when baseline distance is large. In active stereo vision, structured light is additionally used. The stereovision method may be enhanced by using a structured light source (active stereovision). As mentioned before, the stereovision method may use two sensors and was an early technique to produce 3D images (see, e.g., the Fujifilm FinePix Real 3D).
In the structured-light technique, a three-dimensional object is illuminated by projecting a pattern of light onto the surface of the object. One example of the pattern of light are parallel stripes of light. Due to the three-dimensional structure of the object, the illumination pattern looks different on the surface of the three-dimensional object than on a planar surface, i.e., the illumination pattern is reflected differently from the three-dimensional object than from the planar surface. From the way the projected light pattern looks in the image reconstruction of the three-dimensional structure of the object can be achieved. The structured-light technique may employ a light source made of points with a known geometry. The structure-light technique was used in, e.g., the iPhone 8
A further known technique is time-of-flight (TOF) imaging, in which light with a known temporal profile impinges on a three-dimensional object and is reflected to a camera. The time for the light to reach the camera depends on the distance of the point of reflection on the surface of the object to the camera. Knowledge of this time enables the three-dimensional structure of the object to be inferred. The light may be modulated. By sampling the reflected light at a high frequency, the camera or sensor measures the phase of the reflected light with respect to the emitted light. The phase is related to the distance of the point on the object from which light was reflected. This method is called indirect TOF or i-TOF.
More recently, and thanks to the improvements in avalanche detectors, silicon photomultipliers (SiPMs), avalanche photodiodes (APDs), or single-photon avalanche diodes (SPADs), direct TOF (dTOF) has gained increased importance in 3D imaging methods. It is known that avalanche detectors can generate a large voltage signal when capturing a single electron. This voltage signal can then be used to precisely measure the time of arrival of one photon. With the equation
where c is the speed of light and Δt is the interval between the instant at which the light pulse is generated and the reflected pulse is detected on the sensor, the distance d at which the object is located can be calculated.
The SPADs have been available based on CMOS technology, which enables the fabrication of large arrays of SPADs. The performance of CMOS-based SPADs still does not meet the performance levels of the SPADs, made by an optimized fabrication technology, but has been continuously improving and enables the creation of small arrays.
The technology of the SPADs is described, for example, in the international patent application No. WO 2021/067377 A1 (Sense Photonics/University of Edinburgh). This patent application discloses a light detection and ranging (LIDAR) system including an emitter array comprising a plurality of emitter units operable to emit optical signals, a detector array comprising a plurality of detector pixels (such as SPADs) operable to detect light for respective strobe windows between pulses of the optical signals, and one or more control circuits.
However, the SPADs suffer from drawbacks. The fill factor, i.e., the ratio of a SPAD's light sensitive area to the SPAD's total area, can be limited. The SPADs require use of surface area on a chip for the drive circuitry to be operative. A further drawback is that is that the quantum efficiency is low, especially in the near infrared (NIR) region in which dTOF operates for eye-safety reasons and improvements of the signal-to-noise ratio (SNR) by reducing background light. Furthermore, SPADs have a relatively high-power consumption.
A technique called Temporal Pixel Multiplexing (TPM) is described in the international patent application No. WO 2008/138543 A1. Bub et al. (see, e.g., Nature Methods volume 7, pages 209-211 (2010)) implemented TPM by using digital micromirror devices (DMD) and a CCD image sensor. High-speed imaging can be achieved by grouping pixels into pixel subsets having serial exposure timings to generate a frame per pixel subset. This results in a set of time-separated low-resolution images. The low-resolution images can then be viewed in sequence as a movie or can be combined to form a single high-resolution image.
In a further development, the international patent application No. WO 2012/013918 A1 discloses a pixel design for CMOS sensors which allows for selection of individual pixels. The TPM can be implemented electronically on chip using the disclosed CMOS sensors, without the use of any electro-mechanical moving parts. A pixel circuit is described which has as inputs a shutter signal, a column select signal, and a row select signal.
International patent application No. WO 2019/239128 A1 discloses a pixel design in which a column sample signal and a row sample signal can be provided to each pixel to read an exposure signal into a storage node of the pixel. The storage node is arranged such that both the column sample signal and the row sample signal need to be active for the transfer to take place. The disclosed pixel design does not require a shutter signal.
The U.S. patent application No. US 2015/319422 A1 (Fereyre) US2015319422A1 discloses a method for producing images of a three-dimensional scene including information on the distance of each point of the scene, this method uses a pulsed light source and an image sensor, the sensor comprising pixels able to accumulate, in a respective storage node of each pixel, charge generated by the light.
The present disclosure aims to measure TOF directly using a new architecture for a 2D sensor.
The present disclosure relates to a depth sensing technique in which pixels of a 2D image sensor are driven with individual timings in order to acquire time information on an arrival of photons generating an image. Furthermore, the present disclosure relates to a depth-scanning image sensor (DSIS) that collects depth information by scanning in the depth direction.
An image sensor comprises a plurality of pixels which are controlled by means of control signals.
If pixel operations related to the exposure timings of an individual one of the pixels are controlled in a temporal manner, e.g., synchronized with the emission of, e.g., a light pulse, the pixel operations can be used to gather information on the TOF of the light pulse, amounting to a dTOF measurement. The gathered TOF information can be used to derive depth information.
In order to simplify an implementation of exposure timings, the pixels may be grouped into subgroups so that fewer of the exposure timings need to be implemented for each image.
Subsequent images may have different schemes of the exposure timings assigned to the subgroups of the pixels. Varying ones of exposure timing schemes enable scanning over varying time ranges.
In general, in order to gather depth information over a depth range, a series of images with different exposure timing schemes is acquired. Suppose that, for a specific application, M different exposure timings are required. These M different exposure timings would be applied to every pixel, which would require M different images. However, 3D images often do not require a level of resolution as corresponding 2D images, for which reason there is a possibility of trading off 2D spatial resolution with the number of images required to reconstruct the depth information. By grouping the pixels into subgroups of, e.g., N pixels and applying the M different exposure timings to the pixels of each group, M/N images (or frames) would be required to scan the depth information. The 2D resolution would correspondingly be reduced by a factor of 1/N.
The parameters of light pulse duration, light pulse intensity, light pulse repetition rate, sensor timing within nonoverlapping sub-blocks in the photosensing array are controlled by signals that use different timings. The exposure timings together with a light pulse are used to compute distances of objects in a scene by means of Time of Flight (ToF) measurement together with standard 2D imaging.
An imaging device for 3D imaging according to the present disclosure comprises a pixel array comprising a plurality of pixels. The plurality of pixels comprises a plurality of subgroups of pixels and at least one of the plurality of subgroups of pixels comprising depth scanning pixels. The imaging device comprises further); a signal generator for generating an exposure timing scheme defining exposure durations for the least one subgroup of the pixels, wherein ones of the exposure durations cover at least portions between a frame start time and a frame end time of predefined frame; a synchronizer for synchronizing a generation of a light pulse with the exposure timing scheme; and readout circuitry for readout of charges accumulated in the at least one subgroup of the pixels.
The imaging device may further comprise a light source for generating the light pulse.
The at least one subgroup of pixels may comprise a photodiode, a transistor for resetting the photodiode, a floating diffusion, a transistor TX for transferring a charge from the photodiode to the floating diffusion, a transistor for resetting the floating diffusion, a source follower for converting the charge at the floating diffusion into a voltage, and a select transistor for providing the voltage to an output line.
The imaging device may comprise a storage node connected to the input transistor.
At least some of the exposure durations may cover distinct portions between the frame start time and the frame end time.
The imaging device may further comprise a processor for processing the accumulated charges after readout.
The processor may be configured to determine a temporal position of a peak of a reflected light pulse.
The imaging device may further comprise a 2D/3D logic switch for switching between 2D and 3D imaging modes.
The signal generator may comprise a shift register comprised of a series of base blocks, the base block being connected in series and ones of the base block comprising flip-flops.
The shift register further comprises two signal generators connected in series to both ends of the series of base blocks, wherein the two signal generators comprise flip-flops.
A method of 3D imaging of an object using a pixel array is also disclosed. As already note, The pixel array comprises a plurality of subgroups of pixels and at least one of the plurality of subgroups of pixels comprises depth scanning pixels. The method comprises generating a light pulse; exposing the at least one subgroup of pixels of a pixel array, grouped into at least one subgroup of the pixels, to the reflected light pulse, reflected from the object, according to an exposure timing scheme defining exposure durations for the at least one subgroup of the pixels, wherein ones of the exposure durations cover at least portions between a frame start time and a frame end time of predefined frame; reading out charges accumulated in the at least one subgroup of the pixels; determining a temporal position of a peak of the reflected light pulse.
The method may further comprise defining at least some of the exposure durations to cover distinct portions between the frame start time and the frame end time.
The method may further comprise repeating the exposing before the reading-out of the charges.
The method may further comprise repeating the exposing, according to a modified one of the exposure timing scheme and followed by the reading-out of the charges.
The determining of the temporal position of the peak of the reflected light pulse comprises using a linear center-of-gravity formula.
The determining the temporal position of the peak of the reflected light pulse comprises using a non-linear formula.
The non-linear formula may be a non-linear center-of-gravity formula.
One or both of the frame start time and the frame end time of the frame may be defined based on one or more of at least one wavelength of the light pulse, a power of the light pulse, a duration of the light pulse, a repetition rate of the light pulse, a resolution to be achieved, noise, and background light.
The invention will now be described on the basis of the drawings. It will be understood that the embodiments and aspects of the invention described herein are only examples and do not limit the protective scope of the claims in any way. The invention is defined by the claims and their equivalents. It will be understood that features of one aspect or embodiment of the invention can be combined with a feature of a different aspect or aspects and/or embodiments of the invention.
The present disclosure relates to an imaging device 100 (see
Let us consider a semiconductor imaging device 100 having a pixel array 10 with a number X×Y of pixels 20. The pixels 20 may be of the same type and be constructed in the same way. In the example of
The light pulse 50 is generated by a light source 107. The light source 107 is connected to a laser driver 150 and receives control signals from the laser driver 150 relating to generation of the light pulse 50. The light pulse 50 reflected by the object 105 to be imaged is detected by the pixel array 10 by exposing the pixels 20 to the reflected light pulse 50. In one aspect of the disclosure, the imaging device 100 comprises the light source 107. In a further aspect, the imaging device 100 may comprise light guiding means to guide the light pulse 50 in a selected direction to the object 105 to be imaged or a portion thereof.
The light source 107 has characteristics such as at least one wavelength including at least one central wavelength or a main wavelength; a narrow band; a peak power; a light pulse duration; a pulse repetition rate; and an average power. These characteristics may be chosen using the laser driver 150. In one aspect of the disclosure, the semiconductor imaging device 100 is constructed on a silicon substrate, and the at least one wavelength of the light source 107 may be chosen to lie between approximately 0.8 μm and approximately 1.05 μm. In another aspect, the imaging device 100, the range of the at least one wavelength can be extended to approximately 1.55 μm or 1.6 μm. In other words, the range of the at least one wavelength extends from approximately 0.8 μm to 1.6 μm. The light pulse 50 thus may be composed of infrared light (IR). In order to improve the detection of IR radiation, in one aspect of the disclosure, the material of the substrate may be chosen to be a material with higher detection efficiency, e.g., Germanium. Using IR light to generate the light pulse 50 enables human interaction without sensation of the light pulse 50 as the human eye has a low sensitivity for IR light. Furthermore, solar irradiation at ground level may render impossible detecting the light pulse 50 having the at least one wavelength in the visible frequency range. Due to atmospheric absorption there are dips in the IR range of the solar irradiation spectrum at ground level, enabling decreasing a level of background noise from solar irradiation when using light in the IR range for the generation of the light pulse 50. Settings of an average power of the light pulse 50, a peak power of the light pulse 50 a duration of the light pulse 50, and a repetition rate of the light pulse 50 may depend on each other and may furthermore be set in consideration of the object 105 to be imaged and/or imaging conditions such as a distance range, a resolution, and eye safety. In one aspect of the disclosure, the light pulse 50 has a relatively short duration with a relatively high peak power and a repetition rate in the order of the kHz. Furthermore, a field of view (FOV) of the light source 107 may match a FOV of the imaging device 100. The light pulse 50 may illuminate the entire FOV of the imaging device 100, which obviates any need of using the light guiding means.
The X×Y pixels 20 are grouped into one or more subgroups 40 in the pixel array 10. The one or more subgroups 40 are termed depth scanning pixels or DS pixels. It is assumed that M different timings are required to gain sufficient depth information on the object 105 to be imaged. The number M depends on a range of depths (or distances) to be detected (or depth range) and a resolution chosen for the imaging application. Suppose that the depth range is identified by its boundaries d1 and d2, i.e., d1≤d≤d2. The imaging device 100 will detect ones of the reflected light pulse 50 arriving at the imaging device 100, i.e. at the pixel array 10, in a time range
relative to an instant at which the light pulse 50 leaves the light source 107 and travels toward the object 105 to be detected. In order to detect a single one of the reflected pulse 50, repeated exposures of the imaging device 100 may be performed depending on the shape of the reflected light pulse 50, e.g., a time interval spanned by the reflected light pulse 50, and on an exposure timing scheme 45 (explained below). However, it is conceivable that, apart from the single one of the light pulse 50 that repeatedly irradiates the object 105 to be imaged, further ones of the light pulse 50 are used to irradiate the object 105 to be imaged. The further light pulses 50 may have, e.g., different shapes. Depending on the exposure timing scheme 45 and the previously defined time range, the imaging device 100 will perform a certain number (M) of exposures with different exposure timings, as further explained below.
The number of DS pixels 40 in the pixel array 10 can be different from, e.g., smaller than or equal to, the number of pixels 20 in the pixel array 10. If the pixels 20 are divided into a number of DS pixels 40, e.g., A×B DS pixels 40, with A≤X; B≤Y, ones of the DS pixels 40 have a number of (X×Y)/(A×B) pixels 20. We define n=X/A and m=Y/B. A resulting DS resolution of the pixel array is n×m.
In the afore-mentioned example, the pixel array 10 is divided regularly into DS pixels 40, i.e., the DS pixels 40 have a same number of pixels 20. In one aspect of the disclosure, the number X×Y of pixels 20 of the pixel array 10 may be divided into DS pixels 40 in an irregular manner, in which n and m vary across the pixel array 10. In this aspect, the pixel array 10 comprises DS pixels 40 of different sizes, resulting in an image that has a varying DS resolution.
The subgroups 40 (or DS pixels 40) have n×m pixels 20. In the example of
In the example shown in
Other ones of the exposure timing scheme 45 may be adopted. In one aspect, it is possible to start the exposure of the i-th pixel, i=1, . . . , 6, at a common time and distribute or scatter the times when the exposure for the i-th pixel ends. For example, the exposure duration for the first pixel, i=1, may stop first and the exposure duration for the sixth pixel, i=6, may stop last.
In another aspect, the end time of the exposure of the i-th pixel may correspond to the start time of the exposure of another one of the pixels 10 of the DS pixel 40, e.g., but not limited to, of the i+1-th pixel.
Once the exposure timing scheme 45 has been chosen, it is possible to calculate the amount of light received by the DS pixel 40 during the i-th exposure duration texp(i), i=1, . . . , 6. In the example shown in
In general, this method yields n×m exposure durations. The number n×m also indicates the reduction in spatial resolution, when switching from an X×Y pixel array 10 to an A×B DS pixel array (as shown in
To obtain a higher number of exposure durations, it is possible to expose the DS pixel 40 to a subsequent one of the reflected light pulse 50 during a subsequent exposure, after the object 105 has been exposed during a prior exposure. In one aspect, during the subsequent exposure, the exposure timing scheme 45 may be the same as the exposure timing scheme 45 used during the prior exposure. In another aspect, during the subsequent exposure, the exposure timing scheme 45 may be different from the exposure timing scheme 45 used during the prior exposure.
Exposing the DS pixel 40 to a prior one of the light pulse 50 and subsequently to a subsequent one of the light pulse 50 enables detecting the reflected light pulse 50 during a longer period without further reducing the spatial resolution of the pixel array 10. Depending on the size of the pixel array 10, i.e., the number of pixels 20 in the pixel array 10, there is a trade-off between the number of pixels 20 per DS pixel 40, i.e., the reduction in spatial resolution, and the number of irradiations required to gain sufficient depth information on the object 105 to be imaged.
In the aspect of the disclosure shown in
The repeated exposure enables covering a longer time range, e.g., a time range covering at least the time course of the generated light pulse 50 and/or covering at least the time course of the reflected light pulse 50, as shown in
In some cases, the readout of the sensor takes longer than the acquisition of a single DSIS frame. For example, for a measurement over a 10 m range, the maximum time-of-flight is 66.7 ns. State-of-art readout rates are around 25 billion pixels per second. Hence a megapixel sensor could be read at 40,000 frames per second, which is equivalent to a readout period of 25 μs>>66.7 ns. Most of the power of the sensor is consumed during the readout. The power consumed depends on the detailed architecture of the sensor and may range in about few Watts for very high-speed sensors with many pixels. As explained above, the power of the light pulse 50 needs to be kept low in order to limit the danger risks. Furthermore, an overall power to operate a system 200 (see
The repeated exposure increases the signal-to-noise ratio (SNR). Increasing the SNR is of advantage, e.g., when the available power of the light source 107 is low. In other circumstances, for eye safety considerations, the output power of the light source 107 has to be restricted, and this restriction of the output power may result in results of inferior quality. The repetition of the irradiation of the object 105 with the light pulse 50 enables overcoming situations in which a limited amount of light impinges on the pixel array 10 by increasing the SNR.
The time range covered by the repetitions of the irradiation may depend on the distance d (or on a range of a two-dimensional distribution of distances d when scanning across the object 105 in two dimensions) at which the object 105 to be imaged is arranged from the pixel array 10 and/or from the light source 107. As explained above, the formula
relates the distance d between the light source 107 and/or the pixel array 10 as well as the time Δt the light pulse needs to travel twice the distance d. Thus, if the object 105 has a distance of 1 m, Δt=6.7 ns; if the object 105 has a distance of 100 m, Δt=670 ns. Repeating the measurement 1000 times will thus lead to a covered time range of 6.7 μs for a distance of 1 m and to a time range of 0.67 s for a distance of 100 m. Accumulating the charge in the pixels 10 over repeated irradiations thus can be achieved relatively fast Furthermore, charge accumulation reduces time spent for readout of the pixels 10 which is relatively time-consuming, as explained above.
In one aspect, ones of the pixels 10 have an architecture in which a charge storage device is arranged within the pixel architecture. Charge representing the amount of light falling on the pixel is transferred to the charge storage device. The imaging device 100 may further comprise a control device for controlling the transfer of charge to the charge storage device.
In another aspect, a charge storage device may be arranged outside the pixels 20, e.g., at the periphery of the pixel array 10.
An example of a method of imaging the object 105 to be imaged will now be described. However, variations of the described algorithm are possible.
The method according to the disclosure may be improved by the fact that the light pulse may last longer than the accumulated exposure durations of the exposure timing scheme 45.
The method uses interpolation between pixels to achieve a TOF resolution which is better than the exposure time intervals used during exposure of the pixel array 10. The present disclosure is directed to a method that may be referred to as non-linear Center-Of-Gravity (COG). The non-linear COG represents an improvement over the linear COG as it can cope with light pulses 50 of any arbitrary shape, while the linear COG only works well with rectangular light pulses 50. The non-linear COG method is used to determine the temporal position of a peak of the reflected pulse 50. Based on the determined peak position, depth information regarding the object 105 to be imaged is measured, as explained below.
In
In this way a series of integrated charges Q(j), j=0, . . . , n×m−1 is obtained. The series of charges Q(j) corresponds to exposure time steps (or exposure time intervals) having time lengths equal to ΔT(j)=texp(j+1)−texp(j) represented in
The series of charges Q(j) can be regarded as representing a time course of charges accumulated in the DS pixel 40 for the exposure time steps ΔT(j), j=0, . . . , n×m−1, relative to the onset tdelay(k). In this case, the exposure time steps ΔT(j) can be regarded as “time points”.
In the case shown in
Accounting more generally for the onset tdelay(k) of the exposure timing scheme 45, a set of charges Q(j, k), corresponding to exposure time intervals ΔT(j, k) is obtained. As indicated in
More generally, it is possible that the exposure time steps ΔT(j, k) overlap. The meaning of the exposure time step ΔT(j, k) being prior to the exposure time step ΔT(l, m) is that at least some of time points within the exposure time step (or exposure time interval) ΔT(j, k) occur earlier than the time points within the exposure time step ΔT (l, m). For the sake of simplicity, the case of overlapping exposure time steps ΔT(j, k) is not analyzed below.
When repeatedly exposing the pixel array 10 to the reflected light pulse 50, using a series of the onset tdelay(k), the series of charges Q(j, k) can be regarded as representing a time course of charges accumulated in the DS pixel 40 for the exposure times t(j, k)=tdelay(k)+Σi=0i=j−1 ΔT(i, k) with j=0, . . . , m×n−1. In case the exposure time steps (or intervals) are constant, these exposure times become t(j, k)=tdelay(k)+j·ΔT. In this aspect of the disclosure, the exposure times t(j, k) are defined with respect to the beginnings of the exposure time steps ΔT(j, k).
In another aspect shown in
In this aspect, the exposure times become
in case the exposure time steps (or intervals) are constant.
Let an index z=1, 2, . . . represent the pairs of indices j and k for a temporally ordered series of ΔT(j, k). In other words, (i, j)→z such that ΔT(z)>ΔT(z′) if z>z′ for all z=1, 2, . . . . The mapping also defines a corresponding series of accumulated charges Q(z).
To find the temporal position of the reflected light pulse 50, e.g., the temporal position T0 of the peak (see
For example, linear filtering can be applied in a first step by convolving the series of accumulated charges Q(z) with a known filter with temporal expression h(z) in order to improve the SNR of the measurement. In a step of selecting, those elements Q(z0), Q(z1), . . . of the series Q(z) which are above a certain threshold are selected. The elements Q(zi), i=1, 2, . . . are further processed in a step of determining neighboring elements (zi±1), Q(zi±2), . . . for the Q(zi). Based on the determined neighboring elements of Q(zi±1), Q(zi±2), . . . it is possible in a detection step to detect the presence of the reflected light pulse 50, e.g., of the peak of the reflected light pulse 50.
In the case of peak detection, the corresponding element of the series of charges Q(z) may be denoted Q(zm). Based merely on the element Q(zm) and on the element Q(zm′)=max{Q(zm−1), Q(zm+1)}, i.e., the one of the two adjacent elements Q(zm±1) having the larger charge value, the temporal position T0 of the peak (see
In one aspect, the step of determining the peak position comprises calculating the temporal position T0 of the peak based on the COG method, i.e. using the linear COG formula T0=(Q(zm)·t(zm)+Q(zm′)·t(zm′))/(Q(zm)+Q(zm′)).
In the case of constant exposure time steps (or exposure time intervals), i.e., ΔT(j, k)=ΔT and using the center of the exposure time steps as reference time point, the linear COG formula can be simplified. Defining Qleft to be the element of Q(zm) and Q(zm′) that has the smaller index (which is denoted p) and Qright to be the element of Q(zm) and Q(zm′) that has the larger index (which is denoted p+1). With these definitions, the linear COG formula results in
With the definition η=Qright/(Qleft+Qright), the result can be written as
The foregoing results can be simplified if the rewritten with respect to a reference time
In this case, T0=ΔT(p+η).
If the time course of the reflected light pulse 50 is completely included within one exposure time interval (or exposure time step), Q(zm′) is zero, which results in η=1 (if zm′<zm) or η=0 (if zm′>zm). If the peak of the reflected light pulse 50 is on the border between adjacent ones of the exposure time intervals then Q(zm)=Q(zm′) (within statistical uncertainty/noise fluctuations) due to the symmetry of the Gaussian form of the reflected light pulse 50, which results in η=0.5.
In implementations of the described method that deviate from the afore-mentioned situations, the variable η assumes different values depending on the relative length in time of the exposure time intervals and the time course of the reflected light pulse 50. In particular, as the size of the exposure time interval shrinks relative to the time course of the reflected light pulse 50, the range in which the defined parameter η varies shrinks accordingly. This means that the above-mentioned case of 0≤η≤1], when the time course of the reflected light pulse 50 is completely included within one exposure time interval, constitutes an upper limit in size (or length) of the range in which the parameter η can vary.
In
Depending on the relative length in time of the exposure time intervals and the time course of the reflected light pulse 50, known corrections to the linear relationship will be applied.
According to the linear COG formula for the peak position,
The linear relationship between η and the relative peak position T0′ is thus η=T0′/ΔT.
In the more general, non-linear case, the relationship between η and the relative peak position T0′ is η=F(T0′/ΔT). By inverting the function F, we obtain T0′/ΔT=F−1(η). As η is measured and F is known, one can calculate
The function F may be determined by calibration of the imaging device 100. The time course of the light pulse 50 may be predetermined for the purpose of the calibration. Furthermore, a structure of the object 105 to be imaged may be predetermined for the purpose of the calibration.
In a step of determining the depth (or distance), the measured temporal position T0 is converted into a depth (or distance) based on the TOF formula.
For the purpose of illustration, in a statistical simulation, the function F is numerically calculated based on an ideal reflected light pulse 50 without noise. During the statistical simulation, noise is repeatedly added and a distribution of the peak position T0 is determined. Corresponding results, i.e., the standard deviation of the determined distribution for different calculation methods, are plotted in the
For a more detailed view of the achieved time resolution, only the standard deviations for the temporal position T0 of the peak that are obtained with the non-linear COG method are shown in
In the example described above, two charge values Q(zm) and Q(zm′) from the series Q(z) are used in the non-linear COG method. This should not be considered as a requirement. It is possible to use more samples from the series Q(z) to improve the determination of the time position T0 of the peak. For example, one could use two or more charge values to the right and the left of the peak and calculate the temporal position (x-axis) of the peak of the reflected light pulse 50 based on an adapted non-linear correction formula.
The present disclosure relates to the architecture of the pixel 20 of the pixel array 10. In one aspect of the disclosure, the pixel 20 is a 5T pixel (see
The pixel 20 provides correlated double sampling (CDS) that allows to remove an undesired offset in measuring voltages and/or currents. CDS further enables noise reduction, e.g., a reduction of thermal noise such as kTC noise.
The pixel 20 further provides a global shutter mode and/or a rolling shutter mode. In the global shutter mode, all of the pixels 20 in the pixel array 10 having equal exposure timings and exposure time intervals. The 2D imaging may be performed in rolling shutter mode, enabling increased spatial resolution.
In one aspect of the disclosure, 3D imaging may be performed in a DSIS global shutter operation. An example of the DSIS global shutter operation is shown in the exposure timing diagram of
Other possible embodiments could differ regarding the schematic level or on the layout having a different implementation at process level. At schematic level, a storage node may be added after the source-follower input transistor SFI of the source follower. The storage node allows to store results of the method of the present disclosure and to perform readout of the results while initiating a further implementation of the method.
At a pixel layout and a fabrication level, photodiode doping profiles and shapes can be engineered to make the charge collection faster. For infrared detection using infrared light, i.e., wavelengths of the light pulse 50 and/or of the reflected light pulse 50 ranging approximately from 0.8 um to 1.6 um, which is a preferred wavelength range for 3D imaging or 3D measurements, an absorption length is generally larger than for visible light. The larger absorption length results in at least some of the photons of the light pulse 50 penetrating deeper into a substrate of the photodiode PD. The deeper penetration leads to the charge generated by the light pulse 50 travelling larger distances to a collection node. The faster charge collection enables faster operating of the pixel 20, which is of advantage for 3D imaging.
3D imaging may be performed in the following way. The transistor RST performs the reset of the floating diffusion FD at the start of the frame. Thereafter, the transistor AB and the transistor TX implement timings, such as the exposure timing scheme 45.
In
During the GRST phase, the floating diffusion FD of the i-th pixel, i=1, . . . , 6, is reset by turning on the RST transistor. Furthermore, the photodiode PD of the i-th pixel, i=1, . . . , 6, is freed of any residual collected charge and kept at its pinning voltage by means of the transistors AB.
During the TRIG phase, a trigger is generated internally or externally to start the DSIS phase.
After receiving the trigger, during the DSIS phase, a DSIS synchronizer 110 generates a start pulse signal for a DSIS signal generator 120 to generate the exposure timing scheme 45. The generating of the start pulse for the DSIS signal generator 120 accounts for the onset tdelay(k) that has been previously set. The DSIS signal generator 120 generates a signal timing for the transistor AB and the transistor TX of the i-th pixel, i=1, . . . , 6, to reproduce the exposure timing scheme 45, e.g., the exposure timing scheme 45 shown
As can be seen from
Furthermore, upon generation of the start pulse signal by the DSIS synchronizer 110 during the DSIS phase, a TX signal goes high. The moment at which the TX signal goes high depends on an operation mode. In a Mode 0, the TX signal goes high shortly after the start pulse signal and before the AB signal goes low for the i-th pixel, i=1, . . . , 6. In a Mode 1, the TX goes high after the AB signal goes low for the i-th pixel, i=1, . . . , 6.
During the DSIS phase, the DSIS synchronizer 110 generates the stop pulse signal which is sent to the DSIS signal generator 120 to end the exposure. The stop pulse signal results in the TX signal going low and charge being transferred from the photodiode PD node to the floating diffusion FD. The AB signal goes high for the i-th pixel, i=1, . . . , 6, which resets the photodiode PD node.
In the RO rolling phase, the i-th pixel, i=1, . . . , 6, is read in a rolling shutter mode. In the rolling shutter mode, rows may be read in pairs.
In an aspect of the disclosure, a CDS (correlated double sampling) phase may be included in the timing after the GRST phase and before the TRIG phase. The CDS phase includes reading the reset level, which is different among the pixels 20 in the pixel array 10 and depends on: (i) the voltage applied at the gate of the RST transistor; (ii) VDD; (iii) charge injection and feedthrough on the FD node; (iv) reset noise (which is characteristic of the CMOS imaging device 100 and due to the fact that a resetting operation effectively samples a voltage on the photodiode node PD); (iv) dark current (which is generally negligible in pixel designs that involve the use of pinned ones of the photodiode PD).
In the aspect shown in
2D imaging may be performed in the following exemplary way that implements a rolling shutter mode in which light integration (or charge integration) and pixel readout are both controlled row by row. The 2D imaging includes CDS, an example of which is shown in
During a SINT (start integration) phase, the address counter 140 is activated and starts generating the addresses to be sent to the row decoder 145. The row decoder 145 decodes the address and generates a row pointer signal which is sent to the row drivers to start the integration on ones of the pixels 20 belonging to selected consecutive rows of the pixel array 10, e.g., first on rows 0 and 1 and subsequently on rows 2 and 3 in the example of
During the RST RO (reset readout) phase, a reset of the floating diffusion FD for the ones of the pixels 20 belonging to the selected rows is performed, when during the RST RO phase the RST signal goes high and subsequently goes low again for the ones of the pixels 20 belonging to the selected rows. A voltage at the floating diffusion node FD due to the accumulated charge is provided to the column output line OUT for readout, when during the RST RO phase for the ones of the pixels 20 belonging to the selected rows the SEL signal goes high and during the SIGNAL RO (signal readout) phase the SEL signals goes low.
After the reset value has been read, for the ones of the pixels 20 belonging to the selected rows an accumulated or integrated charge is transferred from the photodiode PD to the floating diffusion FD for the selected rows by pulsing the transistor TX for the selected rows. Subsequently, the accumulated or integrated charge is transferred to the output line OUT in order to be read, while the SEL signal continues to be high for the ones of the pixels 20 belonging to the selected rows.
After the selected rows have been read out, the foregoing phases will be applied further selected consecutive ones of the rows of the pixel array 10. The selection and readout of rows of the pixel array 10 may be repeated until the pixel array 10 has been completely read. Once the sensor has been read completely the process may start over again from the rows initially selected.
The signal timing described with reference to
The imaging device 100 comprises the DSIS synchronizer 110. The DSIS synchronizer 110 synchronizes the laser driver 150 with the exposure timing scheme 45. The DSIS synchronizer 110 is programmable to account for the onset tdelay(k) of the exposure timing scheme 45.
The comparator coupled to the first register 1520 for storing the light-pulse-generation delay generates a Laser_start pulse and passes the Laser_start pulse to the laser driver 150, when the counter reaches the value of the light-pulse-generation delay. The comparator coupled to the second register 1530 for storing the start-of-exposure delay generates a DSIS_start pulse and passes the DSIS_start pulse to a basic state machine to turn on the signal generation for the pixel array 10, when the counter reaches the value of the start-of-exposure delay. The comparator coupled to the third register 1540 for storing the end-of-exposure delay generates a DSIS_stop pulse and passes the DSIS_stop pulse to the basic state machine to turn off the signal generation for the pixel array 10, when the counter reaches the value of the end-of-exposure delay.
The state machine comprises a multiplexer and a D flip-flop. An example of an operation of the state machine is illustrated by means of a timing of output waveforms, shown in
The imaging device 100 further comprises the DSIS signal generator 120, as shown in
A DSIS_start_stop_AB-generator 123 receives a DSIS_start_stop signal from the DSIS synchronizer 110. The DSIS_start_stop_AB-generator 123 generates a DSIS_start_stop_AB signal, which is received by the several AND gates 122 connecting the several base blocks 121 of the DSIS signal generator 120. The DSIS_start_stop_AB signal is the result of an inverted and delayed one of the DSIS_start_stop signal from the DSIS synchronizer 110, as shown in
The DSIS signal generator 120 has an asynchronous reset async_rst. The base blocks 121, as mentioned above, are connected in series, and form a signal generator shift register. The signal generator shift register generates, by means of an inverting of an output of a flip-flop (e.g., a D flip-flop) of the base block, output signals referred to as AB<i>, i=1, . . . , 6, for the i-th pixel, i=1, . . . , 6. As explained above, the number of pixels 20 per DS pixel 40 is not restricted to 2×3=6, but is generally n×m. As shown in
The DSIS_start_stop_AB signal is used as a synchronous reset for the DSIS signal generator 120. The DSIS_start_stop_AB signal is received by the several AND gates 122 connecting the several base blocks 121 and sets the outputs of the DSIS signal generator 120 to a known state when the DSIS_start_stop_AB signal is low. The outputs of the DSIS signal generator 120 are inverted ones of the outputs of each flip flop which means that: (i) When DSIS_start_stop_AB is low, the outputs are all kept at high voltage level. (ii) When DSIS_start_stop_AB is high, the output of the flip flop of the base block depends on the output of the flip-flop of the previous base block. Point (ii) describes the phase in which AB signals go low, as explained with respect to
In fact, when DSIS_start_stop_AB is high and as soon as the first rising edge of the clock is hit, the output of the first flip flop goes low. When the second rising edge comes, the output of the second flip flop, which depends on the output of the previous one, goes low. The output is then propagated to the successive flip flop until the end of the series is reached. At this point all the outputs are kept low until the DSIS_start_stop_AB signal turns low again. This event makes all the outputs go high at the same time as soon as the first rising edge of the clock is hit.
Regarding TX signals two operation modes are allowed: In Mode 0, the TX signal goes high from the beginning, before the first AB signal goes low. In Mode 1, the TX signal goes high after the last AB signal goes low.
AB and TX timings are strictly correlated to each other. In one aspect, a TX signal generator generates the TX signal. In this aspect, the TX signal generator is integrated into the previously presented shift register. For example, two blocks 127, 128 comprising a flip flop 127 and an AND gate 128 connected to the flip flop input are added (appended) to the series of the base block 121 of the DSIS signal generator 120. One of the blocks 127, 128 is arranged at one end of the series of the base block 121 for TX_mode<0>. Another one of the block 127, 128 is arranged at the other end of the series of the base block 121 for TX_mode<1> (see
The imaging device 100 further comprises a 2D/3D logic switch 130, as shown in
The device 100 according to the present disclosure may comprise a processor for processing the accumulated charges after readout.
| Number | Date | Country | Kind |
|---|---|---|---|
| 22382171.1 | Feb 2022 | EP | regional |
| Filing Document | Filing Date | Country | Kind |
|---|---|---|---|
| PCT/EP2023/055026 | 2/28/2023 | WO |