This application claims the priority benefit of European Application for Patent No. 20306680.8, filed on Dec. 23, 2020, the content of which is hereby incorporated by reference in its entirety to the maximum extent allowable by law.
The present disclosure relates generally to image sensors and, more particularly, to time of flight sensors.
Image sensors of the time of flight type are known. Among these sensors, indirect time of flight sensors are configured to determine a dephasing between periodic light emitted by the sensor towards a scene to capture, and light received by pixels of the sensor, the received light corresponding to the light reflected by the scene when illuminated by the sensor. Based on the dephasing determined for each pixel of the sensor, a distance between this pixel and a conjugated point of the scene may be calculated. From the determined distance for each pixel, a depth map of the scene may be generated.
There is a need to overcome all or some of the drawbacks of known indirect time of flight sensors.
Embodiments herein address all or some of the drawbacks of known indirect time of flight sensors.
One embodiment provides an indirect time of flight sensor comprising: a matrix of pixels wherein each pixel comprises a photoconversion region and at least two sets each comprising a charge storage region and a controllable transfer device for transferring charge from the photoconversion region towards said storage region; first conductive lines parallel to each other, configured to transmit first control signals to the transfer devices; a first circuit configured to provide the first signals to the first conductive lines; an illumination device for illuminating a scene to capture; and a second circuit configured to control the illumination device. The scene is divided into first areas and the illumination device and the second circuit are configured to successively illuminate each first area. The matrix is divided into second areas each comprising adjacent lines of pixels, parallel to the first conductive lines, wherein a disposition of the matrix and of the illumination device is configured such that each first area corresponds to one of the second areas. The first circuit is configured to provide different first signals to the different second areas.
According to one embodiment, the illumination device comprises an array of laser sources and an optical device configured to direct light emitted by the array of laser sources towards the scene. The array is divided into sets of laser sources, each set being configured to illuminate a corresponding first area, the second circuit being configured to control said sets one after the other. The optical device is configured to direct the emitted light differently depending on a control signal, the second circuit being configured to provide, at each illumination of a first area, said control signal corresponding to a directing of the light towards said first area.
According to one embodiment: the sensor comprises second conductive lines parallel to the first conductive lines and configured to receive output signals of the pixels; each pixel comprises a selection device configured to selectively couple output(s) of said pixel to at least one corresponding second conductive line; and the first circuit is configured to provide second control signals to the selection devices via third conductive lines perpendicular to the second conductive lines.
According to one embodiment, the first circuit is configured to control, by means of the second signals, a reading of all the pixels after each illumination of a first area, before an illumination of a next first area.
According to one embodiment: the sensor comprises second conductive lines parallel to each other and perpendicular to the first conductive lines, the second conductive lines being configured to receive output signals of the pixels. Each pixel comprises a selection device configured to selectively couple output(s) of said pixel to at least one corresponding second conductive line; and the first circuit is configured to provide second control signals to the selection devices via third conductive lines perpendicular the second conductive lines.
According to one embodiment, the second circuit is configured, before each reading of all the pixels controlled by the first circuit, to control several successive illumination cycles each comprising a unique illumination of each first area, and to control an absence of light emission by the illumination device during said reading.
According to one embodiment, the first circuit is configured to control, after each illumination of a first area, a reading of only the pixels of the second area corresponding to said first area.
According to one embodiment, the second circuit is configured to control an absence of light emission by the illumination device when the first circuit controls the reading of the pixels of a second area.
According to one embodiment: the matrix is divided into first and second halves, a separation between first and second halves being parallel to the first lines, and the second conductive lines of each half ending at said separation. The first circuit is configured to simultaneously control charge transfers in the pixels of a second area of one of the halves and a reading of the pixels of a second area of the other one of the halves. A first part of a semiconductor substrate comprises the first half of the matrix and a second part of said semiconductor substrate comprises the second half of the matrix; insulation structures passing through the semiconductor substrate insulate said parts of the semiconductor substrate from each other. A reference voltage provided to the first part of the semiconductor substrate is electrically decoupled from a reference voltage provided to the second part of the semiconductor substrate.
According to one embodiment, for each voltage level intended to be provided to at least one pixel of the first half of the matrix, and, simultaneously, to at least one pixel of the second half of the matrix, the sensor comprises a generator of said voltage level for the first half and a generator of said voltage level for the second half, the generators being electrically decoupled from each other.
According to one embodiment, the sensor comprises a first reading circuit coupled the second conductive lines of the first half of the matrix, and a second reading circuit coupled to the second conductive lines of the second half of the matrix, a reference voltage of the first reading circuit being electrically decoupled from a reference voltage of the second reading circuit.
According to one embodiment, the first reading circuit is disposed along a first edge of the matrix, on the side of the first half, and the second reading circuit is disposed along a second edge of the matrix, on the side of the second half, the first and second edges being parallel.
According to one embodiment: the semiconductor substrate comprising the matrix of pixels lies on another semiconductor substrate comprising commutators, the commutators being preferably disposed below the separation between the halves of the matrix; each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs; and the sensor comprises a reading circuit connected to the output of each commutator, the reading circuit preferably belonging to the same semiconductor substrate as the commutators.
According to one embodiment: the semiconductor substrate comprising the matrix of pixels lies on another semiconductor substrate comprising commutators, the commutators being preferably disposed below the separation between the halves of the matrix; each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs; the pixels of the matrix are arranged in column parallel to the second conductive lines; each commutator connected to second conductive lines of an odd column has its output connected to a first reading circuit; each commutator connected to second conductive lines of an even column has its output connected to a second reading circuit; and the first and second reading circuits preferably belonging to the same semiconductor substrate as the commutators.
According to one embodiment, the sensor comprises a control circuit for controlling the commutators such that the output of each commutator is coupled to the first input of said commutator during a reading of pixels of the first half of the matrix, and to the second input of said commutator during a reading of pixels of the second half of the matrix.
The foregoing features and advantages, as well as others, will be described in detail in the following description of specific embodiments given by way of illustration and not limitation with reference to the accompanying drawings, in which:
Like features have been designated by like references in the various figures. In particular, the structural and/or functional features that are common among the various embodiments may have the same references and may dispose identical structural, dimensional and material properties.
For the sake of clarity, only the operations and elements that are useful for an understanding of the embodiments described herein have been illustrated and described in detail. In particular, usual electronic systems and applications in which an indirect time of flight sensor may be provided are not described in detail, the described embodiments being compatible with these usual systems and applications.
Unless indicated otherwise, when reference is made to two elements connected together, this signifies a direct connection without any intermediate elements other than conductors, and when reference is made to two elements coupled together, this signifies that these two elements can be connected or they can be coupled via one or more other elements.
In the following disclosure, unless indicated otherwise, when reference is made to absolute positional qualifiers, such as the terms “front”, “back”, “top”, “bottom”, “left”, “right”, etc., or to relative positional qualifiers, such as the terms “above”, “below”, “higher”, “lower”, etc., or to qualifiers of orientation, such as “horizontal”, “vertical”, etc., reference is made to the orientation shown in the figures.
Unless specified otherwise, the expressions “around”, “approximately”, “substantially” and “in the order of” signify within 10%, and preferably within 5%.
Pixel 1 comprises a photoconversion region, or photosensitive region PD, for example a photodiode, preferably a pinned photodiode. The photoconversion region PD has an electrode, for example its anode, which is connected to a node 100 configured to receive a reference voltage, for example the ground GND. The photoconversion region PD is configured such that charges are generated therein when light is received by the region PD.
Pixel 1 further comprises two identical memory circuit sets E1 and E2, delimited by dashed lines in
Each set E1, E2 of the pixel 1 comprises a charge storage region mem1, mem2 and a controllable charge transfer device TGmem1, TGmem2.
Device TGmem1, respectively TGmem2, is connected between the region PD and the region mem1, respectively mem2. Device TGmem1, respectively TGmem2, is configured to transfer charges from the region PD to the region mem1, respectively mem2. More precisely, device TGmem1, respectively TGmem2, is configured to transfer charges from the region PD to the region mem1, respectively mem2, when its control signal TG1, respectively TG2, is active, for example at a high level, and to block any charge transfer between the region PD and the region mem1, respectively mem2, when this control signal is inactive, for example at a low level. Each device TGmem1, TGmem2 is, for example, a transfer gate transistor.
Region mem1, respectively mem2, is configured to store charges which are transferred therein by the transfer device TGmem1, respectively TGmem2, until these charges are transferred elsewhere in the pixel 1 during a reading phase. Each region mem1, mem2 is, for example, a pinned diode. Each pinned diode mem1, mem2 has an electrode, for example its anode, connected to the node 100, and another electrode 104, for example its cathode, coupled to the electrode 102 of the region PD by the corresponding transfer device TGmem1, TGmem2.
Pixel 1 has an output 106. During a reading phase of the pixel 1, output signals of the pixel 1 are available on the output 106.
Pixel 1 comprises a selection device 108, for example a Metal Oxide Semiconductor (MOS) transistor. The device 108 is connected between the output 106 and a reading conductive line Vx. The selection device 108 is configured to selectively couple the output 106 of the pixel 1 to the line Vx. More precisely, during a reading phase of the pixel 1, for example when a control signal RD of the device 108 is active, for example at a high level, the device 108 couples the output 106 to line Vx, and outside of a reading phase of the pixel 1, for example when signal RD is inactive, for example at a low level, the device 108 isolates output 106 from line Vx.
For example, in known time of flight sensors comprising a matrix of pixels 1 arranged in rows and columns, a line Vx is shared by all the pixels 1 which belong to the same column. To read a given pixel of the matrix, all the pixels of the row to which belongs this pixel are selected by activating signal RD for this row of pixels.
Pixel 1 comprises a controllable output circuit 110, delimited in dashed lines in
For example, during a reading phase of the pixel, when a first signal RD1 is active, for example at a high level, the circuit 110 provides a signal, for example a voltage referenced to node 100, indicative of the number of charges stored in region mem1, and, when a second signal RD2 is active, for example at a high level, the circuit 110 provides a signal, for example a voltage referenced to node 100, indicative of the number of charges stored in region mem2.
In the particular example of
The pixel 1, for example, further comprises a transistor AB connected between the electrode 102 of the region PD and a node 118 configured to receive a bias voltage VAB. The transistor AB is controlled by a signal TGAB. The transistor AB is configured, when off, to operate as an antiblooming device for the region PD, and, when on, to reset the region PD, that is to say to evacuate all the photo-generated charges accumulated in the region PD towards the node 118.
In a usual indirect time of flight sensor comprising a matrix of pixels 1 arranged in rows and columns, during an integration phase, all the transfer devices TGmem1 and TGmem2 of all the pixels 1 of the matrix are driven simultaneously to transfer charges photo-generated in the region PD of each pixel towards regions alternatively mem1 and mem2 of this pixel. Further, during the integration phase, the scene to capture is illuminated by the sensor in a flash manner, that is to say that each time the sensor emits light, the whole scene is illuminated. During the integration phase, the light is, for example, emitted under the form of a burst of successive periodic pulses of light. After an integration phase, all the pixels 1 of the matrix are read. More particularly, during the reading of all the pixels 1 of the matrix, the rows of pixels are selected the one after the other with the signals RD, and all the pixels 1 of a selected row are read simultaneously.
Although in the example of
Further, although in the example of
More generally, many different pixels known by those skilled in the art may be used in a matrix of pixels of an indirect time of flight sensor, and the pixel 1 of
In the following description, unless indicated otherwise, when reference is made to a pixel of an indirect time of flight sensor, this means a reference to the pixel 1 of
It is here proposed to capture a scene with an indirect time of flight sensor by successively illuminating different areas of the scene, only one area of the scene being illuminated at a time. Said otherwise, the scene is divided into a plurality of area, and the scene is fully illuminated by successively illuminating each area of scene, each of these areas being illuminated at least once.
The sensor 2 comprises a matrix 200 of pixels 1, only one pixel 1 being referenced on
The sensor 2 comprises a reading circuit READOUT. Circuit READOUT is configured to received output signals of the pixels of the matrix 200 which are coupled to the Vx lines when these pixels are selected. In other words, circuit READOUT is configured to received output signals of the pixels having their outputs 106 coupled to corresponding lines Vx thank to their selection devices 108 (
The sensor 2 comprises a control circuit CTRL1. The control circuit CTRL1 is configured to control reading phases and integration phases for the pixels of the matrix 200.
To provide control signals TG1 and TG2 to the transfer devices TGmem1 and TGmem2 of each pixel 1 (
In the embodiment of
To provide control signal RD to the selection device 108 of each pixel 1 (
In this embodiment, the lines 206 are perpendicular to the lines Vx. Each line 206 is, for example, shared by all the pixels of a corresponding row of the matrix. In
Although not shown in
The sensor 2 comprises an illumination device 205. The illumination device 205 is configured to illuminate a scene to capture. The sensor 2 further comprises a control circuit CTRL2 configured to control the illumination device 205. For example, the control circuit CTRL2 provides a control signal cmd to the device 205. The signal cmd is, for example, a digital signal comprising several bits.
As indicated above, the scene to capture is divided into a plurality of areas, and it is here proposed to successively illuminate each area of the scene, by illuminating only one area at a time, being understood that, in practice, parts of the scene which are adjacent to the illuminated area may also receive some light. Said in other words, the device 205 and its control circuit CTRL2 are configured to successively illuminate each area of the scene. For example, the device 205 is configured to illuminate different areas of the scene to capture, the area which is illuminated by the device 205 being determined by the signal cmd.
Control circuits CTRL1 and CTRL2 are synchronized, for example by means of a synchronization circuit SYNC which couples circuits CTRL1 and CTRL2. Said in other words, circuit SYNC receives and/or sends synchronization signals to and/or from circuits CTRL1 and CTRL2.
In a similar manner to the scene, the matrix 200 is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of
Each area M1, M2, M3, M4 comprises adjacent lines of pixels 1, these lines of pixels being parallel to the conductive lines 204. In the embodiment of
The matrix 200 and the device 205 are disposed relative to each other such that each area M1, M2, M3, M4 of the matrix 200 corresponds to an area of the scene, taken among the areas the scene is divided into and which are successively illuminated. Said in other words, the matrix 200 and the device 205 are disposed relative to each other such that, each time an area of the scene, taken among the plurality of areas the scene is divided into, is illuminated by the device 205, the light reflected by this area of the scene is received by the pixels 1 of the corresponding area M1, M2, M3 or M4 of the matrix 200, being understood that, in practice, some other pixels of the matrix, which are disposed near this corresponding area M1, M2, M3 of M4, may also receive part of the light reflected by the scene. The implementation of this disposition of the matrix 200 and the device 205 relative to each other is in the abilities of those skilled in the art.
The sensor 2 allows a scanned illumination of the scene to capture. For a given power supply provided to device 205 during an illumination of an area of the scene, all the light generated by the device 205 is directed towards this area of the scene. This differs from usual indirect time of flight sensors in which this given power supply is used to provide a flash illumination of the whole scene to capture. As a result, the signal-to-noise ratio of the light received by the sensor 2 is increased compared to that of the light received by these usual sensors. Indeed, for a given power supply, with a flash illumination, the light received by each area of scene carries less optical power than the light received by the only area of the scene which is illuminated by the sensor 2 during a scanned illumination.
The control circuit CTRL1 is further configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200. Said in other words, the control circuit CTRL1 is configured to control the charge transfers independently in each area M1, M2, M3, M4 of the matrix 200, or, said differently, independently between the areas M1, M2, M3 and M4. For example, control circuit CTRL1 comprises a different sub-circuit (not shown on
For example, the control circuit CTRL1 is configured to control an integration phase for the pixels of any one of the areas M1, M2, M3 and M4, while the control circuit CTRL1 controls no integration phase for the pixels of the other areas. More particularly, when an area of the scene is illuminated by the device 205, and the light reflected by this area of the scene is received by the corresponding area M1, M2, M3 or M4 of the matrix 200, control signals TG1, TG2 are maintained, by control circuit CTRL1, at the inactive state for the other areas of the matrix 200. The control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area M1, M2, M3 or M4 which is receiving light. Said in other words, control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area of the matrix 200 corresponding to the area of the scene which is illuminated, such that in each pixel of said area of the matrix 200, charges are alternatively transferred, from the region PD, to each storage regions mem1, mem2 of the pixel.
In practice, each commutation of the signal TG1, respectively TG2, corresponds to a charge or a discharge of a capacitance, typically the gate capacitance of the charge transfer device TGmem1, respectively TGmem2. Thus, by reducing the number of pixels for which signals TG1 and TG2 simultaneously commute, a power consumption of the sensor 2 is reduced compared to that of a usual indirect time of flight sensor, in which signals TG1, respectively TG2, commute simultaneously in all the pixels of the sensor.
The illumination device 205 comprises an array 300 of laser sources 301, only one laser source being referenced in
Device 205 further comprises an optical device (or element) 302, represented in the form of a block in
In this embodiment, the array 300 is divided into a plurality of sets of laser sources. In the example of
Each set A1, A2, A3, A4 is configured to illuminate a corresponding area of the scene to capture. Indeed, the laser sources 301 of the array can be each controlled independently from the other laser sources 301. For example, the array 300 is controlled such that, when laser sources 301 of a given sets A1, A2, A3 or A4 of the array 300 is emitting light, the laser sources 301 of the other sets are emitting no light. For example, the laser sources 301 which are emitting light and those which are emitting no light are determined by the signal cmd.
The control circuit CTRL2 (
For each set A1, A2, A3, A4, when the laser sources 301 of the set are emitting light, the emitted light is directed towards a corresponding area of the scene to capture by the device 302, the illuminated area of the scene being different for each set A1, A2, A3, A4 of the array 300 of laser sources 301.
For example, in
For example, the device 205 comprises a control circuit CTRL3 configured to control the emission of light by each light source 301 of the array 300 based on signal cmd.
In the device 205 of
The device 205 of
However, in the embodiment of
Further, in the embodiment of
For example, in
The device 302, for example, comprises mirror(s) and/or one or several lenses, the orientation of which being controllable by the signal cmd. Preferably, the optical device 302 comprises at least one controllably movable micro-mirror, or, in other words, a controllably movable MicroElectroMechanical System (MEMS) micro-mirror. The implementation of the optical device 302 is in the abilities of those skilled in the art.
In the device 205 of
The embodiments of
Between an instant t0 and an instant t1 posterior to instant t0, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, by commutating the control signals TG1, TG2 of the charge transfer devices TGmem1, TGmem2 of these pixels between their active and inactive states, at a frequency upper than that of the emitted light.
Between the instant t1 and an instant t2 posterior to instant t1, no light is emitted by the device 205 and the pixels of the area M1 are read. Because the lines 204 are parallel to the lines Vx (
Between the instant t2 and an instant t3 posterior to instant t2, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only.
Between the instant t3 and an instant t4 posterior to instant t3, no light is emitted by the device 205 and the pixels of the area M2 are read, by reading all the pixels of the matrix (“all matrix”), similarly to what has been done between instants t1 and t2.
Between the instant t4 and an instant t5 posterior to instant t4, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by the area S3 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the area M3 only.
Between the instant t5 and an instant t6 posterior to instant t5, no light is emitted by the device 205 and the pixels of the area M3 are read, by reading all the pixels of the matrix (“all matrix”).
Between the instant t6 and an instant t7 posterior to instant t6, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area A4 of the matrix, and an integration phase is performed in the area M4 only.
At the instant t7, all the areas S1, S2, S3, S4 of the scene have been illuminated once during the scanned illumination of the scene.
Between the instant t7 and an instant t8 posterior to instant t7, no light is emitted by the device 205 and the pixels of the area M4 are read, by reading all the pixels of the matrix (“all matrix”).
At the instant t8, the output signals of the pixels of the area M1 read after the illumination of the area M1 (between instants t1 and t2), the output signals of the pixels of the area M2 read after the illumination of the area M2 (between instants t3 and t4), the output signals of the pixels of the area M3 read after the illumination of the area M3 (between instants t5 and t6), and the output signals of the pixels of the area M4 read after the illumination of the area M4 (between instants t7 and t8) may be used to generate, or compute, an image, or depth map, of scene.
At the instant t8, a new scanned illumination of the scene begins, by illuminating, with the device 205, the area M1 of the scene.
In the operating mode described in relation with
Preferably, when capturing a scene, during the successive illuminations of the areas of the scene, the device 205 is supplied with an average power supply having a given peak power, which is equal to an average power, having the same peak power, provided to an illumination device of a usual sensor during a flash illumination of the scene. In this case, the duration T of the illumination phase of each area of the scene during a scanned illumination is preferably equal to the duration of the flash illumination divided by the number of areas of the scene. This allows to further increase the signal-to-noise ratio in the sensor 2, compared to a usual sensor, without modifying the power supply used to illuminate the scene to capture.
Between an instant t10 and an instant t11 posterior to instant t10, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, by commutating the control signals TG1, TG2 of the charge transfer devices TGmem1, TGmem2 of these pixels between their active and inactive states, at a frequency upper than that of the emitted light.
Between the instant t11 and an instant t12 posterior to instant t11, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only.
Between the instant t12 and an instant t13 posterior to instant t12, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by the area S3 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the pixels of the area M3 only.
Between the instant t13 and an instant t14 posterior to instant t13, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the pixels of the area M4 only.
As illustrated in
At the instant t17, the control circuit CTRL1 (
In the operation described in relation with
Compared to the operation described in relation with
Preferably, in
In the embodiments described in relation with
The sensor 2′ comprises, as the sensor 2 (
Instead of the control circuit CTRL1, the sensor 2′ comprises a control circuit CTRL1′. The control circuit CTRL1′ is configured to control reading phases and integration phases for the pixels of the matrix 200. The control circuit CTRL1′ is configured to provide the control signals TG1 and TG2 (
In the embodiment of
Although not shown on
Control circuits CTRL1′ and CTRL2 are synchronized, for example by means of a synchronization circuit SYNC which couples circuits CTRL1′ and CTRL2. Said in other words, circuit SYNC receives and/or sends synchronization signals to and/or from circuits CTRL1′ and CTRL2.
As with the sensor 2, the matrix 200 of sensor 2′ is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of
Each area M1, M2, M3, M4 comprises adjacent lines of pixels 1, these lines of pixels being parallel to the conductive lines 204. In the embodiment of
As already described for the sensor 2, in the sensor 2′ the matrix 200 and the device 205 are disposed relative to each other such that each area M1, M2, M3, M4 of the matrix 200 corresponds to an area of the scene.
The sensor 2′ allows, as with the sensor 2 of
The control circuit CTRL1′ is configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200. Said in other words, the control circuit CTRL1′ is configured to control the charge transfers independently in each area M1, M2, M3, M4 of the matrix 200. For example, control circuit CTRL1′ comprises a different sub-circuit (not shown on
For example, the control circuit CTRL1′ is configured to control an integration phase for the pixels of any one of the areas M1, M2, M3 and M4, while the control circuit CTRL1′ controls no integration phase for the pixels of the other areas. More particularly, when an area of the scene is illuminated by the device 205, and the light reflected by this area of the scene is received by the corresponding area M1, M2, M3 or M4 of the matrix 200, control signals TG1, TG2 are maintained, by control circuit CTRL1′, at the inactive state for the other areas of the matrix 200. The control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area M1, M2, M3 or M4 which is receiving light. Said in other words, control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area of the matrix 200 corresponding to the area of the scene which is illuminated, such that in each pixel of said area of the matrix 200, charges are alternatively transferred, from the region PD, to each storage regions mem1, mem2 of the pixel. As a result, a power consumption of the sensor 2′ is reduced compared to that of a usual indirect time of flight sensor.
An advantage of the sensor 2′ compared to the sensor 2 is that the pixels of a given area M1, M2, M3 or M4 of the matrix 200 of sensor 2′ may be read without performing a full reading of the matrix 200, by reading ones after the other only the rows of this area.
The chronograms of
More specifically, in the
In the sensor 2′, the duration of the reading of the pixels of a given area of the matrix is reduced compared to that of the sensor 2, because it is not needed anymore to read the all the pixels of the matrix to read the pixels of a given area of the matrix.
In an alternative embodiment, the sensor 2′ operates as described in relation with
To take profit of the fact that lines 204 are perpendicular to lines Vx, it is here proposed to read the pixels of an area M1, M2, M3 or M4 of the matrix 200 whereas another area of the matrix 200 is receiving light. However, when pixels of a given area M1, M2, M3 or M4 receiving light are in an integration phase and when pixels of another area are simultaneously in a reading phase, it has been shown that the high frequency commutations of the signals transmitted using lines 204 to the pixels of in the integration phase generate noise in the output signals of the pixels in the reading phase, the output signals being available on the Vx lines. This noise is, for example, transmitted via the reference voltage GND which is provided to the different circuits and to all the pixels of the sensor, and/or by the cross coupling between lines Vx and lines 204.
To suppress this noise, it is here proposed a split ground and bias strategy to minimize unwanted coupling. More specifically, the pixels matrix is split into two insulated halves. Further, separated, or electrically decoupled, supply voltage, reference voltage, bias voltages and control signals are provided to each matrix half. It is then possible to read pixels of one half of the matrix while pixels of the other half are integrating, without generating noise. Different embodiments of indirect time of flight sensors implementing this strategy will be now described.
In sensor 2″, the matrix 200 is split into two halves P1 and P2. More specifically, a separation between parts P1 and P2 of the matrix 200 is parallel to the lines 204.
The parts P1 and P2 of the matrix 200 are adjacent, the part P1 being disposed along the part P2. More specifically, each column comprises a first portion, or half, belonging to part P1, and a second portion, or half, belonging to part P2 and being aligned with the first portion of the column. For example, the parts P1 and P2 have a common edge, which corresponds to the separation between parts P1 and P2.
Further, the lines Vx, which are parallel to the column of the matrix and perpendicular to lines 204, are interrupted at the separation between parts P1 and P2 of the matrix 200. Said in other words, the lines Vx of the part P1 of the matrix 200 and the lines Vx of the part P2 of the matrix end at the separation between parts P1 and P2 of the matrix 200. Said differently, the lines Vx of part P1 of the matrix are insulated from the lines Vx of part P2 of the matrix, and the lines Vx of part P1, respectively P2, do not extend above or below the part P2, respectively P1. In
A line Vx of the part P2 corresponds to a line Vx of the part P1 when these two lines Vx belong to the same column of the matrix 200. For example, in each column of the matrix 200, a line Vx of the part P2 corresponds to a line Vx of the part P1 when the line Vx of the part P1 is selectively coupled to given outputs of the pixels of the part P1 disposed in this column, and the line Vx is selectively coupled to the corresponding outputs of the pixels of the part P2 disposed in this column.
The part P1 of the matrix 200 is electrically decoupled from the part P2 of the matrix 200. More specifically, a semiconductor substrate to which the pixels 1 of the matrix 200 belong has a first part which comprises the part P1 of the matrix 200 and a second part which comprises the part P2 of the matrix 200. Said in other words, the first part of the substrate comprises the half P1 of the matrix and a second part of the substrate comprises the half P2 of the matrix.
The first and second parts of the substrate are insulated from each other using insulation structures passing through the substrate, the insulation structures being preferably insulation structures provided between pixels to insulate the pixels from each other.
In the example of
In this example, the region PD of each pixel 1 is laterally delimited by a capacitive deep trench insulation 1005, for example a U-shaped insulation structure 1005 in the view of
In this example, each pixel 1 further comprises transfer devices TGmem1 and TGmem2, the coupling devices TGRD1 and TGRD2, the transistor 112 and the selection device 108, the transistors 112 and 108 being shared by the two adjacent pixels.
The example shown in
Further, although in the example of
Referring back to the
The reference voltage GND which is provided to the first part of the substrate and the reference voltage GND which is provided to the second part of the substrate are electrically decoupled from each other. For example, the reference voltage GND provided to the first part of the substrate, or, in other words, to each pixel of the part P1 of the matrix, is provided by a first bonding pad 900 of the sensor 2″, and the other reference voltage GND provided to the second part of the substrate, or, in other words, to each pixel of the part P2 of the matrix, is provided by a second bonding pad 902 of the sensor 2″. Each bonding pad 900, 902 receives an off-chip reference voltage GND. Each bonding pad 900, 902 acts as a low-pass filter, as it is schematically represented in
Preferably, the insulation structures 1000 are CDTI. In this case, it is preferable to provide a bias voltage to structure 1000 delimiting the part P1 of the matrix 200, which is electrically decoupled from a bias voltage provided to structure 1000 delimiting the part P2 of the matrix 200. For example, in
Instead of the control circuit CTRL1′, the sensor 2″ comprises a control circuit CTRL1″. The control circuit CTRL1″ is configured to control reading phases and integration phases for the pixels of the matrix 200. The control circuit CTRL1″ is configured to provide the control signals TG1 and TG2 (
In the embodiment of
Although not shown on
Control circuits CTRL1″ and CTRL2 (not shown in
As for sensor 2′, the matrix 200 of sensor 2″ is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of
The control circuit CTRL1″ is configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200, in a way similar to that described for the control circuit CTRL1′ (
Preferably, for each voltage level which is provided to at least one pixel 1 of the part P1 of the matrix 200, and simultaneously to at least one pixel 1 of the other part P2 of the matrix, the sensor 2″ comprises a voltage generator configured to provide this voltage level to the part P1 of the matrix, and a voltage generator configured to provide this voltage level to the other part P2 of matrix. These two generators are electrically decoupled form each other.
In
Although the provision of two generators which are electrically decoupled from each other and configured to provide simultaneously the same voltage level to both parts P1 and P2 of the matrix 200 is here illustrated only for the voltage level TGmemL, those skilled in the art are capable to implement other pairs of electrically decoupled voltage generator for generating any voltage level which is provided simultaneously to both parts P1 and P2 of the matrix.
According to one embodiment, which is illustrated by
Circuit READOUT1, respectively READOUT2, is configured to received output signals of the pixels of the part P1, respectively P2, of matrix 200 which are coupled to the Vx lines of part P1, respectively P2, when these pixels are selected. Each reading circuit READOUT1 and READOUT2 for example comprises a plurality of analog-to-digital converters (ADC), preferably one ADC for each Vx line coupled to this reading circuit.
The circuit READOUT1 receives a reference voltage, in this example the ground GND, and the circuit READOUT2 receives a reference voltage, in this example the ground GND. The reference voltage GND of the circuit READOUT1 is electrically decoupled from that of the circuit READOUT2. For example, the reference voltage GND applied to the circuit READOUT1 is provided by a third bonding pad 912 of the sensor 2″, and the other reference voltage GND applied to the circuit READOUT2 is provided by a fourth bonding pad 914 of the sensor 2″. Each bonding pad 912, 914 receives the off-chip reference voltage GND. Each bonding pad 912, 914 acts as a low-pass filter as schematically represented in
Between an instant t20 and an instant t21 posterior to instant t20, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of part P1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, thus only in part P1 of the matrix.
Between the instant t21 and an instant t22 posterior to instant t21, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by this area S3 is received by the corresponding area M3 of part P2 of the matrix. An integration phase of the received light is done in the pixels of the area M3 only, thus only in part P2 of the matrix. In the same time, the area M1 of the part P1 of the matrix is read. More specifically, the reading of the pixels of the area M1 is controlled by control circuit CTRL1″ and is completed by reading the rows of pixels of the area M1 ones after the other.
Between the instant t22 and an instant t23 posterior to instant t22, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only, thus only in part P1 of the matrix. In the same time, the area M3 of the part P2 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22.
Between the instant t23 and an instant t24 posterior to instant t23, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area M4 of the matrix, and an integration phase is performed in the pixels of the area M4 only, thus only in part P2 of the matrix. In the same time, the area M2 of the part P1 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22.
Between the instant t24 and an instant t25 posterior to instant t24, the area M4 of part P2 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22. At the instant t25, a depth map of the scene may be computed. More specifically, the depth map is generated based on the output signals of the pixels of the area M1 read between the instants t21 and t22, of the area M2 read between the instants t22 and t23, of the area M3 read between the instants t23 and t24, and of the area M4 read between the instants t24 and t25.
As it is represented in
In this embodiment, circuit READOUT1 is disposed along a first edge of the matrix 200, on the side of the half P1 of the matrix, circuit READOUT2 being disposed along a second edge of the matrix, on the side of the half P2. The first and second edges are parallel. More specifically, the first and second edges are perpendicular to the lines Vx (not shown on
This disposition of the circuits READOUT1 and READOUT2 relative to the matrix 200 is, for example, used when the circuits READOUT1 and READOUT2 belongs to the same semiconductor substrate than the matrix 200.
In the embodiment of
The lines Vx of the part P1 of the matrix 200 are coupled to the circuit READOUT1, for example thanks to an interconnection structure (not shown) which is sandwiched between the first and second substrates. Similarly, the lines Vx of the part P2 of the matrix 200 are coupled to the circuit READOUT2, for example thanks to same interconnection structure. In
Preferably, as shown in
The embodiment of
Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuits READOUT1 and READOUT2 in order to generate a depth map of a scene.
In this alternative embodiment, a first semiconductor substrate comprises the matrix 200, and lies on a second semiconductor substrate. In other words, the two substrates are stacked one over the other.
The sensor 2″ further comprises commutators 1500, only one of the commutators 1500 being referenced in
Each commutator 1500 comprises a first input 1501, a second input 1502, an output 1503 and is controlled by a signal Sel. Each commutator 1500 is configured to electrically couple its input 1501 to its output 1503 when signal Sel is in a first state, and to couple its input 1502 to its output 1503 when signal Sel is in a second state.
In the alternative embodiment illustrated by
In this alternative embodiment, instead of the two circuits READOUT1 and READOUT2, the sensor 2″ comprises only one reading circuit READOUT3. Preferably, the circuit READOUT3 belongs to the same substrate as the commutators 1500. Although in
Each commutator 1500 has its outputs 1503 coupled, preferably connected, to the circuit READOUT3. The circuit READOUT3, for example, comprises an ADC for each commutator 1500.
A control circuit, for example the control circuit CTRL1″, is configured to control the commutators 1500 such that the output 1503 of each commutator is coupled to the input 1501 of this commutator during a reading of pixels of the half P1 of the matrix, and to the input 1502 of this commutator during a reading of pixels of the half P2 of the matrix. Said in other words, the circuit for controlling the commutators, in this example the control circuit CTRL1″, is configured to provide the signal Sel at its first state during a reading of pixels of the half P1 of the matrix, and at its second state during a reading of pixels of the half P2 of the matrix.
In sensor 2″ of the
Compared to the sensor 2″ described in relation with
As already indicated in relation with
The lines Vx of the parts P1 and P2 of the matrix 200 are, for example, conductive lines of an interconnection structure which is sandwiched between the first and second substrates, only one line Vx of the part P1 and one corresponding line Vx of the part P2 being represented in
The commutators 1500 are disposed below the separation between the parts P1 and P2 of matrix 200, or, said in other words, below the common edge of the parts P1 and P2 of the matrix 200.
In this particular embodiment, the circuit READOUT3 belongs to the same substrate as the commutators 1500. The circuit READOUT3 is preferably disposed below the matrix 200, for example below the part P2 of the matrix as represented in
Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuit READOUT3 in order to generate a depth map of a scene.
The embodiments described in relation with
In this alternative embodiment, the sensor 2″ comprises two reading circuits READOUT4 and READOUT5 instead of the reading circuit READOUT3. Preferably, the circuits READOUT4 and READOUT5 belong to the same substrate as the commutators 1500. Although in
As in
However, in the embodiment of
As already indicated in relation with
In the sensor 2″ of
Preferably, the commutators 1500 are disposed below the separation between the parts P1 and P2 of the matrix 200. Preferably, the circuit READOUT4 is disposed below one of the parts P1 and P2 of the matrix 200, the circuit READOUT5 being disposed below the other one of the parts P1 and P2.
As already indicated in relation with
The lines Vx of the parts P1 and P2 of the matrix 200 are, for example, conductive lines of an interconnection structure which is sandwiched between the first and second substrates, only one line Vx of the part P1 and one corresponding line Vx of the part P2 being represented in
The commutator 1500 are disposed below the separation between the parts P1 and P2 of matrix 200, or, said in other words, below the common edge of the parts P1 and P2 of the matrix 200.
In this particular embodiment, circuits READOUT4 and REDAOUT5 belong to the same substrate as the commutators 1500. The circuit READOUT4 is disposed below one of the parts P1 and P2 of the matrix 200, the circuit READOUT5 being disposed below the other one of the parts P1 and P2. In the example of
Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuits READOUT4 and READOUT5 in order to generate a depth map of a scene.
The embodiments described in relation with
Various embodiments and variants have been described. Those skilled in the art will understand that certain features of these embodiments can be combined, and other variants will readily occur to those skilled in the art. In particular, although in the above described embodiments, the scene to capture divided into only four areas S1, S2, S3 and S4, the illumination device 205 is configured to direct the light towards each of the areas S1, S2, S3, S4 of the scene by illuminating only one area at a time, and the matrix 200 is divided into four corresponding areas M1, M2, M3 and M4, those skilled in the art are capable to implement embodiment wherein the scene is divided into more (or less) than four areas, the device 205 is configured to independently illuminate each of these areas of the scene, and the matrix 200 is divided into areas such that each area of the matrix corresponds to an area of the scene. Further, those skilled in the art are capable of implementing embodiments in which the pixels of the matrix 200 are different from pixel 1 described in relation with
Finally, the practical implementation of the embodiments and variants described herein is within the capabilities of those skilled in the art based on the functional description provided hereinabove.
Number | Date | Country | Kind |
---|---|---|---|
20306680.8 | Dec 2020 | EP | regional |