INDIRECT TIME OF FLIGHT SENSOR

Information

  • Patent Application
  • 20220196835
  • Publication Number
    20220196835
  • Date Filed
    December 21, 2021
    2 years ago
  • Date Published
    June 23, 2022
    2 years ago
Abstract
An indirect time of flight sensor includes a matrix of pixels, wherein each pixel includes at least two controllable transfer devices. First conductive lines transmit first control signals to the transfer devices, these first signals being provided by a first circuit. A device is provided for illuminating a scene that is divided into at least two first areas. The device successively illuminates each first area. The matrix is similarly divided into at least two second areas. The matrix and illumination device are disposed such that each first area corresponds to one second area. The first circuit provides different first signals to the different second areas.
Description
PRIORITY CLAIM

This application claims the priority benefit of European Application for Patent No. 20306680.8, filed on Dec. 23, 2020, the content of which is hereby incorporated by reference in its entirety to the maximum extent allowable by law.


TECHNICAL FIELD

The present disclosure relates generally to image sensors and, more particularly, to time of flight sensors.


BACKGROUND

Image sensors of the time of flight type are known. Among these sensors, indirect time of flight sensors are configured to determine a dephasing between periodic light emitted by the sensor towards a scene to capture, and light received by pixels of the sensor, the received light corresponding to the light reflected by the scene when illuminated by the sensor. Based on the dephasing determined for each pixel of the sensor, a distance between this pixel and a conjugated point of the scene may be calculated. From the determined distance for each pixel, a depth map of the scene may be generated.


There is a need to overcome all or some of the drawbacks of known indirect time of flight sensors.


SUMMARY

Embodiments herein address all or some of the drawbacks of known indirect time of flight sensors.


One embodiment provides an indirect time of flight sensor comprising: a matrix of pixels wherein each pixel comprises a photoconversion region and at least two sets each comprising a charge storage region and a controllable transfer device for transferring charge from the photoconversion region towards said storage region; first conductive lines parallel to each other, configured to transmit first control signals to the transfer devices; a first circuit configured to provide the first signals to the first conductive lines; an illumination device for illuminating a scene to capture; and a second circuit configured to control the illumination device. The scene is divided into first areas and the illumination device and the second circuit are configured to successively illuminate each first area. The matrix is divided into second areas each comprising adjacent lines of pixels, parallel to the first conductive lines, wherein a disposition of the matrix and of the illumination device is configured such that each first area corresponds to one of the second areas. The first circuit is configured to provide different first signals to the different second areas.


According to one embodiment, the illumination device comprises an array of laser sources and an optical device configured to direct light emitted by the array of laser sources towards the scene. The array is divided into sets of laser sources, each set being configured to illuminate a corresponding first area, the second circuit being configured to control said sets one after the other. The optical device is configured to direct the emitted light differently depending on a control signal, the second circuit being configured to provide, at each illumination of a first area, said control signal corresponding to a directing of the light towards said first area.


According to one embodiment: the sensor comprises second conductive lines parallel to the first conductive lines and configured to receive output signals of the pixels; each pixel comprises a selection device configured to selectively couple output(s) of said pixel to at least one corresponding second conductive line; and the first circuit is configured to provide second control signals to the selection devices via third conductive lines perpendicular to the second conductive lines.


According to one embodiment, the first circuit is configured to control, by means of the second signals, a reading of all the pixels after each illumination of a first area, before an illumination of a next first area.


According to one embodiment: the sensor comprises second conductive lines parallel to each other and perpendicular to the first conductive lines, the second conductive lines being configured to receive output signals of the pixels. Each pixel comprises a selection device configured to selectively couple output(s) of said pixel to at least one corresponding second conductive line; and the first circuit is configured to provide second control signals to the selection devices via third conductive lines perpendicular the second conductive lines.


According to one embodiment, the second circuit is configured, before each reading of all the pixels controlled by the first circuit, to control several successive illumination cycles each comprising a unique illumination of each first area, and to control an absence of light emission by the illumination device during said reading.


According to one embodiment, the first circuit is configured to control, after each illumination of a first area, a reading of only the pixels of the second area corresponding to said first area.


According to one embodiment, the second circuit is configured to control an absence of light emission by the illumination device when the first circuit controls the reading of the pixels of a second area.


According to one embodiment: the matrix is divided into first and second halves, a separation between first and second halves being parallel to the first lines, and the second conductive lines of each half ending at said separation. The first circuit is configured to simultaneously control charge transfers in the pixels of a second area of one of the halves and a reading of the pixels of a second area of the other one of the halves. A first part of a semiconductor substrate comprises the first half of the matrix and a second part of said semiconductor substrate comprises the second half of the matrix; insulation structures passing through the semiconductor substrate insulate said parts of the semiconductor substrate from each other. A reference voltage provided to the first part of the semiconductor substrate is electrically decoupled from a reference voltage provided to the second part of the semiconductor substrate.


According to one embodiment, for each voltage level intended to be provided to at least one pixel of the first half of the matrix, and, simultaneously, to at least one pixel of the second half of the matrix, the sensor comprises a generator of said voltage level for the first half and a generator of said voltage level for the second half, the generators being electrically decoupled from each other.


According to one embodiment, the sensor comprises a first reading circuit coupled the second conductive lines of the first half of the matrix, and a second reading circuit coupled to the second conductive lines of the second half of the matrix, a reference voltage of the first reading circuit being electrically decoupled from a reference voltage of the second reading circuit.


According to one embodiment, the first reading circuit is disposed along a first edge of the matrix, on the side of the first half, and the second reading circuit is disposed along a second edge of the matrix, on the side of the second half, the first and second edges being parallel.


According to one embodiment: the semiconductor substrate comprising the matrix of pixels lies on another semiconductor substrate comprising commutators, the commutators being preferably disposed below the separation between the halves of the matrix; each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs; and the sensor comprises a reading circuit connected to the output of each commutator, the reading circuit preferably belonging to the same semiconductor substrate as the commutators.


According to one embodiment: the semiconductor substrate comprising the matrix of pixels lies on another semiconductor substrate comprising commutators, the commutators being preferably disposed below the separation between the halves of the matrix; each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs; the pixels of the matrix are arranged in column parallel to the second conductive lines; each commutator connected to second conductive lines of an odd column has its output connected to a first reading circuit; each commutator connected to second conductive lines of an even column has its output connected to a second reading circuit; and the first and second reading circuits preferably belonging to the same semiconductor substrate as the commutators.


According to one embodiment, the sensor comprises a control circuit for controlling the commutators such that the output of each commutator is coupled to the first input of said commutator during a reading of pixels of the first half of the matrix, and to the second input of said commutator during a reading of pixels of the second half of the matrix.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing features and advantages, as well as others, will be described in detail in the following description of specific embodiments given by way of illustration and not limitation with reference to the accompanying drawings, in which:



FIG. 1 illustrates an example of a circuit of a pixel of an indirect time of flight sensor;



FIG. 2 illustrates an indirect time of flight sensor according to one embodiment;



FIG. 3 illustrates an illumination device of an indirect time of flight sensor according to one embodiment;



FIG. 4 illustrates an illumination device of an indirect time of flight sensor according to one alternative embodiment;



FIG. 5 shows chronograms illustrating operation of the sensor of FIG. 2 according to one embodiment;



FIG. 6 shows chronograms illustrating operation of the sensor of FIG. 2 according to one alternative embodiment;



FIG. 7 illustrates an indirect time of flight sensor according to a further embodiment;



FIG. 8 shows chronograms illustrating operation of the sensor of FIG. 7 according to one embodiment;



FIG. 9 illustrates an indirect time of flight sensor according to a further embodiment;



FIG. 10 shows a very schematic top view of two adjacent pixels of the sensor of FIG. 9;



FIG. 11 shows a very schematic cross section view along plan AA of FIG. 10;



FIG. 12 shows chronograms illustrating operation of the sensor of FIG. 7 according to one embodiment;



FIG. 13 illustrates, in a very schematic manner, an implementation of the sensor of FIG. 9;



FIG. 14 illustrates, in a very schematic manner, another implementation of the sensor of FIG. 9;



FIG. 15 illustrates an alternative embodiment of the indirect time of flight sensor of FIG. 9;



FIG. 16 illustrates, in a very schematic manner, an implementation of the sensor of FIG. 15;



FIG. 17 illustrates another alternative embodiment of the indirect time of flight sensor of FIG. 9; and



FIG. 18 illustrates, in a very schematic manner, an implementation of the sensor of FIG. 17.





DETAILED DESCRIPTION

Like features have been designated by like references in the various figures. In particular, the structural and/or functional features that are common among the various embodiments may have the same references and may dispose identical structural, dimensional and material properties.


For the sake of clarity, only the operations and elements that are useful for an understanding of the embodiments described herein have been illustrated and described in detail. In particular, usual electronic systems and applications in which an indirect time of flight sensor may be provided are not described in detail, the described embodiments being compatible with these usual systems and applications.


Unless indicated otherwise, when reference is made to two elements connected together, this signifies a direct connection without any intermediate elements other than conductors, and when reference is made to two elements coupled together, this signifies that these two elements can be connected or they can be coupled via one or more other elements.


In the following disclosure, unless indicated otherwise, when reference is made to absolute positional qualifiers, such as the terms “front”, “back”, “top”, “bottom”, “left”, “right”, etc., or to relative positional qualifiers, such as the terms “above”, “below”, “higher”, “lower”, etc., or to qualifiers of orientation, such as “horizontal”, “vertical”, etc., reference is made to the orientation shown in the figures.


Unless specified otherwise, the expressions “around”, “approximately”, “substantially” and “in the order of” signify within 10%, and preferably within 5%.



FIG. 1 illustrates an example of a circuit of a pixel 1 of an indirect time of flight sensor.


Pixel 1 comprises a photoconversion region, or photosensitive region PD, for example a photodiode, preferably a pinned photodiode. The photoconversion region PD has an electrode, for example its anode, which is connected to a node 100 configured to receive a reference voltage, for example the ground GND. The photoconversion region PD is configured such that charges are generated therein when light is received by the region PD.


Pixel 1 further comprises two identical memory circuit sets E1 and E2, delimited by dashed lines in FIG. 1. Each set E1, E2 is coupled to the region PD, and more particularly to the electrode 102 of the region PD which is not connected to the node 100.


Each set E1, E2 of the pixel 1 comprises a charge storage region mem1, mem2 and a controllable charge transfer device TGmem1, TGmem2.


Device TGmem1, respectively TGmem2, is connected between the region PD and the region mem1, respectively mem2. Device TGmem1, respectively TGmem2, is configured to transfer charges from the region PD to the region mem1, respectively mem2. More precisely, device TGmem1, respectively TGmem2, is configured to transfer charges from the region PD to the region mem1, respectively mem2, when its control signal TG1, respectively TG2, is active, for example at a high level, and to block any charge transfer between the region PD and the region mem1, respectively mem2, when this control signal is inactive, for example at a low level. Each device TGmem1, TGmem2 is, for example, a transfer gate transistor.


Region mem1, respectively mem2, is configured to store charges which are transferred therein by the transfer device TGmem1, respectively TGmem2, until these charges are transferred elsewhere in the pixel 1 during a reading phase. Each region mem1, mem2 is, for example, a pinned diode. Each pinned diode mem1, mem2 has an electrode, for example its anode, connected to the node 100, and another electrode 104, for example its cathode, coupled to the electrode 102 of the region PD by the corresponding transfer device TGmem1, TGmem2.


Pixel 1 has an output 106. During a reading phase of the pixel 1, output signals of the pixel 1 are available on the output 106.


Pixel 1 comprises a selection device 108, for example a Metal Oxide Semiconductor (MOS) transistor. The device 108 is connected between the output 106 and a reading conductive line Vx. The selection device 108 is configured to selectively couple the output 106 of the pixel 1 to the line Vx. More precisely, during a reading phase of the pixel 1, for example when a control signal RD of the device 108 is active, for example at a high level, the device 108 couples the output 106 to line Vx, and outside of a reading phase of the pixel 1, for example when signal RD is inactive, for example at a low level, the device 108 isolates output 106 from line Vx.


For example, in known time of flight sensors comprising a matrix of pixels 1 arranged in rows and columns, a line Vx is shared by all the pixels 1 which belong to the same column. To read a given pixel of the matrix, all the pixels of the row to which belongs this pixel are selected by activating signal RD for this row of pixels.


Pixel 1 comprises a controllable output circuit 110, delimited in dashed lines in FIG. 1. The circuit 108 is configured to selectively generate, on the output 106, an output signal indicative of the number of charges stored in the charge storage region mem1 of the pixel or an output signal indicative of the number of charges stored in the charge storage region mem2 of the pixel.


For example, during a reading phase of the pixel, when a first signal RD1 is active, for example at a high level, the circuit 110 provides a signal, for example a voltage referenced to node 100, indicative of the number of charges stored in region mem1, and, when a second signal RD2 is active, for example at a high level, the circuit 110 provides a signal, for example a voltage referenced to node 100, indicative of the number of charges stored in region mem2.


In the particular example of FIG. 1, the circuit 100 comprises, for each set E1, E2, a controllable coupling device TGRD1, TGRD2, for example a transfer gate. Device TGRD1, respectively TGRD2, is connected to the set E1, respectively E2, and, more precisely, to region mem1, respectively mem2, for example to the electrode 104 of the region mem1, respectively mem2. The device TGRD1, respectively TGRD2, is configured to couple the region mem1, respectively mem2, to a node 111 when the signal RD1, respectively RD2, is active, and to insulate the region mem1, respectively mem2, from node 111 when the signal RD1, respectively RD2, is inactive. Circuit 110 further comprises a source follower MOS transistor 112 having its gate connected to node 111, its source connected to output 106 and its drain connected to a node 114 configured to receive a supply voltage Vdd.


The pixel 1, for example, further comprises a transistor AB connected between the electrode 102 of the region PD and a node 118 configured to receive a bias voltage VAB. The transistor AB is controlled by a signal TGAB. The transistor AB is configured, when off, to operate as an antiblooming device for the region PD, and, when on, to reset the region PD, that is to say to evacuate all the photo-generated charges accumulated in the region PD towards the node 118.


In a usual indirect time of flight sensor comprising a matrix of pixels 1 arranged in rows and columns, during an integration phase, all the transfer devices TGmem1 and TGmem2 of all the pixels 1 of the matrix are driven simultaneously to transfer charges photo-generated in the region PD of each pixel towards regions alternatively mem1 and mem2 of this pixel. Further, during the integration phase, the scene to capture is illuminated by the sensor in a flash manner, that is to say that each time the sensor emits light, the whole scene is illuminated. During the integration phase, the light is, for example, emitted under the form of a burst of successive periodic pulses of light. After an integration phase, all the pixels 1 of the matrix are read. More particularly, during the reading of all the pixels 1 of the matrix, the rows of pixels are selected the one after the other with the signals RD, and all the pixels 1 of a selected row are read simultaneously.


Although in the example of FIG. 1 the pixel comprises only two identical sets E1 and E2, in other examples not illustrated, the pixel may comprise more than two identical sets, for example four identical sets.


Further, although in the example of FIG. 1 the pixel 1 has only one output 106, in other examples not illustrated, the pixel may comprise more than one output 106. For example, the pixel may comprise one output 106 for each set E1, E2, the circuit 110 being then connected between the sets E1, E2 and the outputs 106. The selection device 108 is then configured to selectively couple the outputs 106 to at least one corresponding line Vx. For example, the output 106 associated to the set E1 is selectively coupled to a first line Vx by the device 108, and the output 106 associated to the set E2 is selectively coupled to a second line Vx by the device 108.


More generally, many different pixels known by those skilled in the art may be used in a matrix of pixels of an indirect time of flight sensor, and the pixel 1 of FIG. 1 is only one example of these known pixels. Further, usual controls of these different pixels during an integration phase and during a reading phase are well known by those skilled in the art.


In the following description, unless indicated otherwise, when reference is made to a pixel of an indirect time of flight sensor, this means a reference to the pixel 1 of FIG. 1. However, those skilled in the art will be capable of adapting the following description to other pixels, for example to pixels comprising more than two identical sets and/or more than one output 106.


It is here proposed to capture a scene with an indirect time of flight sensor by successively illuminating different areas of the scene, only one area of the scene being illuminated at a time. Said otherwise, the scene is divided into a plurality of area, and the scene is fully illuminated by successively illuminating each area of scene, each of these areas being illuminated at least once.



FIG. 2 illustrates an indirect time of flight sensor 2 according to one embodiment.


The sensor 2 comprises a matrix 200 of pixels 1, only one pixel 1 being referenced on FIG. 2 to avoid complicating the drawing. Pixels 1 are arranged in rows (horizontally on FIG. 2) and columns (vertically on FIG. 2). In the example of FIG. 2, the matrix 200 comprises 8 rows and 8 columns, although, in practice, the matrix 200 may comprise hundreds of rows and hundreds of columns.


The sensor 2 comprises a reading circuit READOUT. Circuit READOUT is configured to received output signals of the pixels of the matrix 200 which are coupled to the Vx lines when these pixels are selected. In other words, circuit READOUT is configured to received output signals of the pixels having their outputs 106 coupled to corresponding lines Vx thank to their selection devices 108 (FIG. 1). As it is usual in indirect time of flight sensors, in the sensor 2, the Vx lines are arranged parallel to the columns of the matrix 200, or, said in other words, the Vx lines are vertical on FIG. 2. Each Vx line is coupled, preferably connected, to the circuit READOUT. In order to avoid complicating the FIG. 2, only one Vx line is fully represented, in dashed lines, in this Figure. As it can be seen on FIG. 2, each Vx line is shared by several pixels, and, more particularly, by all the pixels of a corresponding column in the embodiment of FIG. 2. The reading circuit READOUT, for example, comprises a plurality of analog-to-digital converters (ADC), preferably one ADC for each Vx line.


The sensor 2 comprises a control circuit CTRL1. The control circuit CTRL1 is configured to control reading phases and integration phases for the pixels of the matrix 200.


To provide control signals TG1 and TG2 to the transfer devices TGmem1 and TGmem2 of each pixel 1 (FIG. 1), the sensor 2 comprises parallel conductive lines 204. Lines 204 are connected to control circuit CTRL1. The control circuit CTRL1 is configured to provide the control signals TG1 and TG2 (FIG. 1) to the lines 204.


In the embodiment of FIG. 2, the lines 204 are parallel to the lines Vx. Each line 204 is, for example, shared by all the pixels of a corresponding column of the matrix. In FIG. 2, only one line 204 is fully represented, in dashed lines, in order to avoid complicating the Figure. Further, in order to avoid complicating the Figure, only one line 204 by column is represented in FIG. 2. However, in practice, each pixel receives control signals TG1 and TG2 (FIG. 1) via two corresponding lines 204, and each column is thus associated to a line 204 for transmitting signal TG1 to all the pixels of the column, and to another line 204 for transmitting signal TG2 to all these pixels.


To provide control signal RD to the selection device 108 of each pixel 1 (FIG. 1), the sensor 2 further comprises parallel conductive lines 206. Lines 206 are connected to control circuit CTRL1. The control circuit CTRL1 is configured to provide the control signals RD to the lines 206.


In this embodiment, the lines 206 are perpendicular to the lines Vx. Each line 206 is, for example, shared by all the pixels of a corresponding row of the matrix. In FIG. 2, only one line 206 is fully represented in dashed lines in order to avoid complicating the Figure.


Although not shown in FIG. 2, the other control signals provided to the pixels of the matrix 200 are preferably provided by the control circuit CTRL1. As usual in indirect time of flight sensors, the sensor 2 comprises other conductive lines (not shown) to provide other control signals and voltages to the pixels of the matrix 200. For example, in the embodiment of FIG. 2, the sensor 2 comprises: for each row of the matrix 200, a conductive line for transmitting voltage GND (FIG. 1) to all the pixels of the row; for each column of the matrix 200, a conductive line for transmitting signal TGAB (FIG. 1) to each pixel of the column; for each column of the matrix 200, a conductive line for transmitting bias voltage VAB (FIG. 1) to all the pixels of the column; for each row of the matrix 200, a conductive line for transmitting signal RD (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting signal RD1 (FIG. 1) to all the pixels of the row; and for each row of the matrix 200, a conductive line for transmitting signal RD2 (FIG. 1) to all the pixels of the row.


The sensor 2 comprises an illumination device 205. The illumination device 205 is configured to illuminate a scene to capture. The sensor 2 further comprises a control circuit CTRL2 configured to control the illumination device 205. For example, the control circuit CTRL2 provides a control signal cmd to the device 205. The signal cmd is, for example, a digital signal comprising several bits.


As indicated above, the scene to capture is divided into a plurality of areas, and it is here proposed to successively illuminate each area of the scene, by illuminating only one area at a time, being understood that, in practice, parts of the scene which are adjacent to the illuminated area may also receive some light. Said in other words, the device 205 and its control circuit CTRL2 are configured to successively illuminate each area of the scene. For example, the device 205 is configured to illuminate different areas of the scene to capture, the area which is illuminated by the device 205 being determined by the signal cmd.


Control circuits CTRL1 and CTRL2 are synchronized, for example by means of a synchronization circuit SYNC which couples circuits CTRL1 and CTRL2. Said in other words, circuit SYNC receives and/or sends synchronization signals to and/or from circuits CTRL1 and CTRL2.


In a similar manner to the scene, the matrix 200 is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of FIG. 2, the matrix 200 is divided into four areas M1, M2, M3 and M4.


Each area M1, M2, M3, M4 comprises adjacent lines of pixels 1, these lines of pixels being parallel to the conductive lines 204. In the embodiment of FIG. 2, each area M1, M2, M3, M4 comprises two adjacent lines of pixels 1 which are parallel to the lines 204, or, said in other words, each area M1, M2, M3, M4 comprises two adjacent columns of pixels 1.


The matrix 200 and the device 205 are disposed relative to each other such that each area M1, M2, M3, M4 of the matrix 200 corresponds to an area of the scene, taken among the areas the scene is divided into and which are successively illuminated. Said in other words, the matrix 200 and the device 205 are disposed relative to each other such that, each time an area of the scene, taken among the plurality of areas the scene is divided into, is illuminated by the device 205, the light reflected by this area of the scene is received by the pixels 1 of the corresponding area M1, M2, M3 or M4 of the matrix 200, being understood that, in practice, some other pixels of the matrix, which are disposed near this corresponding area M1, M2, M3 of M4, may also receive part of the light reflected by the scene. The implementation of this disposition of the matrix 200 and the device 205 relative to each other is in the abilities of those skilled in the art.


The sensor 2 allows a scanned illumination of the scene to capture. For a given power supply provided to device 205 during an illumination of an area of the scene, all the light generated by the device 205 is directed towards this area of the scene. This differs from usual indirect time of flight sensors in which this given power supply is used to provide a flash illumination of the whole scene to capture. As a result, the signal-to-noise ratio of the light received by the sensor 2 is increased compared to that of the light received by these usual sensors. Indeed, for a given power supply, with a flash illumination, the light received by each area of scene carries less optical power than the light received by the only area of the scene which is illuminated by the sensor 2 during a scanned illumination.


The control circuit CTRL1 is further configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200. Said in other words, the control circuit CTRL1 is configured to control the charge transfers independently in each area M1, M2, M3, M4 of the matrix 200, or, said differently, independently between the areas M1, M2, M3 and M4. For example, control circuit CTRL1 comprises a different sub-circuit (not shown on FIG. 2) for each area M1, M2, M3, M4 of the matrix, each sub-circuit being configured to provide control signals for charge transfer in the pixels of the area M1, M2, M3 or M4 this sub-circuit is associated with.


For example, the control circuit CTRL1 is configured to control an integration phase for the pixels of any one of the areas M1, M2, M3 and M4, while the control circuit CTRL1 controls no integration phase for the pixels of the other areas. More particularly, when an area of the scene is illuminated by the device 205, and the light reflected by this area of the scene is received by the corresponding area M1, M2, M3 or M4 of the matrix 200, control signals TG1, TG2 are maintained, by control circuit CTRL1, at the inactive state for the other areas of the matrix 200. The control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area M1, M2, M3 or M4 which is receiving light. Said in other words, control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area of the matrix 200 corresponding to the area of the scene which is illuminated, such that in each pixel of said area of the matrix 200, charges are alternatively transferred, from the region PD, to each storage regions mem1, mem2 of the pixel.


In practice, each commutation of the signal TG1, respectively TG2, corresponds to a charge or a discharge of a capacitance, typically the gate capacitance of the charge transfer device TGmem1, respectively TGmem2. Thus, by reducing the number of pixels for which signals TG1 and TG2 simultaneously commute, a power consumption of the sensor 2 is reduced compared to that of a usual indirect time of flight sensor, in which signals TG1, respectively TG2, commute simultaneously in all the pixels of the sensor.



FIG. 3 illustrates, in a very schematic manner, the illumination device 205 according to one embodiment.


The illumination device 205 comprises an array 300 of laser sources 301, only one laser source being referenced in FIG. 3 in order to avoid complicating the Figure. Each laser source 301 is, preferably, a VCSEL (“Vertical-Cavity Surface-Emitting Laser”). In the example of FIG. 3, the array 300 comprises 8×2 laser sources 301, although the number of light sources 301 of the array can be different in other examples.


Device 205 further comprises an optical device (or element) 302, represented in the form of a block in FIG. 3. Optical device 302 is configured to direct, or orientate, the light emitted by the array 300 of laser sources 301 towards the scene to capture.


In this embodiment, the array 300 is divided into a plurality of sets of laser sources. In the example of FIG. 3, the array 300 is divided into four sets A1, A2, A3 and A4 of laser sources 301. Preferably, the number of sets of the array 300 is equal to the number of areas of the scene, and to the number of areas M1, M2, M3, M4 of the matrix 200 (FIG. 2).


Each set A1, A2, A3, A4 is configured to illuminate a corresponding area of the scene to capture. Indeed, the laser sources 301 of the array can be each controlled independently from the other laser sources 301. For example, the array 300 is controlled such that, when laser sources 301 of a given sets A1, A2, A3 or A4 of the array 300 is emitting light, the laser sources 301 of the other sets are emitting no light. For example, the laser sources 301 which are emitting light and those which are emitting no light are determined by the signal cmd.


The control circuit CTRL2 (FIG. 2) is configured to control, with the signal cmd, an emission of light by sets A1, A2, A3 and A4 the one after the other. More precisely, the set A1, A2, A3 or A4 which emits light depends on the value of the signal cmd.


For each set A1, A2, A3, A4, when the laser sources 301 of the set are emitting light, the emitted light is directed towards a corresponding area of the scene to capture by the device 302, the illuminated area of the scene being different for each set A1, A2, A3, A4 of the array 300 of laser sources 301.


For example, in FIG. 3, the optical device 302, for example a lens or an objective, is configured to direct the light emitted by the laser sources 301 of the respective set A1, A2, A3 or A4 in a respective direction O1, O2, O3 or O4. Thus, when set A1 (respectively A2, A3 or A4) emits light, a first (respectively a second, a third or a fourth) area of the scene is illuminated and reflected light is received by the area M1 (respectively M2, M3 or M4) of the matrix 200 (FIG. 2).


For example, the device 205 comprises a control circuit CTRL3 configured to control the emission of light by each light source 301 of the array 300 based on signal cmd.


In the device 205 of FIG. 3, a given power supply provided to the array 300 is shared, or split, between those of the light sources 301 which are emitting light. Thus, for a given power supply provided to the array 300, the optical power of the light received by an area of the scene is greater when only the light sources of the set A1, A2, A3 or A4 corresponding to this area are emitting light (scanned illumination), than when all the light sources 301 are emitting light simultaneously (flash illumination).



FIG. 4 illustrates the illumination device 205 according to one alternative embodiment.


The device 205 of FIG. 4 comprises, as the one of FIG. 3, the array 300 of laser sources 301, and the optical device 302.


However, in the embodiment of FIG. 4, the array 300 is not divided into a plurality of sets of light independently controllable. For example, depending on the signal cmd, all the light sources 301 emit light, or do not emit any light. For example, the device 205 comprises the control circuit CTRL3 configured to control the emission of light by all light sources 301 of the array 300 based on signal cmd.


Further, in the embodiment of FIG. 4, the optical device 302 is controllable. More precisely, the direction in which the light emitted by the array 300 is directed by the device 302 is controllable. Said in other words, the device 302 is configured to direct the emitted light differently depending on signal cmd. The control circuit CTRL2 (FIG. 2), which provides the control signal cmd to the device 205, is configured to provide, at each illumination of an area of the scene to capture, the control signal cmd which corresponds to a directing of the light, by the device 302, towards this area of the scene.


For example, in FIG. 4, the optical device 302 is configured to direct the light emitted by the array of laser sources 301 in four different directions O1, O2, O3 or O4, each corresponding to a different area of the scene. Thus, when signal cmd is at a first (respectively a second, a third or a fourth) value, a first (respectively a second, a third or a fourth) area of the scene is illuminated and reflected light is received by the area M1 (respectively M2, M3 or M4) of the matrix 200 (FIG. 2).


The device 302, for example, comprises mirror(s) and/or one or several lenses, the orientation of which being controllable by the signal cmd. Preferably, the optical device 302 comprises at least one controllably movable micro-mirror, or, in other words, a controllably movable MicroElectroMechanical System (MEMS) micro-mirror. The implementation of the optical device 302 is in the abilities of those skilled in the art.


In the device 205 of FIG. 4, a given power supply which is provided to the array 300 during an illumination phase is shared between all the light sources 301. However, all the light emitted by the array 300 is concentrated towards a given area of the scene by the optical device 302. This differs from a flash illumination for which the light emitted by the array 300 is directed, or spread, towards the whole scene to capture. Thus, for a given power supply provided to the array 300, a scanned illumination of the scene allows to improve the optical power of the light successively received by each area of the scene, compared to that of the light received simultaneously by all the areas of the scene during a flash illumination.


The embodiments of FIGS. 3 and 4 may be combined. Further, the described embodiments of indirect time of flight sensors are not limited to the embodiments of the device 205 described in relation with FIGS. 3 and 4. Those skilled in the art are capable of using other illumination devices which are controllable, such that the emitted light is directed only towards an area of the scene to capture, selected in a controllable manner among a plurality of areas of the scene.



FIG. 5 shows chronograms (i.e., timing diagrams) illustrating operation of the sensor of FIG. 2 according to one embodiment. More specifically, in this example the scene to capture is divided into four areas S1, S2, S3 and S4, and the FIG. 5 shows, depending on time t, the light (“light”) emitted by the illumination device 205 (FIG. 2), which area S1, S2, S3 or S4 receives the light (“illuminated area of the scene”), which corresponding area M1, M2, M3 or M4 of the matrix 200 (FIG. 2) receives the reflected light and has its pixels in an integration phase (“integrated area”), and which pixels of the matrix are read (“read”). In this example, the device 205 emits light under the form of a burst of periodic pulses of light.


Between an instant t0 and an instant t1 posterior to instant t0, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, by commutating the control signals TG1, TG2 of the charge transfer devices TGmem1, TGmem2 of these pixels between their active and inactive states, at a frequency upper than that of the emitted light.


Between the instant t1 and an instant t2 posterior to instant t1, no light is emitted by the device 205 and the pixels of the area M1 are read. Because the lines 204 are parallel to the lines Vx (FIG. 2), the reading of the pixels of the area M1 implies that all the pixels of the matrix are read (“all matrix”). Thus, between instants t1 and t3, the control circuit CTRL1 is configured to control, by means of signals RD, a reading of all the pixels of the matrix 200, by reading the rows of the matrix ones after the other.


Between the instant t2 and an instant t3 posterior to instant t2, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only.


Between the instant t3 and an instant t4 posterior to instant t3, no light is emitted by the device 205 and the pixels of the area M2 are read, by reading all the pixels of the matrix (“all matrix”), similarly to what has been done between instants t1 and t2.


Between the instant t4 and an instant t5 posterior to instant t4, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by the area S3 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the area M3 only.


Between the instant t5 and an instant t6 posterior to instant t5, no light is emitted by the device 205 and the pixels of the area M3 are read, by reading all the pixels of the matrix (“all matrix”).


Between the instant t6 and an instant t7 posterior to instant t6, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area A4 of the matrix, and an integration phase is performed in the area M4 only.


At the instant t7, all the areas S1, S2, S3, S4 of the scene have been illuminated once during the scanned illumination of the scene.


Between the instant t7 and an instant t8 posterior to instant t7, no light is emitted by the device 205 and the pixels of the area M4 are read, by reading all the pixels of the matrix (“all matrix”).


At the instant t8, the output signals of the pixels of the area M1 read after the illumination of the area M1 (between instants t1 and t2), the output signals of the pixels of the area M2 read after the illumination of the area M2 (between instants t3 and t4), the output signals of the pixels of the area M3 read after the illumination of the area M3 (between instants t5 and t6), and the output signals of the pixels of the area M4 read after the illumination of the area M4 (between instants t7 and t8) may be used to generate, or compute, an image, or depth map, of scene.


At the instant t8, a new scanned illumination of the scene begins, by illuminating, with the device 205, the area M1 of the scene.


In the operating mode described in relation with FIG. 5, after each illumination of an area S1, S2, S3 or S4 of the matrix 200, all the pixels of the matrix are read to obtain the output signals of the pixels of the area M1, M2, M3 or M4 of the matrix 200 corresponding to the illuminated area. More precisely, after each illumination of an area S1, S2, S3 or S4, all the pixels of the matrix are read before the next area S1, S2, S3 or S4 is illuminated.


Preferably, when capturing a scene, during the successive illuminations of the areas of the scene, the device 205 is supplied with an average power supply having a given peak power, which is equal to an average power, having the same peak power, provided to an illumination device of a usual sensor during a flash illumination of the scene. In this case, the duration T of the illumination phase of each area of the scene during a scanned illumination is preferably equal to the duration of the flash illumination divided by the number of areas of the scene. This allows to further increase the signal-to-noise ratio in the sensor 2, compared to a usual sensor, without modifying the power supply used to illuminate the scene to capture.



FIG. 6 shows chronograms illustrating operation of the sensor of FIG. 2 according to one alternative embodiment. In this example the scene to capture is divided into four areas S1, S2, S3 and S4, and the FIG. 6 shows, depending on time t, the light (“light”) emitted by the illumination device 205, which area S1, S2, S3 or S4 receives the light (“illuminated area of the scene”), which corresponding area M1, M2, M3 or M4 of the matrix 200 receives the reflected light and has its pixels in an integration phase (“integrated area”), and which pixels of the matrix are read. In this example, the device 205 emits light under the form of a burst of periodic pulses of light.


Between an instant t10 and an instant t11 posterior to instant t10, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, by commutating the control signals TG1, TG2 of the charge transfer devices TGmem1, TGmem2 of these pixels between their active and inactive states, at a frequency upper than that of the emitted light.


Between the instant t11 and an instant t12 posterior to instant t11, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only.


Between the instant t12 and an instant t13 posterior to instant t12, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by the area S3 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the pixels of the area M3 only.


Between the instant t13 and an instant t14 posterior to instant t13, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area M3 of the matrix, and an integration phase is performed in the pixels of the area M4 only.


As illustrated in FIG. 6, a cycle of successive illuminations of the areas S1, S2, S3 and S4, in which each area S1, S2, S3, S4 is illuminated once, may be then repeated several times before a reading of all the pixels of the matrix (“all matrix”). In the example of FIG. 6, before the reading, the cycle of successive illuminations of the area S1, S2, S3 and S4 is performed four times, once between the instants t10 and t14, once between the instant t14 and an instant t15 posterior to instant t14, once between the instant t15 and an instant t16 posterior to instant t15, and once between the instant t16 and an instant t17 posterior to instant t16.


At the instant t17, the control circuit CTRL1 (FIG. 2) controls, by means of signals RD, a reading of all the pixels of the matrix (“all matrix”), by reading the rows of the matrix ones after the other. No light is emitted during this reading phase. At the end of the reading phase, a depth map of the scene can be generated, or computed, based on the output signals of the pixels read during the reading phase.


In the operation described in relation with FIG. 6, before each reading of all the pixels, the reading being controlled by the control circuit CTRL1, the control circuit CTRL2 is configured to control several successive illumination cycles, each comprising a unique illumination of each area S1, S2, S3 and S4. The control circuit CTRL1 is further configured to control an absence of emission of light by the illumination device 205 during the reading.


Compared to the operation described in relation with FIG. 5 to capture a full scene, only one reading of all the pixels of the matrix is performed in the operation described in relation FIG. 6, which results in a decrease of the time needed to capture the scene.


Preferably, in FIG. 6, the duration T1 of each illumination phase of each area S1, S2, S3 and S4 is equal to the duration T of the illumination phase of each area S1, S2, S3 and S4 described in relation with FIG. 5, divided by the number of times the illumination cycle of the areas S1, S2, S3 and S4 is repeated before a full reading of the matrix. Said in other words, in this example, the illumination duration T1 is equal to a quarter of the illumination duration T (FIG. 5). As a result, the power supply provided to device 205 for capturing the scene is the same in the operation mode of FIG. 6 and in the operation mode of FIG. 5. Further, in case the device 205 is implemented as described in relation with FIG. 3, the operation described in relation with FIG. 6 allows to mitigate the temperature elevation in the array 300 of the device 205, compared to the operation described in relation with FIG. 5.


In the embodiments described in relation with FIGS. 2 to 6, the lines 204 for providing the control signals TG1, TG2 to the transfer devices TGmem1, TGmem2 of the pixels are parallel to the lines Vx. Other embodiments will be described below, in which lines 204 are perpendicular to the lines Vx.



FIG. 7 illustrates an indirect time of flight sensor 2′ according to a further embodiment, in which lines 204 are perpendicular to lines Vx.


The sensor 2′ comprises, as the sensor 2 (FIG. 2), the matrix 200 of pixels 1, the circuit READOUT, the lines Vx coupled to the circuit READOUT, the lines 206, and the illumination device 205 and its control circuit CTRL2, which will not be described again.


Instead of the control circuit CTRL1, the sensor 2′ comprises a control circuit CTRL1′. The control circuit CTRL1′ is configured to control reading phases and integration phases for the pixels of the matrix 200. The control circuit CTRL1′ is configured to provide the control signals TG1 and TG2 (FIG. 1) to the lines 204. The control circuit CTRL1′ is further configured to provide the control signals RD to the lines 206.


In the embodiment of FIG. 7, the lines 204, which are each connected to control circuit CTRL1′, are perpendicular to the lines Vx. Each line 204 is shared by all the pixels of a corresponding row of the matrix. In FIG. 7, only one line 204 is fully represented in dashed lines, in order to avoid complicating the Figure. Further, in order to avoid complicating the figure, only one line 204 by row is represented in FIG. 7. However, in practice, each pixel receives control signals TG1 and TG2 (FIG. 1) via two corresponding lines 204, and each row is thus associated to a line 204 for transmitting signal TG1 to all the pixels of the row, and to another line 204 for transmitting signal TG2 to all these pixels.


Although not shown on FIG. 7, the other control signal provided to the pixels of the matrix 200 are preferably provided by the control circuit CTRL1′. Further, similarly to what has been described for the sensor 2 of FIG. 2, the sensor 2′ comprises other conductive lines (not shown) to provide other control signals and voltages to the pixels of the matrix 200. For example, in the embodiment of FIG. 7, the sensor 2′ comprises: for each row of the matrix 200, a conductive line for transmitting, or providing, control signal TGAB (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting bias voltage VAB (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting voltage GND (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting signal RD (FIG. 1) to each pixel of the row; for each row of the matrix 200, a conductive line for transmitting signal RD1 (FIG. 1) to all the pixels of the row; and for each row of the matrix 200, a conductive line for transmitting signal RD2 (FIG. 1) to all the pixels of the row.


Control circuits CTRL1′ and CTRL2 are synchronized, for example by means of a synchronization circuit SYNC which couples circuits CTRL1′ and CTRL2. Said in other words, circuit SYNC receives and/or sends synchronization signals to and/or from circuits CTRL1′ and CTRL2.


As with the sensor 2, the matrix 200 of sensor 2′ is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of FIG. 7, the matrix 200 is divided into four areas M1, M2, M3 and M4.


Each area M1, M2, M3, M4 comprises adjacent lines of pixels 1, these lines of pixels being parallel to the conductive lines 204. In the embodiment of FIG. 7, each area M1, M2, M3, M4 comprises two adjacent lines of pixels 1 which are parallel to the lines 204, or, said in other words, each area M1, M2, M3, M4 comprises two adjacent rows of pixels 1.


As already described for the sensor 2, in the sensor 2′ the matrix 200 and the device 205 are disposed relative to each other such that each area M1, M2, M3, M4 of the matrix 200 corresponds to an area of the scene.


The sensor 2′ allows, as with the sensor 2 of FIG. 2, a scanned illumination of the scene to capture. As a result, the signal-to-noise ratio of the light received by the sensor 2′ is increased compared to that of the light received by the usual sensors.


The control circuit CTRL1′ is configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200. Said in other words, the control circuit CTRL1′ is configured to control the charge transfers independently in each area M1, M2, M3, M4 of the matrix 200. For example, control circuit CTRL1′ comprises a different sub-circuit (not shown on FIG. 7) for each area M1, M2, M3, M4 of the matrix, each sub-circuit being configured to provide control signals for charge transfer in the pixels of the area M1, M2, M3 or M4 with which this sub-circuit is associated.


For example, the control circuit CTRL1′ is configured to control an integration phase for the pixels of any one of the areas M1, M2, M3 and M4, while the control circuit CTRL1′ controls no integration phase for the pixels of the other areas. More particularly, when an area of the scene is illuminated by the device 205, and the light reflected by this area of the scene is received by the corresponding area M1, M2, M3 or M4 of the matrix 200, control signals TG1, TG2 are maintained, by control circuit CTRL1′, at the inactive state for the other areas of the matrix 200. The control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area M1, M2, M3 or M4 which is receiving light. Said in other words, control signals TG1, TG2 are repeatedly commuted between active and inactive states only for the pixels 1 of the area of the matrix 200 corresponding to the area of the scene which is illuminated, such that in each pixel of said area of the matrix 200, charges are alternatively transferred, from the region PD, to each storage regions mem1, mem2 of the pixel. As a result, a power consumption of the sensor 2′ is reduced compared to that of a usual indirect time of flight sensor.


An advantage of the sensor 2′ compared to the sensor 2 is that the pixels of a given area M1, M2, M3 or M4 of the matrix 200 of sensor 2′ may be read without performing a full reading of the matrix 200, by reading ones after the other only the rows of this area.



FIG. 8 shows chronograms illustrating operation of the sensor 2′ of FIG. 7 according to one embodiment. More specifically, in this example the scene to capture is divided into four areas S1, S2, S3 and S4, and the FIG. 8 shows, depending on time t, the light (“light”) emitted by the illumination device 205 (FIG. 7), which area S1, S2, S3 or S4 receives the light (“illuminated area of the scene”), which corresponding area M1, M2, M3 or M4 of the matrix 200 (FIG. 7) receives the reflected light and has its pixels in an integration phase (“integrated area”), and which pixels of the matrix are read (“read”). In this example, the device 205 emits light under the form of a burst of periodic pulses of light.


The chronograms of FIG. 8 are identical to those of FIG. 5, except for the reading phase. Indeed, in the operation of FIG. 8, each illumination of an area S1, S2, S3 or S4 of the scene is followed by a reading of the pixels of only the area M1, M2, M3 or M4 of the matrix which corresponds to this area of the scene. Said in other words the control circuit CTRL1′ is configured to control, after each illumination of an area S1, S2, S3 or S4, a reading of only the pixels of the area M1, M2, M3 or M4 corresponding to this area S1, S2, S3 or S4, before an illumination of a next area of the scene. During the reading of the pixels of a given area M1, M2, M3 or M4 of the matrix, the control circuit CTRL2 is configured to control an absence of light emission by the device 205.


More specifically, in the FIG. 8: the pixels of the area M1 only are read between the instants t1 and t2, the pixels of the area M2 only are read between the instants t3 and t4, the pixels of the area M3 only are read between the instants t5 and t6, and the pixels of the area M4 only are read between the instants t7 and t8.


In the sensor 2′, the duration of the reading of the pixels of a given area of the matrix is reduced compared to that of the sensor 2, because it is not needed anymore to read the all the pixels of the matrix to read the pixels of a given area of the matrix.


In an alternative embodiment, the sensor 2′ operates as described in relation with FIG. 6. In such alternative embodiment, the control circuit CTRL2 is configured, before each reading of all the pixels, which is controlled by the control circuit CTRL1′, to control several successive illumination cycles each comprising a unique illumination of each area S1, S2, S3 and S4 of the scene. The control circuit CTRL2 is further configured to control an absence of light emission by the illumination device 205 during the reading of the matrix.


To take profit of the fact that lines 204 are perpendicular to lines Vx, it is here proposed to read the pixels of an area M1, M2, M3 or M4 of the matrix 200 whereas another area of the matrix 200 is receiving light. However, when pixels of a given area M1, M2, M3 or M4 receiving light are in an integration phase and when pixels of another area are simultaneously in a reading phase, it has been shown that the high frequency commutations of the signals transmitted using lines 204 to the pixels of in the integration phase generate noise in the output signals of the pixels in the reading phase, the output signals being available on the Vx lines. This noise is, for example, transmitted via the reference voltage GND which is provided to the different circuits and to all the pixels of the sensor, and/or by the cross coupling between lines Vx and lines 204.


To suppress this noise, it is here proposed a split ground and bias strategy to minimize unwanted coupling. More specifically, the pixels matrix is split into two insulated halves. Further, separated, or electrically decoupled, supply voltage, reference voltage, bias voltages and control signals are provided to each matrix half. It is then possible to read pixels of one half of the matrix while pixels of the other half are integrating, without generating noise. Different embodiments of indirect time of flight sensors implementing this strategy will be now described.



FIG. 9 illustrates an indirect time of flight sensor 2″ according to a further embodiment. The sensor 2″ is similar to the sensor 2′ of FIG. 7, and only the difference between these two sensors will be described in detail. In FIG. 9, the illumination device 205 and its control circuit CTRL2 are not shown.


In sensor 2″, the matrix 200 is split into two halves P1 and P2. More specifically, a separation between parts P1 and P2 of the matrix 200 is parallel to the lines 204.


The parts P1 and P2 of the matrix 200 are adjacent, the part P1 being disposed along the part P2. More specifically, each column comprises a first portion, or half, belonging to part P1, and a second portion, or half, belonging to part P2 and being aligned with the first portion of the column. For example, the parts P1 and P2 have a common edge, which corresponds to the separation between parts P1 and P2.


Further, the lines Vx, which are parallel to the column of the matrix and perpendicular to lines 204, are interrupted at the separation between parts P1 and P2 of the matrix 200. Said in other words, the lines Vx of the part P1 of the matrix 200 and the lines Vx of the part P2 of the matrix end at the separation between parts P1 and P2 of the matrix 200. Said differently, the lines Vx of part P1 of the matrix are insulated from the lines Vx of part P2 of the matrix, and the lines Vx of part P1, respectively P2, do not extend above or below the part P2, respectively P1. In FIG. 9, in order to not complicating the Figure, only one line Vx of the part P1 is represented in dashed line, and only one corresponding line Vx of the part P2 is represented in dashed line.


A line Vx of the part P2 corresponds to a line Vx of the part P1 when these two lines Vx belong to the same column of the matrix 200. For example, in each column of the matrix 200, a line Vx of the part P2 corresponds to a line Vx of the part P1 when the line Vx of the part P1 is selectively coupled to given outputs of the pixels of the part P1 disposed in this column, and the line Vx is selectively coupled to the corresponding outputs of the pixels of the part P2 disposed in this column.


The part P1 of the matrix 200 is electrically decoupled from the part P2 of the matrix 200. More specifically, a semiconductor substrate to which the pixels 1 of the matrix 200 belong has a first part which comprises the part P1 of the matrix 200 and a second part which comprises the part P2 of the matrix 200. Said in other words, the first part of the substrate comprises the half P1 of the matrix and a second part of the substrate comprises the half P2 of the matrix.


The first and second parts of the substrate are insulated from each other using insulation structures passing through the substrate, the insulation structures being preferably insulation structures provided between pixels to insulate the pixels from each other.



FIG. 10 shows a very schematic top view of two adjacent pixels 1 of the sensor of FIG. 9, according to an example. FIG. 11 shows a very schematic cross section view along plan AA of FIG. 10. In this example, the two adjacent pixels 1 belong to the same column of the matrix, but to two different adjacent rows. The pixels 1 are disposed in and on a semiconductor substrate 1003.


In the example of FIGS. 10 and 11, the two adjacent pixels are laterally delimited, or surrounded, by an insulation structure 1000, which is schematically represented by a simple line in FIG. 10. The insulation structure 1000 passes through the substrate 1003. As it can be seen in more detail in FIG. 11, the insulation structure 1000 is preferably a capacitive deep trench insulation (CDTI), that is to say a trench filled with a conductive material 1001, insulated from the semiconductor substrate 1003 by an insulative layer 1002. Preferably, the conductive material is a metal, for example tungsten or aluminum, or a metal alloy. Indeed, the use of a metal or metal alloy allows to reduce the optical cross-talk.


In this example, the region PD of each pixel 1 is laterally delimited by a capacitive deep trench insulation 1005, for example a U-shaped insulation structure 1005 in the view of FIG. 10. The storage region mem1 and mem2 of each pixel 1 are defined, or delimited, by a portion of the structure 1000 and a portion of the structure 1005 which is opposite and parallel to this portion of the structure 1000. Said in other words, each storage region mem1, mem2 is laterally delimited, in a direction perpendicular to its length, by two parallel portions of the respective structures 1000 and 1005.


In this example, each pixel 1 further comprises transfer devices TGmem1 and TGmem2, the coupling devices TGRD1 and TGRD2, the transistor 112 and the selection device 108, the transistors 112 and 108 being shared by the two adjacent pixels.


The example shown in FIGS. 10 and 11 is not limitative. For example, the pixels of the matrix 200 can be arranged by groups of four pixels, the pixels of each group sharing the same transistors 112 and 108. In another example, each pixel of matrix 200 has its own transistors 112 and 108. Further, the storage region mem1 and mem2 of each pixel may be delimited by CDTI which are not a portion of the insulation structure 1000 which laterally delimitate the pixel.


Further, although in the example of FIGS. 10 and 11, the insulation structure 1000 is of the CDTI type, in other example, this insulation structure may be a deep trench insulation (DTI), that is to say a trench filled with an insulating material, the DTI passing through the substrate.


Referring back to the FIG. 9, for example, the set of all the pixels 1 of the part P1 of the matrix are surrounded by an insulation structure 1000, which delimits the first part of the substrate, and the set of all the pixels of the part P2 of the matrix are surrounded by another insulation structure 1000, which delimits the second part of the substrate.


The reference voltage GND which is provided to the first part of the substrate and the reference voltage GND which is provided to the second part of the substrate are electrically decoupled from each other. For example, the reference voltage GND provided to the first part of the substrate, or, in other words, to each pixel of the part P1 of the matrix, is provided by a first bonding pad 900 of the sensor 2″, and the other reference voltage GND provided to the second part of the substrate, or, in other words, to each pixel of the part P2 of the matrix, is provided by a second bonding pad 902 of the sensor 2″. Each bonding pad 900, 902 receives an off-chip reference voltage GND. Each bonding pad 900, 902 acts as a low-pass filter, as it is schematically represented in FIG. 9 by a resistance R and an inductance L series-connected in each bonding pad.


Preferably, the insulation structures 1000 are CDTI. In this case, it is preferable to provide a bias voltage to structure 1000 delimiting the part P1 of the matrix 200, which is electrically decoupled from a bias voltage provided to structure 1000 delimiting the part P2 of the matrix 200. For example, in FIG. 9, the bias voltage of the CDTI 1000 of the part P1 of the matrix 200 is provided by a voltage generator 904, and the bias voltage of the CDTI 1000 of the part P2 of the matrix 200 is provided by a voltage generator 906, which is electrically decoupled form the generator 904.


Instead of the control circuit CTRL1′, the sensor 2″ comprises a control circuit CTRL1″. The control circuit CTRL1″ is configured to control reading phases and integration phases for the pixels of the matrix 200. The control circuit CTRL1″ is configured to provide the control signals TG1 and TG2 (FIG. 1) to the lines 204. The control circuit CTRL1″ is further configured to provide the control signals RD to the lines 206.


In the embodiment of FIG. 9, the lines 204, which are each connected to control circuit CTRL1″, are parallel to the lines Vx. Each line 204 is shared by all the pixels of a corresponding row of the matrix. In FIG. 9, only one line 204 for each part P1, P2 of the matrix is fully represented in dashed lines, in order to avoid complicating the Figure. Further, in order to avoid complicating the Figure, only one line 204 by row is represented in FIG. 9. However, in practice, each pixel receives control signals TG1 and TG2 (FIG. 1) via two corresponding lines 204, and each row is thus associated to a line 204 for transmitting signal TG1 to all the pixels of the row, and to another line 204 for transmitting signal TG2 to all these pixels.


Although not shown on FIG. 9, the other control signal provided to the pixels of the matrix 200 are preferably provided by the control circuit CTRL1″. Further, similarly to what has been described for the sensor 2′ of FIG. 7, the sensor 2″ comprises other conductive lines (not shown) to provide other control signals and voltages to the pixels of the matrix 200. For example, in the embodiment of FIG. 9, the sensor 2″ comprises: for each row of the matrix 200, a conductive line for transmitting, or providing, control signal TGAB (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting bias voltage VAB (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting voltage GND (FIG. 1) to all the pixels of the row; for each row of the matrix 200, a conductive line for transmitting signal RD (FIG. 1) to each pixel of the column; for each row of the matrix 200, a conductive line for transmitting signal RD1 (FIG. 1) to all the pixels of the row; and for each row of the matrix 200, a conductive line for transmitting signal RD2 (FIG. 1) to all the pixels of the row.


Control circuits CTRL1″ and CTRL2 (not shown in FIG. 9) are synchronized, for example by means of a synchronization circuit SYNC (not shown in FIG. 9), which couples circuits CTRL1″ and CTRL2.


As for sensor 2′, the matrix 200 of sensor 2″ is divided into a plurality of areas, the total number of areas of the matrix being, preferably, equal to the total number of areas of the scene. In the example of FIG. 9, the matrix 200 is divided into four areas M1, M2, M3 and M4. Each area M1, M2, M3, M4 comprises adjacent lines of pixels 1, parallel to the conductive lines 204. In the embodiment of FIG. 9, each area M1, M2, M3, M4 comprises two adjacent lines of pixels 1 which are parallel to the lines 204, or, said in other words, each area M1, M2, M3, M4 comprises two adjacent rows of pixels 1. As already described for the sensors 2 and 2′, in the sensor 2″ the matrix 200 and the device 205 (not shown in FIG. 9) are disposed relative to each other such that each area of the scene corresponds to an area M1, M2, M3, M4 of the matrix 200. In the example of FIG. 9, areas M1 and M2 belong to part P1 of the matrix 200, areas M3 and M4 belonging to part P2 of the matrix 200.


The control circuit CTRL1″ is configured to provide different control signals TG1 and TG2 to the different areas M1, M2, M3 and M4 of the matrix 200, in a way similar to that described for the control circuit CTRL1′ (FIG. 7). Compared to the control circuit CTRL1′ of the sensor 2′ of FIG. 7, the control circuit CTRL1″ is further configured to simultaneously control charge transfers in the pixels of an area of one of the halves P1 and P2 of the matrix 200, and a reading of the pixels of an area of the other one of the halves P1 and P2. For example, when the pixels of the area M1 or M2 of the part P1 are in a reading phase (respectively in an integration phase) controlled by the control circuit CTRL1″, the pixels of the area M3 or M4 of the part P1 are in an integration phase (respectively in a reading phase) controlled by the control circuit CTRL1″.


Preferably, for each voltage level which is provided to at least one pixel 1 of the part P1 of the matrix 200, and simultaneously to at least one pixel 1 of the other part P2 of the matrix, the sensor 2″ comprises a voltage generator configured to provide this voltage level to the part P1 of the matrix, and a voltage generator configured to provide this voltage level to the other part P2 of matrix. These two generators are electrically decoupled form each other.


In FIG. 9, this is, for example, illustrated for the signals TG1 and TG2 provided to lines 204 by the control circuit CTRL1″. More specifically, when a pixel 1 of the matrix is in a reading phase, there is no charge transfer between the region PD (FIG. 1) and the regions mem1 and mem2 (FIG. 1) of this pixel. Thus, the control signals TG1 and TG2 provided to the transfers devices TGmem1 and TGmem2 (FIG. 1) of this pixel, via corresponding lines 204, are maintained at an inactive state, which corresponds to a low voltage level TGmemL in this example. The same occurs when the pixel 1 is neither in a reading phase, nor in an integration phase. However, when a pixel 1 of the matrix is in an integration phase, charge transfers between the region PD (FIG. 1) and the regions mem1 and mem2 (FIG. 1) are performed. Thus, the control signals TG1 and TG2 provided to the transfers devices TGmem1 and TGmem2 (FIG. 1) of this pixel, via corresponding lines 204, are repeatedly commuted between their inactive state (the low voltage level TGmemL in this example) and their active state, which corresponds to a high voltage level TGmemH in this example. As a result, in the sensor 2″, the voltage level TGmemH of the signals TG1 and TG2 is never provided simultaneously to a pixel of the part P1 and to a pixel of the part P2, whereas the low voltage level TGmemL is provided simultaneously to a pixel of the part P1 and to a pixel of the part P2. Thus, the sensor 2″ comprises a voltage generator 910 configured to provide the voltage level TGmemL to part P1 of the matrix 200, and a voltage generator 912 configured to provide the voltage level TGmemL to the part P2 of the matrix 200. Further, as illustrated by FIG. 9, the sensor 2″ may comprise only one voltage generator 908 configured to provide the voltage level TGmemH, which is for example alternatively to the part P1 and to the part P2 of the matrix by the control circuit CTRL1″.


Although the provision of two generators which are electrically decoupled from each other and configured to provide simultaneously the same voltage level to both parts P1 and P2 of the matrix 200 is here illustrated only for the voltage level TGmemL, those skilled in the art are capable to implement other pairs of electrically decoupled voltage generator for generating any voltage level which is provided simultaneously to both parts P1 and P2 of the matrix.


According to one embodiment, which is illustrated by FIG. 9, the sensor 2″ comprises a first reading circuit READOUT1 coupled the lines Vx of the half P1 of the matrix 200, and a second reading circuit READOUT2 coupled to the lines Vx of the half P2 of the matrix 200.


Circuit READOUT1, respectively READOUT2, is configured to received output signals of the pixels of the part P1, respectively P2, of matrix 200 which are coupled to the Vx lines of part P1, respectively P2, when these pixels are selected. Each reading circuit READOUT1 and READOUT2 for example comprises a plurality of analog-to-digital converters (ADC), preferably one ADC for each Vx line coupled to this reading circuit.


The circuit READOUT1 receives a reference voltage, in this example the ground GND, and the circuit READOUT2 receives a reference voltage, in this example the ground GND. The reference voltage GND of the circuit READOUT1 is electrically decoupled from that of the circuit READOUT2. For example, the reference voltage GND applied to the circuit READOUT1 is provided by a third bonding pad 912 of the sensor 2″, and the other reference voltage GND applied to the circuit READOUT2 is provided by a fourth bonding pad 914 of the sensor 2″. Each bonding pad 912, 914 receives the off-chip reference voltage GND. Each bonding pad 912, 914 acts as a low-pass filter as schematically represented in FIG. 9 by a resistance R and an inductance L series-connected in each bonding pad.



FIG. 12 shows chronograms illustrating operation of the sensor of FIG. 9 according to one embodiment. More specifically, in this example the scene to capture is divided into four areas S1, S2, S3 and S4, and the FIG. 12 shows, depending on time t, the light (“light”) emitted by the illumination device 205 of the sensor 2″, which area S1, S2, S3 or S4 receives the light (“illuminated area of the scene”), which corresponding area M1, M2 of part P1 or M3, M4 of part P2 receives the reflected light and has its pixels integrating light (“integrated area of P1” and “integrated area of P2”), and which area M1, M2 of part P1 or M3, M4 of part P2 is read (“read area of P1” and “read area of P2”). In this example, the device 205 emits light under the form of a burst of periodic pulses of light.


Between an instant t20 and an instant t21 posterior to instant t20, device 205 emits light with the direction O1, towards the area S1 of the scene. The light reflected by this area S1 is received by the corresponding area M1 of part P1 of the matrix. An integration phase of the received light is done in the pixels of the area M1 only, thus only in part P1 of the matrix.


Between the instant t21 and an instant t22 posterior to instant t21, device 205 emits light with the direction O3, towards the area S3 of the scene. The light reflected by this area S3 is received by the corresponding area M3 of part P2 of the matrix. An integration phase of the received light is done in the pixels of the area M3 only, thus only in part P2 of the matrix. In the same time, the area M1 of the part P1 of the matrix is read. More specifically, the reading of the pixels of the area M1 is controlled by control circuit CTRL1″ and is completed by reading the rows of pixels of the area M1 ones after the other.


Between the instant t22 and an instant t23 posterior to instant t22, device 205 emits light with the direction O2, towards the area S2 of the scene. The light reflected by the area S2 is received by the corresponding area M2 of the matrix, and an integration phase is performed in the pixels of the area M2 only, thus only in part P1 of the matrix. In the same time, the area M3 of the part P2 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22.


Between the instant t23 and an instant t24 posterior to instant t23, device 205 emits light with the direction O4, towards the area S4 of the scene. The light reflected by the area S4 is received by the corresponding area M4 of the matrix, and an integration phase is performed in the pixels of the area M4 only, thus only in part P2 of the matrix. In the same time, the area M2 of the part P1 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22.


Between the instant t24 and an instant t25 posterior to instant t24, the area M4 of part P2 of the matrix 200 is read, similarly to the manner the area M1 was read between instants t21 and t22. At the instant t25, a depth map of the scene may be computed. More specifically, the depth map is generated based on the output signals of the pixels of the area M1 read between the instants t21 and t22, of the area M2 read between the instants t22 and t23, of the area M3 read between the instants t23 and t24, and of the area M4 read between the instants t24 and t25.


As it is represented in FIG. 12, between the instants t24 and t24, device 205 may emit light with the direction O1, towards the area S1 of the scene, such that the reflected light is integrated by the area M1 only. This allows to start a new acquisition of the scene to capture, similar to that described between the instants t20 and t25. In another example, a blanking time is provided after the instant t25, and before a new acquisition of the scene implemented as described between instants t20 and t25.



FIG. 13 illustrates, in a very schematic manner, an implementation of the sensor of FIG. 9, according to one embodiment. FIG. 13 is a top view of the disposition of the circuit READOUT1 and READOUT2 relative to the matrix 200.


In this embodiment, circuit READOUT1 is disposed along a first edge of the matrix 200, on the side of the half P1 of the matrix, circuit READOUT2 being disposed along a second edge of the matrix, on the side of the half P2. The first and second edges are parallel. More specifically, the first and second edges are perpendicular to the lines Vx (not shown on FIG. 13).


This disposition of the circuits READOUT1 and READOUT2 relative to the matrix 200 is, for example, used when the circuits READOUT1 and READOUT2 belongs to the same semiconductor substrate than the matrix 200.



FIG. 14 illustrates, in a very schematic manner, an implementation of the sensor of FIG. 9, according to one alternative embodiment. FIG. 14 is a perspective view of the disposition of the circuit READOUT1 and READOUT2 relative to the matrix 200.


In the embodiment of FIG. 14, the matrix belongs to a first semiconductor substrate, and the circuits READOUT1 and READOUT2 belong to a second semiconductor substrate. The first substrate is stacked over the second substrate.


The lines Vx of the part P1 of the matrix 200 are coupled to the circuit READOUT1, for example thanks to an interconnection structure (not shown) which is sandwiched between the first and second substrates. Similarly, the lines Vx of the part P2 of the matrix 200 are coupled to the circuit READOUT2, for example thanks to same interconnection structure. In FIG. 14, only one line Vx is represented in dashed line, in each part P1, P2 of the matrix 200.


Preferably, as shown in FIG. 14, the circuit READOUT1 is disposed below the part P1 of the matrix 200, the circuit READOUT2 being disposed below the part P2 of the matrix.


The embodiment of FIG. 14 allows to obtain a more compact sensor 2″.


Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuits READOUT1 and READOUT2 in order to generate a depth map of a scene.



FIG. 15 illustrates an alternative embodiment of the indirect time of flight sensor 2″ of the FIG. 9. Only the differences between the sensor 2″ of FIG. 9 and the sensor 2″ of FIG. 15 are detailed. In FIG. 15, the two parts P1 and P2 of the matrix 200 are spaced from each other to simplify the illustration of the sensor 2″, although, in practice, these two parts P1 and P2 are adjacent to each other, the part P1 being disposed along the part P2, similarly to what has been described in relation with FIG. 9.


In this alternative embodiment, a first semiconductor substrate comprises the matrix 200, and lies on a second semiconductor substrate. In other words, the two substrates are stacked one over the other.


The sensor 2″ further comprises commutators 1500, only one of the commutators 1500 being referenced in FIG. 15 in order to avoid complicating the Figure. The commutators 1500 belong to the second substrate. Preferably, the commutators 1500 are disposed below the separation between the two parts P1 and P2 of the matrix 200. The sensor 2″ comprises as much commutators 1500 as the half P1 of the matrix 200 comprises lines Vx.


Each commutator 1500 comprises a first input 1501, a second input 1502, an output 1503 and is controlled by a signal Sel. Each commutator 1500 is configured to electrically couple its input 1501 to its output 1503 when signal Sel is in a first state, and to couple its input 1502 to its output 1503 when signal Sel is in a second state.


In the alternative embodiment illustrated by FIG. 15, each commutator 1500 has its input 1501 connected to a line Vx of the part P1 of the matrix 200, and its input 1502 connected to a corresponding line Vx of the part P2 of the matrix 200. A line Vx of the part P2 corresponds to a line Vx of the part P1 when these two lines belong to the same column of the matrix 200. In each column of the matrix 200, a line Vx of the part P2 of the matrix 200 corresponds to a line Vx of the part P1 of the matrix 200, for example, when the line Vx of the part P1 is selectively coupled to given outputs of the pixels of the part P1 disposed in this column, and the line Vx of part P2 is selectively coupled to corresponding outputs of the pixels of the part P2 disposed in said column.


In this alternative embodiment, instead of the two circuits READOUT1 and READOUT2, the sensor 2″ comprises only one reading circuit READOUT3. Preferably, the circuit READOUT3 belongs to the same substrate as the commutators 1500. Although in FIG. 15, the lines Vx of the part P1 seems to pass through the circuit READOUT3, as represented by portions of the lines Vx in dashed lines, in practice this is not the case. Preferably, a reference voltage GND applied to the circuit READOUT3 is provided by a bonding pad 1505 of the sensor 2″, which receives the off-chip reference voltage GND and acts as a low-pass filter as schematically represented in FIG. 15 by a resistance R and an inductance L series-connected in the bonding pad 1505.


Each commutator 1500 has its outputs 1503 coupled, preferably connected, to the circuit READOUT3. The circuit READOUT3, for example, comprises an ADC for each commutator 1500.


A control circuit, for example the control circuit CTRL1″, is configured to control the commutators 1500 such that the output 1503 of each commutator is coupled to the input 1501 of this commutator during a reading of pixels of the half P1 of the matrix, and to the input 1502 of this commutator during a reading of pixels of the half P2 of the matrix. Said in other words, the circuit for controlling the commutators, in this example the control circuit CTRL1″, is configured to provide the signal Sel at its first state during a reading of pixels of the half P1 of the matrix, and at its second state during a reading of pixels of the half P2 of the matrix.


In sensor 2″ of the FIG. 15, when reading pixels of the part P1, respectively P2, of the matrix 200, each line Vx of part P1, respectively P2, is coupled by a corresponding commutator 1500 to the circuit READOUT3 which then receives output signals of theses pixels. Further, when reading pixels of the part P1, respectively P2, of the matrix 200, the circuit READOUT3 is insulated from the lines Vx of part P2, respectively P1, by the commutators 1500.


Compared to the sensor 2″ described in relation with FIG. 9, the sensor 2″ of FIG. 15 is more compact as it comprises only one reading circuit.



FIG. 16 illustrates, in a very schematic manner, an implementation of the sensor 2″ of FIG. 15 according to one embodiment. FIG. 16 is a perspective view of the disposition of the circuit READOUT3 and the commutators 1500 relative to the matrix 200.


As already indicated in relation with FIG. 15, the matrix 200 belongs to a first semiconductor substrate (not represented in FIG. 16), and the commutators 1500 belong to a second semiconductor substrate (not represented in FIG. 16), the first substrate being stacked over the second substrate.


The lines Vx of the parts P1 and P2 of the matrix 200 are, for example, conductive lines of an interconnection structure which is sandwiched between the first and second substrates, only one line Vx of the part P1 and one corresponding line Vx of the part P2 being represented in FIG. 16 in order to avoid complicating the Figure.


The commutators 1500 are disposed below the separation between the parts P1 and P2 of matrix 200, or, said in other words, below the common edge of the parts P1 and P2 of the matrix 200.


In this particular embodiment, the circuit READOUT3 belongs to the same substrate as the commutators 1500. The circuit READOUT3 is preferably disposed below the matrix 200, for example below the part P2 of the matrix as represented in FIG. 16.


Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuit READOUT3 in order to generate a depth map of a scene.


The embodiments described in relation with FIGS. 15 and 16, for example, correspond to a case where the pitch of the inputs of the circuit READOUT3, which are each connected to an output 1503 of a corresponding commutator 1500, is equal to or narrower than the pitch of the pixels 1 of the matrix 200 between two adjacent column of the matrix 200.



FIG. 17 illustrates another alternative embodiment of the sensor 2″ of the FIG. 9. Only the differences between the sensor 2″ of FIG. 15 and the sensor 2″ of FIG. 17 are here detailed.


In this alternative embodiment, the sensor 2″ comprises two reading circuits READOUT4 and READOUT5 instead of the reading circuit READOUT3. Preferably, the circuits READOUT4 and READOUT5 belong to the same substrate as the commutators 1500. Although in FIG. 17, the lines Vx of the part P1, respectively P2, seems to pass through the circuit READOUT4, respectively READOUT5, as represented by portions of the lines Vx in dashed lines, in practice this is not the case. Preferably, a reference voltage GND applied to the circuit READOUT4 is provided by a bonding pad 1700 of the sensor 2″, and a reference voltage GND applied to the circuit READOUT5 is provided by a bonding pad 1702 of the sensor 2″. Each bonding pad 1700 and 1702 receives the off-chip reference voltage GND and acts as a low-pass filter, as schematically represented in FIG. 17 by a resistance R and an inductance L series-connected in each bonding pad.


As in FIG. 15, each commutator 1500 has its input 1501 connected to a line Vx of the part P1 of the matrix and its input 1502 connected to a corresponding line Vx of the part P2 of the matrix.


However, in the embodiment of FIG. 17, each commutator 1500 connected to lines Vx of an odd column of the matrix 200 has its output 1503 connected to the circuit READOUT4, whereas each commutator 1500 connected to lines Vx of an even column of the matrix 200 has its output 1503 connected to the circuit READOUT5. Each circuit READOUT4, REDAOUT5 for example comprises an ADC for each commutator 1500 coupled, preferably connected, to this circuit.


As already indicated in relation with FIG. 15, a control circuit, for example the control circuit CTRL1″, is configured to control the commutators 1500 such that the output 1503 of each commutator is coupled to the first input 1501 of this commutator during a reading of pixels of the half P1 of the matrix, and to the second input 1502 of this commutator during a reading of pixels of the half P2 of the matrix.


In the sensor 2″ of FIG. 17, when reading pixels of the part P1, respectively P2, of the matrix 200, each line Vx of part P1, respectively P2, is coupled by a corresponding commutator 1500 to the circuit READOUT4 when this line Vx belongs to an odd column of the matrix 200 and to the circuit READOUT5 when this line Vx belongs to an even column of the matrix 200, such that each output signal of each of these pixels is received either by the circuit READOUT4 or the circuit READOUT 5. Further, during a reading of pixels of part P1, respectively P2, circuits READOUT4 and READOUT5 are insulated from the lines Vx of part P2, respectively P1, by the commutators 1500.


Preferably, the commutators 1500 are disposed below the separation between the parts P1 and P2 of the matrix 200. Preferably, the circuit READOUT4 is disposed below one of the parts P1 and P2 of the matrix 200, the circuit READOUT5 being disposed below the other one of the parts P1 and P2.



FIG. 18 illustrates, in a very schematic manner, an implementation of the sensor 2″ of FIG. 17 according to one embodiment. FIG. 18 is perspective view of the disposition of the circuits READOUT4 and READOUT5 and of the commutators 1500 relative to the matrix 200.


As already indicated in relation with FIG. 17, the matrix 200 belongs to a first semiconductor substrate (not represented in FIG. 18), and the commutators 1500 belong to a second semiconductor substrate (not represented in FIG. 18), the first substrate being stacked over the second substrate.


The lines Vx of the parts P1 and P2 of the matrix 200 are, for example, conductive lines of an interconnection structure which is sandwiched between the first and second substrates, only one line Vx of the part P1 and one corresponding line Vx of the part P2 being represented in FIG. 18 in order to avoid complicating the Figure.


The commutator 1500 are disposed below the separation between the parts P1 and P2 of matrix 200, or, said in other words, below the common edge of the parts P1 and P2 of the matrix 200.


In this particular embodiment, circuits READOUT4 and REDAOUT5 belong to the same substrate as the commutators 1500. The circuit READOUT4 is disposed below one of the parts P1 and P2 of the matrix 200, the circuit READOUT5 being disposed below the other one of the parts P1 and P2. In the example of FIG. 18, the circuit READOUT4 is disposed below the part P1 of the matrix, the circuit READOUT5 being disposed below the part P2 of the matrix.


Preferably, the second substrate further comprises digital circuits, for example in CMOS technology, for example a circuit for processing signals provided by the circuits READOUT4 and READOUT5 in order to generate a depth map of a scene.


The embodiments described in relation with FIGS. 17 and 18, for example, correspond to a case where the pitch of the inputs of the circuits READOUT4 and READOUT5, which are each connected to an output 1503 of a corresponding commutator 1500, is larger than the pitch of the pixels 1 of the matrix 200 between two adjacent column of the matrix 200.


Various embodiments and variants have been described. Those skilled in the art will understand that certain features of these embodiments can be combined, and other variants will readily occur to those skilled in the art. In particular, although in the above described embodiments, the scene to capture divided into only four areas S1, S2, S3 and S4, the illumination device 205 is configured to direct the light towards each of the areas S1, S2, S3, S4 of the scene by illuminating only one area at a time, and the matrix 200 is divided into four corresponding areas M1, M2, M3 and M4, those skilled in the art are capable to implement embodiment wherein the scene is divided into more (or less) than four areas, the device 205 is configured to independently illuminate each of these areas of the scene, and the matrix 200 is divided into areas such that each area of the matrix corresponds to an area of the scene. Further, those skilled in the art are capable of implementing embodiments in which the pixels of the matrix 200 are different from pixel 1 described in relation with FIG. 1, and in more for a specific example in relation with FIGS. 10 and 11.


Finally, the practical implementation of the embodiments and variants described herein is within the capabilities of those skilled in the art based on the functional description provided hereinabove.

Claims
  • 1. An indirect time of flight sensor, comprising: a matrix of pixels, wherein each pixel in the matrix comprises a photoconversion region coupled to at least two memory circuit sets, each memory circuit set comprising a charge storage region and a controllable transfer device for transferring charge from the photoconversion region to said storage region;first conductive lines extending parallel to each other and configured to transmit first control signals to the controllable transfer devices;a first circuit configured to provide the first control signals to the first conductive lines;an illumination device configured to illuminate a scene that is divided into a plurality of first areas; anda second circuit configured to control the illumination device to successively illuminate each first area;wherein the matrix of pixels is divided into a plurality of second areas, each second area comprising adjacent lines of pixels which are parallel to the first conductive lines, and wherein each first area corresponds to one of the second areas; andwherein the first circuit is configured to provide different first signals to the different second areas with the first signals repeatedly commutated between active and inactive states only for pixels within the second area corresponding to the first area which is illuminated.
  • 2. The sensor according to claim 1, wherein the illumination device comprises an array of laser sources and an optical device configured to direct light emitted by the array of laser sources towards the scene, and wherein: the array of laser sources is divided into a plurality of sets of laser sources, each set of laser sources configured to illuminate a corresponding one of the first areas, the second circuit being configured to control said sets of laser sources one after the other.
  • 3. The sensor according to claim 1, wherein the illumination device comprises an array of laser sources and an optical device configured to direct light emitted by the array of laser sources towards the scene, and wherein: the optical device is configured to direct the emitted light differently depending on a control signal, the second circuit being configured to provide, at each illumination of one of the first areas, said control signal causing a directing of the light towards said one of the first areas.
  • 4. The sensor according to claim 1, wherein: the sensor comprises second conductive lines extending parallel to the first conductive lines and configured to receive output signals of the pixels;each pixel comprises a selection device configured to selectively couple an output of said pixel to at least one corresponding second conductive line; andthe first circuit is configured to provide second control signals to the selection devices via third conductive lines extending perpendicular to the second conductive lines.
  • 5. The sensor according to claim 4, wherein the first circuit is configured to control, using the second signals, a reading of all the pixels after each illumination of one of the first areas before an illumination of a next one of the first areas.
  • 6. The sensor according to claim 4, wherein the second circuit is configured, before each reading of all the pixels controlled by the first circuit, to control several successive illumination cycles each comprising a unique illumination of each first area, and to control an absence of light emission by the illumination device during said reading.
  • 7. The sensor according to claim 1, wherein: the sensor comprises second conductive lines extending parallel to each other and perpendicular to the first conductive lines, the second conductive lines configured to receive output signals of the pixels;each pixel comprises a selection device configured to selectively couple an output of said pixel to at least one corresponding second conductive line; andthe first circuit is configured to provide second control signals to the selection devices via third conductive lines perpendicular the second conductive lines.
  • 8. The sensor according to claim 7, wherein the second circuit is configured, before each reading of all the pixels controlled by the first circuit, to control several successive illumination cycles each comprising a unique illumination of each first area, and to control an absence of light emission by the illumination device during said reading.
  • 9. The sensor according to claim 7, wherein the first circuit is configured to control, after each illumination of one of the first areas, a reading of only the pixels of the second area corresponding to said one of the first areas.
  • 10. The sensor according to claim 9, wherein the second circuit is configured to control an absence of light emission by the illumination device when the first circuit control the reading of the pixels of a second area.
  • 11. The sensor according to claim 9, wherein: the matrix is divided into a first half and a second half, a separation between the first half and the second half being parallel to the first conductive lines, and the second conductive lines of each half ending at said separation;the first circuit is configured to simultaneously control charge transfers in the pixels of a second area of one of the first and second halves and a reading of the pixels of a second area of the other one of the first and second halves;a first part of a semiconductor substrate comprises the first half of the matrix and a second part of said semiconductor substrate comprises the second half of the matrix;insulation structures passing through the semiconductor substrate to insulate said first and second parts of the substrate from each other; anda reference voltage provided to the first part of the semiconductor substrate that is electrically decoupled from a reference voltage provided to the second part of the semiconductor substrate.
  • 12. The sensor according to claim 11, wherein, for each voltage level provided to at least one pixel of the first half of the matrix and, simultaneously, to at least one pixel of the second half of the matrix, the sensor comprises a first generator of said voltage level for the first half and a second generator of said voltage level for the second half, the first and second generators being electrically decoupled from each other.
  • 13. The sensor according to claim 11, comprising a first reading circuit coupled the second conductive lines of the first half of the matrix, and a second reading circuit coupled to the second conductive lines of the second half of the matrix, a reference voltage of the first reading circuit being electrically decoupled from a reference voltage of the second reading circuit.
  • 14. The sensor according to claim 13, wherein the first reading circuit is disposed along a first edge of the matrix, on a side of the first half, and the second reading circuit is disposed along a second edge of the matrix, on a side of the second half, the first and second edges being parallel.
  • 15. The sensor according to claim 11, wherein: the semiconductor substrate comprising the matrix of pixels lies above another semiconductor substrate comprising commutators, the commutators disposed below the separation between the first and second halves of the matrix;each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs; andthe sensor comprises a reading circuit connected to the output of each commutator, the reading circuit provided on the another semiconductor substrate.
  • 16. The sensor according to claim 15, comprising a control circuit configured to control the commutators such that the output of each commutator is coupled to the first input of said commutator during a reading of pixels of the first half of the matrix, and to the second input of said commutator during a reading of pixels of the second half of the matrix.
  • 17. The sensor according to claim 11, wherein: the semiconductor substrate comprising the matrix of pixels lies above another semiconductor substrate comprising commutators, the commutators disposed below the separation between the first and second halves of the matrix;each commutator comprises a first input connected to one of the second conductive lines of the first half, a second input connected to a corresponding second conductive line of the second half, and an output configured to be selectively coupled to one of said inputs;the pixels of the matrix are arranged in columns parallel to the second conductive lines;each commutator connected to second conductive lines of an odd column has its output connected to a first reading circuit;each commutator connected to second conductive lines of an even column has its output connected to a second reading circuit; andthe first and second reading circuits are one the another semiconductor substrate.
  • 18. The sensor according to claim 17, comprising a control circuit configured to control the commutators such that the output of each commutator is coupled to the first input of said commutator during a reading of pixels of the first half of the matrix, and to the second input of said commutator during a reading of pixels of the second half of the matrix.
Priority Claims (1)
Number Date Country Kind
20306680.8 Dec 2020 EP regional