The present invention relates to a method for measuring depth and a Time-of-Flight (ToF) depth sensor.
Depth sensors have become increasingly prevalent in electronic devices. They see heavy use in automotive applications and the manufacturing industry, as well as in Augmented Reality systems and in security applications for mobile devices. Time-of-Flight (ToF) depth sensing technology is widely used in modern depth sensors.
Time-of-Flight sensors work by illuminating a scene with modulated light, typically infrared light, and measuring the light reflected from the scene in order to determine depth values within the scene. Although Time-of-Flight sensors typically make use of modulated light, Time-of-Flight sensors that work using other types of emission, such as ultrasound are in use as well. As Time-of-Flight depth sensors have developed, a desire for higher resolution has led to increased illumination intensities, increased modulation frequency of the illumination light, and increased frequency of measurement of reflected light. The increased measurement frequency and resolutions have in turn increased the data processing required to calculate the depth values. Accordingly, the illuminator and a processor within the Time-of-Flight sensor consume increasing amounts of power, especially when operating with higher light intensity and frequency of measurement. Such higher intensities and frequencies are necessary in certain applications (e.g. rapidly moving devices, distant objects, outdoor environments) and unnecessary in others (e.g. static devices, indoor environment). In some applications, such as employing depth sensors in a headset for extended reality (XR), augmented reality (AR) and mixed reality (MR) applications, the power available may be limited.
A method and apparatus to balance power consumption against required performance is therefore desirable.
According to a first aspect there is provided a time-of-flight (ToF) depth sensor apparatus, comprising an emitter configured to emit a signal towards a scene, which emitter is controllable to adjust at least one of an intensity and a modulation frequency of the signal output from the emitter; at least one signal sensor, configured to detect an intensity of the signal from the emitter that has been reflected by the scene; and a controller configured to receive context information about the scene for depth capture by the time-of-flight depth sensor and to adjust at least one of the intensity and modulation frequency of a signal output by the emitter in dependence on the context information.
According to a second aspect there is provided a method for measuring depth using a Time-of-Flight (ToF) depth sensor apparatus, the method comprising: receiving context information about a scene for depth capture by the time-of-flight depth sensor; adjusting, in dependence on the received context information, a modulation frequency or intensity of a signal to be output from at least one emitter; emitting a signal towards the scene using the at least one emitter with the adjusted intensity or modulation frequency of output signal; detecting, using at least one signal sensor, the intensity of light from the emitter that has been reflected by the scene; and calculating at least one depth measurement based on the measured intensity of the detected signal that has been reflected by the scene.
According to a third aspect there is provided a non-transitory computer readable storage medium comprising instructions that, when executed by a time-of-flight sensor apparatus, cause the time-of-flight sensor apparatus to perform a method comprising: receiving context information about a scene for depth capture by the time-of-flight depth sensor; adjusting, in dependence on the received context information, a modulation frequency or intensity of a signal to be output from at least one emitter; emitting a signal towards the scene using the at least one emitter with the adjusted intensity or modulation frequency of output signal; detecting, using at least one signal sensor, the intensity of signal from the emitter that has been reflected by the scene; and calculating at least one depth measurement based on the measured intensity of the detected signal that has been reflected by the scene.
Embodiments will now be described, by way of example only, with reference to the accompanying drawings in which:
Before discussing particular embodiments with reference to the accompanying figures, the following description of embodiments is provided.
In a first embodiment there is provided a time-of-flight (ToF) depth sensor apparatus comprising an emitter configured to emit a signal towards a scene, which emitter is controllable to adjust at least one of an intensity and a modulation frequency of the signal output from the emitter; at least one signal sensor, configured to detect an intensity of the signal from the emitter that has been reflected by the scene; and a controller configured to receive context information about the scene for depth capture by the time-of-flight depth sensor and to adjust at least one of the intensity and modulation frequency of a signal output by the emitter in dependence on the context information.
In some embodiments the signal may be light. In other embodiments, the signal may be sound waves.
In the context of this application, the term “light” is understood to encompass both visible and nonvisible electromagnetic radiation. In many embodiments this radiation will be in the infrared (IR) wavelength (700 nm to 1 mm). In many embodiments, the intensity of the radiation is controlled not exceed the bounds of the range permitted by appropriate safety regulations.
In some embodiments, the emitter may be configurable to capture images periodically at one of a plurality of refresh rates of the sensor. In such embodiments, the controller may be configured to control the refresh rate of the signal sensor in dependence upon the context information. The refresh rate of the signal sensor may be controlled to be different from that of the emitter.
In some embodiments, the emitter is a single emission source. Alternatively, the emitter may comprise a plurality of emission sources. The controller may be configured to control at least one of the intensity and modulation frequency of each emission source separately. These emission sources may be configured to direct signals to different portions of the scene presented for depth capture, or to not emit signals to one or more portions of the scene. In some embodiments the emitter may comprise a single emission source with a plurality of reflective elements arranged to direct the signal to different portions of the scene.
In some embodiments, the ToF depth sensor apparatus also comprises at least one context sensor. The context sensor may be at least one of: an accelerometer, an image-sensing component, and a location sensing component. In some embodiments, the controller is configured to receive a signal from the context sensor and to adjust at least one of the intensity and frequency of the signal output by the emitter in dependence upon the signal received from the context sensor. Where the context sensor is an accelerometer, the controller may be configured to control the illuminator to increase the modulation frequency in response to a detection of movement by the accelerometer. The ToF depth sensor apparatus, including the context sensor, may form part of a device such as a mobile telecommunications device or a headset device.
In some embodiments in which the context sensor is an image-sensing component, the context sensor may be configured to detect motion in the scene, and the controller may be configured to increase the modulation frequency in response to a detection of movement in the scene.
In some embodiments in which the context sensor is an image-sensing component, the ToF depth sensor apparatus may be configured to, in response to determining that a change between successive images captured by the image-sensing component is greater than a threshold value, perform the aforementioned steps of emitting a signal towards the scene, detecting the intensity of the signal that has been reflected by the scene, and determining at least one depth measurement.
In some embodiments, the at least one signal sensor may comprise multiple raster scan signal sensors configured to detect the signal from the emitter that has been reflected by the scene.
In other embodiments, the time-of-flight sensor apparatus may comprise a single signal sensor comprising a plurality of detectors, which plurality of detectors are grouped across the surface of the sensor into tiles of detectors. The time-of-flight sensor apparatus may be configured to read-out data from the signal sensor on a tile-by-tile basis. In such embodiments, the time-of-flight sensor may be configured to compress pixel values read out from the signal sensor on a tile-by-tile basis. The time-of-flight sensor may be configured to calculate depth values based on a plurality of tile readouts from the signal sensor, each readout occurring at a different time, wherein the calculation of depth values is performed on a tile-by-tile basis.
In a second embodiment there is provided a method for measuring depth using a Time-of-Flight (ToF) depth sensor, the method comprising: receiving context information about a scene for depth capture by the time-of-flight depth sensor; adjusting, in dependence on the received context information, a modulation frequency or intensity of a signal to be output from at least one emitter; emitting a signal towards the scene using the at least one emitter with the adjusted intensity or modulation frequency of output signal; detecting, using at least one sensor, the intensity of a signal from the emitter that has been reflected by the scene; and calculating at least one depth measurement based on the measured intensity of the detected signal that has been reflected by the scene.
The context information may comprise information regarding content of the scene received from an image-sensing device component. In such embodiments, the method may include the step of, in response to determining that a change between successive images captured by the image-sensing device is greater than a threshold value, performing the aforementioned steps of emitting a signal towards the scene, detecting the intensity of a signal that has been reflected by the scene, and determining at least one depth measurement. In cases where a change in the scene is detected, the method may comprise determining a portion of the captured scene in which the change is detected. In such embodiments, one of a plurality of emitters may be controlled in dependence on the portion of the scene in which the change is detected.
In some embodiments, the ToF depth sensor comprises at least one context sensor. The context sensor may be at least one of: an accelerometer, an image-sensing component, and a location sensing component. In some embodiments, the method comprises receiving a signal from the context sensor and adjusting at least one of the intensity and frequency of the signal output by the emitter in dependence upon the signal received from the context sensor. Where the context sensor is an accelerometer, the method may include controlling the emitter to increase the modulation frequency in response to a detection of movement by the accelerometer.
In some implementations, the time-of-flight sensor comprises a single signal sensor having a plurality of detectors, which plurality of detectors are grouped across the surface of the signal sensor into tiles of detectors, wherein the method comprises reading out data from the sensor on a tile-by-tile basis. The method in such implementations may further comprise compressing the readout data from the sensor on a tile-by-tile basis.
The method may also comprise calculating depth values for pixels based on pixel values of a plurality of readouts of the illumination sensor, by obtaining a plurality of tiles of data read out from the same tile of detectors at different times, and calculating the depth values using the obtained plurality of tiles of data.
In a third embodiment there is provided a non-transitory computer readable storage medium comprising instructions that, when executed by a time-of-flight sensor apparatus, cause the time-of-flight sensor apparatus to perform a method comprising: receiving context information about a scene for depth capture by the time-of-flight depth sensor; adjusting, in dependence on the received context information, a modulation frequency or intensity of a signal to be output from at least one emitter; emitting a signal towards the scene using the at least one emitter with the adjusted intensity or modulation frequency of output signal; detecting, using at least one signal sensor, the intensity of signal from the emitter that has been reflected by the scene; and calculating at least one depth measurement based on the measured intensity of the detected signal that has been reflected by the scene.
According to a further embodiment there is provided a signal sensor having a plurality of detectors, which plurality of detectors are grouped across the surface of the sensor into tiles of detectors, wherein the time-of-flight sensor is configured to read-out data from the sensor on a tile-by-tile basis.
The signal sensor may be configured to read-out from the detectors within each tile in row major order. In some embodiments the tiles may be read-out in row major order.
The signal sensor may be an image sensor for a time-of-flight sensor apparatus.
According to a further embodiment, there is provided a method for a signal sensor having a plurality of detectors, which plurality of detectors are grouped across the surface of the sensor into tiles of detectors, the method comprising reading-out data from the sensor on a tile-by-tile basis.
The method may further comprise compressing data read-out from the signal sensor on a tile-by-tile basis.
Particular embodiments will now be described with reference to the Figures.
A further implementation of a ToF depth sensor is shown in
The illumination sensor 12 is contains an array of photodetectors, such as phototransistors or photodiodes. These photodetectors convert the photonic energy of the reflected light beam 26 into electrical current, allowing the sensor to measure intensity of the reflected light beam 26. The photodetectors accumulate charge during an exposure phase and are periodically read-out, during which read-out process charge is drained from each photodetector and measured.
The illumination sensor 12 measures charge from each photodetector. These measured amounts of the electrical current correspond to the photonic energy of the reflected light beam 26 during the exposure time corresponding to the measurement taken.
The illuminator 11 is configured to emit a modulated light beam. The light beam 24 is modulated with a sine or square wave. The modulation is intensity modulation, such that the intensity of the light emitted by the illuminator 11 varies over time.
The ToF sensor 10 receives reflected light beam 26 and, using the illumination sensor 12, samples the modulated light at four separate points in each complete cycle of the modulated emitted beam 24. Each sample point is 90 degrees removed from the point before, so the four samples are evenly spaced throughout the timing cycle of the emitted light beam 24. One example of evenly spaced sample points would be samples taken at 90, 180, 270 and 360 degrees. The values of each sample may be designated Q1 through Q4, with each Q-value representing a separate sample.
Using the sample values Q1 through Q4, the phase shift between emitted light beam 24 and reflected light beam 26, φ, may be calculated as follows:
The phase shift φ is then used to calculate the distance d between system 10 and object 25, as follows:
where c is the speed-of-light constant and f is the modulation frequency of the emitted light beam 24.
In step S31, the depth sensor system emits a first light beam. As shown in
In step S34, the ToF depth sensor 10A adjusts the peak intensity of light emitted by the illuminator 11 in dependence on the maximum distance measurement determined in step S33. This may be done by comparison with a threshold. For example, if the maximum distance within the scene is under five (5) meters, the ToF depth sensor 10A may reduce the peak intensity of the modulated light to 50%. This reduction of illuminator intensity in close-range applications—where greater intensities are unnecessary—allows for a reduction in power consumption by the ToF sensor 10A without compromising accuracy. Any practical scheme for adjusting illuminator intensity in dependence on context information may be used. Such methods may for example include use of multiple thresholds, where illuminator intensity is reduced to a given percentage unless a particular threshold maximum distance is present in the depth measurements
In step S35, the ToF depth sensor 10A emits a second light beam. This second light beam is modulated with an intensity that was set in step S34. In the example provided above, the second light beam is emitted at 50% intensity. As described for steps S31 and S32, and in
In step S37, the ToF depth sensor 10A uses the second reflected light beam received in step S36, along with the second light beam generated in step S35, to calculate the phase shift φ of the second reflected light beam. This phase shift is then used to calculate the distance d between the ToF depth sensor 10A and the object within the scene, as described in relation to
The above example describes controlling peak intensity of the modulated light emitted by the illuminator 11 in response to the maximum distance measurement. It is noted here that the frequency of the light (e.g. infrared light) is not varied, but that the light is intensity modulated so that the frequency/wavelength can be controlled. In other implementations, the modulation frequency may be varied in dependence upon the determined maximum distance measurement. This can be done in order to preserve accuracy of readings. When the maximum distance in the scene presented is such that a full period of the light beam generated elapses before the light beam is reflected, the system may be unable to accurately calculate the depth of the scene. Reducing the modulation frequency will alleviate this, as the period of the light beam will be longer, meaning that greater depths can be accurately measured.
A higher modulation frequency (shorter modulation wavelength) allows for more accurate distance measurements, but at the expense of more frequent sampling and a greater number of calculations to be performed. Inversely, using a lower frequency (longer modulation wavelength) reduces the number of samples per unit time and reduces overall processing, but at the expense of less frequent measurements. Accordingly, in cases where the measured maximum distance is longer, the modulation frequency may be reduced because measurement accuracy at larger distances will be reduced in any case. The reduction in modulation frequency may be done in combination with increasing the peak modulation intensity in order to maintain intensity of the measured reflected light beam 26 when the light travels a further distance.
In further embodiments, other context information may be used in addition to the maximum distance. In some embodiments, the context information may comprise information regarding the reflectivity of objects in the scene, which may be determined by comparing the intensity of the received reflected light to the light intensity emitted from the illuminator 11. If the intensity of the received reflected light is low for a particular distance of object (the object has a low reflectivity), the intensity of light emitted by the illuminator may be increased.
In other embodiments, the context information may comprise information about the background noise of the scene, which may be determined by analyzing the signal-to-noise ratio of the received light. The light received by the illumination sensor 12 will include background light from the scene in addition to the light emitted from the illuminator 11. It is desirable that the received intensity of the background light is not too large relative to the intensity of the reflected light from the illuminator 11. If the intensity of the background light is large (a low signal to noise ratio), the illuminator 11 may be controlled to increase the intensity of emitted light. Such embodiments may be useful for applications of the ToF depth sensor system in outdoor environments in which the levels of background light are likely to be higher.
In step S41, the ToF depth sensor 10B receives information from a context sensor 14B. The context sensor 14B is an accelerometer, and the context information received from the context sensor is information regarding device motion. In step S42, motion of the device is determined based on a signal from the accelerometer.
In step S43, the ToF depth sensor 10B adjusts the modulation frequency of the light to be emitted by the illuminator 11 in dependence on the context information determined in step S42. The adjustment of the modulation frequency may be performed by comparison with a threshold value. For example, if the device determined to be moving at 5 m/s based on measurements from the accelerometer, the modulation frequency may be increased to 100% of the maximum frequency because as explained above, the sampling frequency is determined by the modulation frequency and a higher modulation frequency therefore allows more frequent measurements. This is only one example, and any practical scale or means of relating modulation frequency of light emitted by the illuminator 11 to device movement may be used.
From step S44 onward, the method is very similar to that described in relation to
Some embodiments utilizing multiple illuminators will now be described. As described, in some implementations a single illuminator 11 is used causing the entire scene presented for depth capture to be illuminated with the same intensity and modulation frequency, potentially wasting power. For example, a scene mostly containing objects within five meters of the sensor, but with one quadrant of the scene only having objects fifteen meters away would, with a single illuminator 11, be lit in its entirety at the higher intensity demanded by the greater distance. Accordingly, some embodiments that will now be described include multiple illuminators, each configured to have adjustable intensity and frequency of the modulated light signal emitted, which allow power consumption to be improved. In these embodiments, each illuminator 11 may have illumination parameters configured separately to optimize the balance between power and accuracy.
Each illuminator 51A is positioned to selectively illuminate a particular portion of the scene. In the example shown, each illuminator is positioned to selectively illuminate one quadrant of the scene presented. The lower half of
The arrangement of illuminators presented above is only one embodiment of the invention. Other embodiments may have different numbers of illuminators in different arrangements. One such embodiment may have an illuminator configured to illuminate an upper third of the scene presented for depth capture, an illuminator configured to illuminate a lower third of the same scene, and multiple illuminator configured to illuminate separate areas of a middle third.
In further embodiments, a single illuminator may be used in combination with a plurality of mirrors, which are used to direct the light to different parts of the scene. For example, a micro-mirror array could be used to selectively direct light to different parts of the scene. Such embodiments may allow energy waste due to illuminating unnecessary parts of the scene for which depth measurements are not required to be reduced.
There may be overlap regions of the quadrants 51B that are illuminated by more than one illuminator 52A. Such regions could be ignored in the depth measurement processing or the additional illumination may be taken into account the processing of the measured reflected light.
The embodiments described above control the peak modulation intensity and the modulation frequency in dependence upon the context information received. The modulation frequency selected is a tradeoff between power consumption and accuracy. A higher modulation frequency necessitates more frequent measurements by the illumination sensor and more power consumed performing depth calculations but provides more accurate distance measurements because the wavelength of the square or sinewave is shorter. In some embodiments, depth measurements may be performed at a first modulation frequency by the ToF depth sensor 10 and then periodically the ToF depth sensor 10 may perform measurements with a higher modulation frequency. Such a scheme may be useful for saving power in applications in which the scene is not expected to vary by much. The lower modulation frequency measurements may be used to confirm that the scene hasn't changed, and the higher modulation frequency measurements may be used to provide a more accurate measurement.
In other implementations, the ToF sensor 10 may by default make measurements using the first lower modulation frequency mentioned above and only make a measurement at the higher second modulation frequency if a predetermined difference is determined between infrared images measured by the illumination sensor at the first lower modulation frequency.
In a yet further implementation, the second context sensor 14B may be a video camera and a depth measurement at the higher second modulation frequency may be triggered by a detection of motion in the video camera feed.
In implementations in which multiple illuminators 52A, such as the four illuminators shown in
Referring to
Light from the incoming light pulse strikes the photosensitive element 61 and is converted into current, which travels through the gates G1 and G2. Depending on the delay of the reflected light, each gate will receive a different proportion of the current for each pulse. The current is directed through the relevant gate G1 or G2 and passed to respective capacitor S1 or S2. The capacitors S1 and S2 act as summation elements, storing the electrons collected by the photosensitive element 61 and passed through the relevant gate G1 or G2. The voltage in each capacitor S1 or S2 at the end of the integration period is the signal produced by the photodetector and is used to calculate phase delay.
The switches (gates G1 and G2 above) are operated such that a first reading of the photosensitive element occurs at the same time that a light pulse is emitted and a second reading of the photosensitive element is made during a time period adjacent and following the time period during which the pulse of light is emitted. These two time periods can be seen on the third and fourth lines of
The bottom part of
The above example is simplified and a ToF sensor 10 would typically look at four or eight time periods to determine the phase of the reflected pulse of light. The circuit shown in
The method described above in connection with
As mentioned in connection with
71A shows a conventional raster readout of the type commonly found in the prior art, wherein each line of photodetectors is read out sequentially. Processing measurements from multiple lines of photodetectors often requires storing these lines of data on local memory. Examples which follow make use of a tiled readout that allows data from the sensor to be processed in tiles. Processing the data in tiles allows memory and processing efficiencies that can lower power consumption.
72A and 73A of
The use of tile-based illumination sensors within a ToF sensor 10 allows for tile-by-tile processing by the ToF depth sensor. As noted above, a first feature of such tile-based image sensors is that the use of delay lines may be reduced because when processing on a tile-by-tile basis it is only necessary to be concerned about time delay of read-out within the tile rather than across the whole sensor when performing a conventional full image sensor raster scan.
A second feature of tiled readout patterns is the ability to improve compression of the readout data for data storage and retrieval within the ToF sensor 10. Several tile-based image compression algorithms are known, such as AFBC (ARM Frame Buffer Compression) and other compression algorithms including lossy compression algorithms. A reason for compressing on a tile-by-tile basis is that there is likely to be greater spatial correlation between measured values within a 2D region, for example a tile, due to greater spatial locality between the photodetectors within a tile compared to a conventional raster scan. This tends to lead to better compression performance.
The use of a tiled readout pattern may also improve image processing and reduce resource requirements such as buffers. Image processing is likely to be performed on a 2D region of data. When performing such image processing using image data stored in a conventional raster scan format, multiple lines of data will need to be read out and buffered. In contrast, using a tile format significantly reduces buffering requirements and allows specific selected tiles to be processed separately from the rest of the image, preventing wasted processing power. For example, when independently illuminating quadrants of the scene as described previously, the quadrants of the scene may be selectively retrieved from memory and processed. Furthermore, denoising and filtering operations may be performed more readily on tiles of data than on lines of data. More specifically for the ToF sensor 10, the depth calculations may be performed on a tile-by-tile basis allowing depth calculations to be performed selectively for selected parts of a scene being captured by the illumination sensor 12. As explained above, multiple images from different times are required to determine the phase shift and perform a distance calculation. If this calculation is performed on a tile-by-tile basis it is possible to retrieve data only for the selected tile, thereby reducing memory requirements when performing the calculations and avoiding the need to retrieve, for example, four complete image sensor read outs to perform a calculation.
Embodiments described above use a single sensor and perform a tile-based readout from that sensor before processing the read-out data in tiles. However, in other implementations, multiple image sensors could be arrayed and each image sensor in the array could be readout and processed sensor-by-sensor in a similar manner to the tile-by-tile processing described above.
The ToF depth sensors described above have been sensors that use a phase shift in the returned light relative to the emitted light to calculate distance. The teaching above may be applied to other types of depth sensor, such as direct time-of-flight sensors which directly measure the time of flight of a laser pulse that leaves the sensor and reflects back to the sensor. Such direct time-of-flight sensors may not have a modulation frequency of the emitted light, but the intensity of the emitted light may be varied in accordance with the techniques described above.
Embodiments have been described above in which the frequency of measurements by the illumination sensor is controlled to correspond to the modulation frequency of the light emitted by the illuminator. In some further embodiments an illuminator may illuminate a scene and a plurality of illumination sensors or portions of an illumination sensor may be controlled to measure distance with differing measurement frequencies for different parts of the scene. For example, if motion is detected in a particular quadrant of a scene, a portion of an illumination sensor corresponding to that quadrant may read out measurements or calculate depth values for that quadrant at a higher frequency than read out measurements or calculation of depth information for the other quadrants. For example, the higher frequency may correspond to a depth measurement performed at the same frequency as the modulation frequency of light emitted by the illuminator, whereas the lower frequency, associated with the portions of the illumination sensor recording light from other quadrants, may correspond to a readout or calculation every other cycle of the modulation frequency of the emitted light.
Further, although light-based ToF depth sensors have been described above, other embodiments may make use of, for example, an ultrasound emitter and receiver to measure the distance.
Various time-of-flight sensors have been described above. The time-of-flight sensors may be included in various hardware including, but not limited to, extended reality headsets, mobile phones, vehicle sensors, robotics and surveillance cameras. In cases where processing to map the surrounding environment is desired, depth information from the time-of-flight sensors may be used with software, such as simultaneous location and mapping (SLAM), to form a map and keep track of location within that map.
The methods described herein may be embodied in software, wholly in hardware or in any combination thereof. Where a software implementation is used, examples may comprise a computer-readable medium, which may be a non-transitory computer-readable medium, comprising computer-executable instructions that, when executed by a processor instruct the processor to carry out the method.