The present disclosure relates to the field of robotic vision, and in particular to low power high resolution three-dimensional (3D) robotics vision techniques.
The field of robotic vision has expanded tremendously recently. In both home and industrial markets, robotic vision is helping improve efficiency, safety and mobility. From home devices such as robotic vacuum cleaners to industrial assembly line robots, there is a need for 3D robotic vision. Autonomous vehicles such as drones and self-driving automobiles also have a great need for 3D robotic vision. While optical 3D sensing has the potential to provide the highest resolution compared with other sensing modalities such as ultrasound and millimeter wave technology, current 3D optical sensor devices rely on large sensor arrays and consume significant amounts of power, but produce results with limited resolution.
In one example, a system for three-dimensional robotic vision includes a light source and a transmit optical element. The transmit optical element is configured to project a virtual array pattern of laser points that illuminates a scene. The system also includes a receive imaging lens configured to receive reflected light from the scene and project the light onto a pixel array. The system further includes a processing element coupled to the pixel array and configured to generate a three-dimensional map of the scene from information generated by the pixel array responsive to receiving the reflected light.
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate an implementation of apparatus and methods consistent with the present disclosure and, together with the detailed description, serve to explain advantages and principles consistent with the disclosure. In the drawings,
In this description: (a) references to numbers without subscripts are understood to reference all instance of subscripts corresponding to the referenced number; and (b) reference to “one embodiment” or to “an embodiment” means that a particular feature, structure or characteristic described in connection with the embodiments is included in at least one embodiment of the invention, and multiple references to “one embodiment” or “an embodiment” should not be understood as necessarily all referring to the same embodiment.
Although some of the following description is written in terms that relate to software or firmware, embodiments can implement the features and functionality described herein in software, firmware, or hardware as desired, including any combination of software, firmware, and hardware. References to daemons, drivers, engines, modules, or routines should not be considered as suggesting a limitation of the embodiment to any type of implementation.
This kind of flood illumination has an optical power loss proportional to
in both the transmit and receive paths. Low intensity light may be used for safety and power consumption reasons, but low intensity light is optically power inefficient compared to using high intensity light illumination techniques described below. In addition, using the large pixel array 140, which is required to obtain the required resolution, is electrically power inefficient. Each sensor pixel in the large pixel array 140 would need a dedicated analog front end (AFE). Digitizing the large pixel array 140 is computationally intensive, which means that the power utilization of this prior art technique is large. In addition, the amount of time required to process all 4800 pixels means that slow data access can cause image distortion and artifacts in the digital images produced from the pixel data.
Other approaches such as lidar systems use motorized or optical scanning of a scene that require complex mechanical apparatus, making them inappropriate for many type of 3D robotic vision applications.
DOEs 320 are optical elements that diffract the light passing through them from a single coherent light source into a pattern of points. DOEs 320 are typically made of plastic or glass, and can be made inexpensively. DOEs 320 can be configured with any desired pattern, such as regular array pattern 330 or irregular pseudo-random pattern 340 illustrated in
Although described here in terms of laser light sources and DOEs, other light sources and other techniques for generating a virtual array of points can be used. For example, display devices based on optical micro-electro-mechanical technology using digital micro-mirrors, such as a DLP® imaging device from Texas Instruments, can be used instead of a DOE. Phased arrays can also be used for steering a light source to form a virtual array instead of a DOE.
Instead of flooding the scene, light from each of the four lasers can be diffracted into a collection of points, such as 4×4 array 225. Light reflected back from the scene from each of the 4×4 arrays of points 225 is captured by single pixel of a small pixel array 230, such as a 300 pixel 20×15 array in one embodiment. At any given moment in time in this embodiment, the pixel sees only the four points from one of the four lasers of the laser array 310, but with a computational technique described below, all four points can be resolved.
Although there is optical loss, by using a laser dot projection technique such as this, the optical power loss is proportional to
in only me receive pain, so mere is less optical power loss than in the technique illustrated in
A further aspect relates to the time modulation of the light, which is explained in
E=N·S·Tp
where N is the number of pulses, S is the peak optical signal power in watts and Tp is the duration of one pulse (half power, full width). For diodes, the peak optical power is proportional to current. As a result, when the noise in the system is dominated by the analog circuitry (e.g. photodiode shot noise, AFE noise, etc.), the SNR ∝ S2·N·Tp. The safety of a light source is related to the total number of photons per second transmitted by the light source (i.e. the amount of optical power than can be transmitted onto a given field-of-view). As a result, the amplitude of the signal S can be calculated as
where E=the energy of the light transmitted. Finally, the precision of the detector P is proportional to the ratio of the pulse width to the square root of the SNR, or
As a result, one can perform system tradeoffs for the two illumination techniques.
As an example of flood illumination, a robotic vision system requires an amplitude S=5 mA, a number of pulses N=20 k, and a pulse width Tp=20 ns, leading to a total time Tf=4 ms. In contrast, the proposed discrete spatial modulation system as illustrated in
A multipath resolution algorithm may be used to identify objects in the scene. One such algorithm is the Multi-Signal Classification (MUSIC) algorithm, but other multipath resolution algorithms may be used. The MUSIC technique involves eigenvalue decomposition on the sample covariance matrix:
Eigenvalue decomposition forms an orthonormal basis to separate the signal space as orthogonal to the noise space. Forming a one-dimensional searching (steering) vector:
The signal space, which is orthogonal to the noise subspace can be spanned, forming a MUSIC spectrum:
Peaks in the MUSIC spectrum correspond to signal sources. The resolution depends upon the pulse rate of the signal source. For example, in
By combining a low-cost laser array, with a low cost optical element such as a DOE, and using a multipath algorithm for processing the information received by a small pixel array, a high-resolution, low-power system for 3D robotic vision can be achieved.
Modifications are possible in the described embodiments, and other embodiments are possible, within the scope of the claims.
This application claims priority to U.S. Provisional Application No. 62/786,906, filed Dec. 31, 2018, which is hereby incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
20160306032 | Schwarz | Oct 2016 | A1 |
20170003392 | Bartlett | Jan 2017 | A1 |
20170328989 | Bartlett | Nov 2017 | A1 |
20170328990 | Magee | Nov 2017 | A1 |
20200018837 | Baron | Jan 2020 | A1 |
20200166616 | East-Lavoie | May 2020 | A1 |
20200209984 | Li | Jul 2020 | A1 |
20210124047 | Egawa | Apr 2021 | A1 |
Number | Date | Country | |
---|---|---|---|
20200209359 A1 | Jul 2020 | US |
Number | Date | Country | |
---|---|---|---|
62786906 | Dec 2018 | US |