The present disclosure is related to light detection and ranging (LIDAR) systems.
Conventional Frequency-Modulated Continuous-Wave (FMCW) LIDAR systems include several possible sources of noise or phase impairments such as laser phase noise, circuitry phase noise, flicker noise, drift over temperature/weather, and chirp rate offsets. These impairments can make detection of secondary peaks challenging, increase false alarm and range/velocity bias, and increase the error in estimated target range/velocity.
The present disclosure describes various embodiments of LIDAR systems and methods that, among other things, facilitate selecting multiple returns or peaks. One embodiment of the present disclosure relates to a light detection and ranging (LIDAR) system comprising: an optical beam source to transmit a first optical beam to a target, a photo detector to receive a return from the target, and signal processing circuitry. The signal processing circuitry is to threshold a frequency domain waveform to identify a plurality of peaks above a threshold level corresponding to a plurality of targets, apply a primary peak selection to the frequency domain waveform to identify a primary peak from the plurality of peaks, apply a secondary peak selection to a portion of the frequency domain waveform outside a guard-band area to identify a secondary peak from the plurality of peaks that is outside the guard-band area, and determine range and velocity information corresponding to the primary peak and secondary peak.
Another embodiment relates to a method of selecting multiple returns in a light detection and ranging (LIDAR) system. The method comprises thresholding a frequency domain waveform to identify a plurality of peaks above a threshold level corresponding to a plurality of targets and applying a primary peak selection to the frequency domain waveform to identify a primary peak from the plurality of peaks. The method further comprises applying a secondary peak selection to a portion of the frequency domain waveform outside a guard-band area to identify a secondary peak from the plurality of peaks that is outside the guard-band area. The method further comprises determining range and velocity information corresponding to the primary peak and secondary peak.
Other embodiments of the present disclosure relate to a light detection and ranging (LIDAR) system comprising: an optical beam source to transmit a first optical beam to a target, a photo detector to receive a return from the target, and signal processing circuitry including a memory storing instructions which, when executed, cause the signal processing circuitry to: threshold a frequency domain waveform to identify a plurality of peaks above a threshold level corresponding to a plurality of targets and apply a primary peak selection to the frequency domain waveform to identify a primary peak from the plurality of peaks. The signal processing circuitry is further to apply a secondary peak selection to a portion of the frequency domain waveform outside a guard-band area to identify a secondary peak from the plurality of peaks that is outside the guard-band area and determine range and velocity information corresponding to the primary peak and secondary peak; and feed the range and velocity information to a point cloud.
For a more complete understanding of the various examples, reference is now made to the following detailed description taken in connection with the accompanying drawings in which like identifiers correspond to like elements.
The present disclosure describes various examples of LIDAR systems and methods to detect multiple returns in FMCW LIDAR systems. According to some embodiments, the described LIDAR system may be implemented in any sensing market, such as, but not limited to, transportation, manufacturing, metrology, medical, augmented reality, virtual reality, and security systems. According to some embodiments, the described LIDAR system is implemented as part of a front-end of frequency modulated continuous-wave (FMCW) device that assists with spatial awareness for automated driver assist systems, or self-driving vehicles.
Free space optics 115 may include one or more optical waveguides to carry optical signals, and route and manipulate optical signals to appropriate input/output ports of the active optical circuit. The free space optics 115 may also include one or more optical components such as taps, wavelength division multiplexers (WDM), splitters/combiners, polarization beam splitters (PBS), collimators, couplers or the like. In some examples, the free space optics 115 may include components to transform the polarization state and direct received polarized light to optical detectors using a PBS, for example. The free space optics 115 may further include a diffractive element to deflect optical beams having different frequencies at different angles along an axis (e.g., a fast-axis).
In some examples, the LIDAR system 100 includes an optical scanner 102 that includes one or more scanning mirrors that are rotatable along an axis (e.g., a slow-axis) that is orthogonal or substantially orthogonal to the fast-axis of the diffractive element to steer optical signals to scan an environment according to a scanning pattern. For instance, the scanning mirrors may be rotatable by one or more galvanometers. Objects in the target environment may scatter an incident light into a return optical beam or a target return signal. The optical scanner 102 also collects the return optical beam or the target return signal, which may be returned to the passive optical circuit component of the optical circuits 101. For example, the return optical beam may be directed to an optical detector by a polarization beam splitter. In addition to the mirrors and galvanometers, the optical scanner 102 may include components such as a quarter-wave plate, lens, anti-reflective coated window or the like.
To control and support the optical circuits 101 and optical scanner 102, the LIDAR system 100 includes LIDAR control systems 110. The LIDAR control systems 110 may include a processing device for the LIDAR system 100. In some examples, the processing device may be one or more general-purpose processing devices such as a microprocessor, central processing unit, or the like. More particularly, the processing device may be complex instruction set computing (CISC) microprocessor, reduced instruction set computer (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets. The processing device may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like.
In some examples, the LIDAR control systems 110 may include a signal processing unit 112 such as a DSP. The LIDAR control systems 110 are configured to output digital control signals to control optical drivers 103. In some examples, the digital control signals may be converted to analog signals through signal conversion unit 106. For example, the signal conversion unit 106 may include a digital-to-analog converter. The optical drivers 103 may then provide drive signals to active optical components of optical circuits 101 to drive optical sources such as lasers and amplifiers. In some examples, several optical drivers 103 and signal conversion units 106 may be provided to drive multiple optical sources.
The LIDAR control systems 110 are also configured to output digital control signals for the optical scanner 102. A motion control system 105 may control the galvanometers of the optical scanner 102 based on control signals received from the LIDAR control systems 110. For example, a digital-to-analog converter may convert coordinate routing information from the LIDAR control systems 110 to signals interpretable by the galvanometers in the optical scanner 102. In some examples, a motion control system 105 may also return information to the LIDAR control systems 110 about the position or operation of components of the optical scanner 102. For example, an analog-to-digital converter may in turn convert information about the galvanometers' position to a signal interpretable by the LIDAR control systems 110.
The LIDAR control systems 110 are further configured to analyze incoming digital signals. In this regard, the LIDAR system 100 includes optical receivers 104 to measure one or more beams received by optical circuits 101. For example, a reference beam receiver may measure the amplitude of a reference beam from the active optical component, and an analog-to-digital converter converts signals from the reference receiver to signals interpretable by the LIDAR control systems 110. Target receivers measure the optical signal that carries information about the range and velocity of a target in the form of a beat frequency, modulated optical signal. The reflected beam may be mixed with a signal from a local oscillator. The optical receivers 104 may include a high-speed analog-to-digital converter to convert signals from the target receiver to signals interpretable by the LIDAR control systems 110. In some examples, the signals from the optical receivers 104 may be subject to signal conditioning by signal conditioning unit 107 prior to receipt by the LIDAR control systems 110. For example, the signals from the optical receivers 104 may be provided to an operational amplifier for amplification of the return signals and the amplified signals may be provided to the LIDAR control systems 110.
In some applications, the LIDAR system 100 may additionally include one or more imaging devices 108 configured to capture images of the environment, a global positioning system 109 configured to provide a geographic location of the system, or other sensor inputs. The LIDAR system 100 may also include an image processing system 114. The image processing system 114 can be configured to receive the images and geographic location, and send the images and location or information related thereto to the LIDAR control systems 110 or other systems connected to the LIDAR system 100.
In operation according to some examples, the LIDAR system 100 is configured to use nondegenerate optical sources to simultaneously measure range and velocity across two dimensions. This capability allows for real-time, long range measurements of range, velocity, azimuth, and elevation of the surrounding environment.
In some examples, the scanning process begins with the optical drivers 103 and LIDAR control systems 110. The LIDAR control systems 110 instruct the optical drivers 103 to independently modulate one or more optical beams, and these modulated signals propagate through the passive optical circuit to the collimator. The collimator directs the light at the optical scanning system that scans the environment over a preprogrammed pattern defined by the motion control system 105. The optical circuits 101 may also include a polarization wave plate (PWP) to transform the polarization of the light as it leaves the optical circuits 101. In some examples, the polarization wave plate may be a quarter-wave plate or a half-wave plate. A portion of the polarized light may also be reflected back to the optical circuits 101. For example, lensing or collimating systems used in LIDAR system 100 may have natural reflective properties or a reflective coating to reflect a portion of the light back to the optical circuits 101.
Optical signals reflected back from the environment pass through the optical circuits 101 to the receivers. Because the polarization of the light has been transformed, it may be reflected by a polarization beam splitter along with the portion of polarized light that was reflected back to the optical circuits 101. Accordingly, rather than returning to the same fiber or waveguide as an optical source, the reflected light is reflected to separate optical receivers. These signals interfere with one another and generate a combined signal. Each beam signal that returns from the target produces a time-shifted waveform. The temporal phase difference between the two waveforms generates a beat frequency measured on the optical receivers (photodetectors). The combined signal can then be reflected to the optical receivers 104.
The analog signals from the optical receivers 104 are converted to digital signals using ADCs. The digital signals are then sent to the LIDAR control systems 110. A signal processing unit 112 may then receive the digital signals and interpret them. In some embodiments, the signal processing unit 112 also receives position data from the motion control system 105 and galvanometers (not shown) as well as image data from the image processing system 114. The signal processing unit 112 can then generate a 3D point cloud with information about range and velocity of points in the environment as the optical scanner 102 scans additional points. The signal processing unit 112 can also overlay a 3D point cloud data with the image data to determine velocity and distance of objects in the surrounding area. The system also processes the satellite-based navigation location data to provide a precise global location.
The embodiment shown in
In one example embodiment, the thresholding metric used can be SNR, while the peak selection metric is intensity. In such an embodiment, SNR is used to eliminate all frequencies below the threshold value MT(f), and then the primary peak 603 is selected as the highest intensity peak for the frequencies that passed the thresholding operation. After selecting a primary peak 603, a secondary peak 601 can be selected using the same peak selection metric, or a different metric. For example, the primary peak can be selected based on intensity with thresholding using SNR, while the secondary peak can be selected based on SNR and with thresholding using SNR. For third or more returns, the search can exclude all the frequencies corresponding to peaks that have already been selected. In order to prevent detecting side lobes of the primary peak, a secondary peak search could be performed on all frequencies that are outside of a guard band 605 for the primary peak 603.
In one example embodiment, the peak selection operation could be formulated as the following optimization where different likelihood metrics are used for thresholding M2(f) and peak selection M1(f):
Subject to M2(f)≥MT(f), where MT(f) is the threshold required. Similarly, the multiple peak calculation could be determined based on the same metrics:
Subject to M2(f)≥MT(f), where f∉P1 indicates that the search is performed on all frequencies except the ones that are part of the primary peak P1. In practice, P1 can be defined based on a guard band, as detailed below.
According to another embodiment, a highest frequency peak 707 with the highest-frequency detection above a given threshold can be selected as the secondary peak. This peak corresponds to the farthest peak detected from the sensors. This peak selection strategy is biased towards detections to targets that appear farther away in range, and may favor true targets in the presence of fog or other sparse targets (e.g., chain link fences). If the farthest peak is desired, the process can begin by searching the highest frequency detections working backward, and stop after the first peak is encountered, since this would correspond to the highest-frequency detection outside of the primary peak.
According to another embodiment, the lowest frequency peak 701 with the lowest-frequency detection above a given threshold can be selected as the secondary peak. This peak corresponds to the closest peak detected from the sensors. This peak selection strategy is biased towards detections to targets appearing closest to the sensor in range, and may be useful to detect occlusions or blockages on the sensor window, or other close-range objects. If the closest peak is desired, the process can begin by searching the lowest frequency detections and stop after the first peak is encountered, since this would correspond to the lowest-frequency detection outside of the primary peak.
In one embodiment, the techniques described in
In an embodiment, the technique disclosed in
The techniques described in
In an example where three peaks are selected (P=3), let N=1024, M=64, and G=16. The memory requirements for such an operation would be: 64+2*16=96. This is a significant memory saving compared to the full-band operation which would require 1024. Specifically, the memory saving in such an example would be: 1024−96=928.
For each sample the new input magnitude is compared with the current peaks. If the new input is less than all three peaks, nothing is done. For a case where the new input is less than P1 and P2, but greater than P3, if the new value is outside the guard band for P1 or P2, the top of the buffer for P3 is updated with the new input. For a case where the new input is less than P1, but greater than P2 and P3, if the new value is outside the guard band of P1, the top of the buffer for P2 is updated with the new input, and the top of the buffer for P3 is updated with the bottom of P2. For a case where the new input is greater than all three peaks, the top of the buffer for P1 is updated with the new input, the top of the buffer for P2 is updated with the bottom of P1, and the top of the buffer for P3 is updated with the bottom of P2. The buffers are then pushed down by one sample, and another input can be analyzed.
According to the processes disclosed in
At operation 1703 primary peak selection is applied to the frequencies that pass the thresholding operation at 1701. The peak selection can be based on SNR values, intensity values, or other metrics in other embodiments.
At operation 1705, a primary peak is identified based on the primary peak selection.
At operation 1707, a secondary peak selection is applied. The secondary peak selection can be based on SNR values, intensity values, or other metrics, and can be applied to frequencies outside of a guard band of the primary peak identified at operation 1705.
At operation 1709, a secondary peak is identified based on the secondary peak selection. A secondary peak can be identified based on the highest intensity peak outside of the primary peak, the lowest frequency peak outside of the primary peak, or the highest intensity peak outside of the primary peak, in various embodiments.
At operation 1711, range and velocity information is determined for each of the peaks detected. Each of the peaks identified corresponds to a target or object. Thus, the techniques disclosed herein allow the system to determine range and velocity information (as well as other information such as reflectivity, etc.) for a number of targets or objects. In some embodiments, the range and velocity information can be fed into a point cloud.
In some embodiments, where more than two peaks are desired, the process can continue applying the same or different peak selection techniques in order to identify additional peaks, as discussed above. Thus, the techniques disclosed herein can be used to detect multiple targets, and to determine range and velocity data for any number of targets. The range and velocity data for each target can be fed into a point cloud, in some embodiments.
The various operations and methods described in the present application can be performed, in some embodiments, using the signal processing unit 112, signal conversion unit 106, or the signal conditioning unit 107 described above in
The preceding description sets forth numerous specific details such as examples of specific systems, components, methods, and so forth, in order to provide a thorough understanding of several examples in the present disclosure. It will be apparent to one skilled in the art, however, that at least some examples of the present disclosure may be practiced without these specific details. In other instances, well-known components or methods are not described in detail or are presented in simple block diagram form in order to avoid unnecessarily obscuring the present disclosure. Thus, the specific details set forth are merely exemplary. Particular examples may vary from these exemplary details and still be contemplated to be within the scope of the present disclosure.
Any reference throughout this specification to “one example” or “an example” means that a particular feature, structure, or characteristic described in connection with the examples are included in at least one example. Therefore, the appearances of the phrase “in one example” or “in an example” in various places throughout this specification are not necessarily all referring to the same example.
The term “coupled,” along with its derivatives, is used to indicate that two or more elements interact with each other. These coupled elements may or may not be in direct physical or electrical contact with each other.
Although the operations of the methods herein are shown and described in a particular order, the order of the operations of each method may be altered so that certain operations may be performed in an inverse order or so that certain operation may be performed, at least in part, concurrently with other operations. Instructions or sub-operations of distinct operations may be performed in an intermittent or alternating manner.
The above description of illustrated implementations of the invention, including what is described in the Abstract, is not intended to be exhaustive or to limit the invention to the precise forms disclosed. While specific implementations of, and examples for, the invention are described herein for illustrative purposes, various equivalent modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize. The words “example” or “exemplary” are used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “example” or “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs. Rather, use of the words “example” or “exemplary” is intended to present concepts in a concrete fashion. As used in this application, the term “or” is intended to mean an inclusive “or” rather than an exclusive “or”. That is, unless specified otherwise, or clear from context, “X includes A or B” is intended to mean any of the natural inclusive permutations. That is, if X includes A; X includes B; or X includes both A and B, then “X includes A or B” is satisfied under any of the foregoing instances. In addition, the articles “a” and “an” as used in this application and the appended claims should generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form. Furthermore, the terms “first,” “second,” “third,” “fourth,” etc. as used herein are meant as labels to distinguish among different elements and may not necessarily have an ordinal meaning according to their numerical designation.