The present invention relates to an object-position detector apparatus and method for estimating and detecting a position of an object based on radar information from a radar apparatus, for example.
In general, a radar apparatus detects a moving object as a target by applying peak detection (CFAR (Constant False Alarm Rate)) or cluster analysis to a reflection intensity and Doppler velocity in an observation area generated from a received radar signal.
For example, Patent Document 1 discloses a radar apparatus capable of suppressing erroneous detection and extracting only a target object with high accuracy. The radar apparatus transmits and receives pulses or continuous waves to create range-Doppler (RD) data from N times (N≥1) of coherent pulse interval (CPI) data and extracts a cell of a range-Doppler axis exceeding a predetermined threshold for the RD data. Then, the radar apparatus selects a representative value of a cluster to be a target candidate by analyzing the cluster using the extracted cell, extracts range-Doppler of the target from the representative value, performs at least any one of range measurement, speed measurement, and angle measurement processing, and then, outputs a target observation value.
However, as compared with an optical system sensor, since it is difficult for the radar apparatus to increase the angle separation resolution due to the restriction of the number of antennas, such a problem is caused that it is difficult to detect an object in such a situation that a plurality of adjacent objects or structures are close to each other such as reflected waves interfere with each other by a method such as CFAR or cluster analysis.
For example, when the CFAR method is used, in a case where two objects are close to each other at the same range and at a range narrower than the radar-specific angular resolution, reflected waves from the two objects interfere with each other, and it becomes difficult to detect the peak of the reflected wave source derived from the target. In addition, when the clustering method is used, it is difficult to separate clusters in a case where two objects move close to each other at the same speed.
An object of the present invention is to solve the above problems and to provide an object-position detector apparatus and method capable of detecting an object with high accuracy even in such a situation that a plurality of adjacent objects or structures are close to each other such that reflected waves interfere with each other, as compared with the prior art.
According to one aspect of the present invention, an object-position detector apparatus is provided that detects a position of an object based on a radar signal from a radar apparatus. The object-position detector apparatus includes a position estimator unit configured to estimate presence or absence of and a position of the object using a machine learning model learned by predetermined training image data representing the position of the object, based on image data including the radar signal, and output image data representing the presence or absence of and the position of the estimated object.
Therefore, according to the object-position detector apparatus and the like of the present invention, it is possible to detect an object with high accuracy even in such a situation that a plurality of adjacent objects or structures are close to each other such that reflected waves interfere with each other.
Hereinafter, embodiments according to the present invention will be described with reference to the drawings. It is noted that the same or similar components are denoted by the same reference numerals.
The present inventors found such a problem that counting does not work well in an environment where multipath fading or clutter occurs in a tunnel, a place where a soundproof wall is close, or the like in an application of measuring the number of passing vehicles using a radar apparatus, for example, and have devised an embodiment of the present invention as means for solving the problem.
In the present invention, a method for estimating a wave source of a radar signal by machine learning is devised instead of a method for combining CFAR and clustering as means for estimating a position of a target, and an image in which a label corresponding to the position of the target is drawn is generated by using time difference time-series information generated by signal processing and a machine learning model for image recognition. As a result, it is characterized in that, even in such a situation of problem caused in CFAR or clustering, the position of each target is estimated with high accuracy. Hereinafter, embodiments will be described.
In the object estimation of
Referring to
Further, a wireless receiver unit of the radar apparatus 1 receives a wireless signal reflected by a target using, for example, an array antenna including a plurality of antennas, and then mixes the wireless signal with a transmission wireless signal of a wireless transmitting unit and performs low-pass filtering to calculate a plurality of beat signals corresponding to respective antennas and outputs the beat signals to the signal processor unit 2.
Referring to
Referring to
It is noted that, in the processing of the time difference calculator unit 31, for example, as described above, the backward difference calculation is sequentially performed for each predetermined number of frames in time series on the two-dimensional image data of each pair of time frames temporally adjacent to each other among the image data including the information of the reflection signal intensity as the pixel value, and the estimation frequency is once with respect to the number of frames related to the predetermined number of frames in time series. On the other hand, the present invention is not limited thereto, and the processing of the time difference calculator unit 31 may be performed, for example, as described above, while sequentially shifting the backward difference calculation for the two-dimensional image data of each pair of time frames temporally adjacent to each other among the image data including the information of the reflection signal intensity as the pixel value for each frame in time series. In this case, the estimation frequency is once for one frame.
In addition, in the processing of the time difference calculator unit 31, the backward difference calculation is performed on the two-dimensional image data of each pair of time frames temporally adjacent to each other among the image data including the information of the reflected signal intensity (indicating the presence or absence of and the position of the target) as the pixel value in the two-dimensional image data of the range with respect to the azimuth, and then, a time-series three-dimensional image data is generated that has the dimensions of the azimuth, the range, and has a plurality of frames of the range with respect to the azimuth in terms of times. In this case, as alternative means for the backward difference calculation, Doppler FFT may be performed on the two-dimensional image data of the range with respect to the azimuth to calculate a zero Doppler component in order to suppress clutter, and the zero Doppler component may be subtracted from the original two-dimensional image data. At this time, based on the subtracted two-dimensional force image data, backward difference calculation is performed on the two-dimensional image data of each pair of time frames temporally adjacent to each other to generate time-series three-dimensional image data having the dimensions of the azimuth, the range, and the time, where the time-series three-dimensional image data has a plurality of frames of the range with respect to the azimuth in terms of time.
On the other hand, the output processor unit 5A stores two-dimensional training image data (as will be described later with reference to
Table 1 below shows signal processing (types of signal data) and output data formats of the signal processor unit 2 and the input processor unit 3.
The object detector unit 4 of
In the object detection in
The three-dimensional convolutional encoder 41 of
The input data of the three-dimensional convolutional encoder 41 has, for example, the following data format.
(Time,range,azimuth,channel)=(80,128,80,1)
In addition, the output data of the two-dimensional convolutional decoder 42 has, for example, a data format of the following equation.
(Range,azimuth,channel)=(128,80,1)
For example, additional processing such as deep drop or batch normalization of general deep learning may be performed on the basic configuration of the machine learning model 40A of
The two-dimensional convolutional encoder 51 of
The input data of the two-dimensional convolutional encoder 51 has, for example, the following data format.
(Range,azimuth,channel)=(128,80,1)
In addition, the output data of the two-dimensional convolutional decoder 52 has, for example, a data format of the following equation:
(Range,azimuth,channel)=(128,80,1).
The machine learning model 40B of
Next, an example of image data of the machine learning model 40 of
As shown in
According to the simulations and experiments by the inventors, empirically, it is desirable to set the width in the azimuth direction of the graphic representing the position of the target in the range of 0.25 to 0.75 times the main lobe width of the radar apparatus 1 in order to achieve both convergence and resolution of learning. In the above implementation example, the best estimation accuracy is obtained when the number of pixels is about 11 (the width in the horizontal direction of the white portion in the central portion in
As described with reference to
(2) a size of the object label in each dimension direction in the training image data is set as an upper limit, with a main lobe width in each dimension direction determined from a number of channels in an azimuth direction and a radar bandwidth of the radar apparatus.
As described above, the machine learning model 40 used in the present embodiment is preferably configured as follows.
(1) The input data is image data representing a heat map obtained by performing backward difference calculation on a radar image including information of reflection intensities in each azimuth and each range in a time frame.
(2) The range of the range and azimuth can be arbitrarily set according to the application.
(3) The input image data does not necessarily need to be three-dimensional and may be inferred or estimated in one frame without time series.
(4) The ranges of the input and output ranges and azimuths (image sizes) are not necessarily the same.
(5) It is also possible to increase the modal of the input data, and for example, the reflection intensities image before the time difference calculation as the second modal as the second channel like the RGB image.
(1) The training image data and the output image data include, for example, a grayscale image expressed in an output range of 0 to 1.
(2) The background or the like other than the target to be detected is expressed as 0.
(3) The target to be detected is set to 1 and is expressed by an object label of a graphic having a predetermined shape of a plurality of pixels.
(4) The position of the target is determined based on a specific frame in the time series, and in the implementation examples of
(5) The center coordinates of the object label indicate position information of the target.
(6) The size of the object label is represented by a constant size without depending on the size of the target.
Next, a configuration example and a processing example of the object coordinate detector unit 6 in
The template matching unit 61 performs pattern matching processing by obtaining a cross-correlation value between the pattern of the object label of the target and the pattern of the object label of the correct target of the target stored in advance among the input image data, so that the degree of coincidence is calculated, and the image data of the similarity map is calculated. Next, the peak search unit 62 performs the following peak search processing on the image data of the similarity map output from the template matching unit 61.
(1) The maximum value filter (the filter is a 6×6 filter, and the size can be arbitrarily determined according to the correct label size and the application) is used to retrieve the peak position of the maximum value, and image data having the pixels of the search result is obtained.
(2) Next, a mask processing is performed on the obtained image data to obtain image data including only remaining elements having the same values as those of the original array.
(3) Further, in order to remove noise from the obtained image data, image data is obtained by excluding pixels having a low peak equal to or less than a predetermined threshold value.
Further, the object coordinate detector unit 6 outputs and displays the number of detection points of the target and the coordinate position (range, azimuth) of the target to and on the display unit 7 together with the image data obtained by peak search unit 62.
As described above, the object coordinate detector unit 6 can obtain the number and coordinates of targets included in the observation area of the radar apparatus 1 by combining the processing of the template matching unit 61 and the processing of the peak search unit 62. In addition to the processing used here, a general method for obtaining the center coordinates of a graphic can be applied, and the basic effect of the invention is not impaired.
As described above, according to the present embodiment, the image data in which the object label corresponding to the position of the target is drawn can be generated by using the time difference time-series information generated by the signal processing and the machine learning model for image recognition instead of the method for combining CFAR and clustering as the means for estimating the position of the target. As a result, even in such a situation of problem caused in CFAR or clustering, the position of each target can be estimated with high accuracy as compared with the prior art. In this case, such a unique advantageous effect can be obtained that the number of wave sources can be correctly counted, and the position can be detected in an object proximity situation where counting is not successful in Doppler FFT, CFAR, and cluster analysis which are prior arts.
In the present embodiment, in particular, for the machine learning model 40 for image recognition, the time-series time-difference radar image data including the reflection intensities information in each azimuth and each range processed by the signal processor unit 2 is input to the machine learning model 40 of the object detector unit 4, so that the image data in which the object label corresponding to the position of the target is drawn can be generated. As a result, the user can simultaneously detect the position of each moving object with high accuracy as compared with the prior art even in an environment where a plurality of moving targets are close to each other. In addition, for example, even in a situation in a tunnel and where walls such as soundproof walls of a highway are close to each other and clutter occurs in radar data, the position of each moving object can be detected simultaneously with high accuracy as compared with the prior art.
The present embodiment further has the following unique advantageous effects:
(1) In the signal processor unit 2, the Doppler velocity FFT processing for separating signals becomes unnecessary. It is noted that, in the prior art, two-dimensional FFT processing of the range (range) and the Doppler velocity (relative velocity) is generally performed.
(2) CFAR and cluster analysis are not required.
(3) There is no need to suppress clutter and side lobes.
(4) Even if a reflection signal from a desired target is weak, it can be detected.
(5) In the object detection processing according to the embodiment, the information of the radar image is not lost.
(6) The detection can be performed with high accuracy even if the number of antenna elements is small.
Further, for example, a radar apparatus that does not have a channel with respect to an azimuth and detects the position of a target only with respect to one-dimensional range is also conceivable, and the position may be used as the range. In this case, when the present invention is applied to estimate the wave source position on the range-speed plane, a more accurate target position can be obtained. Therefore, the image data according to the present invention may be image data having at least two dimensions of range, azimuth, and speed.
Application examples of the object-position detector apparatus according to the present embodiment will be described below.
The radar apparatus is mounted on a ceiling in a tunnel and is used for monitoring a traffic flow (traffic traffic) in the tunnel. According to the embodiment of the present invention, it is possible to correctly detect a position of a target (moving object) existing in an observation area without being affected by a virtual image formed by an inner wall or vehicles. In addition, adding tracking to the results can be applied to monitoring of a traffic flow and control of notification.
(2) Monitoring Traffic Flow in Environment with Large Structure Close to Road
A radar apparatus is mounted on a roadside machine in an environment where a large structure close to a road such as a soundproof wall (sound insulation wall) of a highway exists and is used for monitoring a traffic flow. According to the embodiment of the present invention, it is possible to correctly detect a position of a target (moving object) existing in an observation area with no influence due to a virtual image formed by a soundproof wall or vehicles. In addition, adding tracking to the results can be applied to monitoring of a traffic flow and control of notification.
The radar apparatus is mounted on a roadside machine at an intersection and used for monitoring a traffic flow. According to the embodiment of the present invention, a position of a target (moving object) present in an observation area is correctly detected with no influence due to a virtual image by humans, a utility pole, a building, or the like, and the results thereof are used for safety monitoring of a pedestrian and control of notification.
The radar apparatus is mounted on an automatic conveyance vehicle, or a self-propelled robot operating in a factory and is used to detect an obstacle. According to the embodiment of the present invention, a position of a target (obstacle) present in an observation area is correctly detected without being affected by a virtual image by a machine, a worker, or the like in a manufacturing line in a factory, and the results thereof are used for travel control of an automatic conveyance vehicle or a self-propelled robot.
Differences from Prior Art and Unique Advantageous Effects
In the embodiment of the present invention, the machine learning model 40 is characterized by outputting information corresponding to the position of the object as image data. In this case, the input data of the machine learning model 40 is image data obtained by time-serializing the image data of the radar signal including the information of the reflection intensity at each position of the range with respect to the azimuth after calculating time differences in time frames on the image data of the radar signal. In addition, the teacher output data of the machine learning model 40 is characterized in that the positional information of the object is represented by a plurality of pixels. As a result, for example, as compared with the prior art using the tensor of the heat map of the range azimuth Doppler (RAD) and the class and position of the correct answer box, since the information regarding the presence or absence of and the position of the plurality of objects is output in the form of one piece of image data, the calculation cost is extremely small. In addition, the structure of the machine learning model 40 is also simple as compared with that of the prior art.
As described above in detail, according to the present invention, it is possible to detect an object with high accuracy even in such a situation that a plurality of adjacent objects or structures are close to each other such that reflected waves interfere with each other, as compared with the prior art. The object-position detector apparatus of the present invention can be applied to, for example, a counter apparatus that counts a vehicle or a pedestrian, a traffic counter apparatus, an infrastructural radar apparatus, and a sensor apparatus that detects an obstacle of an automatic conveyance vehicle in a factory.
Number | Date | Country | Kind |
---|---|---|---|
2021-038305 | Mar 2021 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2022/010041 | 3/8/2022 | WO |