This application claims foreign priority benefits under 35 U.S.C. §119(a)-(d) to DE 10 2014 204 360.3, filed Mar. 10, 2014, which is hereby incorporated by reference in its entirety.
This disclosure relates to a method and device for estimating the distance between a moving vehicle and an object. The estimation may advantageously be used in particular in conjunction with a system for automatic emergency braking or else with a system for adaptive speed regulation.
Systems such as automatic emergency braking, which automatically brake a motor vehicle in order to avoid or reduce the effect of a traffic accident or a collision can, in principle, contribute to reducing the rate of traffic accidents and possibly reducing the amount of damage caused by accidents. However, this may require a real-time measurement of the distance between the moving motor vehicle and the respective object or obstacle.
Various approaches are known for measuring the distance between a moving motor vehicle and an object or obstacle. Thus, for example, LIDAR systems are known, which emit laser pulses and detect the light scattered back from the object in order to establish the distance to the object. Here, the measured distance is a function of the time interval elapsed between the emission of the laser pulse and the detection thereof. However, this approach does not allow a determination of form and type of the object.
A further approach is based upon stereoscopic imaging or recording, wherein the distance to the relevant object is determined from the parallax between two images of the same situation, wherein these two images are recorded by means of two cameras aligned with respect to one another.
A further method is based on measuring the distance between a moving vehicle and an object using a monocular camera; however, this requires the complete and correct compensation of the camera movement (in respect of tilt or inclination angle, pitch angle, etc.) and the carriageway inclination.
It is an object to provide a method and a device for estimating the distance between a moving vehicle and an object, which enable an estimate which is as precise as possible using a simple and robust approach.
A method for estimating the distance between a moving vehicle and an object, wherein the vehicle includes a camera, comprises the following steps:
In particular, certain embodiments are based on the concept of establishing the distance between a moving motor vehicle and an object or obstacle on the basis of the scaling which is estimated using two successive images of the object.
Here, the distance to the object from the scaling (s(t)) can be estimated according to the relationship
where Tz(t) denotes the z-component of the translation vector T between successive individual images.
According to one embodiment, the scaling value is estimated using a Fourier-Mellin transform (FMT). In the following, this Fourier-Mellin transform is discussed briefly; it corresponds to a two-dimensional Fourier transform after a transform into logarithmic coordinates and a transform into polar coordinates.
In the Fourier-Mellin transform, the transform into logarithmic coordinates transforms a scaling in real space into a translation in the frequency domain. Moreover, as a result of the transform into polar coordinates, a rotation in real space is transformed into a translation in the frequency domain. Here, the embodiment makes use of the fact that the Fourier-Mellin transform is not only invariant in relation to translation but that, furthermore, changes in the rotation and in the scaling also respectively appear as an addition of a pure phase shift and an amplitude change proportional to the change in scaling.
The Fourier-Mellin transform of a function f, wherein this Fourier-Mellin transform is subsequently referred to as Mf, therefore emerges from a Fourier transform of the angle coordinate and a Mellin transform of the radial component as
where u is the Mellin transform parameter and v is the Fourier transform parameter.
Image registration (i.e. determining the parameters for aligning two images in image processing) constitutes a basic method in image processing when superposing two or more images. In the image registration method, the parameters t, S and R are determined, where R denotes the rotation matrix in the form
S denotes a scaling matrix, representing scaling in the x- and y-direction, in the form
which, in the case of equal scaling along the axes reduces to a scaling factor, and t denotes the displacement or translation.
Displacement or translation t, rotation R and scaling S respectively have an equivalent in Fourier space. Fourier-based methods differ from other standard methods by virtue of the fact that an ideal correspondence is sought-after in the frequency domain. Here, the Fourier-based methods make use of the displacement theorem and the rotation theorem of the Fourier transform since these provide invariance in relation to translation, rotation and scaling. According to the displacement theorem, a positional change occurring in real space does not lead to a change in amplitude of the Fourier transform.
According to one embodiment, the time profile of the scaling value is smoothed, which, in particular, can be brought about using a Kalman filter.
In accordance with one embodiment, a monocular camera is used as a camera.
Here, there is a direct or immediate measurement of the variation in the light intensity and of the number of all pixels representing the relevant object in the generated camera images in at least two successive individual images (“frames”). The appropriate group of selected pixels is selected in such a way that these represent the relevant object in the generated image. The measured variation in the intensity can be smoothed using a suitable filter.
The calculation of the distance occurs in real time, wherein the nonlinear relationship between the distance between camera and object on the one hand and the change in the scaling in the object in two successive individual images is used at each point in time. The smoothed scaling value is then obtained using the variation in the intensity and the previously measured number of pixels.
There is, in particular, no need for a light source since the concept for calculating the distance is based on estimating the scaling of the relevant object, the distance of which is intended to be established, from two successive individual images. In other words, the distance to the relevant object is calculated or estimated purely from the established camera data (and on the basis of the scaling as an absolute value).
Here, the method is advantageous in that there is no need for exact compensation of the camera movement (in respect of tilt or inclination angle, pitch angle, etc.) and also in that there is no need for establishing the carriageway incline or carriageway drop.
Certain embodiments include a device for estimating the distance between a moving vehicle and an object, wherein the device is configured to carry out a method comprising the features described above. Further embodiments can be gathered from the description.
As required, detailed embodiments of the present invention are disclosed herein; however, it is to be understood that the disclosed embodiments are merely exemplary of the invention that may be embodied in various and alternative forms. The Figures are not necessarily to scale; some features may be exaggerated or minimized to show details of particular components. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a representative basis for teaching one skilled in the art to variously employ the present invention.
A monocular camera assembled on the vehicle is used for estimating the distance between a moving vehicle and an object, wherein the optical axis of the camera corresponds to the direction of the translational movement of the vehicle. The object may be a vehicle (which is at rest or likewise moving) or a standing or moving pedestrian, or another road user.
The calculation of the distance is then performed using the scaling s(t) established from “tracking” over a plurality of individual images recorded by the camera in accordance with equation (1) already mentioned above
Here, Tz(t) denotes the z-component of the translation vector T between successive individual images, which is established with the aid of inertial sensors. In principle, the pixel x=(x, y) corresponding to the projection of a point X=(X, Y, Z) lying in three-dimensional space emerges from the following
where f denotes the focal length of the camera. To a good approximation, the z-coordinate can be considered to be constant in the following since the variation thereof over the surface of the object or obstacle facing the camera is comparatively small relative to the distance between object and camera.
If the path belonging to a relative movement between camera and object occurring between the time t and the time t+Δt is denoted by T (t, Δt), the following emerges
X(t+Δt)=X(t)+T(t, Δt) (6)
In the case of a purely translational movement under consideration, the following result emerges for the transformation of a pixel
Here, s(t) denotes the scaling between successive images at the time t.
The approach proceeds from the aforementioned equation (7), wherein it is possible to show that, under the given circumstances, the distance can be estimated purely on the basis of estimating the scaling factor of the object images. The image scaling s(t) and the translational image shifts between two successive individual images are estimated using the frequency domain analysis.
In accordance with
After the image sectioning undertaken in step S20 and after the preprocessing of the images in step S30, the scaling and the inherent movement are estimated in step S40 using a Fourier-Mellin transform and, on the basis of this estimate, a distance is calculated in step S50 in a distance calculation module, which is fed both the results from step S40 and the logging of the object or obstacle on the basis of the edge analysis from steps S11 and S12. Here, a Kalman filter can be used for smoothing the time profile of the scaling s(t).
While exemplary embodiments are described above, it is not intended that these embodiments describe all possible forms of the invention. Rather, the words used in the specification are words of description rather than limitation, and it is understood that various changes may be made without departing from the spirit and scope of the invention. Additionally, the features of various implementing embodiments may be combined to form further embodiments of the invention.
Number | Date | Country | Kind |
---|---|---|---|
10 2014 204 360.3 | Mar 2014 | DE | national |