The invention relates to a method and a device for determining a roadway coating by means of a vehicle camera system.
Driver assistance systems support the driver in his driving task and are thus helping to make road traffic safer in future and to reduce accident figures. Camera-based driver assistance systems hereby detect the surroundings of a vehicle. Camera systems which are located behind the windshield detect the area in front of the vehicle according to the driver's visual perception. The functional scopes of such assistance systems extend from intelligent headlamp control to detecting and displaying speed limits to warnings in the event of the vehicle failing to keep to its lane or an impending collision. In addition to camera systems, radar sensors, lidar sensors and/or laser scanners help to detect other vehicles, unprotected road users such as e.g. pedestrians and cyclists and the infrastructure such as e.g. crash barriers and traffic lights. This is therefore creating the requirement to depict the immediate vehicle environment more and more accurately.
The degree of automation of motor vehicles will continue to rise in future. As a result, the level of equipping of vehicles with sensor technology will also increase very significantly. Highly or fully automated vehicles will be equipped with a plurality of different sensors and will thus allow a 360° view, in particular via camera systems.
An essential part of a driver's driving task is correctly assessing the roadway state and, thus, the available friction coefficient between tires and the roadway, in order to then adapt his driving style accordingly. In future, highly and fully automated vehicles will take over the task of driving at least in partial areas. To this end, it is essential for the roadway state to be correctly detected and assessed by the system.
DE 10 2004 018 088 A1 shows a roadway detection system having a temperature sensor, an ultrasonic sensor and a camera. The temperature, roughness and image data (roadway data) obtained from the sensors is filtered and compared with reference data, and a level of security is generated for the comparison. The state of the roadway surface is detected on the basis of the comparison of the filtered roadway data with the reference data. The roadway surface (e.g. concrete, asphalt, dirt, grass, sand or gravel) and the state thereof (e.g. dry, icy, snowy, wet) can be classified in this way.
WO 2012/110030 A2 shows a method and a device for estimating coefficients of friction using a 3D camera, e.g. a stereo camera. At least one image of the environment of the vehicle is recorded by means of the 3D camera. A height profile of the road surface is created in the entire area in front of the vehicle from the image data of the 3D camera. The anticipated local coefficient of friction of the road surface in the area in front of the vehicle is estimated from the height profile. In individual cases, the roadway surface can be classified, e.g. as a blanket of snow or a muddy dirt track, from special detected height profiles.
WO 2013/117186 A1 shows a method and a device for detecting the condition of a roadway surface by means of a 3D camera. By means of the 3D camera, at least one image of the surroundings extending in front of the vehicle is acquired. Height profiles of the roadway surface which extend transversely to the direction of motion of the vehicle are determined from the image data of the 3D camera along a plurality of lines. The condition of the roadway surface is detected from the determined height profiles. In addition to the determined height profiles, 2D image data from at least one monocular camera of the 3D camera is optionally evaluated, e.g. by means of a texture or pattern analysis, and is incorporated into the detection of the condition of the roadway surface.
However, the known methods place high demands on the required sensor technology. Therefore, in the case of the indicated methods and/or devices, either a temperature and ultrasonic sensor are required in addition to a camera, or the camera must be configured as a 3D sensor, so that the classification results are sufficiently robust.
An object of an embodiment of this invention is therefore to indicate a roadway state determination by means of a camera which, when different vehicle camera system configurations are used, including, for example, just one mono camera system, makes possible an optimized roadway state detection and friction coefficient estimation derived therefrom.
One of the main ideas of the invention is to specifically detect indications of the existence of a roadway coating as the roadway is traveled over by a vehicle by means of a camera system and image processing. A roadway coating is e.g. snow, rainwater, ice, leaves, dust, i.e. media/objects which lie flat (blanket, carpet) on the roadway surface (asphalt, tar, concrete, etc.). The phenomenon of lying flat on the roadway surface can be described as a blanket or carpet of the medium or of the objects, however the entire roadway surface does not have to be covered. Different roadway coatings typically show different behaviors when they are traveled over by a vehicle. Some of them can be observed and/or detected with a camera and recognized and/or identified in a subsequent evaluation of the image, from which conclusions regarding the presence and the nature of the roadway coating can be drawn.
A method for determining a roadway coating by means of a vehicle camera system according to the invention comprises the following steps: by means of the vehicle camera system, at least one image of the surroundings of the vehicle is recorded. The at least one image is evaluated in order to detect indications of the presence of a roadway coating as the roadway is traveled over by the vehicle with the vehicle camera system or by another vehicle. The detected indications are taken into consideration in the determination of a roadway coating. The result of the determination of the roadway coating can preferably be combined with a direct roadway state detection by means of a classification or an application of a neural network and output to a driver assistance function, to a vehicle control function or also as information to the driver. Driver assistance functions can in particular comprise collision warnings, emergency braking applications or even emergency steering, whilst vehicle control functions can in particular comprise occupant protection measures (airbag anticipatory control, brake anticipatory control, belt tensioner anticipatory control) as well as (partially) autonomous braking and/or steering interventions. In a preferred use of the method according to the invention, the detected indications and/or the result of the determination of the roadway coating are taken into consideration in the design of driver assistance functions and vehicle control functions.
The result of the determination of the roadway coating can in particular be incorporated into an estimate of the friction coefficient for the roadway area which is shown in the image, or a current friction coefficient which is otherwise detected or a predictively detected friction coefficient. This is because the roadway coating has a significant influence on the actual friction coefficient. The friction coefficient, also known as the friction value, adhesion coefficient, (adhesive) friction coefficient or coefficient of friction indicates the maximum force with reference to the wheel load which can be transmitted between a roadway surface and a vehicle tire (e.g. in the tangential direction) and is thus an essential measure of driving safety. In addition to the roadway state, the properties of the tire are required in order to determine the friction value completely.
According to an advantageous embodiment, the effects of precipitation (for example rain, snow, hail or even fog) in the image, on the roadway and on the vehicles or vehicle windshields are detected as indications during the evaluation of the image.
The type and quantity of precipitation can be determined during the evaluation of the image. The detection direction of the camera system can hereby be advantageously linked with the current direction of motion of the vehicle. Thus, when heavy rainfall is detected by a front camera, it can be concluded during the evaluation of the image that the roadway in front of the vehicle will be wet.
In a preferred embodiment, effects of a roadway coating as it is traveled over by at least one tire of a vehicle are detected as indications during the evaluation of the image. Effects are hereby in particular modifications of the roadway surface including the roadway coating as it is traveled over by a vehicle or, more precisely, as it is traveled over by at least one tire of the vehicle.
At least one region in the at least one image is advantageously determined, from which indications of the presence of a roadway coating are to be inferred. These are in particular image regions which include a vehicle tire travelling over said roadway coating or which are located adjacent to (at the side of or behind) such a vehicle tire.
The region in the at least one image is preferably supplied to a classifier which maps the detected indications onto a quantity of classes which are each assigned to one roadway coating.
According to an advantageous embodiment, the vehicle camera system has a rear and/or lateral detection area of the surroundings of the host vehicle. The host vehicle is hereby the vehicle which has the vehicle camera system. In this way, indications of the presence of a roadway coating as the roadway is traveled over by the host vehicle can be detected during the evaluation of the image. The evaluation of the image can, in the process, advantageously be limited to a fixed image region, in which effects of an existing roadway coating are typically shown, therefore e.g. a region behind a tire or at the back of the host vehicle which is shown by a rear camera or a region adjacent to a tire of the host vehicle which is shown by a side camera.
The advantage with this embodiment is that, in order to determine a roadway coating, the driver is not dependent on other vehicles located in the detection area of the vehicle camera system.
In a preferred embodiment, the vehicle camera system has a detection area located in front of the host vehicle, so that indications of the presence of a roadway coating as the roadway is traveled over by a vehicle travelling in front, a vehicle travelling on a road which intersects with the road on which the host vehicle is travelling or an oncoming vehicle are detected during the evaluation of the image.
To that end, the evaluation of the image may, in particular, be limited to an image region in which the back and/or the side of a vehicle travelling in front or the vehicle front and/or the side of an oncoming vehicle is/are shown. The relevant image region can preferably be determined on the basis of the tire contact zones with the roadway in the image.
One advantage of this embodiment is that a predictive determination of a roadway coating is possible.
The indications of the presence of a roadway coating as the roadway is traveled over advantageously comprise whirled-up components of the roadway coating. These include in particular the detection of spray water or sprayed slush, spray mist or spray, snow turbulence, leaves as well as sand or dust.
According to a preferred embodiment, the indications of the presence of a roadway coating comprise obstructed views in the field of view of the vehicle camera system due to the roadway being traveled over by a vehicle travelling in front, a vehicle travelling on a road which intersects with the road on which the host vehicle is travelling or an oncoming vehicle.
The indications of the presence of a roadway coating as the roadway is traveled over by a vehicle preferably comprise tire ruts in the roadway coating. In the case of special roadway coatings, tire ruts are formed directly behind the tires of a moving vehicle, sometimes the impression of the tire profile can even be detected within a tire rut e.g. in the case of a coating of snow.
During the determination of the presence of a roadway coating, in addition to the indications of the presence of a roadway coating as the roadway is traveled over by a vehicle, typical generic properties of different roadway coatings can, in addition, advantageously be detected and taken into consideration. This means properties which are characteristic of roadway coatings and which can be recognized from a camera image even without a vehicle travelling over the roadway surface, e.g. puddles and reflections in the case of rainwater or the topology of the surface. In particular, headlights of oncoming vehicles are reflected more strongly on a rain-wet or icy roadway than on a dry roadway.
The classifier preferably comprises a neural network which has been trained, in order to be able to assign the detected indications to a roadway coating. The neural network can in particular also continually learn to assign indications to a class of roadway coatings.
According to an advantageous further development of the invention, the result of the determination of the roadway coating is combined with the result of a roadway state detection or classification and the friction coefficient is estimated therefrom. In particular, the roadway state detection or classification determines the material and geometric roadway surface e.g. rough or smooth tar, asphalt or concrete, possibly with existing ruts, and can in particular be based on image data from the same vehicle camera system.
The invention additionally relates to a device for determining a roadway coating comprising a vehicle camera system, an evaluating unit, a determining unit and an output unit. The vehicle camera system is configured to record at least one image of the surroundings of the vehicle. The evaluating unit is configured to evaluate the at least one image, in order to detect indications of the presence of a roadway coating by precipitation and/or as the roadway is traveled over by a vehicle (E, F). The determining unit is configured to determine a roadway coating taking into consideration the detected indications. The output unit is set up to output the result detected by the determining unit. The evaluating, determining and output units can in particular be part of the control unit of the vehicle camera system or of other vehicle control units.
The invention will be explained in more detail below by means of figures and embodiment examples, where:
If slush is present on the roadway, spraying slush is produced in a comparable way, which can be detected and recognized as such. Depending on the roadway coating, e.g. in the case of a damp instead of a wet roadway, and depending on the structure of the roadway surface, no spray water will be produced as the road coating is traveled over, instead only spray mist or spray or both will be produced. This can also be detected by a vehicle camera and can be recognized by image processing as spray mist or spray.
When roads covered with new snow or powder snow are traveled over, typical snow turbulence occurs, which spreads to the sides of the vehicle and, as a light covering of snow, to the rear of the vehicle. Other roadway coatings such as e.g. a layer of leaves or a sandy surface on a paved roadway are also whirled up when they are traveled over and can be detected and recognized with a vehicle camera.
In addition, in the case of some roadway coatings, additional indications in the camera image can be detected and identified. Thus, as wet or snow-covered roadways are traveled over or as slush is also traveled over, tire ruts are formed behind a moving vehicle, sometimes the impression of the tire profile within a tire rut can even be detected e.g. when there is a thin layer of snow. In addition, precipitation in the form of rain, snow or hail, and even fog, can be detected as indications of the roadway state in the camera image.
Different indications of the existence of a roadway coating detected from camera images are therefore advantageously taken into consideration in the classification of the nature of the roadway coating.
The different classes can be coarsely divided (no roadway coating/water/snow), but can also be more finely divided (e.g. in the case of snow: powder snow up to 5 mm high, powder snow >5 mm high, solid snow, slush). This subdivision can in particular be made as a function of an assignment of the class to an average friction coefficient, and/or the change of the assigned average friction coefficient during the transition from one subclass to another.
Indications of the presence of a roadway coating as the roadway is traveled over by vehicles travelling in front of the host vehicle (F) or by oncoming vehicles (not shown) can be detected with camera sensors which comprise a forward-facing detection area (1a, 1b, 2) in the direction of travel.
This offers the advantage of a predictive determination of the roadway coating. A stereo camera can preferably be used as a front camera, as a result of which a spatial resolution of the recorded image data is achieved. Objects such as e.g. spray water drops can be analyzed with respect to their spatial form and extent, which improves the classification of an existing roadway coating.
The image processing for determining a roadway coating can be advantageously limited to one or more regions of the camera image, in which a tire contact zone of the roadway is located.
In the case of a vehicle travelling in front (F) an image region can be evaluated, in which the rear of the vehicle of the vehicle travelling in front is located.
Indications of the presence of a roadway coating as the roadway is traveled over by the tires of the host vehicle (E) can be detected with camera sensors, which comprise a detection area (3, 1e, 1f, 1c, 1d) facing backwards or to the side in the direction of travel.
The advantage with this is that the driver is not dependent on other vehicles (F) travelling in front or offset but, thanks to the detection and determination of the effects caused by the host vehicle (E) by means of rear and/or laterally oriented sensor technology, can determine a currently relevant roadway coating, independently of other vehicles (F). Since vehicles will in future be increasingly equipped with 360° camera sensor technology which records a surround view which can be displayed to the driver e.g. as a “top view” in a bird's eye view, the reliable determination of a roadway coating is realistic.
Number | Date | Country | Kind |
---|---|---|---|
10 2013 223 367 | Nov 2013 | DE | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/DE2014/200601 | 10/28/2014 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2015/070861 | 5/21/2015 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
4712635 | Sumiya et al. | Dec 1987 | A |
5351192 | Tsuyama et al. | Sep 1994 | A |
5586028 | Sekine et al. | Dec 1996 | A |
5774821 | Eckert | Jun 1998 | A |
5944392 | Tachihata et al. | Aug 1999 | A |
5963148 | Sekine et al. | Oct 1999 | A |
6597980 | Kogure | Jul 2003 | B2 |
6636258 | Strumolo | Oct 2003 | B2 |
6807473 | Tran | Oct 2004 | B1 |
7315777 | Ono | Jan 2008 | B2 |
7411488 | Gern et al. | Aug 2008 | B2 |
7702446 | Hiwatashi | Apr 2010 | B2 |
7808523 | Nishida et al. | Oct 2010 | B2 |
8180527 | Mueller-Schneiders et al. | May 2012 | B2 |
8306747 | Gagarin et al. | Nov 2012 | B1 |
8310545 | Fujita et al. | Nov 2012 | B2 |
8666562 | Tuononen | Mar 2014 | B2 |
8957949 | Randler et al. | Feb 2015 | B2 |
9081387 | Bretzigheimer et al. | Jul 2015 | B2 |
20020007661 | Takahashi | Jan 2002 | A1 |
20020087251 | Kogure et al. | Jul 2002 | A1 |
20020176608 | Rose | Nov 2002 | A1 |
20030101805 | Raab | Jun 2003 | A1 |
20040016870 | Pawlicki et al. | Jan 2004 | A1 |
20040138831 | Watanabe et al. | Jul 2004 | A1 |
20040204812 | Tran | Oct 2004 | A1 |
20050060069 | Breed | Mar 2005 | A1 |
20050085987 | Yokota et al. | Apr 2005 | A1 |
20050172526 | Taylor et al. | Aug 2005 | A1 |
20070050121 | Ammon et al. | Mar 2007 | A1 |
20080027607 | Ertl et al. | Jan 2008 | A1 |
20080110556 | Kawasaki | May 2008 | A1 |
20100118116 | Tomasz et al. | May 2010 | A1 |
20100253541 | Seder et al. | Oct 2010 | A1 |
20100315505 | Michalke et al. | Dec 2010 | A1 |
20110063097 | Naka et al. | Mar 2011 | A1 |
20110200199 | Wakao | Aug 2011 | A1 |
20110245995 | Schwarz | Oct 2011 | A1 |
20120029783 | Takenaka et al. | Feb 2012 | A1 |
20120078483 | Yajima et al. | Mar 2012 | A1 |
20120167663 | Groitzsch et al. | Jul 2012 | A1 |
20120323444 | Rieger et al. | Dec 2012 | A1 |
20130013163 | Hong | Jan 2013 | A1 |
20130033603 | Suzuki et al. | Feb 2013 | A1 |
20130211720 | Niemz | Aug 2013 | A1 |
20130332028 | Heger | Dec 2013 | A1 |
20130338878 | Fritz et al. | Dec 2013 | A1 |
20140005875 | Hartmann et al. | Jan 2014 | A1 |
20140052325 | Naegele et al. | Feb 2014 | A1 |
20140195112 | Lu et al. | Jul 2014 | A1 |
20140320644 | Hegemann et al. | Oct 2014 | A1 |
20140347448 | Hegemann et al. | Nov 2014 | A1 |
20150153266 | Mack | Jun 2015 | A1 |
20150166072 | Powers et al. | Jun 2015 | A1 |
20150178572 | Omer et al. | Jun 2015 | A1 |
20150224925 | Hartmann | Aug 2015 | A1 |
20150251659 | Fischer et al. | Sep 2015 | A1 |
20150344037 | Siegel et al. | Dec 2015 | A1 |
20150371095 | Hartmann et al. | Dec 2015 | A1 |
20150375753 | Schräbler et al. | Dec 2015 | A1 |
20160121902 | Huntzicker et al. | May 2016 | A1 |
Number | Date | Country |
---|---|---|
198 56 510 | Sep 1999 | DE |
198 54 964 | Jun 2000 | DE |
101 55 488 | May 2003 | DE |
102 56 726 | Jun 2004 | DE |
102004018088 | Feb 2005 | DE |
102004055069 | Feb 2006 | DE |
102004047914 | Mar 2006 | DE |
102004048637 | Apr 2006 | DE |
102005039167 | Feb 2007 | DE |
102006012289 | Sep 2007 | DE |
102008047750 | May 2009 | DE |
102010013339 | Jan 2011 | DE |
102009041566 | Mar 2011 | DE |
102010020688 | May 2011 | DE |
102011100907 | Jan 2012 | DE |
102010045162 | Mar 2012 | DE |
102010063017 | Jun 2012 | DE |
102011081362 | Feb 2013 | DE |
102012221518 | May 2014 | DE |
102016009928 | Feb 2017 | DE |
0 412 791 | Feb 1991 | EP |
0 827 127 | Mar 1998 | EP |
1 201 521 | May 2002 | EP |
2 048 476 | Apr 2009 | EP |
2 521 111 | Nov 2012 | EP |
07-035522 | Feb 1995 | JP |
11-211659 | Aug 1999 | JP |
11-223517 | Aug 1999 | JP |
2002-310896 | Oct 2002 | JP |
2005-162091 | Jun 2005 | JP |
2005-226671 | Aug 2005 | JP |
2007-322231 | Dec 2007 | JP |
2009-237938 | Oct 2009 | JP |
2013-014311 | Jan 2013 | JP |
2013-213793 | Oct 2013 | JP |
1020110032422 | Mar 2011 | KR |
WO 2004021546 | Mar 2004 | WO |
WO 2012110030 | Aug 2012 | WO |
WO 2013007800 | Jan 2013 | WO |
WO 2013009697 | Jan 2013 | WO |
WO 2013117186 | Aug 2013 | WO |
Entry |
---|
International Search Report of the International Searching Authority for International Application PCT/DE2014/200601, dated Mar. 26, 2015, 3 pages, European Patent Office, HV Rijswijk, Netherlands. |
PCT International Preliminary Report on Patentability including English Translation of PCT Written Opinion of the International Searching Authority for International Application PCT/DE2014/200601, dated May 17, 2016, 11 pages, International Bureau of WIPO, Geneva, Switzerland. |
German Search Report for German Patent Application No. 10 2013 223 367.1, dated Sep. 12, 2014, 9 pages, Muenchen, Germany, with English translation, 6 pages. |
Raqib Omer, “An Automatic Image Recognition System for Winter Road Surface Condition Monitoring”, Master's Thesis, University of Waterloo, Ontario, Canada, Feb. 22, 2011, pp. i-xii, 1 to 68 retrieved at https://uwspace.uwaterloo.ca/handle/10012/5799, XP055177353A. |
Maria Jokela et al., “Road Condition Monitoring System Based on a Stereo Camera”, Intelligent Computer Communication and Processing, IEEE 5th International Conference ICCP 2009, Piscataway, NJ, USA, Aug. 27, 2009, XP031545069, pp. 423 to 428. |
J. Chetan et al., “An Adaptive Outdoor Terrain Classification Methodology Using Monocular Camera”, Intelligent Robots and Systems, IEEE International Conference IROS 2010, Piscataway, NJ, USA, Oct. 18, 2010, XP031920567, pp. 766 to 771. |
J. H. Choi et al., “Road Identification in Monocular Color Images Using Random Forest and Color Correlogram”, International Journal of Automotive Technology, vol. 13, No. 6, The Korean Society of Automotive Engineers, Heidelberg, Oct. 2, 2012, XP035120063, pp. 941 to 948. |
Raqib Omer et al., “An Automatic Image Recognition System for Winter Road Surface Condition Classification”, Intelligent Transportation Systems, 13th International IEEE Conference ITSC 2010, Piscataway, NJ, USA, Sep. 19, 2010, XP031792816, pp. 1375 to 1379. |
T. Teshima et al., “Detection of the Wet Area of the Road Surface Based on a Saturated Reflection”, Meeting on Image Recognition and Understanding, 2007, XP002730931, pp. 1218 to 1223, retrieved at http://hvrl.ics.keio.ac.jp/paper/pdf/domestic_Conference/2007/MIRU2007_teshima.pdf. |
Ernst Dieter Dickmanns et al., “Dynamic Monocular Machine Vision”, Machine Vision and Applications, 1988 Springer-Verlag New York Inc., pp. 223 to 240. |
Partial English translation of Japanese Office Action in Japanese Patent Application No. 2016-553707, dated Aug. 15, 2018, 7 pages. |
Wang et al., “Study on Curb Detection Method Based on 3D Range Image by Laser Radar”, 2005 IEEE Intelligent Vehicles Symposium, pp. 845-846. |
Wijesoma et al., “Road-Boundary Detection and Tracking Using Ladar Sensing”, IEEE Transactions on Robotics and Automation, vol. 20, No. 3, Jun. 2004, pp. 456-464. |
Siegemund et al., “Curb Reconstruction Using Conditional Random Fields”, 2010 IEEE Intelligent Vehicles Symposium, University of California, San Diego, CA, USA, Jun. 21-24, 2010, pp. 203-210. |
Oniga et al., “Polynomial Curb Detection Based on Dense Stereovision for Driving Assistance”, 2010 13th International IEEE Annual Conference on Intelligent Transportation Systems, Madeira Island, Portugal, Sep. 19-22, 2010, pp. 1110-1115. |
Number | Date | Country | |
---|---|---|---|
20160379065 A1 | Dec 2016 | US |