The present invention relates to a driving assistance system and a raindrop detection method.
Heretofore, an on-vehicle monitor device has been proposed which includes a camera capable of switching between a first focal distance for detecting a raindrop attached to a vehicle and a second focal distance for capturing an image of surroundings of the vehicle, and which detects whether or not a raindrop is attached from an image captured by the camera at the first focal distance (see Patent Literature 1).
Patent Literature 1: Japanese Patent Application Publication No. 2005-225350
The on-vehicle monitor device described in Patent Literature 1, however, needs to switch the focal distance to detect a raindrop and thereby may lead to a possibility of reduction in the detection accuracy for the surrounding environment. It should be noted that this problem is not limited to the case of detecting a raindrop attached to a vehicle but also may occur in the case of detecting a raindrop attached to other moving objects (automatic navigation robot or the like).
Therefore, the present invention has an objective to provide a driving assistance system and a raindrop detection method thereof which are capable of detecting a raindrop while avoiding reduction in the detection accuracy for the surrounding environment.
A driving assistance system of the present invention provides various kinds of information to a driver of a moving object from an image-capturing result of surroundings of the moving object. A principal feature of this driving assistance system is that the driving assistance system includes: image-capturing means installed on the moving object and configured to capture a surrounding image including a portion of the moving object, first edge line storing means configured to store a first edge line detected from a first surrounding image captured in a normal condition by the image-capturing means: calculating means configured to calculate a matching degree between the first edge line stored in the first edge line storing means and a second edge line detected from a second surrounding image currently captured by the image-capturing means; and raindrop judging means configured to judge that a raindrop is attached to a lens unit of the image-capturing means, in response to a decrease in the matching degree between the first edge line and the second edge line.
In addition, a principal feature of a raindrop detection method of the present invention is that the method includes: an image-capturing step of capturing a surrounding image including a portion of the moving object, performed by an image-capturing means installed on the moving object; a calculating step of calculating a matching degree between a previously-stored first edge line detected from a first surrounding image captured in a normal condition in the image-capturing step, and a second edge line detected from a second surrounding image currently captured; and a raindrop judging step of judging that a raindrop is attached to a lens unit of the image-capturing means, in response to a decrease in the matching degree between the first edge line and the second edge line.
According to the present invention, surrounding images including a portion of the moving object are captured, and the matching degree between the first edge line detected from the first surrounding image captured in normal conditions and the second edge line detected from the second surrounding image currently captured, is captured. Then, in response to a decrease in the matching degree between the first edge line and the second edge line, it is judged that a raindrop is attached to the lens unit of the image-capturing means. Thus, raindrop detection can be performed without changing the focal distance, and raindrop detection can be performed while avoiding reduction in the detection accuracy for the surrounding environment.
[
[
[
[
[
[
[
[
[
[
[
[
[
Hereinafter, preferred embodiments of the present invention are described based on the drawings. It should be noted that the following several embodiments include the same component elements. For this reason, in the following description, common reference signs are attached to the same component elements and the overlapping description is omitted.
The camera 10 illustrated in
The warning device 30 is configured to issue a warning to the driver of the system-equipped vehicle V in a case where the obstruction or the like detected by the computer 20 is likely to come into contact with the system-equipped vehicle V. Warnings can also be issued for other situations. Moreover, in a case where a raindrop is attached to the lens unit, the warning device 30 also issues a warning about the raindrop attachment to the driver. Here, the warning method may be via presentation on an image display or via a voice announcement.
Additionally, in the present embodiment, the camera 10 is configured to capture an image of a location behind the system-equipped vehicle V and cause a bumper, as a portion P of the system-equipped vehicle V, to be included in the image-capturing range. In other words, the camera (image-capturing means) 10 according to the present embodiment is installed on the system-equipped vehicle V (the moving object) and captures surrounding images (a first surrounding image or a second surrounding image) I including the portion P of the system-equipped vehicle V. Here, the portion of the system-equipped vehicle V is not limited to the bumper but may be any portion whose image can be captured stably. For example, the image-capturing range may include a number plate, a rear spoiler, a roof spoiler, a casing of the camera 10, or the like depending on the installation position or the optical axis direction of the camera 10. Various methods may be employed as the method of installing the camera 10 on the system-equipped vehicle V. For example, the camera 10 may be installed on the system-equipped vehicle V in an integrally assembled manner or may be detachably installed on the system-equipped vehicle V.
As illustrated in
The current edge detection unit 21 is configured to detect an edge E of the portion P or the system-equipped vehicle V in an image captured by the camera 10.
In the example illustrated in
Here,
Specifically, the reference edge storage unit 22 stores in advance the edge shape E1 obtained from the portion P of the system-equipped vehicle V in normal conditions such as fine weather or the like (when no raindrops are attached to the lens unit of the camera 10).
As described above, the reference edge storage unit (reference edge storing means: first edge line storing means) 22 is configured to store the reference edge shape (first edge line) E1 detected from the first surrounding image I (predetermined region A) captured in normal conditions (in fine weather or the like) by the camera (image-capturing means) 10.
In contrast, the current edge detection unit (edge detecting means) 21 is configured to detect an edge (second edge line) E2 from a second surrounding image I (predetermined region A) currently captured by the camera (image-capturing means) 10.
The deviation degree calculation unit 23 is configured to calculate a matching degree between the reference edge shape (first edge line) E1 detected from the first surrounding image I (predetermined region A) captured in normal conditions (in fine weather or the like) by the camera (image-capturing means) 10 and the edge (second edge line) E2 detected from the second surrounding image I (predetermined region A) currently captured by the camera (image-capturing means) 10.
More specifically, the deviation degree calculation unit 23 calculates a deviation degree between the edge shape (first edge line) E1 stored in the edge storage unit 22 and the edge (second edge line) E2 detected by the current edge detection unit 21. In this calculation, the edge detected from the second surrounding image I (predetermined region A) currently captured by the camera (image-capturing means) 10 is used as the edge (second edge line) E2.
Firstly, the first method is described. As illustrated in
In the first method, the deviation degree calculation unit 23 firstly extracts a special point P2 on the detected edge (second edge line) E2 and a point P1 on the reference edge (first edge line) E1 which is estimated as corresponding to the special point P2. In this process, the deviation degree calculation unit 23 extracts, as the corresponding point, the point P1 next to the special point P2 in the vertical direction of the image, for example. Then, the deviation degree calculation unit 23 determines how many pixels the extracted two points P1, P2 are shifted from each other. In the example illustrated in
The deviation degree calculation unit 23 determines the deviation degrees between all corresponding points P1, P2. Specifically, the deviation degree calculation unit 23 calculates the deviation degrees between the special points and the corresponding points one by one from the leftmost point to the rightmost point on the detected edge E1 and the reference edge E2 and then calculates the sum of the calculated deviation degrees as a final deviation degree.
Next, the second method is described. In the second method, the deviation degree calculation unit 23 uses luminance gradients. Specifically, the deviation degree calculation unit 23 calculates luminance gradient directions (see reference signs D1, D2 in
Subsequently, the deviation degree calculation unit 23 calculates an angle θ formed by the two luminance gradient directions D1, D2. Then, the deviation degree calculation unit 23 determines the deviation degree between the two points P1, P2 as θ.
The deviation degree calculation unit 23 determines the deviation degrees between all corresponding points P1, P2. Specifically, the deviation degree calculation unit 23 calculates the deviation degrees from the luminance gradient directions one by one from the leftmost point to the rightmost point on the detected edge E1 and the reference edge E2 and then calculate the sum of the calculated deviation degrees as a final deviation degree.
Here,
It should be noted that the matching degree and the deviation degree have a reverse relationship where the edges can be judged as deviating (not matching) when the foregoing deviation degree is equal to or larger than the predetermined value, and can be judged as not deviating (matching) when the deviation degree is equal to or smaller than the predetermined value.
Next, the raindrop detection method is described in reference to a flowchart.
As described above, firstly, the camera (image-capturing means) 10 installed on the system-equipped vehicle V (a moving object) captures a current surrounding image (second surrounding image) I including the portion P of the system-equipped vehicle V (image-capturing step)
Then, as illustrated in
Subsequently, the deviation degree calculation unit 23 calculates the deviation degree between the edge E2 detected in step S1 and the reference edge E1 stores in the reference edge storage unit 22 (S2). Any one of the methods described in reference to
Next, the raindrop judgment unit 24 judges if the deviation degree calculated in step S2 is equal to or larger than the predetermined value (S3). If it is judged that the deviation degree calculated in step S2 is equal to or larger than the predetermined value (S3—YES), the raindrop judgment unit 24 determines that the matching degree has decreased, judges that a raindrop is attached to the less unit, and sends a notification of the judgment result to the warning device 30.
Thus, in step S3, a raindrop judging step is performed; wherein, if the deviation degree calculated in the calculating step is equal to or larger than the predetermined value, the matching degree is determined to have decreased and thereby it is judged that a raindrop is attached to the lens unit of the camera 10.
Here, when the raindrop judgment unit 24 judges that a raindrop is attached to the lens unit of the camera 10, the warning device 30 judges whether the deviation degree calculated in step S2 is equal to or higher than a given value (S4). For this deviation degree judgment, the warning device 30 can be provided with a deviation degree judgment unit, for example. The deviation degree judgment unit can be configured to judge that the deviation degree is high when the deviation degree calculated in step S2 is equal to or larger than the preset given value (a value larger than the predetermined value used for the determination in step S3) or to judge that the deviation degree is low when the deviation degree is equal to or smaller than the given value.
Then, if the deviation is judged as equal to or lower than the given value (the value used for the determination, in step S4) (S4: YES), the sensitivity to detect another vehicle (another moving object) from the current surrounding image (second surrounding image) 1 currently captured in the image-capturing step is lowered (S5). For example, detection sensitivity lowering means provided in the warning device 30 can lower the sensitivity to detect another vehicle by raising a detection threshold used by a not-illustrated vehicle detection unit to detect another moving object (another vehicle) from the current surrounding image (second surrounding image) I.
As for the method of lowering the vehicle detection sensitivity, the whole sensitivity can be lowered by adjusting a threshold for an entire difference or edge, or the sensitivity of the relevant image part (part where a raindrop is attached) can be lowered. Instead, the whole sensitivity can be adjusted first, and then the sensitivity of the relevant part (part where a raindrop is attached) can be further lowered.
Thereafter, if there is another vehicle, the warning device 30 presents the existence of the vehicle (S6). Specifically, the not-illustrated vehicle detection unit performs an operation of detecting another vehicle (another moving object), and if another vehicle (another moving object) is detected, the warning device 30 presents the existence of the vehicle. After that, the processing illustrated in
On the other hand, if the deviation degree is judged as equal to or higher than the given value (the value used for the determination in step S4) (S4: NO), the warning device 30 notifies that the system cannot operate (S7). Specifically, when the deviation degree is equal to or higher than the given value (the value used for the determination in step S4), it is judged that raindrops are attached to the lens unit of the camera 10 so heavily that detection of another vehicle by using the camera 10 is impossible, and a notification unit of the waning device 30 notifies that detection of another vehicle is impossible. After that, the processing illustrated in
Meanwhile, if it is judged that the deviation degree is equal to or smaller than the predetermined value (S3: NO), the raindrop judgment unit 24 determines that the matching degree has not decreased and judges that no raindrop is attached to the lens unit of the camera 10. After that, the processing illustrated in
As described above, according to the driving assistance system 1 and the raindrop detection method thereof of the present embodiment, the current surrounding image I including the portion P of the system-equipped vehicle V (a moving object) is captured, and the matching degree between the reference edge (first edge line) E1 detected from the first surrounding image I captured in normal conditions and the edge (second edge line) E2 detected from the currently-captured second surrounding image I is calculated. Then, in response to a decrease in the matching degree between the reference edge (first edge line) E1 and the edge (second edge line) E2, it is judged that a raindrop is attached to the lens unit of the camera (image-capturing means) 10.
Specifically, the image-capturing range includes not only the surroundings of the system-equipped vehicle V but also the portion P of the system-equipped vehicle V, and the edge of the portion P of the system-equipped vehicle V to be obtained by edge detection is stored as the reference edge shape (first edge line) E1. In addition, the edge (second edge line) E2 is detected for the portion P of the system-equipped vehicle V in the actually captured image, and the deviation degree from the stored edge shape is calculated. If the deviation degree is equal to or larger than the predetermined value, the matching degree is determined as decreased, and a raindrop is determined as attached to the lens unit. Here, if a raindrop is attached to the lens unit, the light is refracted by the raindrop, and thereby the stored edge shape and the detected edge deviate from each other. In this way, raindrop detection can be performed without changing the focal distance, and also raindrop detection can be performed while avoiding reduction in the detection accuracy for the surrounding environment.
In addition, when the raindrop judgment unit (raindrop judging means) 24 judges that a raindrop is attached to the lends unit of the camera (image-capturing means) 10, the sensitivity to detect another vehicle (another moving object) from the second surrounding image I currently captured by the camera (image-capturing means) 10 is lowered. This makes it possible to prevent an object other than another vehicle (another moving object) from being detected as another vehicle (other moving object) when using the camera 10 in which a raindrop is attached to the lens.
Moreover, since the camera 10 is installed with the optical axis directed obliquely downward from the horizontal direction, a raindrop attached to the lens unit can be caused to stay at a certain position in a lower portion of the lens. This makes it possible to prevent a situation where sequential change of the raindrop position makes edge detection difficult.
Additionally, the calculation of the deviation degree from the difference between the luminance gradient directions enables detection of a phenomenon where the luminance gradients change along with the formation of a lens system by the raindrop. Thus, the raindrop detection accuracy can be improved.
The driving assistance system 1 and the raindrop detection method thereof according to the present modified example are basically the same as those in the foregoing first embodiment, but different processing is performed after the raindrop judgment unit 24 makes the raindrop judgment.
Hereinafter, the raindrop detection method according to the present modified example is explained with reference to the flowchart in
To begin with, the same processing as in the above first embodiment is performed in steps S11 to S13. To be more precise, the camera (image-capturing means) 10 installed on the system-equipped vehicle V (a moving object) captures a current surrounding image (second surrounding image) I including the portion P of the system-equipped vehicle V.
Then, the current edge detection unit 21 detects the edge in the predetermined region A in the image captured by the camera 10 (S11).
Next, the deviation degree calculation unit 23 calculates the deviation degree between the edge E2 detected in step S1 and the reference edge E1 stored in the reference edge storage unit 22 (S12). Here, any one of the methods described in reference to
Then, the raindrop judgment unit 24 judges if the deviation degree calculated in step S2 is equal to or larger than the predetermined value (S13).
If it is judged that the deviation degree is equal to or larger than the predetermined value (S13: YES), the raindrop judgment unit 24 determined that the matching degree has decreased, judges that a raindrop is attached to the lens unit, and sends the notification of the judgment result to the warning device 30.
Here, when the raindrop judgment unit 24 judges that a raindrop is attached to the lens unit of the camera 10, raindrop detection sensitivity is lowered for the part which is judged as having the raindrop attached thereto, within the predetermined region A of the surrounding image (second surrounding image) I (S14). For example, raindrop detection sensitivity lowering means provided in the warning device 30 can raise the detection threshold for the part which is judged as having the raindrop attached thereto, within the predetermined region A of the current surrounding image (second surrounding image) I, and thereby can lower the raindrop detection sensitivity for the part in comparison to the remaining part.
Thereafter, the processing illustrated in
Moreover, as for the part with the adjusted sensitivity due to the raindrop attachment, it is preferable to perform processing which includes the excluded part again after a certain period of time has elapsed. This is because the raindrop attachment condition varies over time due to evaporation of raindrops or other reasons.
Meanwhile, if it is judged that the deviation degree is equal to or smaller than the predetermined value (S13: NO), the raindrop judgment unit 24 determines that the matching degree has not decreased and judges that no raindrop is attached to the lens unit of the camera 10. After that, the processing illustrated in
The foregoing modified example can also produce operations and effects similar to those in the above first embodiment.
Moreover, in the present modified example, when the raindrop judgment unit 24 judges that a raindrop is attached to the lens unit of the camera 10, the raindrop detection sensitivity is lowered for the part judged as having the raindrop within the predetermined region A in the surrounding image (second surrounding image) I. The lowering of the raindrop detection sensitivity for the part having a raindrop attached thereto, as described above, makes it easier to detect a raindrop newly attached to the predetermined region A.
A driving assistance system 1A and a raindrop detection method according to the present embodiment are basically the same as those in the aforementioned first embodiment.
Specifically, the driving assistance system 1A is configured to provide various kinds of information to the driver of a system-equipped vehicle V from an image-capturing result of the surroundings of the system-equipped vehicle V, and includes a camera (image-capturing means) 10, a computer 20A, and a warning device 30.
This camera 10 is also installed on the system-equipped vehicle V (a moving object) and is configured to capture surrounding images (first surrounding image and second surrounding image) I including a portion P of the system-equipped vehicle V.
In the present embodiment, the computer 20A includes a current edge detection unit (edge detecting means) 21, a reference edge storage unit (first edge line storing means: reference edge storing means) 22, an edge change degree calculation unit (calculating means) 23A, and a raindrop judgment unit (raindrop judging means) 24, as illustrated in
In the present embodiment, as illustrated in
Then, the edge change degree calculation unit (calculating means) 23A is configured to calculate a matching degree between a reference edge shape (first edge line) E1 detected from a first surrounding image I (predetermined region A1) captured by the camera (image-capturing means) 10 in normal conditions (in fine weather or the like), and an edge (second edge line) E2 detected from a second surrounding image I (predetermined region A1) currently captured by the camera (image-capturing means) 10.
Specifically, the edge change degree calculation unit (calculating means) 23A calculates an edge shape change degree between the edge shape (first edge line) E1 stored in the edge storage unit 22 and the edge (second edge line) E2 detected by the current edge detection unit 21. In this process, an edge detected from the second surrounding image I (predetermined region A1), currently captured by the camera (image-capturing means) 10 is used as the edge (second edge line) E2.
Here also, any one of the methods described in reference to
However, as for the method of calculating an edge shape change degree, a method can be employed in which a part where less edge appears than in the surrounding part is judged as having a raindrop attached thereto.
When this method is employed, the part having a raindrop attached thereto is judged in the following way.
To begin with, the are above the bumper set as the predetermined region A1 is an area where a road surface stably shows up as a steady background object, and the edge intensity (luminance difference between pixels) in this area is usually constant. For this reason, if there is no part having an edge intensity lower than its surrounding part, it can be judged that no raindrop is attached. On the other hand, if there is a part having an edge intensity lower than its surrounding part, the part having the lower edge intensity can be judged as a part to which a raindrop is attached. To be more specific, the integral of a part judged as having a low edge intensity when viewed in time sequence is calculated (for example, a counter for the part is incremented when the edge intensity is equal to or smaller than a predetermined value). Then, when the counter for the part reaches a predetermined value or above, the part is recognized as having a large deviation degree from the surrounding part and thereby judged as having raindrop detection. Note that this method may be employed for the case of detecting the edge (second edge line) E2 where the image of the bumper of the system-equipped vehicle V is captured in the predetermined region A, as in the first embodiment. To put it differently, the above process may be applied to a vehicle body, and when the edge intensity of a portion of the vehicle body becomes lower than the edge intensity of the vehicle body usually observed, the portion having the lower edge intensity may be judged as having a raindrop attached thereto.
Next, the raindrop detection method is explained with reference to a flowchart.
First of all, the camera (image-capturing means) 10 installed on the system-equipped vehicle V (a moving object) captures a current surrounding image (second surrounding image) I including the portion P of the system-equipped vehicle V (image-capturing step).
Next, as illustrated in
Thereafter, the edge change degree calculation unit (calculating means) 23A calculates an edge shape change degree between the edge E2 detected in step S21 and the reference edge E1 stored in the reference edge storage unit 22 (S22). The foregoing method or the like is employed as the method of calculating the edge shape change degree.
After that, the raindrop judgment unit 24 judges if the edge shape change degree calculated in step S22 is equal to or smaller than a predetermined value (S23). Then, if it is judged that the edge shape change degree is equal to or smaller than the predetermined value (S23: YES), the raindrop judgment unit 24 determines that the matching degree has decreased, judges that a raindrop is attached to the lens unit, and sends the notification of the judgment result to the warning device 30.
When the raindrop judgment unit 24 judges that the raindrop is attached to the lens unit of the camera 10 in this step, processing similar to those of steps S4 to S6 described in the foregoing first embodiment is performed.
Specifically, the warning device 30 judges if the deviation degree of the edge shape change degree calculated in step S22 is equal to or smaller than a given value (S24).
Then, if the deviation degree is judged as equal to or lower than the given value (the value used for the determination in step S24) (S4: YES), the sensitivity to detect another vehicle (another moving object) from the current surrounding image (second surrounding image) I currently captured in the image-capturing step is lowered (S25).
Thereafter, if there is another vehicle, the warning device 30 presents the existence of the vehicle (S26). After that, the processing illustrated in
On the other hand, if the deviation degree is judged as equal to or higher than the given value (the value used for the determination in step S24) (S24: NO), the warning device 30 notifies that the system cannot operate (S27). After that, the processing illustrated in
Meanwhile, if it is judged that the edge shape change degree is equal to or larger than the predetermined value (S23: NO), the raindrop judgment unit 24 determines that the matching degree has not decreased, and thereby judges that no raindrop is attached to the lens unit of the camera 10. After that, the processing illustrated in
the aforementioned embodiment can also produce operations and effects similar to those in the foregoing first embodiment.
The driving assistance system 1A and the raindrop detection method thereof according to the present modified example are basically the same as those in the foregoing second embodiment, but different processing is performed after the raindrop judgment unit 24 makes the raindrop judgment.
Hereinafter, the raindrop detection method according to the present modified example is explained with reference to a flowchart in
To begin with, the same processing as in the above second embodiment is performed in steps S31 to S33. Then, in steps S34 and S35, the same processing as in the above modified example of the first embodiment is performed.
To be more precise, the camera (image-capturing means) 10 installed on the system-equipped vehicle V (a moving object) captures a current surrounding image (second surrounding image) I including the portion P of the system-equipped vehicle V.
Then, the current edge detection unit 21 detects the edge for the predetermined region A1 in the image captured by the camera 10 (S31).
Next, an edge change degree calculation unit (calculating means) 23A calculates the edge shape change degree between the edge E2 detected in step S31 and the reference edge E1 stored in the reference edge storage unit 22 (S32). the foregoing method or the like is employed as the method of calculating the edge shape change degree.
Subsequently, the raindrop judgment unit 24 judges if the edge shape change degree calculated in step S32 is equal to or larger than the predetermined value (S33).
Then, if it is judged that the edge shape change degree is equal to or smaller than the predetermined value (S33: YES), the raindrop judgment unit 24 determines that the matching degree has decreased, judges that a raindrop is attached to the lens unit, and sends the notification of the judgment result to the warning device 30.
Here, when the raindrop judgment unit 24 judges that a raindrop is attached to the lens unit of the camera 10, raindrop detection sensitivity is lowered for the part which is judged as having the raindrop attached thereto, within the predetermined region A1 of the surrounding image (second surrounding image) I (S34).
After that, the processing illustrated in
Meanwhile, if it is judged that the edge shape change degree is equal to or larger than the predetermined value (S33: NO), the raindrop judgment unit 24 determines that the matching degree has not decreased and thereby judges that no raindrop is attached to the lens unit of the camera 10. After that, the processing illustrated in
The aforementioned modified example can also produce operations and effects similar to those in the foregoing first embodiment and the modified example thereof, or the foregoing second embodiment.
A driving assistance system 1B and a raindrop detection method thereof according to the present embodiment are basically the same as those in the foregoing first embodiment.
Specifically, the driving assistance system 1B is configured to provide various kinds of information to the driver of a system-equipped vehicle V from an image-capturing result of the surrounding of the system-equipped vehicle V and includes a camera (image-capturing means) 10, a computer 20B, and a warning device 30.
This camera 10 is also installed on the system-equipped vehicle V (a moving object) and is configured to capture surrounding images (first surrounding image and second surrounding image) 1 including a portion P of the system-equipped vehicle V.
As illustrated in
The edge detection unit 21 is configured to detect an edge E for the portion P of the system-equipped vehicle V in an image captured by the camera 10. The captured image includes a captured image of the portion P (for example, the bumper) of the system-equipped vehicle V in addition to a road surface and the like. The edge detection unit 21 detects an edge E in a predetermined region A (that is a region where an image of the bumper of the system-equipped vehicle V is to be captured) within the captured image.
The reference edge storage unit 22 is configured in advance to store, as an initial value, a reference edge shape (first edge line) E1 for the portion P of the system-equipped vehicle V targeted for image capturing by the camera 10. In the present embodiment, the reference edge storage unit 22 stores as the reference edge shape (first edge line) E1 a first edge line E1 detected from a first surrounding image I captured by the camera (image-capturing means) 10 in normal conditions.
Specifically, the reference edge storage unit 22 stores in advance the edge shape E1 obtained from the portion P of the system-equipped vehicle V in normal conditions (in fine weather or the like) where no raindrops are attached to the lens unit of the camera 10.
As described above, the reference edge storage unit (first edge line storing means: reference edge storing means) 22 is configured to store the reference edge shape (first edge line) E1 detected from the first surrounding image I (predetermined region A) captured in normal conditions (in fine weather or the like) by the camera (image-capturing means) 10.
In contrast, the edge detection unit (edge detecting means) 21 is configured to detect an edge (second edge line) E2 from a second surrounding image 1 (predetermined region A) currently captured by the camera (image-capturing means) 10.
The deviation degree calculation unit 23 is configured to calculate a matching degree between the reference edge shape (first edge line) E1 detected from the first surrounding image I (predetermined region A) captured in normal conditions (in fine weather or the like) by the camera (image-captured means) 10 and the edge (second edge line) E2 detected from the second surrounding image I (predetermined A) currently captured by the camera (image-capturing means) 10.
More specifically, the deviation degree calculation unit 23 calculates a deviation degree between the edge shape (first edge line) E1 stored in the edge storage unit 22 and the edge (second edge line) E2 detected by the current edge detection unit 21. In this calculation, an edge detected from the second surrounding image I (predetermined region A) currently captured by the camera (image-capturing means) 10 is used as the edge (second edge line) E2.
When the deviation degree calculated by the deviation degree calculation unit 23 is equal to or larger than a predetermined value, the raindrop judgment unit 24 determines that the matching degree has decreased and judges that a raindrop is attached to the lens unit of the camera 10. As described in reference to
Moreover, in the present embodiment, the reference edge storage unit 22 has a function to update the reference edge E1. To be more specific, if the deviation degree calculated by the deviation degree calculation unit 23 is kept equal to or smaller than a prescribed value over a certain period of time, the reference edge storage unit 22 stores, as the reference edge shape, an edge detected by the edge detection unit 21 within the certain period of time. This configuration also enables handling of a case where even the position or the optical axis of the camera 10 is gradually displaced.
Next, the raindrop detection method is explained with reference to a flowchart
Then, the raindrop judgment unit 24 determines if the deviation degree calculated in step S42 is equal to or larger than a predetermined value (S43). If it is determined that the deviation degree is equal to or larger than the predetermined value (S43: YES), the raindrop judgment unit 24 determines that the matching degree has decreased, determines that a raindrop is attached to the lens unit, and sends a notification of the determination result to the warning device 30. In response to the notification, the warning device 30 issues a warming that the camera view is poor (S44). Then, the processing illustrated in
Meanwhile, if it is determined that the deviation degree is not equal to or larger than the predetermined value (S43: NO), the computer 20B executes update processing of the reference edge E1 in steps S45 to S47. Specifically, the update processing is performed as follows. Firstly, the deviation degree calculation unit 23 calculates the absolute value of the difference between the deviation degree for the current processing cycle and the deviation degree for the previous processing cycle (S45). Then, the deviation degree calculation unit 23 calculates the sum of the absolute values for a predetermined number of past processing cycles and determines if the sum is smaller than a prescribed value (S46).
If the sum is determined as smaller than the prescribed value (S46: YES), it can be said that the deviation degree has stayed small over a predetermined period of time, and it can be assumed that any calculated deviation degree is generated due to a gradual displacement of the position or the optical axis of the camera 10. Thus, the deviation degree calculation unit 23 sends the reference edge storage unit 22 a signal indicating the determination result, and the reference edge storage unit 22 updates the reference edge E1 to the detected edge E2 that is detected by the edge detection unit 21 within the predetermined period of time (S47). After that, the processing illustrated in
Here, the detected edge E2 to which the reference edge E1 is updated may be any of the detected edges E2 detected within the predetermined period of time and may be the latest detected edge E2 or the earliest detected edge E2. Instead, an average of multiple detected edges E2 may be also used.
Meanwhile, if the sum of the deviation degrees is determined as not smaller than the predetermined value (S46: NO), the processing illustrated in
The foregoing embodiment can also produce operations and effects similar to those in the above first embodiment.
Moreover, when the deviation degree is kept equal to or smaller than the prescribed value over the predetermined period of time, and edge detected within the predetermined period of time is stored as the reference edge shape and thereby the reference edge shape is updated. This makes it possible to store the new edge shape corresponding to a deformation of a vehicle component or a change in the installation position or the like of the camera 10.
A driving assistance system 1C and a raindrop detection method according to the present embodiment are basically the same as those in the aforementioned third embodiment.
As illustrated in
For this reason, the deviation degree calculation unit 23 calculates a deviation degree between the detected edge E2 detected in the current processing by the edge detection unit 21 and the past edge (first edge line) stored in the past edge storage unit 25. Then, if the deviation degree is equal to or larger than a predetermined value, the raindrop judgment unit 24 determines that the matching degree has decreased and determines that a rain drop is attached to the lens unit, and the warning device 30 warns that the view of the camera 10 is poor.
Next, the raindrop detection method according to the fourth embodiment is described. In the raindrop detection method according to the fourth embodiment, the reference edge E1 is not updated. Thus, the processing in steps S41 to S44 illustrated in
The foregoing present embodiment can also produce operations and effects similar to those in the above third embodiment.
Moreover, the present embodiment is capable of preventing a situation where sequential change of a raindrop position or shape makes edge detection difficult, and capable of detecting a phenomenon where the luminance gradients change along with the formation of a lens system by the raindrop. Thus, the raindrop detection accuracy can be improved.
Moreover, according to the present embodiment, since the deviation degree from the past edge is calculated, it is possible to detect how the edge shape of a raindrop changes over time depending on airflow during the running of the vehicle. Thus, the raindrop detection accuracy can be improved.
Although the present invention has been described above based on the embodiments, the present invention is not limited to the foregoing embodiments, but may be altered without departing from the spirit of the present invention, or may be implemented in any combination of the embodiments.
For example, in the foregoing embodiments, the driving assistance system is installed in the vehicle, but is not limited to this. The driving assistance system may be installed in a bicycle, an automatic navigation robot, or the like. In addition, the deviation degree may be calculated not only using the methods described above but also using any other method.
Additionally, depending on day or night, the edge detection region (predetermined region) may be switched between a region including the bumper and a region located near the bumper but not including the bumper, for example.
Further, for a dark place, the detection threshold may be set low so that a substance having high luminance can be determined as a raindrop.
Furthermore, the detection threshold may be set low so that a detected edge shape is more likely to be judged as a raindrop as the circularity of the edge shape becomes higher.
The entire contents of Japanese Patent Application No. 2011-088727 (filed on Apr. 13, 2011) are incorporated herein by reference.
According to the present invention, it is possible to provide a driving assistance system and a raindrop detection method thereof which are capable of detecting a raindrop while avoiding reduction in the detection accuracy for the surrounding environment.
1 driving assistance system
10 camera (image-capturing means)
20 computer
21 current edge detection unit (edge detecting means)
22 reference edge storage unit (reference edge storing means: first edge line storing means)
23 deviation degree calculation unit (calculating means)
24 raindrop judgment unit (raindrop determining means)
25 past edge storage unit (past edge storing means: first edge line storing means)
30 warning device
V system-equipped vehicle
Number | Date | Country | Kind |
---|---|---|---|
2011-088727 | Apr 2011 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2012/055900 | 3/8/2012 | WO | 00 | 10/11/2013 |