The present invention relates to an apparatus for monitoring a periphery of a vehicle using an image captured by one or more infrared cameras, more specifically relates to a vehicle periphery monitoring apparatus for extracting an object by a binarization process of the captured image.
Conventionally, an apparatus for capturing an image around a vehicle by an infrared camera that is mounted on the vehicle, and binirizing the captured image to extract an object having higher temperature such as a pedestrian and animal has been proposed. In the patent document 1 below, a method for generating a luminance histogram of an image captured by an infrared camera and determining a threshold value that divides the captured image into a background image and an object image based on the luminance histogram. Through a binarization process using such a threshold value, an object having higher temperature is distinguished and extracted from a background.
Patent Document 1: Japanese patent publication laid-open No. 2003-216949
In addition to living bodies such as pedestrians and animals, artificial structures such as utility poles and walls may exist around a vehicle. In order to distinguish and extract living bodies such as pedestrians and animals from the background as higher temperature objects, it is desirable that such artificial structures are classified into the background in the binarization process. However, depending on the kinds and placement of the artificial structures and the environment such as temperature around the vehicle, the artificial structures may be classified into the higher temperature objects even if the above conventional method is used. Therefore, a technique for distinguishing and extracting a desired object from a background with better accuracy in the binarization process, independently of the environment around the vehicle, is desired.
According to one aspect of the present invention, a vehicle periphery monitoring apparatus for monitoring a periphery of a vehicle using an image captured by an infrared camera mounted on the vehicle detects an outside temperature of the vehicle. A temperature difference between a surface temperature of an object estimated based on the outside temperature and the outside temperature is calculated. Based on a luminance value of the background in the captured image and a luminance difference corresponding to the temperature difference, a luminance value of an object in the captured image is determined. The captured image obtained by the infrared camera is binarized by using the luminance value of the object as a threshold value to extract the object.
A relationship between a surface temperature of an object and an outside temperature is previously determined, and hence the surface temperature can be estimated from the outside temperature. This invention is based on this findings. A temperature difference between the detected outside temperature and a surface temperature of an object estimated based on the outside temperature is calculated. Because the luminance value of the background can be considered as corresponding to the outside temperature, a luminance value corresponding to an object can be determined based on the luminance value of the background and a luminance difference corresponding to the temperature difference. By conducting the binarization using the luminance value thus determined as a threshold value, the object can be separated and extracted well from the background portion that is other than the object. For example, when a pedestrian is extracted as an object, it is prevented that an object such as an artificial structure is erroneously extracted, by previously determining a relationship between the surface temperature of the pedestrian and the outside temperature.
Other features and advantages of the present invention will be apparent from the following detailed description of the present invention and the accompanying drawings.
Preferred embodiments of the present invention will be described referring to the attached drawings.
As shown in
The image processing unit 2 includes an A/D converter circuit for converting input analog signals to digital signals, an image memory for storing digitized image signals, a CPU (central processing unit) for carrying out arithmetic operations, a RAM (Random access memory) used by the CPU for storing data being processed in the arithmetic operations, a ROM (Read Only memory) storing programs executed by the CPU and data (including tables and maps) to be used by the programs, and an output circuit for outputting driving signals to the speaker 3, display signals to the HUD 4, and the like. Output signals from the cameras 1R, 1L and the sensor 5 are converted to digital signals and input into the CPU. As shown in
In steps S11 through S13, output signals (that is, data of captured images) from the cameras 1R, 1L are received, A/D converted and stored in the image memory. Data of images thus stored are gray scale images having luminance values information.
The following steps S14 through S19 are a process for distinguishably extracting a desired object from the background in a binarization process. This embodiment will be described for a case where the desired object is a pedestrian.
In step S14, an outside temperature i (° C.) detected by the outside temperature sensor 5 is obtained. In step S15, a luminance value Tb of the background is determined
The luminance value of the background may be determined by any technique. In this embodiment, a luminance value histogram is created based on the gray scale image. A luminance value having the highest frequency is used as the luminance value Tb of the background. This is because an area occupied by the background is generally largest in the captured image.
In step S16, a map as shown in
The surface temperature fa changes as indicated by a curve 101 with respect to the outside temperature i. The surface temperature fa is higher as the outside temperature i is higher. For a given outside temperature i, a difference of the surface temperature fa(i) with respect to the outside temperature i is indicated by a difference between the curve 101 and a line 103 (which is a straight line indicating fa=i), and is referred to as a surface temperature difference, which is represented by F(i). That is, F(i)=surface temperature fa(i)−outside temperature i. As shown in the figure, there is a tendency that the surface temperature difference F(i) is smaller as the outside temperature i is higher.
In order to improve the accuracy of extracting an object, a predetermined margin range T (° C.) with respect to F(i) is set in this embodiment. An upper limit of the margin range is indicated by a dotted line 101U. A difference between the upper limit and the outside temperature i is represented by F(i)max. A lower limit of the margin range is indicated by a dotted line 101L. A difference between the lower limit and the outside temperature i is represented by F(i)min.
The map as shown in
Alternatively, the upper limit value F(i)max and the lower limit value F(i)min for the surface temperature difference F(i) corresponding to each outside temperature i may be stored in a memory. In this case, determining the surface temperature fa from the outside temperature i can be skipped. The upper limit value F(i)max and the lower limit value F(i)min can be directly determined from the outside temperature i.
Referring back to
dTmax=SiTF×F(i)max
dTmin=SiTF×F(i)min (1)
In step S18, a threshold value for the binarization process is determined. Referring to
Tcmax=Tb +dTmax
Tcmin=Tb +dTmin (2)
The upper limit luminance value Tcmax and the lower limit luminance value Tcmin are set in the threshold values for the binarization process. A region 111 defined by these two threshold values is shown in
In step S19, by using the threshold values thus set in step S18, the binarization process is applied to the gray scale image obtained in step S13 (in this embodiment, the image captured by the camera 1R is used, but alternatively, the image captured by the camera 1L may be used). For each pixel in the captured image, when a luminance value of the pixel is within the luminance region 111, the pixel is set to a white region having a value of 1 because the pixel is determined as constituting an object to be extracted. When a luminance value of the pixel is not within the luminance region 111, the pixel is set to a black region having zero because the pixel is determined as constituting the background.
Here, referring to
In the gray scale image, in addition to a pedestrian 121, artificial structures such as an electric pole 125 and an automobile 127 are captured. According to a conventional method, not only a pedestrian but also these artificial structures 125 and 127 may be extracted as an object or a white region, as shown in (b), depending on threshold values used in the binarization.
In contrast, according to the above technique of the present invention, the surface temperature of an object (in this embodiment, a pedestrian) with respect to the outside temperature is estimated, and a luminance region of the object is established based on a temperature difference of the estimated surface temperature with respect to the outside temperature. Therefore, only a head portion of the pedestrian 121 can be extracted as shown by the white region 131 of (c) (this region is referred to as a head portion region, hereinafter). Even when the artificial structure 125 and the pedestrian 121 are overlapped in the captured image as shown in (a), only the pedestrian 121 can be easily extracted as shown by the white region 131 of (c). Thus, according to the present invention, an object can be better distinguished and extracted from the background portion that is other than the object.
Referring back to
Referring to
In order to achieve this estimation, a general size of a pedestrian in the real space, that is, the width Wa of the head and the height Ha are predetermined. Wa and Ha may be set based on the average value for adults (for example, Wa is 20 centimeters and Ha is a value within a range from 160 to 170 centimeters).
Furthermore, (c) is a diagram where a placement relationship between the camera 1R and the object is represented on an XZ plane. (d) is a diagram where a placement relationship between the camera 1R and the object is represented on a YZ plane. Here, X indicates a direction of the width of the vehicle 10. Y indicates a direction of the height of the vehicle 10. Z indicates a direction of a distance from the vehicle 10 to the object. The camera 1R comprises an imaging element 11R and a lens 12R. f indicates a focal distance of the lens 12R.
Let the distance to the object be Z (centimeters). From the diagram of (c), the distance Z is calculated as shown by the equation (3). Here, pcw indicates an interval between pixels in X direction, that is, a length (centimeters) per pixel in X direction.
Z=Wa×f/(w×pcw) (3)
From the diagram of (d), the height h (centimeters) of the pedestrian in the captured image is calculated using the distance Z, as shown by the equation (4). Here, pch indicates an interval between pixels in Y direction, that is, a length (centimeters) per pixel in Y direction.
h=(Ha/pch)×f/Z (4)
Thus, the size of the pedestrian in the captured image can be estimated as having the width w and the height h. Alternatively, the fact that the width of the body is generally larger than the width of the head can be taken into account. In this case, a value obtained by adding a predetermined margin value to the width w of the head portion region may be used in place of the above-described width w.
Referring back to
Referring back to
If the object is determined as a pedestrian in step S22, the process proceeds to step S23 where a warning determination process is performed. In this process, it is determined whether a warning should be actually output or not to a driver. If the result of this determination is affirmative, the warning is output.
For example, it is determined whether a brake operation is being performed by a driver of the vehicle from the output of a brake sensor (not shown in the figure). If the brake operation is not being performed, the warning may be output. The warning output may be implemented by issuing a warning with voice through the speaker 3 while displaying the image obtained by, for example, the camera 1R on the screen 4a in which the pedestrian is emphatically displayed. The emphatic display is implemented by any technique. For example, the object is emphatically displayed by surrounding the object by a colored frame. Thus, the driver can more surely recognize the pedestrian in front of the vehicle. Alternatively, any one of the warning voice and the image display may be used to implement the warning output.
As another method of the step S20, for example, the height h of the pedestrian in the captured image may be calculated from a height of the head portion region (a height of the rectangle circumscribing the head portion region 131 can be used and is expressed in terms of the number of pixels) and the number of heads tall. For example, if the height of the head portion region 131 is hb and the average height for adults is seven-heads tall, then the height h of the pedestrian can be estimated as h=7×hb.
As yet another method of steps S20 and S21, a method for determining a road surface from luminance values in a region below the head portion region 131 to identify the object region 141 may be employed. This method will be briefly described referring to
In the above embodiments, a luminance value having the highest frequency in the luminance value histogram is set in the luminance value Tb of the background, which is brought into correspondence with the outside temperature. Alternatively, the outside temperature and the temperature of the road surface may be distinguished to determine the luminance value Tb of the background. More specifically, in the case where the camera is placed in the front portion of the vehicle as shown in
The map is referred to based on the detected outside temperature i to determine a corresponding temperature R of the road surface. A temperature difference between the road surface temperature R and the outside temperature i is calculated. The parameter SiTF as described above is used to convert the temperature difference into a luminance difference dTi. Here, referring to
More preferably, because the temperature difference between the road surface temperature and the outside temperature may vary depending on a value of one or more external environment parameters such as a weather condition (sunny or not, wind speed, amount of rainfall, etc.) and/or time passed from sunset, a map may be created and stored for each value of a predetermined external environment parameter. A map according to the external environment parameter value on that day is selected and used.
Similarly, the map of
In the above embodiments, the upper limit value F(i)max and the lower limit value F(i)min that define the margin range T are established for the surface temperature difference F(i) as described referring to
In the above embodiments, a case where an object to be extracted in the binarization process is a pedestrian is described as one example. Alternatively, an object to be extracted may be another living body such as an animal. For example, a map as shown in
The present invention is applicable to an object having a surface temperature that can be pre-defined with respect to the outside temperature via experiments and/or simulations as shown by a map of
Number | Date | Country | Kind |
---|---|---|---|
2008-274483 | Oct 2008 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2009/005261 | 10/8/2009 | WO | 00 | 6/8/2011 |