FUSION OF FAR INFRARED AND VISIBLE IMAGES IN ENHANCED OBSTACLE DETECTION IN AUTOMOTIVE APPLICATIONS

Abstract
A method in computerized system mounted on a vehicle including a cabin and an engine. The system including a visible (VIS) camera sensitive to visible light, the VIS camera mounted inside the cabin, wherein the VIS camera acquires consecutively in real time multiple image frames including VIS images of an object within a field of view of the VIS camera and in the environment of the vehicle. The system also including a FIR camera mounted on the vehicle in front of the engine, wherein the FIR camera acquires consecutively in real time multiple FIR image frames including FIR images of the object within a field of view of the FIR camera and in the environment of the vehicle. The FIR images and VIS images are processed simultaneously, thereby producing a detected object when the object is present in the environment.
Description

BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will become fully understood from the detailed description given herein below and the accompanying drawings, which are given by way of illustration and example only and thus not limitative of the present invention:



FIG. 1 (prior art) illustrates a vehicle with distance measuring apparatus, including a visible light camera and a computer useful for practicing embodiments of the present invention;



FIG. 2 (prior art) graphically demonstrates the error in distance, as result of a two pixel error in horizon estimate, for three different camera heights;



FIG. 3 (prior art) graphically defines the world coordinates of a camera mounted on a vehicle;



FIG. 4
a (prior art) illustrate the focus of expansion phenomenon;



FIG. 4
b (prior art) graphically defines epipolar geometry;



FIG. 5
a is a side view illustration of an embodiment of a vehicle warning system according to the present invention;



FIG. 5
b is atop view illustration of the embodiment of FIG. 5a;



FIG. 6 schematically shows a following vehicle having a distance warning system, operating to provide an accurate distance measurement from a pedestrian in front of the following vehicle, in accordance with an embodiment of the present invention;



FIG. 7
a exemplifies a situation where a pedestrian is on the side-walk, which is not on the ground plane of the host vehicle; and



FIG. 7
b depicts an example of severe curves in road that place the feet of the pedestrian below the road plane as defined by the host vehicle wheels;



FIG. 8 is flow diagram which illustrates an algorithm for refining distance measurements, in accordance with embodiments of the present invention;



FIG. 9 is flow diagram which illustrates an algorithm for refining distance measurements, in accordance with embodiments of the present invention;



FIG. 10 is flow diagram which illustrates an algorithm for tracking based on both FIR and VIS images, in accordance with embodiments of the present invention;



FIG. 11
a is flow diagram which illustrates algorithm steps for vehicle control, in accordance with embodiments of the present invention; and



FIG. 11
b is flow diagram which illustrates algorithm steps for verifying human body temperature, in accordance with embodiments of the present invention.





DESCRIPTION OF THE PREFERRED EMBODIMENTS

The present invention is of a system and method of processing image frames of an obstacle as viewed in real time from two cameras mounted in a vehicle: a visible light (VIS) camera and a FIR camera, whereas the VIS camera is mounted inside the cabin behind the windshield and the FIR camera is mounted in front of the engine, so that the image is not masked by the engine heat. Specifically, at night scenes, the system and method processes images of both cameras simultaneously, whereas the FIR images typically dominate detection and distance measurements in distances over 25 meters, the VIS images typically dominate detection and distance measurements in distances below 10 meters, and in the range of 10-25 meters, stereo processing of VIS images and corresponding FIR images is taking place.


Before explaining embodiments of the invention in detail, it is to be understood that the invention is not limited in its application to the details of design and the arrangement of the components set forth in the following description or illustrated in the drawings. The invention is capable of other embodiments or of being practiced or carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein is for the purpose of description and should not be regarded as limiting.


Embodiments of the present invention are preferably implemented using instrumentation well known in the art of image capture and processing, typically including an image capturing devices, e.g. VIS camera 110, FIR camera 120 and an image processor 130, capable of buffering and processing images in real time. A VIS camera typically has a wider FOV of 3520 -50° (angular) which corresponds to a focal length of 8 mm-6mm (assuming the Micron MT9V022, a VGA sensor with a square pixel size of 6 um), and that enables obstacle detection in the range of 90-50 meters. VIS camera 110 preferably has a wide angle of 42°, f/n. A FIR camera typically has a narrower FOV of 15°-25°, and that enables obstacle detection in the range above 100 meters. FIR camera 120 preferably has a narrow angle of 15°, f/n


Moreover, according to actual instrumentation and equipment of preferred embodiments of the method and system of the present invention, several selected steps could be implemented by hardware, firmware or by software on any operating system or a combination thereof. For example, as hardware, selected steps of the invention could be implemented as a chip or a circuit. As software, selected steps of the invention could be implemented as a plurality of software instructions being executed by a computer using any suitable operating system. In any case, selected steps of the method and system of the invention could be described as being performed by a processor, such as a computing platform for executing a plurality of instructions.


Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The methods, and examples provided herein are illustrative only and not intended to be limiting.


By way of introduction the present invention intends to provide in a vehicle tracking or control system which adequately detects obstacles in ranges of zero and up to and more than 100 meters from the front of the host vehicle. The system detects obstacles in day scenes, lit and unlit night scenes and at different weather conditions. The system includes a computerized processing unit and two cameras mounted in a vehicle: a visible light (VIS) camera and a FIR camera, whereas the VIS camera is mounted inside the cabin behind the windshield and the FIR camera is mounted in front of the engine and detects heat emitting objects. The cameras are mounted such that there is a wide baseline for stereo analysis of corresponding VIS images and FIR images.


It should be noted, that although the discussion herein relates to a forward moving vehicle equipped with VIS camera and FIR camera pointing forward in the direction of motion of host vehicle moving forward, the present invention may, by non-limiting example, alternatively be configured as well using VIS camera and FIR camera pointing backward in the direction of motion of host vehicle moving forward, and equivalently detecting objects and measure the range therefrom.


It should be further noted that lie principles of the present invention are applicable in Collision Warning Systems, such as Forward Collision Warning (FCW) systems based on scale change computations, and other applications such as headway monitoring and Adaptive Cruise Control (ACC) which require knowing the actual distance to the vehicle ahead. Another application is Lane Change Assist (LCA), where VIS camera is attached to or integrated into the side mirror, facing backwards. In the LCA application, a following vehicle is detected when entering a zone at specific distance (e.g. 17 meters), and thus a decision is made if it is safe to change lanes.


Object detecting system 100 of the present invention combines two commonly used object detecting system: a VIS camera based system a FIR camera based system. No adjustment to the mounting location of the two sensors 110 and 120 is required, compared with the prior art location. The added boost in performance of the object detecting system 100 of the present invention, benefits by combining existing outputs of the two sensors 110 and 120 with computation and algorithms, according to different embodiments of the present invention. Reference is now made to FIG. 5a, which is a side view illustration of an embodiment of following vehicle 50 having an object detecting system 100 according to the present invention, and FIG. 5b which is a top view illustration of the embodiment of FIG. 5a. System 100 includes a VIS camera 110, a FIR camera 120 and an Electronic Control Unit (ECU) 130.


The present invention combines the output from a visible light sensor 110 and a FIR sensor 120. Each sensor performs an important application independent of the other sensor, and each sensor is positioned in the vehicle 50 to perform the sensor own function optimally. The relative positioning of the two sensors 110 and 120 has been determined to be suitable for fusing together the information from both sensors and getting a significant boost in performance:

    • a) There is a large height difference between the two sensors 110 and 120, rendering a very wide baseline for stereo analysis. The difference in the height at which the two sensors 110 and 120 are mounted, is denoted by dh, as shown in FIG. 5a.
    • b) Typically, there is also a lateral displacement difference between the sensors 110 and 120, giving additional baseline for stereo analysis performed by processing unit 130. The difference in the lateral displacement is denoted by dx, as shown in FIG. 5b.
    • c) VIS camera 110 has a wider FOV (relative to the FOV of FIR camera 120) and is 1-2 meters behind FIR camera 120 so that VIS camera 110 covers the critical blind spots of FIR camera 120, which also has a narrower FOV (relative to the FOV of VIS camera 110). The blind spots are close to vehicle 50 and are well illuminated even by the vehicle's low beams. In the close range regions (illuminated by the vehicle's low beams), pedestrian detection at night can be done with VIS camera 110. The difference in the range displacement is denoted by dz, as shown in FIG. 5a.


An aspect of the present invention is to enhance range estimates to detected objects 10. FIG. 6 shows a following vehicle 50 having an object detecting system 100, operating to provide an accurate distance D measurement from a detected object, such as a pedestrian 10, in front of the following vehicle 50, in accordance with an embodiment of the present invention. While the use of road geometry (i.e. triangulation) works well for some applications, using road geometry is not really suitable for range estimation of pedestrians 10 with FIR cameras 120, given a camera height of less than 75 centimeters.


The computation of the pedestrian 10 distance D using the ground plane assumption uses equation 1, where f is the focal length of camera used (for example, 500 pixels for FOV=36° of a VIS camera 110), y=ybot (shown in FIG. 6) is the position of bottom 12 of the pedestrian 10 in the image relative to the horizon, and Hcam is a camera height. The assumption here is that pedestrian 10 is standing on road plane 20, but this assumption can be erroneous, as depicted in FIGS. 7a and 7b.


Improving Range Estimation in a System Based on FIR Cameras:

Typically, in FIR images, the head and feet of a pedestrian 10 can be accurately located in the image frame acquired (within 1 pixel). The height of pedestrian 10 is assumed to be known (for example, 1.7 m) and thus, a range estimate can be computed:









D
=


f
*
1.7

h





(
2
)







where h is the height of pedestrian 10 in the image and f is the focal length of the camera.


A second feature that can be used is the change of image size of the target due to the change in distance of the camera from the object, which is herein referred to as “scale change”.


Two assumptions are made here:

    • a) The speed of vehicle 50 is available and is non-zero.
    • b) The speed of pedestrian 10 in the direction parallel to the Z axis of vehicle 50 (see FIG. 3) is small relative to the speed of vehicle 50.


      The height of pedestrian 10 in two consecutive images is denoted as h1 and h2, respectively. The scale change s between the two consecutive images is then defined as:









s
=



h
1

-

h
2



h
1






(
3
)







It can be shown that:









s
=


v





Δ





t

D





(
4
)







or range D is related to scale change by:









D
=


v





Δ





t

s





(
5
)







where v is the vehicle speed and αt is the time difference between the two consecutive images.


The estimation of range D becomes more accurate as host vehicle 50 is closing in on pedestrian 10 rapidly and the closer host vehicle 50 is to pedestrian 10 the more important is range measurement accuracy. The estimation of range D using the scale change method is less accurate for far objects, as the scale change is small, approaching the measurement error. However, measurements of scale change may be made using frames that are more spaced out over time.


For stationary pedestrians or objects 10 not in the path of vehicle 50 (for example pedestrians 10 standing on the sidewalk waiting to cross), the motion of the centroid of object 10 can be used. Equation 5 is used, but the “scale change” s is now the relative change in the lateral position in the FIR image frame of object 10 relative to the focus of expansion (FOE) 90 (see FIG. 4a). Since a lateral motion of pedestrian 10 in the image can be significant, the fact that no leg motion is observed is used to determine that pedestrian 10 is in fact stationary,


Stereo Analysis methods Using FIR and VIS cameras

The range D estimates are combined using known height and scale change. Weighting is based on observed vehicle 50 speed. One possible configuration of an object detecting system 100, according to the present invention, is shown in FIGS. 1 and 2. FIR camera 120 is shown at the vehicle 50 front bumper. VIS camera 110 is located inside the cabin at the top of the windshield, near the rear-view mirror. Both cameras 110 and 120 are connected to a common computing unit or ECU 130.


The world coordinate system of a camera is illustrated in FIG. 3 (prior art). For simplicity, it is assumed that the optical axis of FIR camera 120 is aligned with the forward axis of the vehicle 50, which is the axis Z. Axis X of the world coordinates is to the left and axis Y is upwards.


The coordinate of optical center 126 of FIR camera 120 Pf is thus:










P
f

=

(



0




0




0



)





(
6
)







The optical center 116 of VIS camera 110 is located at the point Pv:









P
v

=

(




X
v






Y
v






Z
v




)





(
7
)







In one particular example the values can be:





Xv=−0.2 meter   (8)





Yv=(1.2−0.5)=0.7 meter   (9)





Zf=−1.5 meter   (10)


In other words, in the above example, VIS camera 110 is mounted near the rearview mirror, 1.5 meter behind FIR camera 120, 0.7 meter above FIR camera 120 a and 0.2 meter off to the right of optical center 126 of FIR camera 120. The optical axes of the two cameras 110 and 120 are assumed to be aligned (In practice, one could do rectification to ensure this alignment).


The image coordinates, in FIR camera 120, of a point P=(X, Y, Z)T in world coordinates is given by the equations:










P
f

=


(




x
f






y
f




)

=

(






f
1


X

Z








f
1


Y

Z




)






(
11
)







where f1 is the focal length of FIR camera 120 in pixels.


In practice, for example, the focal length could be f1=2000.


The image coordinates, in VIS camera 110, of the same point P is given by the equations:










P
v

=


(




x
v






y
v




)

=

(






f
2



(

X
-

X
v


)



Z
-

Z
v










f
2



(

Y
-

Y
v


)



Z
-

Z
v






)






(
12
)







where f2 is the focal length of VIS camera 110 in pixels.


In practice, for example, the focal length could be f2=800, indicating a wider field of view.


Next, for each point in the FIR image, processor 130 finds where a given point might fall in the VIS image. Equation 11 is inverted:









X
=



x
f


D


f
1






(
13
)






Y
=



y
f


D


f
1






(
14
)







The values for X and Y are inserted into equation 12. For each distance D a point pv is obtained. As V varies, p, draws a line called the epipolar line.


A way to project points from the FIR image to the VIS image, as a function of distance D, is now established. For example, given a rectangle around a candidate pedestrian in the FIR image, the location of that rectangle can be projected onto the VIS image, as a function of the distance D. The rectangle will project to a rectangle wit the same aspect ratio. Since the two cameras 110 and 120 are not on the same plane (i.e. Xf≠0) the rectangle size will vary slightly from the ratio








f
1


f
2


.




In order to compute the alignment between patches in the FIR and visible light images, a suitable metric must be defined. Due to very different characteristics of the two images, it is not possible to use simple correlation or sum square differences (SSD) as is typically used for stereo alignment.


A target candidate is detected in the FIR image and is defined by an enclosing rectangle. The rectangle is projected onto the visible image and test for alignment. The following three methods can be used to compute an alignment score. The method using Mutual Information can be used for matching pedestrians detected in the FIR to regions in the visible light image. The other two methods can also be used for matching targets detected in the visible light image to regions in the FIR image. They are also more suitable for vehicle targets.


Alignment using Mutual Information: since a FIR image of the pedestrian 10 is typically much brighter than the background, it is possible to apply a threshold to the pixels inside the surrounding rectangle. The binary image will define foreground pixels belonging to the pedestrian 10 and background pixels. Each possible alignment divides the pixels in the visible light image into two groups: those that are overlaid with ones from the binary and those that are overlaid with the zeros.


Histograms of the gray level values for each of the two groups can be computed. Mutual information can be used to measure the similarity between two distributions or histograms. An alignment which produces histograms which are the least similar are selected as best alignment, In addition to using histograms of gray levels, histograms of texture features such as gradient directions or local spatial frequencies such as the response to a series of Gabor filters and so forth, can be used.


Alignment using Sub-patch Correlation: a method for determining the optimal alignment based on the fact that even though the image quality is very different, there are many edge features that appear in both images. The method steps are as follows:

    • a) The rectangle is split into small patches (for example, 10×10 pixels each).
    • b) For each patch in the FIR that has significant edge like features (one way to determine if there are edge-like features is to compute the magnitude of the gradient at every pixel in the patch. The mean (m) and standard deviation (sigma) of the gradients are computed. The number of pixels that have a gradient which is at least N*sigma over the mean are counted. If the number of pixels is larger than a threshold T then it is likely to have significant edge features. For example: N can be 3 and T can be 3):


c) The patch is scaled according to the ratio








f
1


f
2


.




The location of the center of the patch along the epipolar line is computed in the visible image as a function of the distance D.


For each location of the center of the patch along the epipolar line, the absolute value of the normalized correlation is computed.


Local maxima points are determined.

    • d) The distance D, for which the maximum number of patches have a local maxima is selected as the optimal alignment and the distance D is the estimated distance to the pedestrian 10.


      Alignment using the Hausdorf Distance: The method steps are as follows:
    • a) A binary edge map of the two images (for example, using the canny edge detector) is computed.
    • b) For a given distance D, equations 11 and 12 are used to project the edge points in the candidate rectangle from the FIR image onto the visible light image.
    • c) The Hausdorf distance between the points is computed.
    • d) The optimal alignment (and optimal distance D) is the one that minimizes the Hausdorf distance.


Being able to align and match FIR images and corresponding visible light images, the following advantages of the fusion of the two sensors, can be rendered as follows:


1. Enhanced Range Estimates

Due to the location restrictions in mounting FIR camera 120 there is a significant height disparity between FIR camera 120 and VIS camera 110. Thus, The two sensors can be used as a stereo pair with a wide baseline giving accurate depth estimates.


One embodiment of the invention is particularly applicable to nighttime and is illustrated in tie flow chart of FIG. 8 outlining a recursive algorithm 200. A pedestrian 10 is detected in step 220 in a FIR image acquired in step 210. A distance D to selected points of detected pedestrian 10 is computed in step 230 and then, a matching pedestrian 10 (or part of a pedestrian, typically legs or feet, since these are illuminated even by the low beams) is found in the VIS image after projecting in step 242 the rectangle representing pedestrian 10 in the FIR image onto a location in the VIS image. Matching is performed by searching along the epipolar lines for a distance D that optimizes one of the alignment measures in step 250. Since a rough estimate of the distance D can be obtained, one can restrict the search to D that fall within that range. A reverse sequence of events is also possible. Legs are detected in the visible light image and are then matched to legs in the FIR image, acquired in step 212. After alignment is established in step 250, the distance measurements are refined in step 260, having two sources of distance information (from the FIR image and the VIS image) as well as stereo correspondence information.



FIG. 9 is flow diagram which illustrates a recursive algorithm 300 for refining distance measurements, in accordance with embodiments of the present invention. In a second embodiment of the present invention, in step 320, an obstacle such as a vehicle is detected in the VIS image, acquired in step 312. A distance D to selected points of detected pedestrian 10 is computed in step 330 and then, a corresponding vehicle is then detected in the FIR image, acquired in step 310. In step 340, the rectangle representing pedestrian 10 in the VIS image is projected onto a location in the VIS image. In the next step 350, features between the detected targets in the two images are aligned and matched. During the daytime, the FIR image of the vehicle is searched for the bright spots corresponding to the target vehicle wheels and the exhaust pipe. In the visible light image, the tires appear as dark rectangles extending down at the two sides of the vehicle. The bumper is typically the lowest horizontal line in the vehicle. At night, bright spots in both images correspond to the taillights. The features are matched by searching along the epipolar lines. Matched features are used in step 360 for stereo correspondence.


The accurate range estimates can be used for applications such as driver warning system, active braking and/or speed control. The driver warning system can perform application selected from the group of applications consisting of: detecting lane markings in a road, detecting pedestrians, detecting vehicles, detecting obstacles, detecting road signs, lane keeping, lane change assist, headway keeping and headlights control.


2. Extending the FIR Camera Field of View

In order to provide good visibility of pedestrians 10 at a far distance (typically over 50 meters), FIR camera 120 has a narrow FOV. One drawback of a narrow FOV is that when a vehicle 50 approaches a pedestrian 10 which is in the vehicle 50 path but not imaged at the center of the FIR image, the pedestrian 10 leaves the FIR camera FOV. Thus, in the range of 0-10 meters, where human response time is too long and automatic intervention is required, the target 10 is often not visible by FIR camera 120 and therefore automatic intervention such as braking cannot be applied.


The FOV of VIS camera 110 is often much wider. Furthermore, the VIS camera 110 is mounted inside the cabin, typically 1.5 meter to 2 meters to the rear of FIR camera 120. Thus, the full vehicle width is often visible from a distance of zero in front of the vehicle 50.



FIG. 10 is flow diagram which illustrates a recursive algorithm 400 for tracking based on both FIR and VIS images acquired in respective step 410 and 412, in accordance with embodiments of the present invention. In a third embodiment of the present invention, a pedestrian 10 is detected in step 420, in the FIR image, an accurate range l can be determined in step 430, and matched to a patch in the VIS image (as described above) in step 442. In step 280, the pedestrian 10 is tracked in each camera separately using standard SSD tracking techniques and the match is maintained. As the pedestrian 10 leaves FIR camera 120 FOV as determined in step 270, the tracking is maintained by VIS camera 110 in step 290. Since the absolute distance was determined while still in FIR camera 120 FOV, relative changes in the distance, which can be obtained from tracking, is sufficient to maintain accurate range estimates. The pedestrian 10 can then be tracked all the way till impact and appropriate active safety measures can be applied.


3. Improving Detection Reliability by Mixing Modalities

Visible light cameras 110 give good vehicle detection capability in both day and night. This has been shown to provide a good enough quality signal for Adaptive Cruise Control (ACC), for example. However, mistakes do happen. A particular configuration of unrelated features in the scene can appear as a vehicle. The solution is often to track the candidate vehicle over time and verify that the candidate vehicle appearance and motion is consistent with a typical vehicle. Tracking a candidate vehicle over time, delays the response time of the system. The possibility of a mistake, however rare, also reduces the possibility of using a vision based system for safety critical tasks such as collision avoidance by active braking. FIG. 11a is flow diagram which illustrates algorithm steps for vehicle control, and FIG. 11b is flow diagram which illustrates algorithm steps for verifying human body temperature, in accordance with embodiments of the present invention. The brightness in an FIR image corresponds to temperature of the imaged object. Hence, the brightness in an FIR image can be analyzed to see if the temperature of the imaged object is between 30° C. and 45° C.


In a fourth embodiment of the present invention, a vehicle target, whose distance has been roughly determined from VIS camera 110, is matched to a patch in the FIR image. The patch is then aligned. The aligned patch in the FIR image is then searched for the telltale features of a vehicle such as the hot tires and exhaust. If features are found the vehicle target is approved and appropriate action can be performed sooner, in step 262.


In a fifth embodiment of tee invention, a pedestrian target 10 is detected by VIS camera 110 in a well illuminated scene. A range D estimate is obtained using triangulation with the ground plane 20 and other techniques (described above). Using the epipolar geometry (defined above), the target range and angle (i.e. image coordinates) provide a likely target location in the FIR image. Further alignment between the two images is performed if required. The image brightness in the FIR image is then used to verify that the temperature of the target matches that of a human, in step 264. This provides for higher reliability detection of more difficult targets.


In certain lighting conditions VIS camera 110 cannot achieve good contrast in all parts of the image. VIS camera 110 must then make compromises and tries and optimize the gain and exposure for certain regions of the interest. For example, in bright sunlit days, it might be hard to detect pedestrians in the shadow, especially if the pedestrians are dressed in dark clothes.


Hence, in a sixth embodiment of the invention, a pedestrian target 10 candidate is detected in the FIR image. Information about target angle and rough range is transferred to the visible light system. In stop 244 of algorithm 200 shown in FIG. 8, the camera system optimizes the gain and exposure for the particular part of the image corresponding to the FIR target. The improved contrast means the shape of the target can be verified more reliably.


In another embodiment of the present invention, a Near Infra Red (NIR) camera is used instead of a visible light camera. Since the NIR camera is often located inside the cabin, typically near the rearview mirror, the fusion discussed between visible light cameras and FIR also work between NIR and FIR. The range of the fusion region will of course be larger due to extended night time performance of the NIR camera.


Therefore, the foregoing is considered as illustrative only of the principles of the invention. Further, since numerous modifications and changes will readily occur to those skilled in the art, it is not desired to limit the invention to the exact design and operation shown and described, and accordingly, all suitable modifications and equivalents may be resorted to, falling within the scope of the invention.


While the invention has been described with respect to a limited number of embodiments, it will be appreciated that many variations, modifications and other applications of the invention may be made.

Claims
  • 1. In a computerized system mounted on a vehicle including a cabin and an engine, the system including a visible (VIS/NIR) camera sensitive to visible/near infrared VIS/NIR) light, the VIS/NIR camera mounted inside the cabin, wherein the VIS/NIR camera acquires consecutively in real time a plurality of image frames including respectively a plurality of VIS/NIR images of an object within a field of view of the VIS/NIR camera, a method for detecting the object, the method comprising the steps of: (a) providing a FIR camera mounted on the vehicle in front of the engine, wherein said FIR camera acquires consecutively in real time a plurality of FIR image frames including respectively a plurality of FIR images of an object within a field of view of said FIR camera, wherein at least one of said FIR images is of the environment of the vehicle; and(b) simultaneous processing at least one of said FIR images and at least one of the VIS/NIR images, thereby producing a detected object when the object is present in said environment.
  • 2. The method of claim 1, wherein the VIS/NIR camera includes a visible light blocking and near infrared passing filter, and thereby images only NIR spectrum light.
  • 3. The method of claim 1, further comprising the step of: (c) determining range from the vehicle front to said detected object upon detecting the object.
  • 4. The method of claim 3, wherein said determining of the range to said detected object is performed from images of said VIS/NIR camera, having a field of view (FOV) of more than twenty five degrees.
  • 5. The method of claim 4, wherein said detecting of said detected object and said determining of said range to said detected object is performed from images of said VIS/NIR camera, the method further comprising the steps of: (d) projecting the image location of said detected object in said VIS/NIR image onto a location in the at least one FIR image; and(e) identifying said detected object in a corresponding said at least one FIR image.
  • 6. The method of claim 5, wherein said detected object is a vehicle and wherein said identifying of said detected object in said corresponding at least one FIR image includes locating bright spot features selected from the group of features consisting of the target vehicle wheels (being hot) and the exhaust pipe.
  • 7. The method of claim 4, wherein said VIS/NIR camera is operatively connected to a processor, wherein said processor performs collision avoidance by triggering braking.
  • 8. The method of claim 3, wherein said determining of the range to said detected object is performed from images of said FIR camera, having a field of view (FOV) of less than thirty five degrees, said detected object typically situated more than 10 meters from the vehicle in the vehicle environment.
  • 9. The method of claim 8, wherein said detecting of said detected object and said determining of said range are performed in said at least one FIR image, the method further comprising the steps of; (d) upon said detecting said detected object, projecting an image location of said detected object in said FIR image onto a location in the VIS/NIR image; and(e) identifying said detected object in a corresponding said at least one VIS/NIR image,
  • 10. The method of claim 9, wherein upon said detecting, identifying said detected object in said corresponding said at least one VIS/NIR image, includes alignment of patches of said at least one FIR image with patches of said at least one VIS/NIR image, thereby producing al aligned object.
  • 11. The method of claim 10, wherein said alignment of at least one patch of said at least one FIR image and said at least one VIS/NIR image, said alignment using at least one mechanism selected from the group consisting of histograms of gray scale values, sub-patch correlation and Hausdorf distance computational techniques.
  • 12. The method of claim 10, wherein said alignment of said at least one patch is bounded by an estimate of said range, said estimate computed from said detected object in said at least two FIR image.
  • 13. The method of claim 10, further comprising the steps of: (f) tracking said aligned object by both said VIS/NIR camera and said FIR camera; and(g) if said tracking of said aligned object by either of said VIS/NIR camera and said FIR camera is terminated, the camera that has not terminated said tracking proceeds with said tracking.
  • 14. The method of claim 10, wherein said detected object in said at least one FIR image is a pedestrian and wherein said image brightness in said FIR image is used to verify that said pedestrian temperature matches that of a human.
  • 15. The method of claim 8, further comprising the steps of: (d) said processing optimizes said VIS/NIR camera gain and exposure for said projected image locations of at least one patch of said at least one FIR image location, wherein said VIS image includes at least a portion of said detected object;(e) Identifying said detected object in a corresponding said at least one VIS/NIR image.
  • 16. The method of claim 15, wherein said identifying said detected object in a corresponding said at least one FIR image, includes aligning patches of said at least one FIR image with patches of said at least one VIS/NIR image, thereby producing an aligned object.
  • 17. The method of claim 16, wherein said alignment of at least a patch of said at least one FIR image and said at least one VIS/NIR image, said alignment using at least one algorithm/mechanism/calculation selected from the group consisting of histograms of the gray level values, sub-patch correlation and/or Hausdorf distance computational techniques.
  • 18. The method of claim 16, wherein said alignment of at least one patch is bounded by said estimate of said range, said estimate computed from said detected object in said at least one FIR image.
  • 19. The method of claim 8, wherein said determining of said range is performed in at least two FIR images, the method further comprising the step of: (d) determining the scale change ratio between dimensions of said detected object in the images and using said scale change ratio and the vehicle speed to refine said range.
  • 20. The method of claim 3, wherein said determining of the range to an object is performed from said simultaneous processing of respective images of said VIS/NIR camera and said FIR camera.
  • 21. The method of claim 20, wherein said simultaneous processing includes a stereo analysis of at least a pair of corresponding images acquired by said VIS/NIR camera and said FIR camera, providing refined range estimation.
  • 22. The method of claim 21, wherein said VIS/NIR camera is operatively connected to a processor, wherein said processor said simultaneously performs collision avoidance by triggering braking.
  • 23. The method of claim 3, wherein said determining of said range to said detected object further comprises the step of: (d) determining an accurate lateral distance to said detected object and determining if the object is in the vehicle path and in danger of collision with the moving vehicle.
  • 24. A system mounted on a vehicle including a cabin and an engine, the system including a visible (VIS/NIR) camera sensitive to visible/near infrared (VIS/NIR) light, the VIS/NIR camera mounted inside the cabin, wherein the VIS/NIR camera acquires consecutively in real time a plurality of image frames including respectively a plurality of VIS/NIR images of an object within a field of view of the VIS/NIR camera, the system comprising: (a) a FIR camera mounted on the vehicle in front of the engine, wherein said FIR camera acquires consecutively in real time a plurality of FIR image frames including respectively a plurality of FIR images of an object within a field of view of said FIR camera, wherein at least one of said FIR images is of the environment of the vehicle; and(b) A processor simultaneous processing at least one of said FIR images and at least one of the VIS/NIR images, thereby producing a detected object when the object is present in said environment.
  • 25. The system of claim 24, wherein said processor performs the method of claim 1.
CROSS REFERENCE TO RELATED APPLICATIONS

The present application claims the benefit under 35 USC 119(e) of U.S. provisional application 60/80,356 filed on May 31, 2006 by the present inventors, the disclosure of which is incorporated herein by reference.

Provisional Applications (1)
Number Date Country
60809356 May 2006 US