Lane marker projection method for a motor vehicle vision system

Information

  • Patent Grant
  • 7050908
  • Patent Number
    7,050,908
  • Date Filed
    Tuesday, March 22, 2005
    19 years ago
  • Date Issued
    Tuesday, May 23, 2006
    18 years ago
Abstract
Lane marker coordinate data obtained from a previous frame of a vehicle vision system is projected into the current video frame based on measured translation and rotation of the host vehicle. The projected lane marker coordinate data is consolidated with lane marker coordinate data from the current video frame for lane detection or display applications.
Description
TECHNICAL FIELD

The present invention relates to motor vehicle vision system applications that track or display the location of the vehicle relative to roadway lane markers, and more particularly to a method of consolidating lane marker position information from successively generated video images.


BACKGROUND OF THE INVENTION

Motor vehicle forward vision data generated by a video camera mounted at or near the driver's eye level can be processed to identify various items of interest such as roadway lane markers. The vision system can then determine the location of the vehicle relative to the lane markers, for displaying video information to the driver or for detecting lane changing and/or driving patterns indicative of a drowsy driver. Most of these applications require lane marker detection in a region of about 5–30 meters forward of the vehicle, where the lane markers can be reliably approximated as straight lines. However, dashed or periodic lane markers can have relatively large gaps, and frequently only a fraction of a lane marker is visible to the camera in any given video frame, particularly in mechanizations where a portion of the roadway within the video frame is obscured by the hood or fenders of the vehicle. Since this can degrade the ability of the lane tracking system to perform the intended functions, it would be beneficial if the information obtained from successively generated video images could be consolidated to provide more complete lane marker data, either for display or lane detection purposes.


SUMMARY OF THE INVENTION

The present invention is directed to a method of consolidating lane marker position information by projecting lane marker information from a previously generated video frame into a current video frame. Projecting the lane marker information involves transforming the detected markers from the previous frame to world coordinates, and predicting their position in the current video frame based on measured vehicle rotation and translation parameters. The projected marker coordinates can be combined with world coordinates of lane markers from the current video frame for lane detection applications, or converted to image plane coordinates and combined with lane marker image plane coordinates of the current video frame for driver display purposes.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A and FIG. 1B respectively depict world coordinate and image plane coordinate representations of first and second video frames generated by a vehicle-mounted video camera;



FIG. 2 is a block diagram of a system for carrying out the method of the present invention;



FIG. 3 is a flow diagram depicting the method of this invention; and



FIGS. 4A and 4B respectively depict image plane coordinate and world coordinate representations of lane marker information consolidated from the first and second video frames of FIGS. 1A and 1B according to the method of this invention;





DESCRIPTION OF THE PREFERRED EMBODIMENT

The method of the present invention is carried out in a vehicle-mounted vision system designed, among other things, to capture video images of a scene in the forward path of the vehicle for analysis and/or display to the driver. One of the principle objectives of the video image analysis is to identify lane markers painted on the roadway and the location of the host vehicle relative to the markers. FIG. 1A is a real world or overhead view of a host vehicle 10 equipped with a video camera 12 mounted in the passenger compartment, in the vicinity of an interior rearview mirror, for example. The vehicle 10 is traveling in the direction of arrow 14 down a two-lane roadway 16. The reference numeral 18 designates a continuous lane marker for the edge of roadway 16, and the reference numerals 20a, 20b, 20c designate center lane marker segments dividing the two lanes of roadway 16. FIG. 1B presents an image plane view forward of the vehicle 10 as seen by the camera 12. As illustrated in the figure, the image plane view does not include areas alongside or just forward of the vehicle 10 due to the placement of the camera 12 and obstruction of the vehicle's hood and fenders, designated in FIG. 1B by the reference numeral 10′.


The principle region of interest for purposes of lane marker identification and tracking comprises the portions of the roadway approximately 5–30 meters forward of the vehicle 10. The outside boundary of this region within the real world and image plane views of FIGS. 1A and 1B is depicted by the line 22. Due to the limited depth of the region of interest, the vision system can utilize straight line approximation for lane marker tracking, even on a curved roadway such as depicted in FIG. 1B. However, the obstructed portion of this region (i.e., the portion just forward of the vehicle 10) can often include useful lane marker information, particularly in cases where the center lane marker is dashed.


The present invention enhances the available information by projecting lane marker coordinate data from a previous video frame into the current video frame and consolidating the projected and current lane marker coordinate data to provide a more complete representation of the lane markers and their position with respect to the host vehicle, whether the lane marker data is used for display purposes or lane tracking algorithms. Projection of the previously captured lane marker coordinate data involves characterizing the vehicle's movement in terms of its speed (translation) and yaw rate (rotation) or similar parameters. Thus, a system for carrying out the method of the present invention is represented by the block diagram of FIG. 2, where video data from the camera 12 is supplied along with vehicle speed (VS) and yaw rate (YR) data to a video signal processor (VSP) 30. The VSP 30 analyzes the region of interest in each successive video frame produced by camera 12, identifies lane marker coordinates in the region, and forms a straight-line approximation of the identified markers. According to the method of this invention, the VSP 30 additionally utilizes the measured vehicle movement parameters (VS and YR) to generate the projected lane marker coordinate data and to consolidate it with the lane marker coordinate data from the current video frame. The consolidated lane marker coordinate data can be supplied in image frame format to the video display (VD) 32, and in world format to the block 34, which represents a drowsy driver (DD) and/or collision avoidance (CA) algorithm executed either by VSP 30 or another processor of the system.


Referring to the flow diagram of FIG. 3, the method of the present invention involves, for each successive video frame acquired from camera 12, fetching lane marker coordinates identified in the previous fame of video data (block 50), transforming the prior lane marker coordinates to world coordinates (block 52), obtaining the vehicle movement parameters (block 54), projecting the transformed coordinates based on the obtained vehicle movement parameters (block 56), and consolidating the projected and current lane marker coordinates (block 58). The consolidated coordinates may then be transformed back to the image plane for driver display and/or supplied directly to a lane tracking/fitting algorithm (block 60). While the preferred approach is to consolidate the projected and current lane marker data using world coordinates, the data may alternatively be consolidated using image plane coordinates. At any rate, the video processing techniques involved in identifying lane marker coordinates in a given video frame and transforming image plane coordinates to corresponding world coordinates, and vice-versa, are well understood to those skilled in the video processing art, and therefore are not specifically described herein.


As mentioned above, the projection of lane marker coordinates from a given video frame to the next successive video frame according to this invention involves translating the coordinates based on vehicle speed VS and rotating the coordinates based on yaw rate YR. The starting point is the world coordinate pair (x, y) of a previously identified lane marker, where the x-coordinate represents down-range distance from the vehicle 10 and the y-coordinate represents cross-range distance from the vehicle's central longitudinal axis (as may be represented by the arrow 14 in FIG. 1A). The measured vehicle movement parameters ultimately define the distance traveled d and change in heading φ between successive video frames. In the following discussion the frame notations (“n” for a previously generated image frame and “n+1” for a subsequently generated image frame) are shown as superscripts instead of subscripts to avoid conflict with the subscripted designations applied to some of the parameters. Given the vehicle speed S(n) in m/sec and yaw rate χ(n) in radian/sec at time n, the distance d(n) is given by the product d(n)=T*s(n) where T is the inverse image frame rate of the camera 12 in seconds, and the change in heading φ(n) is given by the product φ(n)=−T*χ(n). Moving at speed s, and yaw rate χ between successive video frames, the host vehicle 10 will translate from coordinate (0, 0) to coordinate (Δx(n), Δy(n), and change its relative heading by φ(n), where the translation coordinates Δx(n) and Δy(n) are given by:

Δx(n)=d(n)cosφ(n), and
Δy(n)=d(n)sinφ(n)

This frame-to-frame origin shift of the vehicle is applied to the lane marker coordinates of the prior video frame. For any such coordinate pair (xi(n), yi(n)), the projected coordinate pair in the next video frame, ({tilde over (x)}i(n+1), {tilde over (y)}i(n+1)), is given by:

{tilde over (x)}i(n+1)=(xi(n)−Δx(n))cosφ(n)+(yi(n)−Δy(n))sinφ(n), and
{tilde over (y)}i(n+1)=−(xi(n)−Δx(n))sinφ(n)+(yi(n)−Δy(n))cosφ(n)

The terms (xi(n)−Δx(n)) and (yi(n)−Δy(n)) account for the vehicle translation, while the functions cosφ(n) and sinφ(n) account for the vehicle rotation.



FIGS. 4A and 4B respectively depict image plane and world representations of the consolidated lane marker coordinates, where current coordinates are designated by the plus signs, projected coordinates are designated by the small triangles, and regions of coordinate overlap are designated by the asterisks. The format of FIG. 4A is most useful for display applications, while the format of FIG. 4B is most useful for video processing applications such as lane tracking.


In summary, the present invention provides simple and cost-effective method of consolidating identified coordinates of images successively generated by a vehicle vision system. The consolidated coordinates provide enhanced display and improved lane marker tracking. While the method of the present invention has been described with respect to the illustrated embodiment, it is recognized that numerous modifications and variations in addition to those mentioned herein will occur to those skilled in the art. Accordingly, it is intended that the invention not be limited to the disclosed embodiment, but that it have the full scope permitted by the language of the following claims.

Claims
  • 1. A method of operation for a vision system that identifies lane marker coordinates in successively generated video images of a roadway in a travel path of a vehicle, comprising the steps of: obtaining world lane marker coordinate data from a previously generated video image;determining a speed and a yaw rate of the vehicle;projecting the obtained world lane marker coordinate data into a subsequently generated video image based on the determined speed and yaw rate and a time interval between the previously and subsequently generated video images;consolidating the projected lane marker coordinate data with lane marker coordinate data from the subsequently generated video image; andprocessing the consolidated lane marker coordinate data.
  • 2. The method of claim 1, wherein the vision system identifies image frame lane marker coordinates, and the step of obtaining world lane marker coordinate data includes the step of: transforming the identified image frame lane marker coordinates from the previously generated video image to corresponding world coordinates.
  • 3. The method of claim 1, including the step of: consolidating the projected world lane marker coordinate data with world lane marker coordinate data from the subsequently generated video image.
  • 4. The method of claim 1, including the steps of: transforming the projected world lane marker coordinate data to corresponding image frame coordinate data for consolidation with image frame coordinate data from the subsequently generated video image.
  • 5. The method of claim 1, where the step of projecting the obtained world lane marker coordinate data into a subsequently generated video image includes the steps of: determining a translation dn of the vehicle between the previously and subsequently generated video images based on a product of the determined speed and said time interval; anddetermining a change in relative heading φn of the vehicle between the previously and subsequently generated video images based on a product of the determined yaw rate and said time interval.
  • 6. The method of claim 5, wherein the projected world lane marker coordinate data (xn+1, yn+1) is given by xn+1=(xn−Δxn)cosφn+(yn−Δyn)sinφn andyn+1=≠(xn−Δxn)sinφn+(yn−Δyn)cosφn
US Referenced Citations (26)
Number Name Date Kind
4970653 Kenue Nov 1990 A
5467634 Brady et al. Nov 1995 A
5473931 Brady et al. Dec 1995 A
5809161 Auty et al. Sep 1998 A
6035253 Hayashi et al. Mar 2000 A
6134509 Furusho et al. Oct 2000 A
6292752 Franke et al. Sep 2001 B1
6647328 Walker Nov 2003 B1
6708087 Matsumoto Mar 2004 B1
6765480 Tseng Jul 2004 B1
6807287 Hermans Oct 2004 B1
6819779 Nichani Nov 2004 B1
6868168 Tsuji Mar 2005 B1
6888447 Hori et al. May 2005 B1
6977630 Donath et al. Dec 2005 B1
20010035880 Musatov et al. Nov 2001 A1
20010056544 Walker Dec 2001 A1
20020198632 Breed et al. Dec 2002 A1
20030023614 Newstrom et al. Jan 2003 A1
20030123705 Stam et al. Jul 2003 A1
20030128182 Donath et al. Jul 2003 A1
20040042638 Iwano Mar 2004 A1
20040049324 Walker Mar 2004 A1
20040066376 Donath et al. Apr 2004 A1
20040143381 Regensburger et al. Jul 2004 A1
20040183905 Comaniciu et al. Sep 2004 A1
Foreign Referenced Citations (2)
Number Date Country
197 49 086 Nov 1997 DE
2003322522 Nov 2003 JP