Imaging system for vehicle

Abstract
An imaging system for a vehicle includes a rear backup camera and a video display screen for displaying video images derived from image data captured by the rear backup camera. The imaging system generates a backup overlay and an alignment overlay that are electronically superimposed on the displayed video images to assist a driver of the vehicle when executing a backup maneuver. The backup overlay includes a pair of side overlays superimposed on the displayed video images and the alignment overlay is superimposed on the displayed video images between the side overlays. The video display screen displays the video images and the backup overlay is generated responsive to the vehicle being shifted into reverse gear. The alignment overlay is generated responsive to a user input.
Description
FIELD OF THE INVENTION

The present invention relates generally to imaging or vision systems, particularly for vehicles which provide a vehicle operator with wide angle scenic information of an area immediately around the vehicle, such as rearward of the vehicle.


BACKGROUND OF THE INVENTION

A long felt need in the art of vehicle vision systems, such as rear-view systems, has been to provide an operator of a vehicle with wide angle scenic information of an area around the vehicle, such as directly rearward of the vehicle, when the vehicle is traveling in reverse. Neither interior rear-view mirrors nor side exterior mirrors allow for visibility of the area immediately rearward of the vehicle's bumper, which is the information most critical to the vehicle operator when backing up. Therefore, various camera-based rear view vision systems have been proposed to provide visibility of this blind spot.


It is also known to provide a graphic overlay on the displayed image of the rearward scene to enhance the driver's perception of the rearward field of view. Examples of such graphic overlays are described in U.S. Pat. Nos. 5,670,935; 5,949,331; 6,222,447 and 6,611,202, which are hereby incorporated herein by reference in their entireties.


Various camera-based rear vision systems for vehicles backing up have been proposed. In one form of these systems, a camera with a conventional, standard lens is located on the rearward portion of the vehicle to provide a view of the area behind the vehicle. However, standard lenses fail to capture a wide angle view of the area, thus failing to provide the vehicle operator with an image of the entire critical area directly rearward of the vehicle. In order to provide a wider angle view of the rearward area, a wide angle lens system may be used with the camera to capture the critical area. Examples of such systems are described in U.S. Pat. Nos. 6,922,292; 6,757,109; 6,717,610; 6,590,719 and 6,201,642, which are hereby incorporated herein by reference in their entireties.


SUMMARY OF THE INVENTION

The present invention is directed toward enhancing exterior visibility to the driver of a vehicle by providing an imaging system that provides a wide angle field of view of the area directly exteriorly of a vehicle, and that provides a graphic overlay superimposed on a displayed image to enhance the driver's viewing and understanding or cognitive awareness of the displayed image of the scene rearward of the vehicle and any object or objects detected therein. The present invention is directed towards enhancing the interpretation of visual information in a rearview vision system by presenting information in a manner which does not require significant concentration of the driver or present distractions to the driver.


The rearview vision system of the present invention has an image capture device or image sensor or camera positioned on the vehicle and directed rearwardly with respect to the direction of travel of the vehicle, and a display for displaying images captured by the image capture device. Image enhancement means may be provided for enhancing the displayed image, such as in the form of graphic overlays superimposed on the displayed image. Such graphic overlays may include indicia of the anticipated path of travel of the vehicle which is useful in assisting the driver in guiding the vehicle in reverse directions. The graphic overlays may be flashed or intermittently activated and/or displayed in different colors depending on a distance to a detected object rearward of the vehicle so as to provide a spectral cue and a spatial cue to the driver as the driver maneuvers the vehicle in a rearward direction. The graphic overlays may comprise graphic overlay segments extending upward and inward along the sides of the image so as to be indicative of segments extending rearward from the sides of the vehicle. The graphic overlay segments may be viewed by the driver as having a virtual height dimension to enhance the driver's cognitive awareness of the graphic overlays. The graphic overlay segments may be displayed as three dimensional graphic overlay segments having a virtual height and a virtual thickness.


The camera or image sensor of the imaging system may include or utilize a wide angle lens that is compact, durable and inexpensive to manufacture. The image sensor and lens assembly of the present invention thus may comprise a wide angle lens or lens assembly, such as a multi-element lens assembly having multiple optics to provide a wide angle field of view with reduced distortion. Optionally, the lens may include multiple optics, such as four to seven optic elements, with an outer element (the element at the rearwardmost end of the lens and exposed at the rear exterior of the vehicle) comprising a glass element and with at least one other element comprising a plastic aspheric element.


These and other objects, advantages, purposes, and features of this invention will become apparent upon review of the following specification in conjunction with drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a perspective view of an imaging device comprising an image sensor and lens in accordance with the present invention;



FIG. 2 is an image as displayed to a driver of the vehicle in accordance with the present invention;



FIG. 3 is another image as displayed to a driver of the vehicle and having a graphic overlay superimposed thereon in accordance with the present invention;



FIG. 4 is another image as displayed to a driver of the vehicle and having a graphic overlay superimposed thereon in accordance with the present invention;



FIG. 4A is a perspective side view of the vehicle and overlay of FIG. 4;



FIG. 5 is another image as displayed to a driver of a vehicle and having a trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 5A is a perspective side view of the vehicle and overlay of FIG. 5;



FIG. 6 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 6A is a perspective side view of the vehicle and overlay of FIG. 6;



FIG. 7 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 8 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 9 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 10 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 10A is a perspective side view of the vehicle and overlay of FIG. 10;



FIG. 11 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 11A is a perspective side view of the vehicle and overlay of FIG. 11;



FIG. 12 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention;



FIG. 13 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention; and



FIG. 14 is an image as displayed to a driver of a vehicle and having another trailer hitch graphic overlay superimposed thereon in accordance with the present invention.





DESCRIPTION OF THE PREFERRED EMBODIMENTS

Referring now to the drawings and the illustrative embodiments depicted therein, a wide angle imaging system provides an operator of a vehicle with scenic information of an area immediately exteriorly, such as rearwardly, of the vehicle, which may be an automobile, a light truck, a van, a large truck, a sport utility vehicle or the like. The imaging system includes an image capture device 12 (FIG. 1) having an imaging sensor and a lens or lens assembly 14 that functions to focus a rearward field of view at an imaging plane of the imaging sensor or camera. Images 16 (FIG. 2) are displayed on a display device or screen with reduced distortion to enhance the driver's viewing and understanding of the displayed images. The rearview imaging system includes an image processor for receiving data signals from the image capture device and superimposing the graphic overlay on the displayed image, as discussed below.


The imaging sensor for the vehicle vision system of the present invention may comprise any suitable sensor, and may utilize various imaging sensors or imaging array sensors or cameras or the like, such as a CMOS imaging array sensor, a CCD sensor or other sensors or the like, such as the types described in U.S. Pat. Nos. 5,550,677; 5,670,935; 5,760,962; 5,715,093; 5,877,897; 6,922,292; 6,757,109; 6,717,610; 6,590,719; 6,201,642; 6,498,620; 5,796,094; 6,097,023; 6,320,176; 6,559,435; 6,831,261; 6,806,452; 6,396,397; 6,822,563; 6,946,978; 7,038,577 and 7,004,606; and/or U.S. patent application Ser. No. 11/315,675, filed Dec. 22, 2005 and published Aug. 17, 2006 as U.S. Patent Publication No. US-2006-0184297; and/or Ser. No. 10/534,632, filed May 11, 2005 and published Aug. 3, 2006 as U.S. Patent Publication No. US-2006-0171704; and/or U.S. provisional applications, Ser. No. 60/845,381, filed Sep. 18, 2007; and Ser. No. 60/837,408, filed Aug. 11, 2006, and/or PCT Application No. PCT/US2003/036177 filed Nov. 14, 2003 and published Jun. 3, 2004 as PCT Publication No. WO 2004/047421, which are all hereby incorporated herein by reference in their entireties. The control may include a lens element or optic between the image sensor and the forward scene to substantially focus the scene at an image plane of the image sensor. Optionally, the optic may comprise a wide angle lens that provides a proportionally distributed central portion of the imaged scene while providing classical optical distortion on the periphery of the imaged scene. Optionally, the optic may comprise a non-flat field curvature, which may focus a generally central portion of the imaged scene onto the image sensor, while providing reduced but acceptable focus on the periphery of the imaged scene. The imaging device may comprise an image sensing module or the like, and may utilize aspects described in U.S. patent application Ser. No. 10/534,632, filed May 11, 2005 and published Aug. 3, 2006 as U.S. Patent Publication No. US-2006-0171704; and/or PCT Application No. PCT/US2006/041709, filed Oct. 27, 2006 and published May 10, 2007 as International Publication No. WO 07/053404; and/or PCT Application No. PCT/US2003/036177 filed Nov. 14, 2003 and published Jun. 3, 2004 as PCT Publication No. WO 2004/047421, which are hereby incorporated herein by reference in their entireties.


Such imaging sensors or cameras are pixelated imaging array sensors having a photosensing array of a plurality of photon accumulating or photosensing light sensors or pixels, which are arranged in a two-dimensional array of rows and columns on a semiconductor substrate. The camera established on the substrate or circuit board includes circuitry that is operable to individually access each photosensor pixel or element of the array of photosensor pixels and to provide an output or image data set associated with the individual signals to the control, such as via an analog to digital converter. As the camera receives light from objects and/or light sources in the target scene, the control may then be operable to process the signal from at least some of the pixels to analyze the image data of the captured image, as discussed below.


Optionally, the imaging sensor may be suitable for use in connection with other vehicle imaging systems, such as, for example, an object detection system or blind spot detection system, where a blind spot indicator may be operable to provide an indication to the driver of the host vehicle that an object or other vehicle has been detected in the lane or area adjacent to the side of the host vehicle. In such a blind spot detector/indicator system, the blind spot detection system may include an imaging sensor or sensors, or ultrasonic sensor or sensors, or sonar sensor, or radar, or LIDAR or sensors or the like. For example, the detection system may utilize aspects of the detection and/or imaging systems described in U.S. Pat. Nos. 7,038,577; 6,882,287; 6,198,409; 5,929,786 and/or 5,786,772, and/or U.S. patent application Ser. No. 11/315,675, filed Dec. 22, 2005 and published Aug. 17, 2006 as U.S. Publication No. US-2006-0184297; and/or Ser. No. 11/239,980, filed Sep. 30, 2005, now U.S. Pat. No. 7,881,496, and/or U.S. provisional applications, Ser. No. 60/696,953, filed Jul. 6, 2006; Ser. No. 60/628,709, filed Nov. 17, 2004; Ser. No. 60/614,644, filed Sep. 30, 2004; and/or Ser. No. 60/618,686, filed Oct. 14, 2004, and/or of reverse or backup aid systems, such as the rearwardly directed vehicle vision systems described in U.S. Pat. Nos. 5,550,677; 5,760,962; 5,670,935; 6,922,292; 6,590,719; 6,201,642; 6,396,397; 6,498,620; 6,717,610; 6,757,109 and/or 7,005,974, and/or of the rain sensors described in U.S. Pat. Nos. 6,250,148 and 6,341,523, and/or of other imaging systems, such as the types described in U.S. Pat. Nos. 7,123,168; 6,353,392 and/or 6,313,454, with all of the above referenced U.S. patents, patent applications and provisional applications and PCT applications being commonly assigned and being hereby incorporated herein by reference in their entireties.


In order to enhance the driver's understanding of what is occurring in the area surrounding the vehicle, the rearview vision system includes a display device or display element having image enhancements (FIG. 3). The display device may comprise any suitable display device or element, such as a video display screen, such as a video display screen utilizing aspects of the display devices described in U.S. Pat. Nos. 5,530,240; 6,329,925; 7,370,983; 7,274,501; 7,255,451; 7,195,381; 7,184,190; 6,690,268; 5,668,663 and/or 5,724,187, and/or in U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US 2006-0050018; Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US 2006-0061008; Ser. No. 10/993,302, filed Nov. 19, 2004, now U.S. Pat. No. 7,338,177; and/or Ser. No. 11/520,193, filed Sep. 13, 2006, now U.S. Pat. No. 7,581,859; and/or PCT Application No. PCT/US03/29776, filed Sep. 9, 2003 and published Apr. 1, 2004 as International Publication No. WO 2004/026633; and/or PCT Application No. PCT/US03/35381, filed Nov. 5, 2003 and published May 21, 2004 as International Publication No. WO 2004/042457; and/or PCT Application No. PCT/US2006/018567, filed May 15, 2006 and published Nov. 23, 2006 as International Publication No. WO 2006/124682; and/or PCT Application No. PCT/US2006/042718, filed Oct. 31, 2005 and published May 10, 2007 as International PCT Publication No. WO 2007/053710; and/or PCT Application No. PCT/US03/40611, filed Dec. 19, 2003 and published Jul. 15, 2004 as International Publication No. WO 2004/058540, which are all hereby incorporated herein by reference in their entireties.


In the illustrated embodiment, the image enhancements of the imaging system and/or display device include graphic overlays 20a, 20b which are lines or segments intended to illustrate to the driver the anticipated path of rearward movement of the vehicle based on the present trajectory of the vehicle. In the illustrated embodiment of FIG. 3, the graphic overlays 20a, 20b indicate the anticipated vehicle motion as a function of the vehicle direction of travel as well as the rate of turn of the vehicle. Optionally, the displayed image may have a textual message 20c alerting the driver of a detected object or otherwise instructing the driver during the rearward maneuver.


The forward or rearward direction of vehicle travel is determined in response to the operator placing the gear selection device (not shown) in the reverse gear position. The degree of turn of the vehicle may be determined by monitoring the movement of the vehicle steering system, monitoring the output of an electronic compass, or monitoring the vehicle differential drive system or the like. Optionally, and desirably, if the vehicle is not in reverse gear position, the graphic overlays are not presented. The imaging system and graphic overlays may utilize aspects of the systems described in U.S. Pat. Nos. 5,670,935; 5,949,331; 6,222,447 and 6,611,202, which are hereby incorporated herein by reference in their entireties. Optionally, the graphic overlays may be static overlays (in other words, graphic overlays that extend generally rearwardly toward a focal point of expansion rearward of the vehicle and that are not dependent on a steering wheel angle or the like), while remaining within the spirit and scope of the present invention.


Optionally, and as shown in FIG. 4, the graphic overlays 30a, 30b may function to enhance the driver's awareness of an object detected rearward of the vehicle when the vehicle is shifted into a reverse gear. Optionally, the graphic overlays may include a centrally located overlay 30c, such as a dashed line or the like, that extends upward toward the focal point of expansion of the displayed image. In the illustrated embodiment, each graphic overlay 30a, 30b may be provided in three different colors, such as a green segment 32, a yellow segment 34 and a red segment 36, with the upper or further out segment 32 being green (or other suitable color or pattern), the middle segment 34 being yellow (or other suitable color or pattern), and the closest or lowest segment 36 being red (or other suitable color or pattern). The individual segments may be activated or overlaid in response to a detection of an object rearward of the vehicle and a detected or determined distance between the rear of the vehicle and the detected object.


For example, the rearward imaging system of the present invention may include an imaging sensor for capturing images in a rearward field of view and a display for displaying the images. The system may include or may operate in conjunction with an object detection system for detecting objects rearward of the vehicle and in the path of the vehicle when the vehicle is traveling in reverse (such as, for example, an ultrasonic sensing system, a lidar sensing system or a radar sensing system or via image processing of the captured images or the like). When the vehicle is reversing and no object is detected by the object detection system (or if an object is detected that is further away from the vehicle than a predetermined threshold distance), the graphic overlay or overlays may not be displayed or may be displayed as non-colored or dark or neutral lines or continuous segment or the like (such as similar to the graphic overlays 20a, 20b shown in FIG. 3). As the vehicle approaches the object so that the object is detected by the object detection system (or when a detected object is within a first threshold distance from the rear of the vehicle but outside of a second threshold distance from the rear of the vehicle), the outer graphic overlay segment 32 may flash or may be displayed as green to alert the driver of the presence of the object, while also notifying the driver that the object is still a safe distance rearward of the vehicle. As the vehicle further approaches the object (for example, when the object is within the second threshold distance from the rear of the vehicle but outside of a third threshold distance from the rear of the vehicle), the middle graphic overlay segment 34 may flash or may be displayed as yellow to alert the driver that the vehicle is approaching the object. As the vehicle further approaches the object (for example, when the object is within the third threshold distance from the rear of the vehicle), the inner graphic overlay segment 36 may flash or may be displayed as red to alert the driver that the vehicle is very near to the object.


Optionally, the system may also provide a tone or audible alert when the vehicle and object are within the third threshold separation distance or when the vehicle and object are less than a fourth threshold distance apart to further alert the driver that the object is very near to the vehicle and that the driver should not back up any further. Optionally, the graphic overlays may also or otherwise vary in shape or pattern or thickness or color to enhance the driver's cognitive awareness of the location of and distance to a detected object rearward of the vehicle. Optionally, as one segment is activated or colored (as the object is detected within the respective range from the rear of the vehicle), the other segment or segments may be deactivated or de-colored, so that the driver can readily discern that the object is detected within the particular range encompassed by the currently activated or colored graphic overlay segments.


Optionally, the graphic overlay may be designed to assist colorblind drivers to distinguish the color segments of the overlay. A majority of the colorblind population are “red-green colorblind” and have problems in either in the red or green opsin gene. For example, people with a mutant red opsin gene are called “protanopes”, while people with a mutant green opsin gene are called “deuteranopes”. It is estimated that for males, about 8 percent of Caucasians, about 5 percent of Asians and about 4 percent of Africans are “red-green” colorblind. Color blindness is typically sex-linked, so a reduced number of females are colorblind. Red-green color blind people have difficulty distinguishing colors between red and green with similar intensity or brightness. To address this problem, for example, the graphic overlays of the present invention may replace the red colored segment or segments in the overlay graphics with a reddish purple or vermilion colored segment or segments, and may replace the green colored segment or segments with a blue or bluish green segment or segments. Such colors may be recognized and discerned by a typical red-green colorblind person (the yellow colored segments may remain yellow since typical red-green colorblind people would be able to recognize and discern the yellow colored segments). Optionally, different textural features or patterns (such as different stippling or cross-hatching or different degrees of transparency or translucency of the segments or the like) can be used to assist colorblind drivers to distinguish different segments of the graphic overlay.


Optionally, the different colors or features or patterns of the segments may be selectively provided in response to the user or driver selecting the desired or appropriate graphic overlay display scheme. For example, a software or hardware toggle switch or other suitable user actuatable input (such as a button or a voice recognition system or the like) may be implemented to allow the user or driver of the vehicle to select a suitable overlay color scheme, such as a display displaying normal colored graphic overlays for non-colorblind drivers or a display displaying modified graphic overlays (such as a different color scheme or a different texture or pattern scheme or the like) for colorblind drivers. The switch may offer different choices for the user, such as a color scheme for a red-green colorblind person or a different color scheme for other types of colorblindness (whereby the system adjusts the color scheme of the graphic overlays accordingly).


Optionally, and with reference to FIGS. 5-14, the present invention provides a graphic overlay to assist the driver of the vehicle to back up to a trailer hitch or the like. Optionally, and desirably, the camera or image sensor at the rear of the subject vehicle may be offset from the centerline of the vehicle and angled so as to have its field of view be generally rearward and toward the centerline of the vehicle. Such a camera orientation allows for a three-dimensional or perspective image of a trailer hitch alignment graphic overlay that may be imaged so as to appear to extend rearward from the rear of the vehicle and generally along the centerline of the vehicle.


For example, and as shown in FIG. 5, a graphic overlay includes a pair of sideward overlays 130a, 130b and a generally centrally located trailer hitch alignment overlay 130c. The side overlays 130a, 130b may be substantially similar to the side overlays 30a, 30b, discussed above, and may have colored segments and the like, such that a detailed discussion of the side overlays need not be included herein. The hitch alignment overlay 130c is generally at or near the center of the displayed image and extends upward toward the focal point of expansion of the displayed image. The side overlays 130a, 130b may function as described above, and may be activated or generated or displayed when the vehicle is shifted into reverse, while the hitch alignment overlay 130c may be similarly activated or generated or displayed, or may be activated or generated or displayed in response to a user input, such as a control input or button or the like that is activated by the driver of the vehicle when the driver is about to back up the vehicle to a trailer (or may be automatically activated or generated or displayed in response to the system detecting or identifying a trailer rearward of the vehicle when the vehicle is traveling in reverse), while remaining within the spirit and scope of the present invention.


As can be seen in FIG. 5, the displayed image is captured by an image sensor that is offset from the centerline of the vehicle and is directed rearward and toward the centerline of the vehicle. In the illustrated embodiment, image sensor is disposed about 305 mm offset from the centerline of the vehicle at about 875 mm above the ground and oriented at an angle of about 40 degrees toward the centerline of the vehicle. The hitch alignment overlay 130c is displayed as a three-dimensional shape that extends upward in the image and generally along the centerline of the vehicle and toward the focal point of expansion of the camera. The hitch alignment overlay 130c is generated so as to appear to be a three dimensional object at the rear of the vehicle (such as an object or alignment element having a height of about 440 mm or thereabouts to approximate a height of a hitch of the vehicle). Thus, the alignment overlay provides two planes, one generally at the ground level and one generally at a typical trailer hitch height.


Other mounting locations and angles for the camera and/or other dimensions of the graphic overlays may be implemented without affecting the scope of the present invention. Because the camera is preferably disposed at a height greater than the virtual height of the alignment element/overlay, the displayed image provides a perspective view rearward and downward toward the three dimensional alignment overlay so that the driver can readily discern and understand the alignment overlay to assist the driver in reversing the vehicle toward a trailer hitch of a trailer or the like. The alignment overlay and the offset camera thus provide a scaled and distorted three dimensional image to provide the driver improved depth perception for driving the vehicle toward the trailer or other targeted object.


In the illustrated embodiment of FIG. 5, alignment overlay 130c is a wire frame or line drawing or representation of a perspective view of a three dimensional rectangular form. However, other shapes or forms may be utilized to provide the desired perspective or view to the driver to assist the driver in reversing the vehicle toward a trailer hitch or other object. For example, and as shown in FIG. 6, the side overlays 130a′, 130b′ may be similar to side overlays 130a, 130b, discussed above, while the alignment overlay 130c′ may be a solid or colored three dimensional rectangular form to further enhance the viewability and discernability of the alignment overlay to the driver of the vehicle. Optionally, for example, and as shown in FIG. 7, the side overlays 230a, 230b may be similar to side overlays 130a, 130b, discussed above, while the alignment overlay 230c may be a generally planar, two dimensional rectangular form (such as only the upper portion or surface of the three dimensional rectangular form of alignment overlay 130c, discussed above), which may be a wire form representation (as shown in FIG. 7), or the alignment overlay 230c′ may be a solid or colored representation (as shown in FIG. 8), depending on the particular application and desired appearance of the alignment overlay in the displayed image. Optionally, and as shown in FIG. 9, the side overlays 330a, 330b may be similar to side overlays 130a, 130b, discussed above, while the alignment overlay 330c may be a prism or wedge shaped form having an upper surface (such as a surface or portion similar to the upper region of the alignment overlay 130c, discussed above) and a shape that tapers or narrows downward to generally a pointed or narrowed lowered portion, to further enhance the perspective view of the alignment overlay. The alignment overlay 330c may be colored or shaded or solid (as shown in FIG. 9), or may be a wire form representation or the like, without affecting the scope of the present invention.


Optionally, the side overlays may also be represented or generated or displayed three dimensionally to further assist the driver of the vehicle during the rearward maneuver. For example, and as shown in FIG. 10, the side overlays 430a, 430b may appear as three dimensional colored or shaded or solid “walls” extending rearward from the vehicle to assist the driver in guiding or driving the vehicle toward the trailer or targeted object. The side overlays 430a, 430b may be otherwise similar to side overlays 30a, 30b, and may have colored segments 432, 434, 436 or the like, such as discussed above (optionally, the alignment overlay 430c may be similarly colored or patterned if desired). Optionally, and as shown in FIG. 11, the side overlays 430a′, 430b′ and alignment overlay 430c′ may be displayed as wire frame representations and/or as partially transparent colored representations so as to provide the desired color and enhanced viewability and discernability while limiting or reducing any potential interference with the driver's view of other objects in the displayed image. Optionally, and as shown in FIG. 12, the side overlays 430a″, 430b″ and alignment overlay 430c″ may be displayed as wire frame representations and/or as partially transparent colored representations, while the side overlays 430a″, 430b″ may include a lower portion that is substantially similar to side overlays 30a, 30b to further enhance the viewability and discernability of the side overlays while limiting or reducing any potential interference with the driver's view of other objects in the displayed image. Optionally, and as shown in FIG. 13, the side overlays 430a″′, 430b″′ and alignment overlay 430c″′ may be displayed as wire frame representations and/or as partially transparent colored representations, while the side overlays 430a″′, 430b″′ may include a lower portion that is substantially similar to side overlays 30a, 30b and an upper portion that is also colored or shaded to generally correspond with the lower portion of the side overlays, in order to provide enhanced viewability and discernability of the side overlays while further limiting or reducing any potential interference with the driver's view of other objects in the displayed image. Optionally, and as shown in FIG. 14, the side overlays 530a, 530b may have an upper portion 531a and a lower portion 531b similar to side overlays 430″′, 430″′, but may not have the wire form or shaded region therebetween. The upper and/or lower portions 531a, 531b of side overlays 530a, 530b may have colored segments to assist in indicating a distance to the targeted object. Optionally, and as also shown in FIG. 14, the hitch alignment overlay 530c may comprise a two dimensional rectangular form (either shaded or colored or wire form), such as described above. Although several optional configurations and combinations of side overlays and center hitch alignment overlays are shown in FIGS. 4-14, clearly other combinations and configurations may be implemented while remaining within the spirit and scope of the present invention.


The side overlays and the alignment overlay may be static overlays or may be adjusted or curved or reconfigured (such as in response to a steering wheel angle of the vehicle or the like) to provide a dynamic overlay when the vehicle is being driven in reverse toward the trailer or targeted object. The side overlays and the centrally located alignment overlay may be activated or generated or displayed in response to a user input, such as a button or other control input that is activated by the driver of the vehicle when it is desired to back up the vehicle toward a targeted object or trailer hitch. Optionally, for example, the displayed image may be activated in response to the vehicle being shifted into reverse and may include side overlays 30a, 30b (discussed above), and may provide the three dimensional side overlays and the centrally located hitch alignment overlay in response to the user input. The side overlays and/or the centrally located hitch alignment overlays may be displayed in different colors or different intensities or may be intermittently displayed or the like based on a distance between the vehicle and the trailer hitch or targeted object, such as discussed above with respect to side overlays 30a, 30b, while remaining within the spirit and scope of the present invention.


Thus, the imaging system and graphic overlays of the present invention provide enhanced cognitive awareness by the driver of any objects detected rearward of the vehicle when the vehicle is shifted into a reverse gear and as the driver drives the vehicle in a reverse direction. The change in color or change in intensity or intermittent actuation of the graphic overlays on a segment-by-segment basis provides both a spatial cue to the driver of the presence of and proximity of an object rearward of the vehicle and a spectral cue to the driver of the presence of and proximity of an object rearward of the vehicle, so that the driver is readily aware of an object that is detected rearward of the vehicle and is readily aware of the proximity of the detected object to the rear of the vehicle. The centrally located hitch alignment overlay further enhances the discernability of the location of trailer hitch or targeted object to further assist the driver in maneuvering the vehicle in a rearward direction toward the trailer hitch or targeted object.


Optionally, and preferably, the image sensor includes or utilizes a wide angle lens that focuses a wide angle rearward field of view (such as a wide angle field of view of preferably greater than 120 degrees wide and more preferably greater than 135 degrees wide and more preferably at least about 145 degrees wide) onto an image plane of the image sensor. The wide angle lens preferably comprises a multi-element lens having multiple lens optics or elements, such as seven optics arranged along an optic path to focus the rearward wide angle field of view at the imaging plane while reducing distortion of the focused image. For example, the wide angle lens may include seven optic elements, with some of the elements being glass elements (such as an outer wide angle element being a glass element) and some of the elements being plastic elements. Preferably, at least one of the optic elements comprises a plastic aspheric element, and more preferably, two of the optic elements comprise plastic aspheric elements and five of the optic elements comprise glass spherical elements. Such a lens assembly may utilize aspects of the lenses described in U.S. Pat. Nos. 6,922,292; 6,757,109; 6,590,719 and 6,201,642, which are hereby incorporated herein by reference in their entireties.


The lens of the present invention thus may have at least one plastic element, which may provide a cost savings and a reduction in overall lens length. The refractive or diffractive lens element or elements may be made by injection molding or other low cost means. The plastic lens elements may comprise a polycarbonate material or acrylic material or any other suitable plastic or polymeric material. The lens includes non-spherical, refractive or diffractive optic elements, and may have a stop or aperture within the lens to correct distortion in the focused image.


For example, a lens assembly may be provided that is approximately 25 to 26 mm in overall length (from the image plane to the outer end of the lens assembly. The lens elements may be provided in a housing or barrel (such as an aluminum barrel) that may be threaded onto the circuit board that has the image sensor established thereon or threaded onto a housing or mounting portion of an image sensor module. The lens may have a cap diameter of about 12 mm to about 23 mm, with about a 1.0 to 2.5 mm radius on the outer most edge of the lens cap. The lens may provide a wide angle field of view of about 130 degrees in the horizontal and about 105 degrees vertical when the image sensor and lens are mounted at the vehicle, and may have an active sensor area having about a 3.584 mm horizontal dimension and about a 2.688 mm vertical dimension and about a 4.480 diagonal dimension.


The image sensor may comprise a pixelated image sensor (such as a CMOS sensor or the like) having a 640×480 array (or other size array depending on the particular application) of pixels that are about 5.6 μm×5.6 μm. The lens may provide an “F-stop” of about 2.0 (or lower)+/−5 percent, and may have a relative illumination at greater than 50 percent at full field. Preferably, the lens provides a reduced or lowered optical distortion (geometrical), preferably better than about −45 percent. The lens may have a modulation transfer function (a measurement of lens resolution quality) of greater than about 0.6 at 45 lp/mm on lens axis (zero degrees) and greater than about 0.25 at 45 lp/mm off-axis (between zero degrees and 60 degrees).


The lens preferably includes an integrated infrared (IR) cutoff coating or filter that provides about 85 percent or greater transmission of light therethrough over the visible spectrum (about 400 nm to about 700 nm) of light, and about 50 percent or lower transmission of light therethrough for light having wavelengths greater than about 700 nm and less than about 400 nm. The lens may provide a wide angle field of view and may be able to focus on an object within about 30 cm or thereabouts from the outer end of the lens. The lens preferably has an anti-reflective (AR) coating on at least some of the surfaces of the optic elements and preferably on all of the surfaces of the optic elements. The lens may have a variation of effective focal length of less than +/−5 percent. The lens preferably has an image circle diameter of greater than about 4.8 mm and the angle between the lens' optical axis and a barrel reference diameter axis is preferably less than about 1.5 degrees.


The lens is robust and capable of withstanding the extreme climate and environmental conditions that it may encounter at the exterior portion of a vehicle. For example, the lens may be operational at up to about 85 degrees Celsius, and may be able to withstand exposure up to 125 degrees Celsius. The lens is capable of withstanding high temperature and high humidity endurance tests and can withstand exposure at 95 degree Celsius and 85 percent relative humidity for 1200 hours or more. The lens is also capable of withstanding mechanical shock tests (such as 6 shock pulses with 100 g and 10 ms half-sine pulses, one in each opposite direction of three perpendicular axes) and vibration tests (such as a vibration test of an RMS acceleration value of about 27.8 m/s2 for about 27 hours in the x-axis, about 27 hours in the y-axis and about 81 hours in the z-axis. Because the outer optic element is exposed to harsh conditions at the rear of the vehicle, the lens comprises a sealed lens and is capable of passing the required OEM testing, such as a car wash spray test and a heavy splash shower test and a mist test and a salt mist test and a dust test and the like. The lens can also withstand a dunk or soak test where the lens is soaked in 95 degrees Celsius water for at least 1 hour and then the upper body of the lens is dunked or soaked in icy water at a depth of 15 cm for at least 30 minutes. The lens is also capable of withstanding exposure to various chemicals. For example, the outer glass surface of the lens and the outer exterior body of the lens may be resistant to automatic transmission fluid, hypoid lubricant, hydraulic fluid, power steering fluid, differential lubricant, central hydraulic fluid, engine oil, engine wax protective, engine coolant/Ethylene Glycol, gasoline, diesel fuel, kerosene, bio-diesel/Methanol based fuel, brake fluid, windshield washer fluid, window glass cleaner, car wash cleaner/soap solution, car wax and silicone protectants, leather wax, battery acid—dilute sulfuric acid, and calcium chloride, and/or other chemicals the lens may encounter at the vehicle assembly plant and/or during its life on the vehicle.


Because the outer glass optic of the lens assembly is positioned toward an exterior of the vehicle such that an outer surface of the outer glass optic may be exposed to the elements outside of the vehicle, the outer surface may be coated with a protective coating to protect the outer optic from deterioration or damage due to exposure to the elements. Such a protective coating may comprise an organic haricot, an inorganic haricot, or an organic/inorganic compound or the like. Additionally, the protective coating may include ultraviolet absorbers or stabilizers to protect the outer optic from UV radiation degradation. Such a coating or other additional coating or coatings substantially reduces the susceptibility of the outer optic to abrasion and/or ultraviolet degradation, thereby providing a substantially clear and durable glass-like appearance to the outer optic. Optionally, the hard coating or protective coating may be of the type described in EPC Application No. 98650039.5, published on Jan. 20, 1999 under Publication No. EP 0892209, the disclosure of which is hereby incorporated herein by reference in its entirety.


Optionally, the outer optic may be heatable to defog or defrost its surfaces. For example, the outer optic may be formed by a glass material and coated with a transparent conductive coating (such as an indium tin oxide (ITO) transparent conductor or a doped tin oxide or the like), or may include a wire mesh or a conductive coating mesh or the like, whereby heat is generated by energizing the conductive coating or mesh via a power source. Optionally, the outer surface of the outer optic may be coated with an anti-soiling or anti-wetting coating, such as a silicone material, which may provide either a hydrophobic or hydrophilic property to the exposed outer surface of the outer optic.


Optionally, the lens body (barrel and optic elements) may be heated electrically to limit or substantially preclude moisture from being trapped inside the lens from condensation. When the lens is assembled in an environment that humidity/moisture level is not tightly controlled below a certain level, the air packet between optic elements inside the lens typically has water molecules in the form of moisture. When the lens is exposed to a temperature that is lower than the dew point temperature of the moisture density level, moisture condensation occurs within the lens, and a small cluster of water droplets may form on the surfaces of the optic elements inside the lens. The condensation or water droplets may blocks or attenuate light passing through the lens and thus may degrade the lens and furthermore may reduce or impair the camera performance. In some situations, the condensation water may drop onto the camera circuitry or circuit board or PCB and may cause an electrical malfunction or electrical shortage. Optionally, a means for measuring the temperature of the lens, such as a temperature sensor, such as a thermal couple or the like, may be used to determine the temperature at or within the lens, while a control means may be responsive to the temperature signal and may activate a heating current or voltage at or in the lens to heat the lens or lens elements above a threshold level in response to a detection or determination that the temperature at or in or of the lens is dropping below the dew point level. Thus, the lens or lens elements may be heated to remain above the dew point temperature so as to limit or substantially avoid moisture condensation from occurring within the lens.


Optionally, a lens body may be filled with a gas that is without water moisture. For example, a dry air, nitrogen gas or helium gas (or other suitable non-moisture gas or air) can be filled inside the lens body. The lens is then sealed to keep the moisture from getting inside lens from outside. The lens can be assembled in an enclosed environment that is filled with the gas without water moisture, such as a dry air, nitrogen or helium gas or the like. Optionally, another means for addressing the moisture issue may provide breathing paths within the lens. For example, the lens body can be designed to have one or more breathing paths by which air and water molecules can travel from all of the inside cavities of the lens to the outside of the lens, such as through the lens bottom or other body wall or housing wall. Optionally, by providing some moisture absorbing material inside the camera cavity, the water molecules that may originally exist inside the lens body can be absorbed by the moisture absorbing material.


The wide angle lens includes corrective refractive and diffractive elements positioned along the optic path between the wide angle outer optic and the image sensor. Additional corrective elements may also be included to further correct color or distortion within the refracted image. Diffractive elements are preferably included to correct the color focusing within the refracted image. The diffractive element or elements may further include an aspheric refractive optic element attached thereto or on an opposite surface of a respective optic element to further correct for distortions and aberrations within the focused image. Optionally, these refractive optic elements may be transparent plastic optic elements, having a varying width, such that the image passing through the refractive lens is refracted in varying degrees, corresponding to the magnitude of distortion in the image received by the refractive optics. Therefore, either by themselves, or in conjunction with the image sensor and/or image processing of the captured images, the refractive and/or diffractive optic elements of the lens function as a means for correcting distortions within the image.


The image system may further reduce image distortion via other distortion reducing means, such as by utilizing non-uniformly distributed pixels at the image array sensor (as discussed below), or such as via an on-imager processing unit (such as a system on chip or the like), such as an FPGA, CPLD, DSP, microprocessor or the like, and associated software, or such as via display means where the distortion correction is done in the displayed image via a non-uniformly distributed pixelated array of the display or via image processing hardware or software of the display device or module.


It is further envisioned that the wide angle imaging system may include electronic image processing or distortion correction to further correct the image displayed on the display. The distortion correction process may be performed by an image processing algorithm or by pixel distribution variation on image capture device and/or an associated display. Preferably, the system may display a vehicle reference, such as a rear bumper of the vehicle, as is seen outside the vehicle, so as to provide a sense of correctness to the image displayed to the driver. For example, an image processing algorithm may be provided which ensures that straight horizontal and/or vertical lines in the scenic information exteriorly of the vehicle are displayed as straight horizontal and/or vertical lines in the image displayed on the display. Such processing allows the imaging system to implement the wide angle lens and still provide a substantially realistic display image to the driver.


Optionally, the image sensor and/or image processing of the captured images may be capable of reducing distortion of the captured image. Such distortion reduction may be accomplished by a non-uniform array of photosensing pixels of the image sensor, such as described in U.S. Pat. No. 5,796,094, which is hereby incorporated herein by reference in its entirety. The non-uniform array may have a coarse distribution of pixels in a center region, with a finer distribution of pixels along the perimeter regions of the captured image. Such non-uniform pixelation of the imaging array receives the focused image and varies correspondingly with the magnitude of distortion in the image refracted through the lens.


Optionally, the image system of the present invention may provide correction of image perspective distortion. Image distortion from the camera with a wide angle lens includes not only lens geometrical distortion (such as pincushion and barrel distortions), but also perspective distortion, which is caused not by lens distortion, but rather is a perspective distortion whereby an object appears in different sizes when it is located at different distances from the camera. This type of distortion may also be corrected via one or more refractive and/or diffractive optic elements of the lens so as to function as lens-based means for correcting distortion. Optionally, the distortion may also or otherwise be corrected via an imager-based means, image processing means and/or display means for correcting distortion, such as those described above.


Optionally, the graphic overlays may be superimposed or projected onto the displayed image based on the image distortion. For example, the system may change or adjust the graphic overlay projection (such as the angle, curvature, thickness, length and/or other characteristics of the lines or overlays) of an overlay pattern in accordance with the position of the overlay pixel at the displayed image.


Thus, the wide angle rearward image system of the present invention provides a wide angle rearward field of view with reduced distortion to the driver of the vehicle so that the images displayed to the driver are readily recognized by the driver when the driver is reversing the vehicle. The image system preferably provides graphic overlays to further enhance the driver's cognitive awareness of detected objects rearward of the vehicle. Although described as a rearward facing imaging system, aspects of the present invention are equally suited for other imaging systems of vehicles, such as side object detection systems or blind spot detection systems or forward facing imaging systems, such as lane departure warning systems and/or the like.


Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.

Claims
  • 1. An imaging system for a vehicle, said imaging system comprising: a rear backup camera, wherein said rear backup camera comprises a two-dimensional array of light sensing photosensor elements, and wherein said rear backup camera has a rearward field of view when mounted at a vehicle equipped with said imaging system;a display device comprising a video display screen for displaying video images derived from image data captured by said rear backup camera;wherein said imaging system generates a backup overlay that is electronically superimposed on the displayed video images to assist a driver of the equipped vehicle when executing a backup maneuver;wherein said backup overlay comprises a pair of side overlays, and wherein said side overlays of said backup overlay are superimposed on the displayed video images so as to extend rearward from respective sides of the rear of the equipped vehicle displayed in the displayed video images;wherein said imaging system generates an alignment overlay that is electronically superimposed on the displayed video images to assist the driver of the equipped vehicle when backing the equipped vehicle towards a trailer hitch of a trailer, and wherein said alignment overlay is superimposed on the displayed video images between said side overlays, and wherein said alignment overlay extends longitudinally along and centrally between said side overlays and is separate and spaced from said side overlays;wherein said alignment overlay is adjusted responsive to a steering angle of the equipped vehicle;wherein said video display screen displays the video images derived from image data captured by said rear backup camera responsive to the equipped vehicle being shifted into reverse gear;wherein said backup overlay is generated and electronically superimposed on the displayed video images responsive to the equipped vehicle being shifted into reverse gear; andwherein said alignment overlay is generated and electronically superimposed on the displayed video images responsive to actuation of an input by the driver of the vehicle to aid back up of the vehicle to the trailer, and wherein the input comprises an actuatable input other than the equipped vehicle being shifted into reverse gear.
  • 2. The imaging system of claim 1, wherein said side overlays of said backup overlay are adjusted responsive to the steering angle of the equipped vehicle.
  • 3. The imaging system of claim 2, wherein said alignment overlay is superimposed along a centerline between said side overlays superimposed on the displayed video images when the steering angle of the equipped vehicle is straight.
  • 4. The imaging system of claim 2, wherein said alignment overlay is superimposed on the displayed video images so as to extend upward toward the focal point of expansion of the displayed video images when the steering angle of the equipped vehicle is straight.
  • 5. The imaging system of claim 1, wherein said alignment overlay comprises a three dimensional alignment overlay.
  • 6. The imaging system of claim 5, wherein said three dimensional alignment overlay is superimposed on the displayed video images so as to appear to be a virtual three dimensional element at the rear of the equipped vehicle.
  • 7. The imaging system of claim 1, wherein said side overlays of said backup overlay comprise three dimensional side overlays.
  • 8. The imaging system of claim 1, wherein said alignment overlay is superimposed on the displayed video images so as to appear as having a different virtual height dimension than said backup overlay to enhance the driver's cognitive awareness of said alignment overlay.
  • 9. The imaging system of claim 1, wherein said rear backup camera is offset from the centerline of the equipped vehicle, and wherein said alignment overlay is superimposed on the displayed video images as a sideward perspective view of a virtual three dimensional alignment overlay to enhance the driver's cognitive awareness of said alignment overlay.
  • 10. The imaging system of claim 1, wherein said alignment overlay is generated and electronically superimposed on the displayed video images responsive to detection, via image processing of image data captured by said rear backup camera, of the trailer rearward of the equipped vehicle during a reversing maneuver of the equipped vehicle.
  • 11. The imaging system of claim 1, wherein each of said side overlays comprises first, second and third overlay zones, and wherein said first overlay zone of said side overlays is indicative of a first distance range from the rear of the equipped vehicle to a first distance from the rear of the equipped vehicle as displayed at said video display screen, and wherein said second overlay zone of said side overlays is indicative of a second distance range from the first distance to a second distance from the rear of the equipped vehicle as displayed at said video display screen, and wherein said third overlay zone of said side overlays is indicative of a third distance range beyond the second distance from the rear of the equipped vehicle as displayed at said video display screen, and wherein, as indicated to the driver viewing the displayed video images at said video display screen when executing a backup maneuver, the first distance is closer to the rear of the equipped vehicle than the second distance and the second distance is closer to the rear of the equipped vehicle than the third distance.
  • 12. The imaging system of claim 11, wherein said first overlay zone of said backup overlay comprises a first color and said second overlay zone comprises a second color and said third overlay zone comprises a third color, and wherein said first color is a different color than said second color and said second color is a different color than said third color.
  • 13. The imaging system of claim 11, wherein said imaging system adjusts said first overlay zone in response to an object being detected rearward of the equipped vehicle and within a first distance threshold from the rear of the equipped vehicle, and wherein said imaging system adjusts said second overlay zone in response to an object being detected rearward of the equipped vehicle and within a second distance threshold from the rear of the equipped vehicle, and wherein said second distance threshold is a larger distance than said first distance threshold.
  • 14. The imaging system of claim 11, wherein, as seen by the driver viewing the displayed video images at said video display screen when executing a backup maneuver, said side overlays of said backup overlay comprise spaced-apart lines extending rearward from the rear of the equipped vehicle, and wherein said first overlay zone comprises a first segment of said spaced-apart lines, said second overlay zone comprises a second segment of said spaced-apart lines, and said third overlay zone comprises a third segment of said spaced-apart lines.
  • 15. The imaging system of claim 14, wherein said backup overlay comprises a first mark where said first overlay zone generally joins with said second overlay zone, and wherein said backup overlay comprises a second mark where said second overlay zone generally joins with said third overlay zone.
  • 16. An imaging system for a vehicle, said imaging system comprising: a rear backup camera, wherein said rear backup camera comprises a two-dimensional array of light sensing photosensor elements, and wherein said rear backup camera has a rearward field of view when mounted at a vehicle equipped with said imaging system;a display device comprising a video display screen for displaying video images derived from image data captured by said rear backup camera;wherein said imaging system generates a backup overlay that is electronically superimposed on the displayed video images to assist a driver of the equipped vehicle when executing a backup maneuver;wherein said backup overlay comprises a pair of side overlays, and wherein said side overlays of said backup overlay are superimposed on the displayed video images so as to extend rearward from respective sides of the rear of the equipped vehicle displayed in the displayed video images;wherein said side overlays of said backup overlay comprise three dimensional side overlays;wherein said imaging system generates an alignment overlay that is electronically superimposed on the displayed video images to assist the driver of the equipped vehicle when backing the equipped vehicle towards a trailer hitch of a trailer, and wherein said alignment overlay is superimposed on the displayed video images between said side overlays and is separate and spaced from said side overlays;wherein said alignment overlay comprises a three dimensional alignment overlay;wherein said alignment overlay is superimposed along a centerline of the displayed video images when a steering angle of the equipped vehicle is straight;wherein said side overlays of said backup overlay and said alignment overlay are adjusted responsive to the steering angle of the equipped vehicle;wherein said video display screen displays the video images derived from image data captured by said rear backup camera responsive to the equipped vehicle being shifted into reverse gear;wherein said backup overlay is generated and electronically superimposed on the displayed video images responsive to the equipped vehicle being shifted into reverse gear; andwherein said alignment overlay is generated and electronically superimposed on the displayed video images responsive to actuation of an input by the driver of the vehicle to aid back up of the vehicle to the trailer, and wherein the input comprises an actuatable input other than the equipped vehicle being shifted into reverse gear.
  • 17. The imaging system of claim 16, wherein said alignment overlay is superimposed on the displayed video images so as to appear as having a different virtual height dimension than said backup overlay to enhance the driver's cognitive awareness of said alignment overlay.
  • 18. The imaging system of claim 16, wherein said rear backup camera is offset from the centerline of the equipped vehicle, and wherein said three dimensional alignment overlay is superimposed on the displayed video images as a sideward perspective view of said three dimensional alignment overlay to enhance the driver's cognitive awareness of said three dimensional alignment overlay.
  • 19. An imaging system for a vehicle, said imaging system comprising: a rear backup camera, wherein said rear backup camera comprises a two-dimensional array of light sensing photosensor elements, and wherein said rear backup camera has a rearward field of view when mounted at a vehicle equipped with said imaging system;a display device comprising a video display screen for displaying video images derived from image data captured by said rear backup camera;wherein said imaging system generates a backup overlay that is electronically superimposed on the displayed video images to assist a driver of the equipped vehicle when executing a backup maneuver;wherein said backup overlay comprises a pair of side overlays, and wherein said side overlays of said backup overlay are superimposed on the displayed video images so as to extend rearward from respective sides of the rear of the equipped vehicle displayed in the displayed video images;wherein said imaging system generates an alignment overlay that is electronically superimposed on the displayed video images to assist the driver of the equipped vehicle when backing the equipped vehicle towards a trailer hitch of a trailer, and wherein said alignment overlay is superimposed on the displayed video images between said side overlays and is separate and spaced from said side overlays;wherein said alignment overlay is superimposed along a centerline between said side overlays of the displayed video images when a steering angle of the equipped vehicle is straight;wherein said side overlays of said backup overlay and said alignment overlay are adjusted responsive to the steering angle of the equipped vehicle;wherein said alignment overlay is superimposed on the displayed video images so as to appear as having a different virtual height dimension than said backup overlay to enhance the driver's cognitive awareness of said alignment overlay;wherein said video display screen displays the video images derived from image data captured by said rear backup camera responsive to the equipped vehicle being shifted into reverse gear;wherein said backup overlay is generated and electronically superimposed on the displayed video images responsive to the equipped vehicle being shifted into reverse gear; andwherein said alignment overlay is generated and electronically superimposed on the displayed video images responsive to detection, via image processing of image data captured by said rear backup camera, of the trailer rearward of the equipped vehicle during a reversing maneuver of the equipped vehicle; andwherein said alignment overlay is generated and electronically superimposed on the displayed video images responsive to actuation of an input by the driver of the vehicle to aid back up of the vehicle to the trailer, and wherein the input comprises an actuatable input other than the equipped vehicle being shifted into reverse gear.
  • 20. The imaging system of claim 19, wherein the input comprises an input selected from the group consisting of (i) a button in the equipped vehicle, (ii) a control input and (iii) a voice recognition system.
CROSS REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. patent application Ser. No. 13/902,042, filed May 24, 2013, now U.S. Pat. No. 9,796,332, which is a continuation of U.S. patent application Ser. No. 12/677,539, filed Mar. 31, 2010, now U.S. Pat. No. 8,451,107, which is a 371 national phase filing of PCT Application No. PCT/US08/76022, filed Sep. 11, 2008, which claims the benefit of U.S. provisional application Ser. No. 60/971,397, filed Sep. 11, 2007, which is hereby incorporated herein by reference in its entirety.

US Referenced Citations (493)
Number Name Date Kind
2632040 Rabinow Mar 1953 A
2827594 Rabinow Mar 1958 A
3349394 Carver Oct 1967 A
3601614 Platzer Aug 1971 A
3612666 Rabinow Oct 1971 A
3665224 Kelsey May 1972 A
3680951 Jordan Aug 1972 A
3689695 Rosenfield et al. Sep 1972 A
3708231 Walters Jan 1973 A
3746430 Brean Jul 1973 A
3807832 Castellion Apr 1974 A
3811046 Levick May 1974 A
3813540 Albrecht May 1974 A
3862798 Hopkins Jan 1975 A
3947095 Moultrie Mar 1976 A
3962600 Pittman Jun 1976 A
3985424 Steinacher Oct 1976 A
3986022 Hyatt Oct 1976 A
4037134 Löper Jul 1977 A
4052712 Ohama et al. Oct 1977 A
4093364 Miller Jun 1978 A
4111720 Michel et al. Sep 1978 A
4161653 Bedini Jul 1979 A
4200361 Malvano Apr 1980 A
4214266 Myers Jul 1980 A
4218698 Bart et al. Aug 1980 A
4236099 Rosenblum Nov 1980 A
4247870 Gabel et al. Jan 1981 A
4249160 Chilvers Feb 1981 A
4266856 Wainwright May 1981 A
4277804 Robison Jul 1981 A
4281898 Ochiai Aug 1981 A
4288814 Talley et al. Sep 1981 A
4355271 Noack Oct 1982 A
4357558 Massoni et al. Nov 1982 A
4381888 Momiyama May 1983 A
4388701 Aichelmann, Jr. et al. Jun 1983 A
4420238 Felix Dec 1983 A
4431896 Lodetti Feb 1984 A
4443057 Bauer Apr 1984 A
4460831 Oettinger et al. Jul 1984 A
4481450 Watanabe et al. Nov 1984 A
4491390 Tong-Shen Jan 1985 A
4512637 Ballmer Apr 1985 A
4529275 Ballmer Jul 1985 A
4529873 Ballmer Jul 1985 A
4546551 Franks Oct 1985 A
4549208 Kamejima et al. Oct 1985 A
4571082 Downs Feb 1986 A
4572619 Reininger Feb 1986 A
4580875 Bechtel Apr 1986 A
4600913 Caine Jul 1986 A
4603946 Kato Aug 1986 A
4614415 Hyatt Sep 1986 A
4620141 McCumber et al. Oct 1986 A
4623222 Itoh Nov 1986 A
4626850 Chey Dec 1986 A
4629941 Ellis Dec 1986 A
4630109 Barton Dec 1986 A
4632509 Ohmi Dec 1986 A
4638287 Umebayashi et al. Jan 1987 A
4647161 Müller Mar 1987 A
4653316 Fukuhara Mar 1987 A
4669825 Itoh Jun 1987 A
4669826 Itoh Jun 1987 A
4671615 Fukada Jun 1987 A
4672457 Hyatt Jun 1987 A
4676601 Itoh Jun 1987 A
4690508 Jacob Sep 1987 A
4692798 Seko et al. Sep 1987 A
4697883 Suzuki Oct 1987 A
4701022 Jacob Oct 1987 A
4713685 Nishimura et al. Dec 1987 A
4717830 Botts Jan 1988 A
4727290 Smith Feb 1988 A
4731669 Hayashi et al. Mar 1988 A
4741603 Miyagi May 1988 A
4768135 Kretschmer et al. Aug 1988 A
4772942 Tuck Sep 1988 A
4789904 Peterson Dec 1988 A
4793690 Gahan Dec 1988 A
4817948 Simonelli Apr 1989 A
4820933 Hong Apr 1989 A
4825232 Howdle Apr 1989 A
4838650 Stewart Jun 1989 A
4847772 Michalopoulos et al. Jul 1989 A
4855822 Narendra et al. Aug 1989 A
4862037 Farber et al. Aug 1989 A
4867561 Fujii et al. Sep 1989 A
4871917 O'Farrell et al. Oct 1989 A
4872051 Dye Oct 1989 A
4881019 Shiraishi et al. Nov 1989 A
4882565 Gallmeyer Nov 1989 A
4886960 Molyneux Dec 1989 A
4891559 Matsumoto et al. Jan 1990 A
4892345 Rachael, III Jan 1990 A
4895790 Swanson et al. Jan 1990 A
4896030 Miyaji Jan 1990 A
4907870 Brucker Mar 1990 A
4910591 Petrossian et al. Mar 1990 A
4916374 Schierbeek Apr 1990 A
4917477 Bechtel et al. Apr 1990 A
4937796 Tendler Jun 1990 A
4953305 Van Lente et al. Sep 1990 A
4956591 Schierbeek Sep 1990 A
4961625 Wood et al. Oct 1990 A
4967319 Seko Oct 1990 A
4970653 Kenue Nov 1990 A
4971430 Lynas Nov 1990 A
4974078 Tsai Nov 1990 A
4987357 Masaki Jan 1991 A
4991054 Walters Feb 1991 A
5001558 Burley et al. Mar 1991 A
5003288 Wilhelm Mar 1991 A
5012082 Watanabe Apr 1991 A
5016977 Baude et al. May 1991 A
5027001 Torbert Jun 1991 A
5027200 Petrossian et al. Jun 1991 A
5044706 Chen Sep 1991 A
5055668 French Oct 1991 A
5059877 Teder Oct 1991 A
5064274 Alten Nov 1991 A
5072154 Chen Dec 1991 A
5086253 Lawler Feb 1992 A
5096287 Kakinami et al. Mar 1992 A
5097362 Lynas Mar 1992 A
5121200 Choi Jun 1992 A
5124549 Michaels et al. Jun 1992 A
5130709 Toyama et al. Jul 1992 A
5140465 Yasui et al. Aug 1992 A
5148014 Lynam Sep 1992 A
5168378 Black Dec 1992 A
5170374 Shimohigashi et al. Dec 1992 A
5172235 Wilm et al. Dec 1992 A
5177685 Davis et al. Jan 1993 A
5182502 Slotkowski et al. Jan 1993 A
5184956 Langlais et al. Feb 1993 A
5189561 Hong Feb 1993 A
5193000 Lipton et al. Mar 1993 A
5193029 Schofield Mar 1993 A
5204778 Bechtel Apr 1993 A
5208701 Maeda May 1993 A
5245422 Borcherts et al. Sep 1993 A
5253109 O'Farrell Oct 1993 A
5276389 Levers Jan 1994 A
5285060 Larson et al. Feb 1994 A
5289182 Brillard et al. Feb 1994 A
5289321 Secor Feb 1994 A
5305012 Faris Apr 1994 A
5307136 Saneyoshi Apr 1994 A
5309137 Kajiwara May 1994 A
5313072 Vachss May 1994 A
5325096 Pakett Jun 1994 A
5325386 Jewell et al. Jun 1994 A
5329206 Slotkowski et al. Jul 1994 A
5331312 Kudoh Jul 1994 A
5336980 Levers Aug 1994 A
5341437 Nakayama Aug 1994 A
5351044 Mathur et al. Sep 1994 A
5355118 Fukuhara Oct 1994 A
5374852 Parkes Dec 1994 A
5386285 Asayama Jan 1995 A
5394333 Kao Feb 1995 A
5406395 Wilson et al. Apr 1995 A
5410346 Saneyoshi et al. Apr 1995 A
5414257 Stanton May 1995 A
5414461 Kishi et al. May 1995 A
5416313 Larson et al. May 1995 A
5416318 Hegyi May 1995 A
5416478 Morinaga May 1995 A
5424952 Asayama Jun 1995 A
5426294 Kobayashi et al. Jun 1995 A
5430431 Nelson Jul 1995 A
5434407 Bauer et al. Jul 1995 A
5440428 Hegg et al. Aug 1995 A
5444478 Lelong et al. Aug 1995 A
5451822 Bechtel et al. Sep 1995 A
5457493 Leddy et al. Oct 1995 A
5461357 Yoshioka et al. Oct 1995 A
5461361 Moore Oct 1995 A
5469298 Suman et al. Nov 1995 A
5471515 Fossum et al. Nov 1995 A
5475494 Nishida et al. Dec 1995 A
5487116 Nakano et al. Jan 1996 A
5498866 Bendicks et al. Mar 1996 A
5500766 Stonecypher Mar 1996 A
5510983 Iino Apr 1996 A
5515448 Nishitani May 1996 A
5521633 Nakajima et al. May 1996 A
5528698 Kamei et al. Jun 1996 A
5529138 Shaw et al. Jun 1996 A
5530240 Larson et al. Jun 1996 A
5530420 Tsuchiya et al. Jun 1996 A
5535314 Alves et al. Jul 1996 A
5537003 Bechtel et al. Jul 1996 A
5539397 Asanuma et al. Jul 1996 A
5541590 Nishio Jul 1996 A
5550677 Schofield et al. Aug 1996 A
5555312 Shima et al. Sep 1996 A
5555555 Sato et al. Sep 1996 A
5568027 Teder Oct 1996 A
5574443 Hsieh Nov 1996 A
5581464 Woll et al. Dec 1996 A
5594222 Caldwell Jan 1997 A
5614788 Mullins Mar 1997 A
5619370 Guinosso Apr 1997 A
5634709 Iwama Jun 1997 A
5642299 Hardin et al. Jun 1997 A
5648835 Uzawa Jul 1997 A
5650944 Kise Jul 1997 A
5660454 Mori et al. Aug 1997 A
5661303 Teder Aug 1997 A
5666028 Bechtel et al. Sep 1997 A
5668663 Varaprasad et al. Sep 1997 A
5670935 Schofield et al. Sep 1997 A
5675489 Pomerleau Oct 1997 A
5677851 Kingdon et al. Oct 1997 A
5699044 Van Lente et al. Dec 1997 A
5703568 Hegyi Dec 1997 A
5724187 Varaprasad et al. Mar 1998 A
5724316 Brunts Mar 1998 A
5737226 Olson et al. Apr 1998 A
5757949 Kinoshita et al. May 1998 A
5760826 Nayer Jun 1998 A
5760828 Cortes Jun 1998 A
5760931 Saburi et al. Jun 1998 A
5760962 Schofield et al. Jun 1998 A
5761094 Olson et al. Jun 1998 A
5765116 Wilson-Jones et al. Jun 1998 A
5781437 Wiemer et al. Jul 1998 A
5786772 Schofield et al. Jul 1998 A
5790403 Nakayama Aug 1998 A
5790973 Blaker et al. Aug 1998 A
5793308 Rosinski et al. Aug 1998 A
5793420 Schmidt Aug 1998 A
5796094 Schofield et al. Aug 1998 A
5798575 O'Farrell et al. Aug 1998 A
5835255 Miles Nov 1998 A
5837994 Stam et al. Nov 1998 A
5844505 Van Ryzin Dec 1998 A
5844682 Kiyomoto et al. Dec 1998 A
5845000 Breed et al. Dec 1998 A
5848802 Breed et al. Dec 1998 A
5850176 Kinoshita et al. Dec 1998 A
5850254 Takano et al. Dec 1998 A
5867591 Onda Feb 1999 A
5877707 Kowalick Mar 1999 A
5877897 Schofield et al. Mar 1999 A
5878370 Olson Mar 1999 A
5883739 Ashihara et al. Mar 1999 A
5884212 Lion Mar 1999 A
5890021 Onoda Mar 1999 A
5896085 Mori et al. Apr 1999 A
5899956 Chan May 1999 A
5914815 Bos Jun 1999 A
5923027 Stam et al. Jul 1999 A
5929786 Schofield et al. Jul 1999 A
5940120 Frankhouse et al. Aug 1999 A
5949331 Schofield et al. Sep 1999 A
5956181 Lin Sep 1999 A
5959367 O'Farrell et al. Sep 1999 A
5959555 Furuta Sep 1999 A
5963247 Banitt Oct 1999 A
5964822 Alland et al. Oct 1999 A
5971552 O'Farrell et al. Oct 1999 A
5986796 Miles Nov 1999 A
5990469 Bechtel et al. Nov 1999 A
5990649 Nagao et al. Nov 1999 A
6001486 Varaprasad et al. Dec 1999 A
6009336 Harris et al. Dec 1999 A
6020704 Buschur Feb 2000 A
6049171 Stam et al. Apr 2000 A
6066933 Ponziana May 2000 A
6084519 Coulling et al. Jul 2000 A
6087953 DeLine et al. Jul 2000 A
6097023 Schofield et al. Aug 2000 A
6097024 Stam et al. Aug 2000 A
6116743 Hoek Sep 2000 A
6118383 Hegyi Sep 2000 A
6124647 Marcus et al. Sep 2000 A
6124886 DeLine et al. Sep 2000 A
6139172 Bos et al. Oct 2000 A
6144022 Tenenbaum et al. Nov 2000 A
6172613 DeLine et al. Jan 2001 B1
6175164 O'Farrell et al. Jan 2001 B1
6175300 Kendrick Jan 2001 B1
6198409 Schofield et al. Mar 2001 B1
6201642 Bos Mar 2001 B1
6222447 Schofield et al. Apr 2001 B1
6222460 DeLine et al. Apr 2001 B1
6243003 DeLine et al. Jun 2001 B1
6250148 Lynam Jun 2001 B1
6259412 Duroux Jul 2001 B1
6266082 Yonezawa et al. Jul 2001 B1
6266442 Laumeyer et al. Jul 2001 B1
6285393 Shimoura et al. Sep 2001 B1
6291906 Marcus et al. Sep 2001 B1
6294989 Schofield et al. Sep 2001 B1
6297781 Turnbull et al. Oct 2001 B1
6302545 Schofield et al. Oct 2001 B1
6310611 Caldwell Oct 2001 B1
6313454 Bos et al. Nov 2001 B1
6317057 Lee Nov 2001 B1
6320176 Schofield et al. Nov 2001 B1
6320282 Caldwell Nov 2001 B1
6326613 Heslin et al. Dec 2001 B1
6329925 Skiver et al. Dec 2001 B1
6333759 Mazzilli Dec 2001 B1
6341523 Lynam Jan 2002 B2
6353392 Schofield et al. Mar 2002 B1
6366213 DeLine et al. Apr 2002 B2
6370329 Teuchert Apr 2002 B1
6396397 Bos et al. May 2002 B1
6411204 Bloomfield et al. Jun 2002 B1
6411328 Franke et al. Jun 2002 B1
6420975 DeLine et al. Jul 2002 B1
6424273 Gutta et al. Jul 2002 B1
6428172 Hutzel et al. Aug 2002 B1
6430303 Naoi et al. Aug 2002 B1
6433676 DeLine et al. Aug 2002 B2
6433817 Guerra Aug 2002 B1
6442465 Breed et al. Aug 2002 B2
6477464 McCarthy et al. Nov 2002 B2
6480104 Wall et al. Nov 2002 B1
6485155 Duroux et al. Nov 2002 B1
6497503 Dassanayake et al. Dec 2002 B1
6498620 Schofield et al. Dec 2002 B2
6513252 Schierbeek et al. Feb 2003 B1
6516664 Lynam Feb 2003 B2
6523964 Schofield et al. Feb 2003 B2
6534884 Marcus et al. Mar 2003 B2
6539306 Turnbull Mar 2003 B2
6547133 DeVries, Jr. et al. Apr 2003 B1
6553130 Lemelson et al. Apr 2003 B1
6559435 Schofield et al. May 2003 B2
6574033 Chui et al. Jun 2003 B1
6578017 Ebersole et al. Jun 2003 B1
6587573 Stam et al. Jul 2003 B1
6589625 Kothari et al. Jul 2003 B1
6593565 Heslin et al. Jul 2003 B2
6594583 Ogura et al. Jul 2003 B2
6611202 Schofield et al. Aug 2003 B2
6611610 Stam et al. Aug 2003 B1
6625587 Erten et al. Sep 2003 B1
6627918 Getz et al. Sep 2003 B2
6631994 Suzuki et al. Oct 2003 B2
6636258 Strumolo Oct 2003 B2
6648477 Hutzel et al. Nov 2003 B2
6650233 DeLine et al. Nov 2003 B2
6650455 Miles Nov 2003 B2
6672731 Schnell et al. Jan 2004 B2
6674562 Miles Jan 2004 B1
6678614 McCarthy et al. Jan 2004 B2
6680792 Miles Jan 2004 B2
6690268 Schofield et al. Feb 2004 B2
6700605 Toyoda et al. Mar 2004 B1
6703925 Steffel Mar 2004 B2
6704621 Stein et al. Mar 2004 B1
6710908 Miles et al. Mar 2004 B2
6711474 Treyz et al. Mar 2004 B1
6714331 Lewis et al. Mar 2004 B2
6717610 Bos et al. Apr 2004 B1
6735506 Breed et al. May 2004 B2
6741377 Miles May 2004 B2
6744353 Sjönell Jun 2004 B2
6757109 Bos Jun 2004 B2
6762867 Lippert et al. Jul 2004 B2
6765607 Mizusawa et al. Jul 2004 B2
6794119 Miles Sep 2004 B2
6795221 Urey Sep 2004 B1
6802617 Schofield et al. Oct 2004 B2
6806452 Bos et al. Oct 2004 B2
6822563 Bos et al. Nov 2004 B2
6823241 Shirato et al. Nov 2004 B2
6824281 Schofield et al. Nov 2004 B2
6831261 Schofield et al. Dec 2004 B2
6847487 Burgner Jan 2005 B2
6882287 Schofield Apr 2005 B2
6889161 Winner et al. May 2005 B2
6891563 Schofield et al. May 2005 B2
6909753 Meehan et al. Jun 2005 B2
6946978 Schofield Sep 2005 B2
6953253 Schofield et al. Oct 2005 B2
6968736 Lynam Nov 2005 B2
6970184 Hirama et al. Nov 2005 B2
6975775 Rykowski et al. Dec 2005 B2
7004593 Weller et al. Feb 2006 B2
7004606 Schofield Feb 2006 B2
7005974 McMahon et al. Feb 2006 B2
7006127 Mizusawa et al. Feb 2006 B2
7038577 Pawlicki et al. May 2006 B2
7062300 Kim Jun 2006 B1
7065432 Moisel et al. Jun 2006 B2
7085637 Breed et al. Aug 2006 B2
7092548 Laumeyer et al. Aug 2006 B2
7116246 Winter et al. Oct 2006 B2
7123168 Schofield Oct 2006 B2
7133661 Hatae et al. Nov 2006 B2
7149613 Stam et al. Dec 2006 B2
7167779 Kashiwada et al. Jan 2007 B2
7167796 Taylor et al. Jan 2007 B2
7202776 Breed Apr 2007 B2
7227459 Bos et al. Jun 2007 B2
7227611 Hull et al. Jun 2007 B2
7311406 Schofield et al. Dec 2007 B2
7325934 Schofield et al. Feb 2008 B2
7325935 Schofield et al. Feb 2008 B2
7339149 Schofield et al. Mar 2008 B1
7344261 Schofield et al. Mar 2008 B2
7375803 Bamji May 2008 B1
7380948 Schofield et al. Jun 2008 B2
7388182 Schofield et al. Jun 2008 B2
7402786 Schofield et al. Jul 2008 B2
7423248 Schofield et al. Sep 2008 B2
7423821 Bechtel et al. Sep 2008 B2
7425076 Schofield et al. Sep 2008 B2
7459664 Schofield et al. Dec 2008 B2
7526103 Schofield et al. Apr 2009 B2
7561181 Schofield et al. Jul 2009 B2
7565006 Stam et al. Jul 2009 B2
7616781 Schofield et al. Nov 2009 B2
7619508 Lynam et al. Nov 2009 B2
7633383 Dunsmoir et al. Dec 2009 B2
7639149 Katoh Dec 2009 B2
7676087 Dhua et al. Mar 2010 B2
7697786 Camus et al. Apr 2010 B2
7720580 Higgins-Luthman May 2010 B2
7843451 Lafon Nov 2010 B2
7855778 Yung et al. Dec 2010 B2
7881496 Camilleri Feb 2011 B2
7914187 Higgins-Luthman et al. Mar 2011 B2
7930160 Hosagrahara et al. Apr 2011 B1
8017898 Lu et al. Sep 2011 B2
8224031 Saito Jul 2012 B2
8237794 Moritz et al. Aug 2012 B2
8451107 Lu et al. May 2013 B2
9796332 Lu et al. Oct 2017 B2
20020015153 Downs Feb 2002 A1
20020044065 Quist et al. Apr 2002 A1
20020113873 Williams Aug 2002 A1
20020159270 Lynam et al. Oct 2002 A1
20030137586 Lewellen Jul 2003 A1
20030222982 Hamdan et al. Dec 2003 A1
20030227777 Schofield Dec 2003 A1
20040012488 Schofield Jan 2004 A1
20040016870 Pawlicki et al. Jan 2004 A1
20040032321 McMahon et al. Feb 2004 A1
20040051634 Schofield et al. Mar 2004 A1
20040114381 Salmeen et al. Jun 2004 A1
20040128065 Taylor et al. Jul 2004 A1
20040200948 Bos et al. Oct 2004 A1
20050078389 Kulas et al. Apr 2005 A1
20050134966 Burgner Jun 2005 A1
20050134983 Lynam Jun 2005 A1
20050146792 Schofield et al. Jul 2005 A1
20050169003 Lindahl et al. Aug 2005 A1
20050195488 McCabe et al. Sep 2005 A1
20050200700 Schofield et al. Sep 2005 A1
20050219852 Stam et al. Oct 2005 A1
20050232469 Schofield et al. Oct 2005 A1
20050237385 Kosaka et al. Oct 2005 A1
20050264891 Uken et al. Dec 2005 A1
20060018511 Stam et al. Jan 2006 A1
20060018512 Stam et al. Jan 2006 A1
20060028731 Schofield et al. Feb 2006 A1
20060050018 Hutzel et al. Mar 2006 A1
20060080005 Lee Apr 2006 A1
20060091813 Stam et al. May 2006 A1
20060103727 Tseng May 2006 A1
20060125919 Camilleri Jun 2006 A1
20060164230 DeWind et al. Jul 2006 A1
20060250501 Wildmann et al. Nov 2006 A1
20070023613 Schofield et al. Feb 2007 A1
20070104476 Yasutomi et al. May 2007 A1
20070109406 Schofield et al. May 2007 A1
20070109651 Schofield et al. May 2007 A1
20070109652 Schofield et al. May 2007 A1
20070109653 Schofield et al. May 2007 A1
20070109654 Schofield et al. May 2007 A1
20070120657 Schofield May 2007 A1
20070176080 Schofield et al. Aug 2007 A1
20070242339 Bradley Oct 2007 A1
20080077882 Kramer Mar 2008 A1
20080147321 Howard et al. Jun 2008 A1
20080180529 Taylor et al. Jul 2008 A1
20080192132 Bechtel et al. Aug 2008 A1
20080266541 Yung Oct 2008 A1
20090113509 Tseng et al. Apr 2009 A1
20090160987 Bechtel et al. Jun 2009 A1
20090190015 Bechtel et al. Jul 2009 A1
20090256938 Bechtel et al. Oct 2009 A1
20100045797 Schofield et al. Feb 2010 A1
20120045112 Lundblad et al. Feb 2012 A1
Foreign Referenced Citations (49)
Number Date Country
4107965 Sep 1991 DE
4118208 Nov 1991 DE
4139515 Jun 1992 DE
4123641 Jan 1993 DE
0416222 Mar 1991 EP
0426503 May 1991 EP
0513476 Nov 1992 EP
1288072 Mar 2003 EP
2641237 Jul 1990 FR
2672857 Aug 1992 FR
2673499 Sep 1992 FR
2726144 Apr 1996 FR
2244187 Nov 1991 GB
2255539 Nov 1992 GB
2327823 Feb 1999 GB
59-114139 Jul 1984 JP
60-79889 Oct 1986 JP
62-72245 Aug 1987 JP
64-014700 Jan 1989 JP
01-123587 May 1989 JP
01-141137 Jun 1989 JP
03-061192 Mar 1991 JP
03-099952 Apr 1991 JP
03-284413 Dec 1991 JP
04-114587 Apr 1992 JP
04-245866 Sep 1992 JP
05-000638 Jan 1993 JP
05-050883 Mar 1993 JP
05-077657 Mar 1993 JP
05-213113 Aug 1993 JP
61-056638 Jun 1994 JP
62-027318 Aug 1994 JP
06-267304 Sep 1994 JP
06-276524 Sep 1994 JP
06-295601 Oct 1994 JP
07-004170 Jan 1995 JP
07-032936 Feb 1995 JP
07-047878 Feb 1995 JP
07-052706 Feb 1995 JP
07-069125 Mar 1995 JP
07-105496 Apr 1995 JP
08-166221 Jun 1996 JP
26-30604 Jul 1997 JP
2003-083742 Mar 2003 JP
3945467 Jul 2007 JP
WO 1986005147 Sep 1986 WO
WO 1996021581 Jul 1996 WO
WO 1997035743 Oct 1997 WO
WO 1998014974 Apr 1998 WO
Non-Patent Literature Citations (24)
Entry
The 2002 Infiniti Q45 Owners Manual.
Article entitled “Generation of Vision Technology,” published by VLSI Vision Limited, publication date unknown.
Article entitled On-Chip CMOS Sensors for VLSI Imaging Systems,: published by VLSI Vision Limited, 1991.
Bow, Sing T., “Pattern Recognition and Image Preprocessing (Signal Processing and Communications)”, CRC Press, Jan. 15, 2002, pp. 557-559.
Dana H. Ballard and Christopher M. Brown, Computer Vision, Prentice-Hall, Englewood Cliffs, New Jersey, 5 pages, 1982.
Decision—Motions—Bd. R. 125(a), issued Aug. 29, 2006 in connection with Interference No. 105,325, which involved U.S. Appl. No. 09/441,341, filed Nov. 16, 1999 by Schofield et al. and U.S. Pat. No. 5,837,994, issued to Stam et al.
G. Wang, D. Renshaw, P.B. Denyer and M. Lu, CMOS Video Cameras, article, 1991, 4 pages, University of Edinburgh, UK.
Hamit, Francis “360-Degree Interactivity: New Video and Still Cameras Provide a Global Roaming Viewpoint,” Advanced Imaging, Mar. 1997, p. 50.
J. Borenstein et al., “Where am I? Sensors and Method for Mobile Robot Positioning”, University of Michigan, Apr. 1996, pp. 2, 125-128.
Johannas, Laura “A New Microchip Ushers in Cheaper Digital Cameras,” The Wall Street Journal, Aug. 21, 1998, p. B1.
Pratt, “Digital Image Processing, Passage—ED.3”, John Wiley & Sons, US, Jan. 1, 2001, pp. 657-659, XP002529771.
Reexamination Control No. 90/007,519, Reexamination of U.S. Pat. No. 6,222,447, issued to Schofield et al.
Reexamination Control No. 90/007,520, Reexamination of U.S. Pat. No. 5,949,331, issued to Schofield et al.
Reexamination Control No. 90/011,477, Reexamination of U.S. Pat. No. 5,949,331, issued to Schofield et al.
Reexamination Control No. 90/011,478, Reexamination of U.S. Pat. No. 6,222,447, issued to Schofield et al.
Tokimaru et al., “CMOS Rear-View TV System with CCD Camera”, National Technical Report vol. 34, No. 3, pp. 329-336, Jun. 1988 (Japan).
Van Leeuwen et al., “Motion Estimation in Image Sequences for Traffic Applications”, vol. 1, May 1, 2000, pp. 354-359, XP002529773.
Van Leeuwen et al., “Motion Estimation with a Mobile Camera for Traffic Applications”, IEEE, US, vol. 1, Oct. 3, 2000, pp. 58-63.
Van Leeuwen et al., “Motion Interpretation for In-Car Vision Systems”, IEEE, US, vol. 1, Sep. 30, 2002, p. 135-140.
Van Leeuwen et al., “Requirements for Motion Estimation in Image Sequences for Traffic Applications”, IEEE, US, vol. 1, May 24, 1999, pp. 145-150, XP010340272.
Van Leuven et al., “Real-Time Vehicle Tracking in Image Sequences”, IEEE, US, vol. 3, May 21, 2001, pp. 2049-2054, XP010547308.
Vlacic et al., (Eds), “Intelligent Vehicle Tecnologies, Theory and Applications”, Society of Automotive Engineers Inc., edited by SAE International, 2001.
International Search Report and Written Opinion dated Nov. 24, 2008 for corresponding PCT Application No. PCT/US2008/076022.
Supplementary European Search Report dated Sep. 4, 2013 from corresponding European Application No. EP08830763.
Related Publications (1)
Number Date Country
20180056874 A1 Mar 2018 US
Provisional Applications (1)
Number Date Country
60971397 Sep 2007 US
Continuations (2)
Number Date Country
Parent 13902042 May 2013 US
Child 15790172 US
Parent 12677539 US
Child 13902042 US