Automatic vehicle exterior light control

Abstract
A system and method of automatically controlling vehicle exterior lights including an image sensor and a controller to generate headlamp control signals.
Description
BACKGROUND OF THE INVENTION

This invention relates generally to vision systems for vehicles and, more particularly, to rearview vision systems which provide the vehicle operator with scenic information in the direction rearward of the vehicle. More particularly, the invention relates to a rearview vision system utilizing image capture devices, such as CMOS imaging arrays and the like.


A long-felt need in the art of vehicle rearview vision systems is to reduce the amount of time spent gathering information of the condition around the vehicle in order to safely carry out a vehicle maneuver, such as a turn or a lane change. It has been determined that up to about 50 percent of maneuver time is spent gathering information with conventional rearview mirrors. This typically requires viewing one or more mirrors, turning the head and body to check blind spots, and double-checking the mirrors prior to executing the maneuver. Some improvement has been obtained by adjusting mirror optics, for example, to increase the observed field of view rearward of the vehicle. However, this is achieved with an increase in image distortion which makes driver awareness of conditions to the side and rear of the vehicle even more difficult.


Another long-felt need in the art of vehicle rearview vision systems has been to eliminate exterior rearview mirrors by utilizing image capture devices, such as cameras, in combination


with dashboard displays. This would be beneficial because it would reduce wind drag on the vehicle, wind noise and vehicle weight. Furthermore, rearview mirrors protrude a substantial distance from the side of the vehicle, which makes maneuvering in tight spaces more difficult. Image capture devices are capable of positioning in a greater variety of locations on the vehicle, providing more flexibility of vehicle styling. It is further expected that camera systems would greatly reduce the blind spots to the sides and rear of the vehicle common with vehicles equipped with conventional rearview mirror systems. The driver cannot perceive vehicles, objects, or other road users in such blind spots without turning his or her body, which interferes with forward-looking visual activities.


Camera-based rearview vision systems for vehicles have not obtained commercial acceptance. One difficulty with proposed systems has been that they present a large amount of visual information in a manner which is difficult to comprehend. This difficulty arises from many factors. In order to significantly reduce blind spots, multiple image capture devices are typically positioned at various locations on the vehicle. The image of an object behind the equipped vehicle is usually captured by more than one image capture device at a time and displayed in multiple images. This may confuse the driver as to whether more than one object is present. When multiple image capture devices are positioned at different longitudinal locations on the vehicle, objects behind the vehicle are at different distances from the image capture devices. This results in different image sizes for the same object. This effect is especially noticeable for laterally extending images, such as a bridge, highway crosswalk markings, the earth's horizon, and the like. Such images are at different vertical angles with respect to the image capture devices. This results in different vertical positions on the display causing the elongated image to appear disjointed.


A camera system provides a monocular view of the scene, compared to the binocular stereoscopic view obtained when the scene is viewed through a rearview mirror. This makes the ability to judge distances in a camera system a problem. This effect is most noticeable at distances close to the vehicle where stereoscopic imaging is relied upon extensively by the driver in judging relative locations of objects. Therefore, known camera systems fail to provide to the driver important information where that information is most needed at small separation distances from surrounding objects.


Another difficulty with camera systems is that, in order to provide a sufficient amount of information, the camera system typically presents the driver with a greatly increased field of view. This improves performance by further reducing blind spots at the side and rear of the vehicle. However, an increased field of view is often obtained by utilizing a wide-angle lens which introduces distortion of the scene and further impairs the ability of the driver to judge distances of objects displayed. The problem with such distortion of the scene is that the driver must concentrate more on the display and take a longer time to interpret and extract the necessary information. This further distracts the driver from the primary visual task of maintaining awareness of vehicles and other objects in the vicinity of the driven vehicle.


Yet an additional difficulty with camera systems is that flat panel displays present the image captured by the rearward-facing image capture device, or devices, at a focal length that approximates the arm length of the vehicle driver. In order to observe the condition of the vehicle utilizing the rearview vision system, the driver must change his or her gaze from the forward field of view of the vehicle to the display. Because the forward field of view of the vehicle is at a focal length that is much greater than the focal length of the displayed image, the eyes of the driver must refocus upon changing gaze. This refocusing further increases the amount of time for the driver to assimilate the information in the displayed image. Furthermore, when the gaze of the driver returns to the forward field of view, the eyes must, again, refocus to the greatly longer distance.


Yet an additional difficulty with camera systems is that of finding adequate space in the crowded area of the vehicle's dashboard for the components making up the display.


SUMMARY OF THE INVENTION

The present invention is directed towards enhancing the interpretation of visual information in a rearview vision system by presenting information in a manner which does not require significant concentration of the driver or present distractions to the driver. This is accomplished according to an aspect of the invention in a rearview vision system having at least two image capture devices positioned on the vehicle and directed generally rearwardly with respect to the direction of travel of the vehicle. A display is provided for images captured by the image capture devices. The display displays an image synthesized from outputs of the image capture devices which approximates a rearward-facing view from a single location. In order to obtain all of the necessary information of activity, not only behind but also along side of the vehicle, the virtual camera should be positioned forward of the driver. The image synthesized from the multiple image capture devices may have a dead space which corresponds with the area occupied by the vehicle. This dead space is useable by the driver's sense of perspective in judging the location of vehicles behind and along side of the equipped vehicle.


The present invention provides techniques for synthesizing images captured by individual, spatially separated, image capture devices into such ideal image, displayed on the display device. This may be accomplished, according to an aspect of the invention, by providing at least three image capture devices. At least two of the image capture devices are side image capture devices mounted on opposite sides of the vehicle. At least one of the image capture devices is a center image capture device mounted laterally between the side image capture devices. A display system displays an image synthesized from outputs of the image capture devices. The displayed image includes an image portion from each of the image capture devices. The image portion from the center image capture device is vertically compressed.


It has been discovered that such vertical compression substantially eliminates distortion resulting from the spatial separation between the cameras and can be readily accomplished. In an illustrated embodiment, the image compression is carried out by removing selective ones of the scan lines making up the image portion. A greater number of lines are removed further away from the vertical center of the image.


The compression of the central image portion produces a dead space in the displayed image which may be made to correspond with the area that would be occupied by the vehicle in the view from the single virtual camera. Preferably, perspective lines are included at lateral edges of the dead space which are aligned with the direction of travel of the vehicle and, therefore, appear in parallel with lane markings. This provides visual clues to the driver's sense of perspective in order to assist in judging distances of objects around the vehicle.


According to another aspect of the invention, image enhancement means are provided for enhancing the displayed image. Such means may be in the form of graphic overlays superimposed on the displayed image. Such graphic overlap may include indicia of the anticipated path of travel of the vehicle which is useful in assisting the driver in guiding the vehicle in reverse directions. Such graphic overlay may include a distance grid indicating distances behind the vehicle of objects juxtaposed with the grid.


According to yet an additional aspect of the invention, a rearview vision system for a vehicle includes at least one image capture device positioned on the vehicle and directed generally rearwardly with respect to the direction of travel of the vehicle. A display system is provided which displays a rear image synthesized from an output of the image capture device. The rear image is substantially contiguous with the forward field of view of the vehicle driver and at a focal length that is forward of the vehicle passenger compartment and preferably within the depth of field of a vehicle driver viewing a distant object. Because the image has a focal length that more closely matches that of the forward field of view observed by the driver, the need for the driver's eyes to refocus from the forward field of view to a much shorter focus distance each time the gaze of the driver is directed at the display system is minimized. This reduces the amount of time required for the driver to gaze at the displayed image and interpret objects displayed in the image. Furthermore, the reduction in the repeated refocusing of the driver's eyes reduces driver fatigue. If there are any near field objects in the periphery of the driver's forward field of view, such as windshield wipers, windshield frame, dashboard, and the like, the display system is preferably positioned in a manner which blocks the view of such near field objects. In this manner, the driver's gaze may shift between the forward field of view and the long focal length display system without being refocused on the near field objects. This is based upon a recognition that the driver's eyes will tend to refocus on the near field object momentarily even though the gaze is being redirected between the forward field of view and the display system.


According to yet an additional aspect of the invention, a rearview vision system for a vehicle includes at least one image capture device positioned on the vehicle and directed generally rearwardly with respect to the direction of travel of the vehicle. A display system is provided for displaying a rear image captured by the image capture device. The displayed image is a unitary image having an aspect ratio that is between approximately 4:1 and approximately 2:1. In a most preferred embodiment, the image has an aspect ratio that is approximately 8:3. The aspect ratio, according to this aspect of the invention, is especially useful where the unitary image is synthesized from a plurality of images which are captured by a plurality of image captured devices and are tiled by the display device.


According to yet an additional aspect of the invention, a rearview vision system for a vehicle includes a plurality of image capture devices positioned on the vehicle and directed generally rearwardly with respect to the direction of travel of the vehicle. A display system which includes at least one image generator and an optical correction system is provided which displays an image synthesized from outputs of the image capture devices as a unitary image. Alternatively, the display system may include a plurality of image generators, each associated with one or more of the image capture devices and an optical correction system which amplifies images generated by the image generators and merges them into a unitary image. The optical correction system additionally increases the focal length, or lengths, of the image, or images, generated by the image generator, or generators. The display system may be an opaque projection display which is positioned approximately at the driver's arm length in front of the driver. Alternatively, the display system may be a view-through heads-up display which projects the unitary image onto a combiner in order to combine the unitary image with the forward field of view of the driver.


These and other objects, advantages, and features of this invention will become apparent by review of the following specification in conjunction with the drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a top plan view of a vehicle having a rearview vision system according to the invention;



FIG. 2 is a side elevation of the vehicle in FIG. 1;



FIG. 3 is a front elevation of a display according to the invention;



FIG. 4 is the same view as FIG. 1 illustrating an alternative embodiment of the invention;



FIG. 5 is a block diagram of an electronic system according to the invention;



FIG. 6 is the same view as FIG. 3 illustrating an alternate mode of operation of the system;



FIG. 7 is the same view as FIG. 2 illustrating an alternative embodiment of the invention;



FIG. 8 is the same view as FIG. 3 illustrating an alternative embodiment of the invention;



FIG. 9 is the same view as FIGS. 1 and 4 illustrating an alternative embodiment of the invention;



FIG. 10 is the same view as FIGS. 3 and 8 illustrating an alternative embodiment of the invention;



FIG. 11 is a chart illustrating the horizontal row of pixels (n1, n2) on which an object will be imaged from two longitudinally separated image capture devices as that object is spaced at different longitudinal distances from the image capture devices;



FIG. 12 is a forward elevation of a vehicle passenger compartment as viewed by a vehicle driver;



FIG. 13 is a sectional view taken along the lines XIII-XIII in FIG. 12;



FIG. 14 is a sectional view taken along the lines XIV-XIV in FIG. 12;



FIG. 15 is the same view as FIG. 14 of an alternative embodiment;



FIG. 16 is the same view as FIG. 14 of another alternative embodiment;



FIG. 17 is an enlarged view of the display system in FIG. 14 illustrating details thereof;



FIG. 18 is a block diagram similar to FIG. 5 of an alternative embodiment of the invention;



FIG. 19 is a side elevation similar to FIG. 2 of an alternative embodiment of the invention;



FIG. 20 is an enlarged side elevation of an image capture device with portions of the housing removed in order to reveal internal structure thereof;



FIG. 21 is a block diagram similar to FIG. 5 of another alternative embodiment of the invention;



FIG. 22 is a block diagram similar to FIG. 5 of another alternative embodiment of the invention;



FIG. 23 is a block diagram similar to FIG. 5 of another alternative embodiment of the invention;



FIG. 24 is a block diagram similar to FIG. 5 of another alternative embodiment of the invention; and



FIG. 25 is a block diagram of a rearview vision system having extended dynamic range capabilities.





DESCRIPTION OF THE PREFERRED EMBODIMENT

Referring now specifically to the drawings, and the illustrative embodiments depicted therein, a vehicle 10, which may be an automobile, a light truck, a sport utility vehicle, a van, a bus, a large truck, or the like includes a rearview vision system, generally illustrated at 12, for providing a driver of the vehicle with a view rearwardly of the vehicle with respect to the direction of travel T of the vehicle (FIG. 1). Vision system 12 includes at least two side image capture devices 14 positioned, respectively, on opposite sides of vehicle 10 and a center image capture device 16 positioned on the lateral centerline of the vehicle. All of the image capture devices are directed generally rearwardly of the vehicle. Rearview vision system 12 additionally includes an image processor 18 for receiving data signals from image capture devices 14, 16 and synthesizing, from the data signals, a composite image 42 which is displayed on a display 20.


As will be set forth in more detail below, the images captured by image capture devices 14, 16 are juxtaposed on display 20 by image processor 18 in a manner which approximates the view from a single virtual image capture device positioned forwardly of the vehicle at a location C and facing rearwardly of the vehicle, with the vehicle being transparent to the view of the virtual image capture device. Vision system 12 provides a substantially seamless panoramic view rearwardly of the vehicle without duplicate or redundant images of objects. Furthermore, elongated, laterally-extending, objects, such as the earth's horizon, appear uniform and straight across the entire displayed image. The displayed image provides a sense of perspective, which enhances the ability of the driver to judge location and speed of adjacent trailing vehicles.


Each of side image capture devices 14 has a field of view 22 and is aimed rearwardly with respect to the vehicle about an axis 24 which is at an angle, with respect to the vehicle, that is half of the horizontal field of view of the image capture device. In this manner, each of the image capture devices 14 covers an area bounded by the side of the vehicle and extending outwardly at an angle defined by the horizontal field of view of the respective side image capture device. Center image capture device 16 has a horizontal field of view 26, which is symmetrical about the longitudinal axis of the vehicle. The field of view of each side image capture device 14 intersect the field of view of center image capture device 16 at a point P which is located a distance Q behind vehicle 10.


Rear blind zones 30 are located symmetrically behind vehicle 10 extending from the rear of the vehicle to point P. Side blind zones 25 located laterally on respective sides of the vehicle extend rearwardly of the forward field of view 36 of the driver to the field of view 22 of the respective side image capture device 14. An object will not be captured by side image capture devices 14 or center image capture device 16 if the object is entirely within one of the blind zones 25, 30. In order for an object, such as another vehicle V or other road user travelling to the side of vehicle 10, to be observed by an operator of vehicle 10, the object must be either at least partially within the forward field of view 36 of the driver or be captured by image capture devices 14, 16 and displayed on display 20. FIG. 4 illustrates vehicle 10 travelling on a three-lane highway having lanes L1, L2, and L3 with the vehicle in lane L2. Another vehicle V is shown positioned mostly within one of the blind zones 25, but with the rearmost portion of the vehicle V extending into field of view 22 where the vehicle image will be captured by one of side image capture devices 14. In the illustrated embodiment, vehicle V is a motorcycle travelling in the center of lanes L1 or L3 and represents a worst case for observing a vehicle travelling at least partially within one of the blind zones 25. In order for a portion of vehicle V to be extending either forwardly or rearwardly of the respective blind zone 25, where the vehicle V may be observed by either the forward field of view 36 of the driver or by the rearview vision system 12, the field of view 22 of side image capture devices 14 must be sufficiently wide to capture a portion of vehicle V as illustrated in FIG. 4. Preferably, the horizontal field of view 22 of side image capture devices 14 is no greater than that required to provide sufficient coverage which would be in the range of between approximately 55 degrees and approximately 70 degrees. In the illustrated embodiment, the horizontal field of view 22 is 61 degrees. In order for a portion of vehicle V to be within a vertical field of view 40 of one of side image capture devices 14, the field of view should extend to the pavement at a plane M which intersects vehicle V (FIG. 2). Preferably, vertical field of view 40 is between approximately 60 degrees and approximately 75 degrees. In the illustrated embodiment, vertical field of view 40 is 66 degrees.


A left overlap zone 32 and a right overlap zone 34 extend rearward from respective points P where the horizontal fields of view of the side image capture devices intersect the field of view of center image capture device 16. Overlap zones 32, 34 define areas within which an object will be captured both by center image capture device 16 and one of the side image capture devices 14. An object in an overlap zone 32, 34 will appear on display 20 in multiple image portions in a redundant or duplicative fashion. In order to avoid the presentation of redundant information to the driver, and thereby avoid confusion and simplify the task of extracting information from the multiple images or combined images on display 20, the object should avoid overlapping zones 32, 34. In practice, this may be accomplished to a satisfactory extent by moving points P away from the vehicle and thereby increasing distance Q. It is desirable to increase distance Q to a length that will exclude vehicles travelling at a typical separation distance behind vehicle 10 from overlapping zones 32, 34. This separation distance is usually a function of the speed at which the vehicles on the highway are travelling. Therefore, optionally, distance Q may be made variable, not fixed. In such embodiment, the faster the vehicles are travelling, the further Q should be moved behind vehicle 10 to keep overlap zones 32 and 34 outside of the recommended vehicle spacing. If, however, the vehicles are travelling at a slower speed, then the generally accepted recommendation for vehicle spacing decreases and it is more likely that a vehicle will be within overlap zone 32, 34. Therefore, the distance Q may be selected to accommodate expected vehicle spacing for an average driving speed of vehicle 10.


Distance Q is a function of the effective horizontal field of view 26 of center image capture device 16. As field of view 26 decreases, points P move further rearward of the vehicle from a distance Q1, to a distance Q2, as best illustrated in FIG. 4. In order to increase distance Q to eliminate redundant and duplicative information displayed on display 20 for most driving conditions of vehicle 10, field of view 26 is preferably less than 12 degrees. In the illustrated embodiment, field of view 26 is between 6 and 8 degrees. Alternatively, distance Q may be dynamically adjusted according to some parameter, such as the speed of vehicle 10. This would allow Q to be greater when the vehicle is travelling at a faster speed, where vehicle separation tends to be larger, and vice versa. Field of view 26 may be adjusted by utilizing a selective presentation of pixels of the captured image in the displayed image.


Referring to FIG. 3, image display device 20 displays a composite image 42 made up of a left image portion 44, a right image portion 46, and a center image portion 48. Each image portion 44-48 is reversed from the image as captured by the respective image capture device 14, 16 utilizing conventional techniques. These techniques include reading the image in reverse with the image capture device, writing the image in reverse to display device 20, or reversing the image in image processor 18. Left image portion 44 is joined with central image portion 48 at a boundary 50. Central image portion 48 is joined with right image portion 46 at a boundary 52. As may best be seen in FIG. 3, the image portions at boundaries 50 and 52 are continuous whereby composite image 42 is a seamless panoramic view rearwardly of the vehicle. As also is apparent from FIG. 3, central image portion 48 is narrower than either left image portion 44 or right image portion 46. This is a result of reducing the horizontal field of view 26 of center image capture device 16 sufficiently to move points P, and thus overlap zones 32 and 34, a sufficient distance behind vehicle 10 to reduce redundant and duplicative images between image portions 44-48. Composite image 42 provides a clear image, which avoids confusion and simplifies the task of extracting information from the multiple image portions 44-48. As also may be seen by reference to FIG. 3, display 20 may additionally include indicia such as the readout of a compass 54, vehicle speed 56, turn signals 58, and the like as well as other graphical or video displays, such as a navigation display, a map display, and a forward-facing vision system. In this manner, rearview vision system 12 may be a compass vision system or an information vision system.


In the embodiment of rearview vision system 12 having a dynamically adjusted value of distance Q, the spacing between boundaries 50 and 52 will dynamically adjust in sequence with the adjustment of distance Q. Thus, as overlap zones 32, 34 move further away from the vehicle; for example, in response to an increase in speed of the vehicle, boundary lines 50 and 52 will move closer together and vice versa. In this manner, composite image 42 is dynamic, having image portions of dynamically adaptive sizes.


Display 20 is of a size to be as natural as possible to the driver. This is a function of the size of the displayed image and the distance between the display and the driver. Preferably, the displayed image simulates an image reflected by a rearview mirror. As such, the size of the displayed image is approximately the combined areas of the three rearview mirrors (one interior mirror and two exterior mirrors) conventionally used with vehicles. As best seen by reference to FIGS. 2 and 12, display 20 is preferably positioned within the driver's physiological field of view forward of the vehicle, generally illustrated at 70, through the windshield 72 without significantly obstructing the forward field of view. It is known that the driver's field of view, with the head and eyes fixed forward, extends further in a downward direction than in an upward direction. Display 20 could be located above the field of view 70 wherein the display may be observed at the upward portion of the driver's field of view. However, the position for the display illustrated in FIGS. 2 and 12 is preferred wherein the display is within the lower portion of the driver's field of view.


Display 20 may be a direct view flat panel display, such as a back-lit or reflective liquid crystal display, a plasma display, a field emission display, a cathode ray tube electroluminescent, light-emitting diode or deformable mirror display. The display may be mounted/attached to the dashboard, facia or header, or to the windshield at a position conventionally occupied by an interior rearview mirror. However, the synthesized image could be displayed using other display techniques such as to provide a projected or virtual image. Alternatively, a virtual image may be displayed on an opaque display adjacent the forward field of view. Alternatively, a virtual image may be displayed on a view-through heads-up display in which the image is superimposed on the forward field of view.


In the embodiment illustrated in FIGS. 12-17, display 20 displays an image at a focal length that is forward of the passenger compartment of vehicle 10. Preferably, the image displayed by display 20 is at a focal length that is within the driver's normal depth of field when viewing a distant object. Display 20 includes an image generator 74, which produces an image captured by one or more image capture devices 14, 16, and an optical correction system 76, which increases the focal distance of the image generated by image generator 74. In the illustrated embodiment, optic correction system 76 increases the focal distance by collimating the rays, generally indicated at X, from diverging rays to generally parallel rays projected from the display. Optical correction system 76 additionally magnifies the image. In the illustrated embodiment, the magnification is a factor of two. In this manner, optical corrective system 76 has the advantage of extending the focal distance of the image generated by image generator 74 and enlarging the image by the factor of magnification thereof. This advantageously allows each image generator 74 to project all or a portion of an image captured by one of image capture devices 14, 16, or a combination of portions of images from one or more image capture devices, by tiling the images or image portions. This is accomplished because the images projected from optical correction system 76 may abut even though the corresponding image generators 74 do not. This provides a convenient technique for joining the images synthesized from the image capture devices into a unitary image which represents a panoramic view rearward of the vehicle.


In the embodiment illustrated in FIG. 14, display 20 is an opaque projection display which projects the image directly toward the driver. In the embodiment illustrated in FIG. 15, a display device 120 is a view-through heads-up display in which the rays X are projected generally upwardly by image generator 74 and optical correction system 76 which are generally vertically aligned, or aligned forward of vertical. The rays X are reflected off a first surface of windshield 72 in the direction of the driver. Windshield 72 acts as a combiner which combines the image displayed by display 120 with a portion of the forward field of view 70 observed by the driver. In the embodiment illustrated in FIG. 15, a combiner other than the windshield may be used. Examples may be holographic or diffractive optical film elements or beam splitters of metal or dielectric thin films. Furthermore, image processor 18 may generate a line in the shape of a polygon, such as a rectangle, around rear image 42. This provides a border around the image to differentiate the rear view from the view forward of the vehicle.


In the embodiment illustrated in FIG. 16, display 20A is oriented at an angle with respect to the forward field of view of driver D. Image rays X are reflected by a mirror 140 toward the driver. Display 20A is an opaque display, with mirror 140 blocking the driver's view of near field objects, such as wipers 98 and the like. Display 20A has the advantage of being capable of location within a forward portion of the dashboard. Additionally, the only portion of the display visible to the driver is mirror 140. This allows near field portions of the display to be significantly reduced.


Because display 20 has a relatively long focus distance, display 20 defines an observation cone, generally designated 78, within which the displayed image can be observed. Therefore, the head of the driver must be properly oriented with respect to observation cone 78 in order to allow the driver to observe the displayed image. However, drivers come in various sizes. Therefore, a driver may be too tall or too short to have his or her head properly positioned within observation cone 78. In order to provide for various size drivers along with various adjustments in seating positions and the like, an accommodation means, generally illustrated at 80, may be provided in order to accommodate variations in the relationship between a driver's head and observation cone 78. In the illustrated embodiment, accommodation means 80 includes adjustment means 82 for adjusting the position of observation cone 78. The adjustment means may adjust the position of the observation cone either vertically, horizontally, or both vertically and horizontally. A vertical adjustment means 82 is illustrated in FIG. 13 in which the adjustment means includes an electric actuator 84 which is joined by linkage 86 with a portion of a housing 88 of display 20. Actuator 84 is electrically interconnected through a reversing switch 90 with a driver-operated actuator 92 which may be positioned on dashboard 94 or other convenient position accessible to the driver. Housing 88 may be adjustably mounted, such as by a pivot 96, in order to allow housing 88 to be adjustably repositioned with respect to dashboard 94. In this manner, by operation of actuator 92, housing 88 may be pivoted upwardly or downwardly with respect to pivot 96 in order to adjust the direction of observation cone 78. In this manner, the location of observation cone 78 may be adjusted in order to coincide with the location of the driver's head. In a similar fashion, the position of observation cone 78 may be adjusted laterally, if desired. If a view-through heads-up display of the type illustrated in FIG. 15 is utilized, the position of the observation cone may be adjusted vertically and laterally, in a similar manner, by mechanical or optical adjustments of display 20.


Accommodation means 80 may include extending the rearward field of view displayed by display 20 laterally outwardly with respect to the bezel 89 beyond that normally observed by a driver, In this manner, a driver's head located generally centrally within observation cone 78 will observe a view generally rearwardly of the vehicle. As the driver's head is moved laterally within observation cone 78, the driver will observe images more laterally to the side of the vehicle as would occur if the driver's head were to be moved with respect to a conventional optical rearview mirror system.


Vehicle 10 may include one or more near field view objects adjacent forward field of view 70. One such object is a windshield wiper 98 of the vehicle. Other such objects may include the top of dashboard 94, the frame around windshield 72, the hoodline, and the like. The housing of display 20 in FIG. 14 and mirror 140 in FIG. 15 are positioned with respect to forward field of view 70 such that housing 88 or mirror 140 covers any near field of view objects in the portion of the forward field of view adjacent display 20, 20A. In this manner, the gaze of the driver can switch between forward field of view 70 and the image displayed on display 20, without the eyes of the driver focusing on any significant near field objects. This is based upon a discovery that, even though the eyes of the driver are switching between the long focal distance of the forward field of view and the long focal distance of the image displayed by display 20, the eyes of the operator will unconsciously momentarily focus on any near field object positioned between the long focal distance views. Therefore, by blocking the driver's gaze of any near field objects, the eyes of the driver will be less stimulated to refocus during the transition from field of view 70 to display 20 and back again.


Image processor 18, which supplies a video signal 100 to image generator 74, may have a second input 102 which modulates the intensity level of the image generated by image generator 74 and displayed by display 20 (FIG. 14). The illumination level of the display is set in response to an ambient light input 104 which is an indication of the ambient light level around vehicle 10. Image processor 18 responds to the value of ambient light input 104 by producing a luminance intensity signal 102 which increases the intensity of the display in response to increases in ambient light level and decreases the intensity of the display in response to decreases in ambient light level. However, the level of display luminance may be limited to vary between upper and lower limits such that, once ambient light reaches a particular upper level, further increases in ambient light level will not result in a further increase in display intensity. Likewise, once the ambient light level decreases below a particular value, further reductions in ambient light level will not result in further reduction in display intensity. Ambient light input 104 may be produced by a separate ambient light sensor of the type which produces a continuously variable output in response to variations in ambient light levels, in which case, the intensity of display 20 may be proportionately adjusted. Alternatively, ambient light input 104 may be produced by a vehicle headlight control system (not shown) which switches the vehicle headlights on, or to a nighttime condition, in response to decreases in ambient light levels and switches the vehicle headlights off, or to a daytime running light condition, in response to increasing ambient light levels. Such system is disclosed in commonly assigned U.S. patent application Ser. No. 08/277,674 filed on Jul. 19, 1994, by Kenneth L. Schierbeek and Niall R. Lynam for an AUTOMATIC REARVIEW MIRROR SYSTEM WITH AUTOMATIC HEADLIGHT ACTIVATION, the disclosure of which is hereby incorporated herein by reference. If the ambient light signal supplied to ambient light input 104 is a binary signal representative of a daytime ambient light level and a nighttime ambient light level, image processor 18 would typically provide a signal on luminance intensity line 102 that would switch the intensity level of display 20 between two intensity levels. Alternatively, ambient light input 104 may be supplied with a signal developed by one or more image capture devices 14, 16. The ambient light signal would be based upon an average intensity value sensed by all, or a group of, pixels in the image capture device or devices. This embodiment eliminates the necessity for a separate ambient light sensor. Alternatively, ambient light input 104 may be responsive to manual actuation of the vehicle's headlights by the driver. Additionally, a comfort level setting may be provided to allow the driver to adjust to a preferred brightness at one ambient light condition. Thereafter, the system automatically adjusts display brightness according to ambient light changes.


In the illustrated embodiment, display 20 incorporates a combined image generator and optical correction system 106 which provides for both image magnification and light ray collimation. In this manner, the image projected from display 20 is larger than the image generated by image generator 74 and has a focal length that is greater than the separation distance between the image generator and the driver and, preferably, is generally at infinity (FIG. 17). Combined image generator and an optical correction system 106 is disclosed in detail in U.S. Pat. No. 5,050,966 for an OPTICAL COMBINER COLLIMATING APPARATUS; U.S. Pat. No. 4,859,031 for an OPTICAL COLLIMATING APPARATUS; U.S. Pat. No. 4,900,133 for a HEADS-UP DISPLAY COMBINER UTILIZING A CHOLESTERIC LIQUID CRYSTAL ELEMENT; U.S. Pat. No. 4,987,410 for a MULTIPLE IMAGE FORMING APPARATUS; and U.S. Pat. No. 5,408,346 for an OPTICAL COLLIMATING DEVICE EMPLOYING CHOLESTERIC LIQUID CRYSTAL AND NON-TRANSMISSIVE REFLECTOR, the disclosures of which are hereby incorporated herein by reference and will not be repeated. Suffice it to say, combined image generator and optical correction system 106 includes a light source 108 which generates broad band white light which is gathered and reflected by a parabolic reflector 110. In the illustrative embodiment, light source 108 is a tungsten halogen incandescent lamp. The light rays then pass through a dielectric green filter 112 which passes light in a specific region of the green portion of the spectrum and through a hot mirror 114 which removes the infrared content of the spectrum. Light then passes through a holographic diffuser 116 which homogenizes and shapes the light pattern. The light rays then pass through a monochrome liquid crystal display with opposing linear polarizers 118 which is supplied with a video signal by image processor 18. Items 108-118 make up image generator 74, which, in the illustrative embodiment, is a transmissive backlit liquid crystal display. However, image generator 74 could additionally be an emissive display or a reflective display, all of which are well known in the art.


Light rays of the image generated by image generator 74 next pass through an anti-reflective coated cover glass 120 which is joined with a left-hand circular polarizer 122 which is bonded to this cover glass. The opposite surface of circular polarizer 122 is bonded to a lens 124 having a 50/50 dielectric coating. Such dielectric coating allows light rays to be both transmitted through the lens and reflected by the lens. The left-hand polarized light X′ transmitted through lens 124 contacts a cholesteric liquid crystal layer (CLC) 126 which is left-hand polarized, which is what gives efficient reflection of left-hand polarized light X′, as illustrated at X″. Fifty percent (50%) of light rays X″ get efficiently reflected by the 50/50 beam splitter on lens 124 as right-hand circular polarized light X′″. Right-hand polarized light X′″ is transmitted by CLC layer 126 and passes through a right-hand circular polarizer 128 and an anti-reflective coated cover glass 130.


As can be seen by reference to FIG. 17, the optical configuration of lens 124 in combination with the left-hand and right-hand circular polarizers 122, 128 and cholesteric liquid crystal layer (CLC) 126, provide image magnification as well as collimate the image light in order to produce a very long focal distance image. Advantageously, this structure allows image portions from multiple image capture devices to be tiled into a unitary image. FIG. 17 illustrates an approach using a single image generator. Merging of multiple image portions would require additional combined image generator and optical correction systems. Although image generators 74 for each of the image portions are laterally spaced apart from each other, the amplification produced by combined image generator and optical correction system 106 causes the image portions to merge at their periphery. FIG. 17 illustrates an approach using a single image generator. Merging of multiple image portions would require additional combined image generators and optical correction systems. Other optical elements such as prisms, or other lenses, may be necessary to merge images to form a unitary image. Although the invention is illustrated with a combined image generator and optical correction system using cholesteric liquid crystal optical processing, other optical correction systems, as are known in the art, may be used. What is required is that the optical system generally collimates the light generated by the image generator and, preferably, provides amplification to the generated image.


In the illustrated embodiment, rear image 42, synthesized from the output of image capture devices 14, 16, has a lateral width verses vertical height aspect ratio that is between approximately 4:1 and 2:1. Most preferably, the aspect ratio of image 42 is 8:3. This allows a panoramic view rearwardly of the vehicle with an optimum information content while reducing display of irrelevant information. The aspect ratio of display 20 may be different from that of the displayed synthesized image 42. The remaining portion of the display, either above or below image 42, may be utilized to display images other than synthesized image 42. For example, the remaining portion of the display can be used to display auxiliary information such as one or more vehicle-operating parameters, such as vehicle speed indicia 56, heading indicia 54, or turn signal indicia 58. Alternatively, the remaining portion of the display can be a reconfigurable high-information content display area to selectively display various types of information. Such information may include incoming facsimile or pager information, phone numbers, and navigational aids including pull-up maps, route guidance information, global positioning system (GPS) data, intelligent vehicle highway system (IVHS) information, as well as radio and environmental system control settings, and the like. Display 20 is especially useful for displaying such alternative data. Because display 20 has a very long focal length, the driver may consult the alternative data by switching the gaze of the driver between forward field of view 70 and to display 20 which does not require extensive refocusing of the driver's eyes. This allows the driver to consult the alternative data quickly with reduced fatigue and distraction. The content of the auxiliary information displayed may be user-selectable by a keypad, trackball, or other input device on the dashboard, steering column, or other position readily accessible to the driver.


Although various camera devices may be utilized for image capture devices 14, 16, an electro-optic, pixilated imaging array, located in the focal plane of an optical system, is preferred. Such imaging array allows the number of pixels to be selected to meet the requirements of rearview vision system 12. The pixel requirements are related to the imaging aspect ratio of the respective image capture devices, which, in turn, are a function of the ratio of the vertical-to-horizontal field of view of the devices, as is well known in the art. In the illustrated embodiment, the imaging aspect ratio of side image capture devices 14 is 2:1 and the image aspect ratio of central image capture device 16 is variable down to 0.1:1. Such aspect ratio will produce images which will not typically match that of commercially available displays. A commercially available display may be used, however, by leaving a horizontal band of the display for displaying alpha-numeric data, such as portions of an instrument cluster, compass display, or the like, as illustrated in FIG. 3.


In the illustrated embodiment, image capture devices 14, 16 are CMOS imaging arrays of the type manufactured by VLSI Vision Ltd. of Edinburgh, Scotland, which are described in more detail in co-pending U.S. patent application Ser. No. 08/023,918 filed Feb. 26, 1993, by Kenneth Schofield and Mark Larson for an AUTOMATIC REARVIEW MIRROR SYSTEM USING A PHOTOSENSOR ARRAY, now U.S. Pat. No. 5,550,677, the disclosure of which is hereby incorporated herein by reference. However, other pixilated focal plane image-array devices, which are sensitive to visible or invisible electromagnetic radiation, could be used. The devices could be sensitive to either color or monochromatic visible radiation or near or far infrared radiation of the type used in night-vision systems. Each image capture device could be a combination of different types of devices, such as one sensitive to visible radiation combined with one sensitive to infrared radiation. Examples of other devices known in the art include charge couple devices and the like.


Preferably, image capture devices 14 and 16 are all mounted at the same vertical height on vehicle 10, although compromise may be required in order to accommodate styling features of the vehicle. The horizontal aim of image capture devices 14 and 16 is preferably horizontal. However, the portion of the image displayed is preferably biased toward the downward portion of the captured image because significantly less useful information is obtained above the horizontal position of the image capture devices.


Each image-capturing device 14, 16 is controlled by appropriate supporting electronics (not shown) located in the vicinity of the imaging array such that, when operating power is supplied, either an analog or a digital data stream is generated on an output signal line supplied to image processor 18. The support electronics may be provided partially on the image chip and partially on associated electronic devices. For each exposure period, a value indicative of the quantity of light incident on each pixel of the imaging array during the exposure period is sequentially outputted in a predetermined sequence, typically row-by-row. The sequence may conform to video signal standards which support a direct view such that, when a scene is viewed by an image-capturing device, the image presented on a display represents directly the scene viewed by the image-capturing devices. However, when looking forward and observing a displayed image of a rearward scene, the driver will interpret the image as if it were a reflection of the scene as viewed through a mirror. Objects to the left and rearward of the vehicle, as viewed by the rearward-looking camera, are presented on the left-hand side of the display and vice versa. If this reversal is effected in image processor 18, it may be by the use of a data storage device, or buffer, capable of storing all of the pixel values from one exposure period. The data is read out of the data storage device in a reversed row sequence. Alternatively, the imaging array electronics could be constructed to provide the above-described reversal at the image-capturing device or at the display.


Data transmission between image capture devices 14, 16 and image processor 18 and/or between image processor 18 and display 20 may be by electrically conductive leads. The leads may comprise either a serial or parallel bus. Alternatively, the data transmission may be via plastic or glass fiber-optic cable or an RF link. It is possible, for particular applications, to eliminate image processor 18 and direct drive display 20 from image capture devices 14, 16 at the pixel level. This may be accomplished by providing an interface between the output of image capture device 14, 16 and display 20 which synchronously maps image pixels captured by the image capture device, or devices, to the display. This synchronous mapping may be accomplished by providing a one-to-one mapping in which each pixel measurement is communicated to the display. Alternatively, the interface may only transmit pixel data which represents changes in the captured image. This allows for a reduction in the communication bandwidth necessary to transmit data between the image capture device, or devices, and the display. This may be accomplished by encoding the pixel data which represents changes in the captured image with additional data which designates the position of the pixel or other relevant information. Communication between the image capture device, or devices, may be multiplexed.


The data streams from image-capturing devices 14, 16 are combined in image processor 18 and directly mapped to the pixel array of display 20. This process is repeated preferably at a rate of at least 30 times per second in order to present an essentially real time video image. The image captured by side image capture device 14 on the right side of the vehicle is presented in right image portion 46 and the image from side image capture device 14 on the left side of the vehicle is displayed on left image portion 44. The image from center image capture device 16 is displayed on central image portion 48. The three image portions 44-48 are presented in horizontal alignment and adjacent to each other. However, the composite image may be positioned at any desired vertical position in the display 20. It is also possible to display image portions 44-48 on separate image devices which are adjacent each other.


In vision system 12, side image capture devices 14 are positioned preferably at a forward longitudinal position on vehicle 10 and center image capture device 16 is positioned at a rearward longitudinal position on the vehicle. As best seen by reference to FIG. 7, this positioning creates a difference in the vertical angle between each side image capture device 14 and center image capture device 16 with respect to a fixed location P1 that is a distance D1 behind the vehicle. This difference in sensing angle will cause each side image capture device 14 to image an object located at P1 on a horizontal row of pixels that is different from the horizontal row of pixels that center image capture device 16 will image the same object. If the image is below the horizontal centerline of the image capture device, it will be imaged on a lower row of pixels by center image capture device 16 than the row of pixels it will be imaged by the side image capture devices 14, as illustrated in FIG. 7. This mismatch between horizontal pixel rows of the captured image is furthermore a function of the distance of the captured image from the rear of the vehicle. This can be understood by reference to FIG. 11 which presents a chart 90 having a first column 92 of pixel lines n1, measured from the array centerline, at which an object will be imaged by side image capture device 14 and a second column 94 of pixel lines n2, measured from the array vertical centerline, at which the same object will be imaged by center image capture device 16. The result is that an object, which is captured by both side and center image capture devices 14, 16, will be vertically disjointed at the boundary of the displayed image, if the object is captured by more than one image capture device. The amount of disjointment will be greater closer to the vehicle and less at further distances. If the object is elongated in the horizontal direction, such as earth's horizon, bridges, or cross-markings on highways, then the object will appear to be either broken or crooked.


In order to provide uniform display of laterally elongated images, a rearview vision system 12′ is provided having a central image portion 48′ which is processed differently from the image display portions 44′ and 46′ produced by the side image capture devices (FIG. 8). Central image portion 48′ is reduced vertically, or compressed, by removing specified scan lines, or pixel rows, from the image captured by center image capture device 16 in a graduated fashion. The difference in the pixel line at which an object will be imaged by each of the side and center image capture devices is a function of the distance D of the object from the rear of the vehicle, with a greater variation occurring at shorter distances and the variation reducing to zero for infinite distances. Therefore, the compression of central image portion 48′ is non-linear, with substantially no compression at the vertical center of the image and greater compression at greater distances above and below the vertical center point of the image. This is accomplished by removing specific lines from the center display in a graduated fashion with a greater number of lines removed further from the vertical center of the image. The removed lines may be merely discarded in order to vertically reduce the image. Alternatively, the data contained in the removed lines may be utilized to modify the value of adjacent pixels above and below the removed line in order to enhance the quality of the compressed image. Averaging, median filtering, or other such known techniques may also be used.


Each of right image portion 46′ and left image portion 44′ includes an upper portion 64 which extends above the compressed upper portion of the central image portion 48′. In the illustrated embodiment, upper portions 64 are deleted in order to present a uniform upper horizontal boundary for display 20′. In the illustrated embodiment, the mismatch between the lower horizontal boundary of central image portion 48′ and each of the left and right image portions provides a dead space 66 which provides a visual prompt to the user of the approximate location of the rearward corners S of vehicle 10. This dead space 66 in the image displayed on display 20′ approximates the footprint occupied by vehicle 10 when viewed from point C. This is particularly useful because it provides a visual indication to the driver that a vehicle passing vehicle 10, as viewed in either left image portion 44′ or right image portion 46′, is at least partially adjacent vehicle 10 if the image of the approaching vehicle is partially adjacent to dead space 66.


In an alternative embodiment, the vertical compression technique may be applied to only a lower vertical portion of central image portion 48′. In most driving situations, objects imaged by rearward-facing image capture devices above the horizon are at a long distance from the vehicle while those below the horizon get progressively closer to the vehicle in relation to the distance below the horizon in the displayed image. Therefore, compression of the upper vertical portion of the central image portion may be eliminated without significant reduction in performance.


Compression of the central image portion may also advantageously be provided horizontally, as well as vertically. Spatial separation of center image capture device 16 from side image capture devices 14 causes similar distortion, as that described above, in the horizontal direction. This effect is spherical in nature and would require a more complex corrective action, such as compressing the image based upon the removal of pixels from an approximation to concentric circles centered on the center of the imaging array, or other techniques which would be apparent to those skilled in the art.


A rearview vision system 12″ includes an image display 20″ having a compressed central image portion 48″ and left and right image portions 44″ and 46″, respectively (FIG. 10). A border 50′ between left side image 44″ and central image 48″ includes a vertical central border portion 50a′, an upper border portion 50b′, and a lower border portion 50c′. Upper border portion 50b′ and lower border portion 50c′ diverge laterally outwardly, vertically away from central portion 50a′. A border 52′ between central image portion 48″ and right image portion 46″ includes a central boundary portion 52a′, an upper boundary portion 52b′, and a lower boundary portion 52c′. Upper boundary portion 52b′ and lower boundary portion 52c′ diverge laterally outwardly vertically away from central portion 52a′. This creates an upper portion of central image portion 48″ and a lower portion of central image portion 48″ which extend beyond the center portion thereof. This configuration is based upon the realization that the surface of the road immediately behind the vehicle is captured by central image capture device 16. Likewise, the horizontal plane above the vehicle, which is symmetrical with the road surface, is captured by the center image capture device. This may be seen by referring to point P in FIG. 10, which illustrate the points where the effective radius 68 of the virtual image capture device intersects dead zones 30 and by referring to point S in FIG. 10 which illustrates the corners or the rear of the vehicle (S).


The image displayed on display 20″ includes a dead space 66′ having diverging lateral sides 68a, 68b. Diverging sides 68a and 68b are configured in order to extend in the direction of travel of vehicle 10 which is parallel to lane markings of a highway on which vehicle 10 is travelling. This further enhances the visual perception of the driver by providing a visual clue of the location of images appearing on display 20″ with respect to the vehicle 10. Side portions 68a, 68b, in the illustrated embodiment, are natural extensions of lower boundary portions 50c′ and 52c′ and extend from point S on each respective side of the vehicle to point R, which represents the intersection of the lower extent of the vertical field of view 40 of each side image capture device 14 with the pavement (FIG. 7).


Rearview vision systems 12′ and 12″ utilize a displayed synthesized image which takes into account the use of perspective in enhancing the driver's understanding of what is occurring in the area surrounding the vehicle. The images produced on displays 20′ and 20″ effectively remove the vehicle bodywork and replace the bodywork with a vehicle footprint as would be viewed by virtual camera C. The image displayed on display 20″ further includes perspective lines which further enhance the roll of perspective in the driver's understanding of what is occurring.


In order to further enhance the driver's understanding of what is occurring in the area surrounding the vehicle, a rearview vision system 12′ includes a display 20′ having image enhancements (FIG. 6). In the illustrative embodiment, such image enhancements include graphic overlays 70a, 70b which are hash marks intended to illustrate to the driver the anticipated path of movement of vehicle 10. In the illustrated embodiment, the anticipated vehicle motion is a function of the vehicle direction of travel as well as the rate of turn of the vehicle. The forward or rearward direction of vehicle travel is determined in response to the operator placing the gear selection device (not shown) in the reverse gear position. The degree of turn of the vehicle may be determined by monitoring the movement of the vehicle steering system, monitoring the output of an electronic compass, or monitoring the vehicle differential drive system. In the embodiment illustrated in FIG. 6, the configuration of graphic overlays 70a, 70b indicates that the vehicle is in reverse gear and that the wheels are turned in a manner that will cause the vehicle to travel toward the driver's side of the vehicle. If the wheels were turned in the opposite direction, graphic overlays 70a, 70b would curve clockwise toward the right as viewed in FIG. 6. If the vehicle's wheels were straight, graphic overlays 70a, 70b would be substantially straight converging lines. If the vehicle is not in reverse gear position, graphic overlays 70a, 70b are not presented. Other types of graphic overlays of the displayed image are comprehended by the invention.


Horizontal grid markings on the display may be provided to indicate distances behind the vehicle at particular markings. Such grid would allow the driver to judge the relative position of vehicles behind the equipped vehicle. In one embodiment, short horizontal lines are superimposed on the displayed image at regular rearward intervals in horizontal positions which correspond to the boundaries of the lane in which the vehicle is travelling. In order to avoid confusion when the vehicle is travelling in a curved path, from a lack of correspondence between the graphic overlay and the road, a signal indicative of the vehicle's rate of turn may be taken into account when generating the graphic overlay. In this manner, the distance indications may be moved laterally, with reduced horizontal separation, to correspond to the positions of the curved lane boundaries and vertically on the image to compensate for the difference between distances along a straight and curved path.


Another image enhancement is to alter the appearance of an object in a particular zone surrounding the vehicle in order to provide an indication, such as a warning, to the driver. As an example, a vehicle that is too close to the equipped vehicle for safe-lane change, may be displayed in a particular color, such as red, may flash, or otherwise be distinguishable from other images on the display. Preferably, the speed of the equipped vehicle 10, which may be obtained from known speed transducers, may be provided as an input to the rearview vision system in order to cause such warning to be a function of the vehicle speed which, in turn, affects the safe separation distance of vehicles. The operation of the turn signal may also be used to activate such highlighting of other road users or to modify the scope of the image displayed. In order to determine the distance of objects behind vehicle 10, a separate distance-measuring system may be used. Such separate system may include radar, ultrasonic sensing, infrared detection, and other known distance-measuring systems. Alternatively, stereoscopic distance-sensing capabilities of side image capture devices 14 may be utilized to determine the separation distance from trailing objects utilizing known techniques.


Thus, it is seen that the image displayed on display 20-20′″ may be different under different circumstances. Such different circumstances may relate to the vehicle's direction of travel, speed, rate of turn, separation from adjacent objects, and the like.


Various other forms of image processing may be utilized with rearview vision system 12-12′″. Luminant and chrominant blending may be applied to the images captured by image capture devices 14, 16 in order to produce equality of the image data whereby the image portions appear as if they were produced by one image capture device. The dynamic range of the image capture devices may be extended in order to provide high quality images under all lighting conditions. Furthermore, individual pixel groups may be controlled in order to selectively compensate for bright or dark spots. For example, anti-blooming techniques may be applied for bright spots. Multiple exposure techniques may be applied to highlight dark areas. Image morphing and warping compensation techniques may additionally be applied. Resolution of the image capture devices and display may be selected in order to provide sufficient image quality for the particular application.


A heater may be applied to each image capture device in order to remove dew and frost that may collect on the optics of the device. Although, in the illustrative embodiment, the optical centerline of the camera coincides with the field of view, particular applications may result in the centerline of the camera pointing in a direction other than the centerline of the field of view. Although, in the illustrative embodiment, the image capture devices are fixed, it may be desirable to provide selective adjustability to the image capture devices or optical paths in particular applications. This is particularly desirable when the system is used on articulated vehicles where automated and coordinated camera aim may be utilized to maintain completeness of the synthesized image.


When operating the vehicle in the reverse direction, it may be desirable to provide additional data concerning the area surrounding the immediate rear of the vehicle. This may be accomplished by utilizing non-symmetrical optics for the center image capture device in order to provide a wide angle view at a lower portion of the field of view. Alternatively, a wide angle optical system could be utilized with the electronic system selectively correcting distortion of the captured image. Such system would provide a distortion-free image while obtaining more data, particularly in the area surrounding the back of the vehicle.


The invention additionally comprehends the use of more than three image capture devices. In addition to side image capture devices positioned at the front sides of the vehicle and a center image capture device positioned at the center rear of the vehicle, additional image capture devices may be useful at the rear corners of the vehicle in order to further eliminate blind spots. It may additionally be desirable to provide an additional center image capture device at a higher elevation in order to obtain data immediately behind the vehicle and thereby fill in the road surface detail immediately behind the vehicle. Such additional detail is particularly useful when operating the vehicle in the reverse direction. Of course, each of the image capture devices could be a combination of two or more image capture devices.


Although the present invention is illustrated as used in a rearview vision system, it may find utility in other applications. For example, the invention may be useful for providing security surveillance in an area where a building or other object obstructs the view of the area under surveillance. Additionally, the invention may find application in night-vision systems and the like. For example, the invention may be applied to forward-facing night-vision systems, or other vision enhancement systems such as may be used in adverse weather or atmospheric conditions such as fog, applied to provide an enhanced display of a synthesized image, which approximates a forward-facing view from a single virtual camera located rearwardly of the driver, taking advantage of the perspective features of the image.


A rearview vision system 150 is provided which, in addition to displaying a rear image on display 20 which is synthesized by image processor 18 from the output of image capture devices 14, 16, also supplies drive signals to an electrically operated optical device such as electro-optic mirror 152, an electro-optic window 154, or both. Even though a panoramic view rearward of the vehicle is displayed on display 20, it may be desired to provide the driver with a rearview mirror of the type which has conventionally been provided on vehicles. One such mirror is an electro-optic mirror, such as an electrochromic mirror, a liquid crystal mirror, or a solenoid-operated prismatic mirror and the like. Additionally, vehicles may be provided with electro-optic windows, such as sunroofs, rear windows, side windows, and the like, which change transmissivity in response to a drive signal to a partial light transmittance level. In U.S. patent application Ser. No. 08/023,918 filed Feb. 26, 1993, by Kenneth Schofield and Mark Larson for an AUTOMATIC REARVIEW MIRROR SYSTEM USING A PHOTOSENSOR ARRAY, now U.S. Pat. No. 5,550,677, the disclosure of which is hereby incorporated herein by reference, a technique is disclosed for producing a drive signal for an electrically operated optical device, such as an electro-optic mirror or window, from the image captured by a rearward-facing array. Utilizing the techniques disclosed therein, image processor 18 produces a drive signal on line 156 in order to control the partial reflectance level of electro-optic mirror 152 and a drive signal on line 158 in order to control the partial light transmittance level of electro-optic window 154.


A rearview vision system 160 is provided which includes a near infrared illumination device 162 in order to enhance an image captured by image capture devices 14, 16 (FIG. 19). In the illustrated embodiment, infrared illumination device 162 illuminates an area immediately behind the vehicle. Preferably, the output of illumination device 162 has a greater near infrared light output than visible light output. This allows an enhanced image to be captured by the image capture device without increasing the visible light perceived by drivers surrounding the vehicle. Infrared illumination device 162 may be actuated in response to the vehicle being placed in reverse gear. This provides backup illumination which is greatly enhanced without having an increased effect on other drivers. Alternatively, infrared illumination devices may be positioned, for example, at other locations on the side or even the front of a vehicle in order to enhance the image captured by the image capture device or devices. This is especially useful in order to utilize rearview vision system 160 with a large truck, such as a trailer truck. This infrared illumination device may flood the area around the trailer with infrared light in order to enhance the image captured by the image capture device, or devices, without distracting other drivers.


Image capture device 14, 16 may include a housing 164 in which an antenna 166 is positioned. This provides a convenient and functional location for a receiving antenna, such as the type used with a global positioning system, cellular telephone, garage door opener, radar distance sensing device, and the like, as disclosed in patent application Ser. No. 08/569,851 filed by Desmond J. O'Farrell, Roger L. Veldman and Kenneth Schofield for a VEHICLE GLOBAL POSITIONING SYSTEM, now U.S. Pat. No. 5,971,552, the disclosure of which is hereby incorporated herein by reference. A heater 168 may be associated with the image capture device in order to stabilize the temperature of the device in low ambient temperature conditions. A similar heater may be supplied in display 20 in order to improve its performance in low ambient temperature conditions. A heater control 170 is provided in order to control the energization of heater 168 and, if utilized, the heater in the display. Heater control 170, preferably, energizes heater 168 prior to the vehicle being started. This allows the temperature of the image capture device to be elevated to a more desirable temperature prior to the driver operating the vehicle. This may be accomplished by heater control 170 being a proximity detector which detects a device carried by the driver as the driver approaches the vehicle. Alternatively, heater control 170 may be responsive to a signal produced by a remote keyless entry device concurrently with the doors being activated. Alternatively, heater control 170 may be responsive to the vehicle device being opened.


A rearview vision system 172 is provided which provides an output, generally referred to at 174, from image processor 18 to display 20. Output 174 provides an indication when an object bears a predetermined relationship to the vehicle. Such object may be of interest because the object is in a blind spot of the vehicle, may be tailgating the vehicle, or may be an object in front of the vehicle which the vehicle is too close to. Display 20 may respond to output 174 by highlighting the displayed vehicle, such as by displaying the vehicle in an artificial color, such as red, by flashing the image of the vehicle, or, otherwise, drawing the attention of the driver to the vehicle. Output 174 may be developed by image processor 18 from the outputs of image capture devices 14, 16. This may be accomplished by, for example, utilizing redundant image portions captured by the image capture devices, even though not displayed by display 20, in order to calculate relative position of the object with respect to the vehicle. Alternatively, an object sensor 176 may be provided in order to supply an output 178 indicative of a predetermined positional relationship of an object sensed by the object sensor with respect to the vehicle. In the illustrated embodiment, object sensor 176 may be a passive infrared sensor which senses the presence of an object in the vehicle's blind spot. Alternatively, object sensor 176 may be a distance-measuring device, such as an active infrared sensor, an ultrasonic sensor, a radar sensor, or the like. Such object sensor is especially useful in determining the separation distance between the vehicle and objects in front of the vehicle. Preferably, object sensor 176 has a sensing field of view that is substantially coincident with the field of view of one or more of the image capture devices 14, 16.


A rearview vision system 178 is provided which has the capability of displaying stereoscopic images rearward of the vehicle. Rearview vision system 178 includes at least one pair of image capture devices 14a, which are closely positioned on the vehicle and have overlapping fields of view. Because the image capture device pairs are closely positioned, they capture substantially the same image but from a slightly different angle. This allows image processor 18 to produce a video signal 100′ with stereoscopic information. This signal is utilized by a stereoscopic display 320 in order to produce a stereoscopic image rearward of the vehicle. Such stereoscopic displays are known in the art. Although one pair of image capture devices are illustrated in FIG. 22, rearview vision system 178 may include multiple pairs of image capture devices. This allows a rear image to be synthesized from the multiple pairs of image capture devices in order produce a panoramic view rearward of the vehicle and stereoscopic image. Preferably, utilizing other aspects of the invention, the stereoscopic image is a rearward-facing view from a single location.


A rearview vision system 180 is provided which produces an indication 182 of road line markings. Indication 182 may also indicate road edges. Image processor 18 detects the road line markings and the road edges from the images captured by image capture devices 14, 16. This feature may be further enhanced by combining it with an infrared illumination device 162 in order to further illuminate areas behind and around the vehicle in order to enhance the image of the road line markings and the road edges. Indication 182 may be utilized by display 20 in order to provide an indication of the vehicle with respect to the road line markings and road edges. The indication may further be utilized by an indicator which indicates the relative position of the vehicle in its lane. Additionally, the indication may be utilized to determine erotic vehicle operation, such as may occur when the driver begins to fall asleep, in order to provide a suitable alarm, or the like.


A rearview vision system 184 is provided with capabilities for infrared communication with other vehicles and stationary beacons. Rearview vision system 184 produces a communication data output 186 which includes communication data decoded from infrared signals detected by image capture device or devices 14, 16. For example, suitable standards may be developed wherein vehicles are equipped with a pair of spaced apart infrared transmitters on a forward portion thereof. Image processor 18 may respond to the temporal and spatial patterns of infrared signals detected by image capture devices 14, 16 in order to determine the speed and distance and, thereby, the separation of the vehicles as well as the rate of change of separation of the vehicles. Such information may be communicated to the trailing vehicle by an infrared transmitter (not shown) in order to control the speed of the trailing vehicle. This feature provides adaptive cruise control in which the speed of the trailing vehicle is controlled according to separation distance with leading vehicles. This allows high-speed convoying between vehicles. The communication system may additionally provide for the identification of emergency vehicles, and the like, which may transmit a unique temporal and/or spatial pattern of an infrared signal. The IR communication signal may additionally be utilized to receive signals from stationary devices, such as location beacons and Intelligent Vehicle Highway System (NHS) data. Because rearview vision system 184 has a field of view which extends generally rearwardly of the vehicle, the system provides the capability for sensing information after the vehicle has passed the beacon. This provides an adjunct to infrared communication systems having a field of view generally forward of, or to the side of, the vehicle.


A rearview vision system 188 is provided having extended dynamic range (FIG. 25). Rearview vision system 188 includes a pair of image capture devices 14 and/or 16, each of which has an image-sensing array 190. Each image capture device is capable of operating in either a color mode, in which a color image is displayed on display 20, or a monochrome mode, in which a monochrome image is displayed on display 20. System 188 includes an image luminance sensor 192 which senses the luminance level of images captured by image capture devices 14, 16. Image luminance sensor 192 may switch the image capture devices between the color mode and the monochrome mode such that, when the image luminance is sufficiently high, the image capture device, or devices, operate in a color mode. During low image luminance conditions, the image capture device, or devices, are operated in a monochromatic mode which does not require as much image luminance. This extends the dynamic range of the system. Rearview vision system 188 may additionally include an exposure control 194 which determines the exposure period for capturing each frame by arrays 190. In order to extend the dynamic range of system 188, exposure control 194 may produce exposure intervals for arrays 190 which vary in length from interval-to-interval. Thus, a series of normal exposure intervals may be occasionally supplanted by a longer exposure interval during which greater detail of the image may be captured. This enhanced image may then be combined with the image captured during the shorter intervals into a merged image of enhanced detail.


Rearview vision system 188 may additionally include a plurality of infrared shutters 196 which are in the optical paths 198 of arrays 190. Each infrared shutter 196 has at least one state in which infrared energy is generally not attenuated to array 190. In another state, the infrared shutter generally blocks infrared radiation from the array. The state of infrared shutters 196 is controlled by image luminance sensor 192. During periods of high image luminance, sensor 192 may switch the infrared shutters 196 to a state blocking near infrared radiation from arrays 190. However, during low image luminance conditions, sensor 198 may switch the infrared shutters 196 to a state in which the near infrared energy is transmitted to arrays 190. The addition of the near infrared radiation at low luminance levels enhances the image luminance sensed by arrays 190. In the illustrated embodiment, infrared shutters 196 are either electrochromic shutters or liquid crystal shutters, both of which are known in the art. Rearview vision system 188 additionally includes means for coordinating the image intensity received from multiple image capture devices and displayed on display 20. This allows a balance composite image to be displayed on the display. This may be provided by a display intensity control 200 which regulates the intensity of the output of both arrays 190 in order to produce color and luminance balancing between the multiple image capture devices. Instead of a separate display intensity control, a direct communication channel may be developed between the image capture devices in order to provide color and luminance balancing.


Each image pixel captured by image capture devices 14, 16 and displayed on display 20 has a resolution which affects the amount of image detail displayed. While it is desirable to have a high degree of detail of displayed image, the increase in image resolution produces a commensurate increase in system cost. While it is desirable to reduce system cost, this should not be achieved by sacrificing necessary detail in the displayed image. According to the invention, it has been determined that sufficient image detail may be obtained at a suitable system cost by having a pixel resolution in the range of between approximately 2 arc minutes and approximately 8 arc minutes. Preferably, system resolution is approximately 6 arc minutes.


It is known in the art to provide imaging array capture devices having mosaic filters which mask image radiation in order to produce pixels which respond respectively to red, green, and blue light. Because such known pixel filter masks do not adequately absorb near infrared radiation, it is known to supply infrared filters in order to block infrared radiation from the pixels so that the pixels respond to only the designed radiation band. However, such additional filters have undesirable characteristics including costs. It has been discovered that a pixel filter mask may be made responsive to red, green, or blue light while filtering out near infrared by adding appropriate dyes to the dyes making up the filter mask.


The heater supplied with each image capture device may include a transparent conductive coating applied to a window covering the device lens. Alternative heater constructions include ITO or a series of fine wire mesh. This provides protection of the lens of the image capture device from physical harm while allowing moisture and frost to be removed from the window.


Various manipulation techniques may be applied to image capture devices 14, 16. For example, when the invention is applied to semi-tractor/trailer combinations, the image capture devices may be motorized and responsive to signals representing the relative position of the trailer and the cab in order to produce full panoramic view during various maneuvers of the vehicle. For example, as the vehicle turns, which may cause the trailer to otherwise block the image captured by a side image capture device, the image capture device may pivot to a different panoramic view which is not blocked by the trailer. Additionally, the panoramic view sensed by the image capture device may be different when the trailer is attached than when the trailer is not attached. In a similar fashion, rearward-mounted image capture device 16 may be motorized to move vertically upwardly and downwardly between a first position when the vehicle is moving in a forward direction and a second position when the vehicle is in reverse gear. In the second position, the motorized camera is directed more downwardly in order to capture images closer to the rear of the vehicle which may be contacted by the vehicle. Image capture devices 14, 16 may be supplied with electrical opto-mechanical zoom devices as well as vibration isolation and compensation stabilizing devices.


Image processor 18 may provide an input to a control for the intensity of the brake lights, turn signals, and the like for the vehicle. In this manner, the image processor may control the intensity of such lights based upon ambient lighting conditions. This allows the intensity of the rearward directed lights to be less distracting for following vehicles.


Additionally, the present invention may be utilized for providing guidance for a tractor/trailer or like vehicle backing to a loading dock. Additionally, the invention may provide guidance for proper alignment of the trailer and tractor which are being joined by backing of the tractor. The present invention may additionally provide accident monitoring by capturing a predetermined continuous stream of images, such as, for example, 15 seconds. This running store of image may be frozen upon impact of the vehicle, in order to record events leading up to the impact, or may be stopped by a manual input. Furthermore, the invention may be utilized to alert the driver to an impending rear-end collision. The vehicle may respond to such indication by deploying an appropriate device, such as a smart headrest or the like.


Thus, it is seen that the present invention enhances the relationship between the driver's primary view and the image presented on the rearview vision system. This is accomplished in a manner which provides ease of interpretation while avoiding confusion so that the driver does not have to concentrate or look closely at the image. In this manner, information presented on the display is naturally assimilated. This is accomplished while reducing blind spots so that other vehicles or objects of interest to the driver will likely be displayed to the driver. Additionally, the use of perspective allows distances to be more accurately determined.


Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.

Claims
  • 1. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor has a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels thain rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller operable, responsive to image processing of a portion of captured image data representative of a portion of the field of view of said image sensor and responsive to at least one vehicle input, to generate at least one vehicle automatic exterior light control signal and an exterior light status indicator signal as a function of said image processing of the portion of captured image data and of said at least one vehicle input; andwherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
  • 2. The vehicle automatic vehicle exterior light control of claim 1 wherein said at least one vehicle input is a speed input.
  • 3. The vehicle automatic vehicle exterior light control of claim 2 wherein said at least one vehicle input is a compass sensor.
  • 4. The vehicle automatic vehicle exterior light control of claim 2 wherein said at least one vehicle input is an auto on/off switch.
  • 5. The vehicle automatic vehicle exterior light control of claim 1 wherein said at least one vehicle input is a light sensor.
  • 6. The vehicle automatic vehicle exterior light control of claim 1 wherein said controller further comprises at least one output.
  • 7. The vehicle automatic vehicle exterior light control of claim 1 wherein said controller comprises at least in part said image processor.
  • 8. The vehicle automatic vehicle exterior light control of claim 1 wherein said controller controls a display in response to an ambient light input supplied with a signal generated by said image sensor.
  • 9. The vehicle automatic vehicle exterior light control of claim 1 wherein said vehicle automatic vehicle exterior light control detects ambient light exterior to the vehicle and generates said exterior light status indicator signal to indicate the status of the detected ambient light.
  • 10. The vehicle automatic vehicle exterior light control of claim 1 wherein said vehicle automatic vehicle exterior light control comprises at least one of (a) a headlight control system responsive to ambient light detected by sensors and (b) a rearview image display control system responsive to ambient light at the vehicle.
  • 11. The vehicle automatic vehicle exterior light control of claim 1 wherein said exterior light status indicator signal is via said image sensor.
  • 12. The vehicle automatic vehicle exterior light control of claim 1 wherein said controller receives a first image input from a first image sensor and a second image input from a second image sensor, and wherein said controller generates a composite image synthesized from at least said first image input and said second image input, and wherein a display system displays said composite image on a single video screen located in a cabin of the equipped vehicle and viewable by a driver of the equipped vehicle when the driver is normally operating the equipped vehicle.
  • 13. The vehicle automatic vehicle exterior light control of claim 12 wherein a first field of view of said first image sensor and a second field of view of said second image sensor are directed generally rearwardly with respect to the direction of travel of the equipped vehicle, and wherein said first field of view has a field of view portion at least partially overlapping a field of view portion of said second field of view.
  • 14. The vehicle automatic vehicle exterior light control of claim 13 wherein generation of said composite image from at least said first and second image inputs for display on said single video screen comprises dynamic adaptation of said first and second image inputs in order to minimize said at least partially overlapping field of view portion.
  • 15. The vehicle automatic vehicle exterior light control of claim 12 wherein said single video screen has a display luminance that is variable responsive to a sensing of an ambient light level.
  • 16. The vehicle automatic vehicle exterior light control of claim 15 wherein the display luminance of said single video screen of said display system is variable responsive to at least one of (i) a vehicle headlight activation control, (ii) an ambient light sensor, and (iii) an indication of ambient light level developed by at least one of said first and second image sensors.
  • 17. The vehicle automatic vehicle exterior light control of claim 12 wherein said single video screen is operable to display at least one of pager information, a telephone number listing, a global positioning system output, a map, route guidance information, intelligent vehicle highway system information, vehicle radio control settings, vehicle environmental system settings, vehicle speed, vehicle heading, and turn signal indicators.
  • 18. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor has a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels than rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller comprising at least one output, wherein said controller is operable, responsive to image processing of a portion of captured image data representative of a portion of the field of view of said image sensor, to control said at least one output and to generate an exterior light status indicator signal as a function of said image processing of the portion of captured image data; andwherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
  • 19. The vehicle automatic vehicle exterior light control of claim 18 wherein said controller further comprises at least one vehicle input.
  • 20. The vehicle automatic vehicle exterior light control of claim 19 wherein said at least one input is a speed input.
  • 21. The vehicle automatic vehicle exterior light control of claim 19 wherein said at least one input is a manual user input switch.
  • 22. The vehicle automatic vehicle exterior light control of claim 18 wherein said at least one output is configured to control at least one exterior light.
  • 23. The vehicle automatic vehicle exterior light control of claim 22 wherein said at least one output is configured to control at least one electrochromic element.
  • 24. The vehicle automatic vehicle exterior light control of claim 18 wherein said at least one output is configured to control at least one headlamp.
  • 25. The vehicle automatic vehicle exterior light control of claim 18 wherein said controller comprises at least in part said image processor.
  • 26. The vehicle automatic vehicle exterior light control of claim 18 wherein said controller controls a display in response to an ambient light input supplied with a signal generated by said image sensor.
  • 27. The vehicle automatic vehicle exterior light control of claim 26 wherein said signal is based on an average intensity value sensed by at least some of said pixels of said image sensor.
  • 28. The vehicle automatic vehicle exterior light control of claim 18 wherein said vehicle automatic vehicle exterior light control detects ambient light exterior to the vehicle and generates said exterior light status indicator signal to indicate the status of the detected ambient light.
  • 29. The vehicle automatic vehicle exterior light control of claim 18 wherein said vehicle automatic vehicle exterior light control comprises at least one of (a) a headlight control system responsive to ambient light detected by sensors and (b) a rearview image display control system responsive to ambient light at the vehicle.
  • 30. The vehicle automatic vehicle exterior light control of claim 18 wherein said vehicle automatic vehicle exterior light control comprises (a) a headlight control system responsive to ambient light detected by sensors and (b) a rearview image display control system responsive to ambient light at the vehicle.
  • 31. The vehicle automatic vehicle exterior light control of claim 18 wherein said exterior light status indicator signal is via said image sensor.
  • 32. The vehicle automatic vehicle exterior light control of claim 18 wherein said vehicle automatic vehicle exterior light control is operable to (a) generate an automatic exterior light control signal responsive to light sensors and (b) generate said exterior light status indicator signal via said image sensor.
  • 33. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor had a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels than rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller comprising at least one output, wherein said output and an exterior light status indicator signal are functions of image processing of at least a portion of captured image data and wherein said controller is further configured to receive at least one vehicle input; andwherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
  • 34. The vehicle automatic vehicle exterior light control of claim 33 wherein said at least one input is a speed input.
  • 35. The vehicle automatic vehicle exterior light control of claim 33 wherein said at least one input is a compass sensor.
  • 36. The vehicle automatic vehicle exterior light control of claim 33 wherein said at least one output is configured to control at least one electrochromic element.
  • 37. The vehicle automatic vehicle exterior light control of claim 33 wherein said at least one output is configured to control at least one exterior light.
  • 38. The vehicle automatic vehicle exterior light control of claim 33 wherein said controller comprises at least in part said image processor.
  • 39. The vehicle automatic vehicle exterior light control of claim 33 wherein said controller controls a display in response to an ambient light input supplied with a signal generated by said image sensor.
  • 40. The vehicle automatic vehicle exterior light control of claim 33 wherein said vehicle automatic vehicle exterior light control detects ambient light exterior to the vehicle and generates said exterior light status indicator signal to indicate the status of the detected ambient light.
  • 41. The vehicle automatic vehicle exterior light control of claim 33 wherein said vehicle automatic vehicle exterior light control comprises at least one of (a) a headlight control system responsive to ambient light detected by sensors and (b) a rearview image display control system responsive to ambient light at the vehicle.
  • 42. The vehicle automatic vehicle exterior light control of claim 33 wherein said exterior light status indicator signal is via said image sensor.
  • 43. The vehicle automatic vehicle exterior light control of claim 33 wherein said vehicle automatic vehicle exterior light control is operable to (a) generate an automatic exterior light control signal responsive to light sensors and (b) generate said exterior light status indicator signal via said image sensor.
  • 44. The vehicle automatic vehicle exterior light control of claim 33 wherein said vehicle automatic vehicle exterior light control detects ambient light exterior to the vehicle and generates said exterior light status indicator signal to indicate the status of the detected ambient light.
  • 45. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor has a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels than rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller operable, responsive to image processing of a portion of captured image data representative of a portion of the field of view of said image sensor and responsive to at least one vehicle input, to generate at least one vehicle automatic exterior light control signal and an exterior light staus indicator signal as function of said image processing of the portion of captured image data and of said at least one vehicle input;wherein said controller comprises at least in part said image processor;wherein said at least one vehicle input comprises at least one of (a) a speed input and (b) an auto on/off switch;wherein said controller further comprises at least one output:wherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
  • 46. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor has a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels than rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller comprising at least one output, wherein said controller is operable, responsive to image processing of portion of captured image data representative of a potion of the field view of said image sensor, to control said at least one output and to generate an exterior light status indicator signal as a function of said image processing of the portion of captured image data;wherein said controller comprises at least in part said image processor;wherein said controller further comprises at least one vehicle input, and wherein said at least one vehicle input comprises at least one of (a) a speed input and (b)a manual user input switch;wherein said at least one output is configured to control at least one headlamp, andwherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
  • 47. A vehicle automatic vehicle exterior light control, said vehicle automatic vehicle exterior light control comprising: an image sensor disposed at a vehicle and having a two-dimensional array of columns and rows of photosensing pixels, wherein said image sensor has a field of view exterior the vehicle and wherein said array of said image sensor comprises more columns of photosensing pixels than rows of photosensing pixels;an image processor operable to process image data captured by said image sensor;a controller comprising at least one output, wherein said controller is responsive to image processing of at least a portion of captured image data representative of a portion of the field of view of said image sensor, and wherein said output and an exterior light status indicator signal are functions of said image processing of the portion of captured image data, wherein said controller is further configured to receive at least one vehicle input;wherein said controller comprises at least in part said image processor;wherein said at least one input is a speed input;wherein said at least one output is configured to control at least one exterior light; andwherein said controller, responsive to image processing of captured image data, is operable to detect road markings on a road along which the vehicle is traveling.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 12/496,357, filed Jul.1, 2009, now U.S. Pat No. 8,462,204, which is a continuation of U.S. patent application Ser. No. 11/122,880, filed May 5, 2005, now U.S. Pat. No. 7,561,181, which is a continuation of U.S. patent application Ser. No. 10/324,679, filed on Dec. 20, 2002, now U.S. Pat. No. 6,891,563, which is a continuation of U.S. patent application Ser. No. 08/952,026, filed under 35 U.S.C.§371, filed Nov. 19, 1997, now U.S. Pat. No. 6,498,620.

US Referenced Citations (1598)
Number Name Date Kind
1472509 Bitter Oct 1923 A
2074251 Braun Jan 1936 A
2166303 Hodny et al. Jul 1939 A
2339291 Paulus et al. Apr 1940 A
2317400 Paulus et al. Aug 1940 A
2331144 Sitter Dec 1940 A
2424288 Severy Mar 1944 A
2414223 DeVirgilis Jan 1947 A
2598420 Onksen et al. May 1952 A
2750583 McCullough Jan 1953 A
2632040 Rabinow Mar 1953 A
2827594 Rabinow Mar 1953 A
3008532 Reed Jan 1958 A
2855523 Berger Oct 1958 A
3011580 Reid Oct 1958 A
2959709 Vanaman et al. Oct 1960 A
3085646 Paufve Apr 1963 A
3141393 Platt Jul 1964 A
3158835 Hipkins Nov 1964 A
3172496 Rabinow et al. Mar 1965 A
3185020 Thelen May 1965 A
3201750 Morin Aug 1965 A
3208070 Boicey Sep 1965 A
3249761 Baumanns May 1966 A
3271577 Griner et al. Sep 1966 A
3280701 Donnelly at al. Oct 1966 A
3325680 Amacher Jun 1967 A
3367616 Bausch et al. Feb 1968 A
3432225 Rock Mar 1969 A
3451741 Manos Jun 1969 A
3453038 Kissa et al. Jul 1969 A
3486066 Hicks et al. Dec 1969 A
3521941 Deb et al. Jul 1970 A
3543018 Barcus et al. Nov 1970 A
3557265 Chisholm et al. Jan 1971 A
3565985 Schrenk et al. Feb 1971 A
3623671 Hargroves Nov 1971 A
3628851 Robertson Dec 1971 A
3673560 Barsh et al. Jun 1972 A
3676668 Collins et al. Jul 1972 A
3689695 Rosenfield et al. Sep 1972 A
3708668 Tilley Jan 1973 A
3711176 Alfrey, Jr. et al. Jan 1973 A
3751711 Schick Aug 1973 A
3781090 Sumita Dec 1973 A
3806229 Schoot et al. Apr 1974 A
3807832 Castellion Apr 1974 A
3821590 Kosman et al. Jun 1974 A
3845572 McCanney Nov 1974 A
3860847 Carley Jan 1975 A
3876287 Sprokel Apr 1975 A
3876940 Wickord et al. Apr 1975 A
3985424 Steinacher Oct 1976 A
4003445 De Bruine Jan 1977 A
4035681 Savage Jul 1977 A
4037134 Loper Jul 1977 A
4044853 Melke Aug 1977 A
4049961 Marcy Sep 1977 A
4052712 Ohama et al. Oct 1977 A
4058796 Oishi et al. Nov 1977 A
4075468 Marcus Feb 1978 A
4088400 Assouline et al. May 1978 A
4093364 Miller Jun 1978 A
4109235 Bouthors Aug 1978 A
4127778 Leitz Nov 1978 A
4139234 Morgan Feb 1979 A
4139801 Linares Feb 1979 A
4143264 Gilbert et al. Mar 1979 A
4174152 Giglia et al. Nov 1979 A
4176728 Otteblad et al. Dec 1979 A
4200361 Malvano Apr 1980 A
4209853 Hyatt Jun 1980 A
4211955 Ray Jul 1980 A
4214266 Myers Jul 1980 A
4218698 Bart et al. Aug 1980 A
4221955 Joslyn Sep 1980 A
4228490 Thillays Oct 1980 A
4236099 Rosenblum Nov 1980 A
4238778 Ohsumi Dec 1980 A
4247870 Gabel et al. Jan 1981 A
4249160 Chilvers Feb 1981 A
4254931 Aikens Mar 1981 A
4257703 Goodrich Mar 1981 A
4266856 Wainwright May 1981 A
4277804 Robison Jul 1981 A
4278142 Kono Jul 1981 A
4281898 Ochiai et al. Aug 1981 A
4281899 Oskam Aug 1981 A
4288814 Talley et al. Sep 1981 A
RE30835 Giglia Dec 1981 E
4306768 Egging Dec 1981 A
4310851 Pierrat Jan 1982 A
4331382 Graff May 1982 A
4338000 Kamimori et al. Jul 1982 A
4348652 Barnes et al. Sep 1982 A
4348653 Tsuzuki et al. Sep 1982 A
4355271 Noack Oct 1982 A
4357558 Massoni et al. Nov 1982 A
4357594 Ehrlich et al. Nov 1982 A
4376909 Tagami et al. Mar 1983 A
4377613 Gordon Mar 1983 A
4381888 Momiyama May 1983 A
4389537 Tsunoda et al. Jun 1983 A
4389639 Torii et al. Jun 1983 A
4398805 Cole Aug 1983 A
4401181 Schwarz Aug 1983 A
4403208 Hodgson et al. Sep 1983 A
4419386 Gordon Dec 1983 A
4420238 Felix Dec 1983 A
4431896 Lodetti Feb 1984 A
4435042 Wood et al. Mar 1984 A
4435048 Kamimori et al. Mar 1984 A
4436371 Wood et al. Mar 1984 A
4441125 Parkinson Apr 1984 A
4443057 Bauer et al. Apr 1984 A
4446171 Thomas May 1984 A
4460831 Oettinger et al. Jul 1984 A
4464789 Sternberg Aug 1984 A
4465339 Baucke et al. Aug 1984 A
4473695 Wrighton et al. Sep 1984 A
4481450 Watanabe et al. Nov 1984 A
4485402 Searby Nov 1984 A
4490227 Bitter Dec 1984 A
4491390 Tong-Shen Jan 1985 A
4495589 Hirzel Jan 1985 A
4499451 Suzuki et al. Feb 1985 A
4512637 Ballmer Apr 1985 A
4521804 Bendell Jun 1985 A
4524941 Wood et al. Jun 1985 A
4529275 Ballmer Jul 1985 A
4529873 Ballmer Jul 1985 A
4532550 Bendell Jul 1985 A
4538181 Taylor Aug 1985 A
4546551 Franks Oct 1985 A
4549208 Kamejima et al. Oct 1985 A
4561625 Weaver Dec 1985 A
4564833 Seko et al. Jan 1986 A
4571082 Downs Feb 1986 A
4572619 Reininger Feb 1986 A
4580196 Task Apr 1986 A
4580875 Bechtel et al. Apr 1986 A
4587522 Warren May 1986 A
4588041 Tsuchuhashi May 1986 A
4588267 Pastore May 1986 A
4599544 Martin Jul 1986 A
4600913 Caine Jul 1986 A
4603946 Kato Aug 1986 A
4614415 Hyatt Sep 1986 A
4620141 McCumber et al. Oct 1986 A
4623222 Itoh Nov 1986 A
4625329 Ishikawa et al. Nov 1986 A
4626850 Chey Dec 1986 A
4629941 Ellis Dec 1986 A
4630109 Barton Dec 1986 A
4630904 Pastore Dec 1986 A
4632509 Ohmi et al. Dec 1986 A
4638287 Umebayashi et al. Jan 1987 A
4645975 Meitzler et al. Feb 1987 A
4646210 Skogler et al. Feb 1987 A
4647161 Müller Mar 1987 A
4652090 Uchikawa et al. Mar 1987 A
4653316 Fukuhara Mar 1987 A
4665321 Chang et al. May 1987 A
4665430 Hiroyasu May 1987 A
4669825 Itoh Jun 1987 A
4671615 Fukada Jun 1987 A
4671619 Kamimori et al. Jun 1987 A
4672457 Hyatt Jun 1987 A
4676601 Itoh Jun 1987 A
4679077 Yuasa et al. Jul 1987 A
4681431 Sims et al. Jul 1987 A
4690508 Jacob Sep 1987 A
4692798 Seko et al. Sep 1987 A
4697883 Suzuki Oct 1987 A
4701022 Jacob Oct 1987 A
4702566 Tukude et al. Oct 1987 A
4712879 Lynam et al. Dec 1987 A
4713685 Nishimura et al. Dec 1987 A
4717830 Botts Jan 1988 A
4727290 Smith et al. Feb 1988 A
4728804 Norsworthy Mar 1988 A
4729076 Masami et al. Mar 1988 A
4731669 Hayashi et al. Mar 1988 A
4733335 Serizawa et al. Mar 1988 A
4733336 Skogler et al. Mar 1988 A
4740838 Mase et al. Apr 1988 A
4741603 Miyagi May 1988 A
4758883 Kawahara et al. Jul 1988 A
4761061 Nishiyama et al. Aug 1988 A
4768135 Kretschmer et al. Aug 1988 A
4772942 Tuck Sep 1988 A
4779095 Guerreri Oct 1988 A
4780752 Angerstein et al. Oct 1988 A
4781436 Armbuster Nov 1988 A
4785280 Fubini et al. Nov 1988 A
4789904 Peterson Dec 1988 A
4793690 Gahan et al. Dec 1988 A
4793695 Wada et al. Dec 1988 A
4799267 Kamejima et al. Jan 1989 A
4807096 Skogler et al. Feb 1989 A
4817948 Simonelli Apr 1989 A
4820933 Hong Apr 1989 A
4825232 Howdle Apr 1989 A
4833469 David May 1989 A
4838650 Stewart Jun 1989 A
4839749 Franklin Jun 1989 A
4843463 Michetti Jun 1989 A
4847489 Dietrich Jul 1989 A
4847772 Michalopoulos et al. Jul 1989 A
4849731 Melocik Jul 1989 A
4855161 Moser et al. Aug 1989 A
4855822 Narendra et al. Aug 1989 A
4859031 Berman et al. Aug 1989 A
4862037 Farber et al. Aug 1989 A
4863130 Marks, Jr. Sep 1989 A
4867561 Fujii et al. Sep 1989 A
4871917 O'Farrell et al. Oct 1989 A
4872051 Dye Oct 1989 A
4881019 Shiraishi et al. Nov 1989 A
4882466 Friel Nov 1989 A
4882565 Gallmeyer Nov 1989 A
4883349 Mittelhauser Nov 1989 A
4884055 Memmola Nov 1989 A
4886960 Molyneux et al. Dec 1989 A
4891559 Matsumoto et al. Jan 1990 A
4892345 Rachael, III Jan 1990 A
4895790 Swanson et al. Jan 1990 A
4896030 Miyaji Jan 1990 A
4900133 Berman Feb 1990 A
4902108 Byker Feb 1990 A
4905151 Weiman et al. Feb 1990 A
4906940 Green et al. Mar 1990 A
4907870 Brucker Mar 1990 A
4910591 Petrossian et al. Mar 1990 A
4916374 Schierbeek Apr 1990 A
4917477 Bechtel et al. Apr 1990 A
4926346 Yokoyama May 1990 A
4930742 Schofield et al. Jun 1990 A
4931937 Kakinami et al. Jun 1990 A
4935665 Murata Jun 1990 A
4936533 Adams et al. Jun 1990 A
4937796 Tendler Jun 1990 A
4949186 Peterson Aug 1990 A
4953305 Van Lente et al. Sep 1990 A
4954962 Evans, Jr. et al. Sep 1990 A
4956591 Schierbeek Sep 1990 A
4959247 Moser et al. Sep 1990 A
4959865 Stettiner et al. Sep 1990 A
4961625 Wood et al. Oct 1990 A
4967319 Seko Oct 1990 A
4970509 Kissinger, Sr. Nov 1990 A
4970653 Kenue Nov 1990 A
4971405 Hwang Nov 1990 A
4971430 Lynas Nov 1990 A
4973844 O'Farrell Nov 1990 A
4974078 Tsai Nov 1990 A
4975703 Delisle et al. Dec 1990 A
4985847 Shioya et al. Jan 1991 A
4987357 Masaki Jan 1991 A
4987410 Berman et al. Jan 1991 A
4991054 Walters Feb 1991 A
4996083 Moser et al. Feb 1991 A
5001558 Burley et al. Mar 1991 A
5003288 Wilhelm Mar 1991 A
5003339 Kikuchi et al. Mar 1991 A
5005213 Hanson et al. Apr 1991 A
5006971 Jenkins Apr 1991 A
5008946 Ando Apr 1991 A
5012082 Watanabe Apr 1991 A
5014167 Roberts May 1991 A
5016977 Baude et al. May 1991 A
5016996 Ueno May 1991 A
5020114 Fujioka et al. May 1991 A
5027001 Torbert Jun 1991 A
5027104 Reid Jun 1991 A
5027200 Petrossian et al. Jun 1991 A
5031101 Kamimura et al. Jul 1991 A
5036437 Macks Jul 1991 A
5038255 Nishihashi et al. Aug 1991 A
5044706 Chen Sep 1991 A
5044956 Behensky et al. Sep 1991 A
5050966 Berman Sep 1991 A
5051906 Evans, Jr. et al. Sep 1991 A
5055668 French Oct 1991 A
5056899 Warszawski Oct 1991 A
5058851 Lawlor et al. Oct 1991 A
5059877 Teder Oct 1991 A
5059947 Chen Oct 1991 A
5063603 Burt Nov 1991 A
5064274 Alten Nov 1991 A
5066108 McDonald Nov 1991 A
5066112 Lynam et al. Nov 1991 A
5072154 Chen Dec 1991 A
5073012 Lynam Dec 1991 A
5075768 Wirtz et al. Dec 1991 A
5076673 Lynam et al. Dec 1991 A
5076674 Lynam Dec 1991 A
5080207 Horneffer Jan 1992 A
5080309 Ivins Jan 1992 A
5081585 Kurami et al. Jan 1992 A
5086253 Lawler Feb 1992 A
5086510 Guenther et al. Feb 1992 A
5087969 Kamada et al. Feb 1992 A
5096287 Kakinami et al. Mar 1992 A
5097362 Lynas Mar 1992 A
5100093 Rawlinson Mar 1992 A
5100095 Haan et al. Mar 1992 A
5101351 Hattori Mar 1992 A
5111289 Lucas et al. May 1992 A
5113721 Polly May 1992 A
5115346 Lynam May 1992 A
5117346 Gard May 1992 A
5121200 Choi Jun 1992 A
5122957 Hattori Jun 1992 A
5124549 Michaels et al. Jun 1992 A
5128799 Byker Jul 1992 A
5130709 Toyama et al. Jul 1992 A
5133605 Nakamura Jul 1992 A
5135298 Feltman Aug 1992 A
5136483 Schöniger et al. Aug 1992 A
5139327 Tanaka Aug 1992 A
5140455 Varaprasad et al. Aug 1992 A
5142407 Varaprasad et al. Aug 1992 A
5144685 Nasar et al. Sep 1992 A
5148014 Lynam Sep 1992 A
5151816 Varaprasad et al. Sep 1992 A
5151824 O'Farrell Sep 1992 A
5155426 Kurami Oct 1992 A
5159557 Ogawa Oct 1992 A
5160200 Cheselske Nov 1992 A
5160201 Wrobel Nov 1992 A
5160780 Ono et al. Nov 1992 A
5160971 Koshizawa et al. Nov 1992 A
5161632 Asayama et al. Nov 1992 A
5163002 Kurami Nov 1992 A
5165108 Asayama Nov 1992 A
5166681 Bottesch et al. Nov 1992 A
5168355 Asayama et al. Dec 1992 A
5168378 Black Dec 1992 A
5170374 Shimohigashi et al. Dec 1992 A
5172235 Wilm et al. Dec 1992 A
5172317 Asanuma et al. Dec 1992 A
5173881 Sindle Dec 1992 A
5177462 Kajiwara Jan 1993 A
5177606 Koshizawa Jan 1993 A
5177685 Davis et al. Jan 1993 A
5178448 Adams et al. Jan 1993 A
5179471 Caskey et al. Jan 1993 A
5182502 Slotkowski et al. Jan 1993 A
5184956 Langlarais et al. Feb 1993 A
5185812 Yamashita et al. Feb 1993 A
5187383 Taccetta et al. Feb 1993 A
5189537 O'Farrell Feb 1993 A
5189561 Hong Feb 1993 A
5193000 Lipton et al. Mar 1993 A
5193029 Schofield et al. Mar 1993 A
5193894 Lietar et al. Mar 1993 A
5197562 Kakinami et al. Mar 1993 A
5204536 Vardi Apr 1993 A
5204778 Bechtel Apr 1993 A
5207492 Roberts May 1993 A
5208701 Maeda May 1993 A
5208750 Kurami et al. May 1993 A
5212468 Adell May 1993 A
5214408 Asayama May 1993 A
5217794 Schrenk Jun 1993 A
5218414 Kajiwara et al. Jun 1993 A
5220508 Ninomiya et al. Jun 1993 A
5223814 Suman Jun 1993 A
5223907 Asayama Jun 1993 A
5225827 Persson Jul 1993 A
5229941 Hattori Jul 1993 A
5230400 Kakinami et al. Jul 1993 A
5233461 Dornan et al. Aug 1993 A
5233527 Shinnosuke Aug 1993 A
5234070 Noah et al. Aug 1993 A
5235178 Hegyi Aug 1993 A
5239405 Varaprasad et al. Aug 1993 A
5239406 Lynam Aug 1993 A
5243417 Pollard Sep 1993 A
5243524 Ishida et al. Sep 1993 A
5245422 Borcherts et al. Sep 1993 A
5246193 Faidley Sep 1993 A
5249126 Hattori Sep 1993 A
5249128 Markandey et al. Sep 1993 A
5249157 Taylor et al. Sep 1993 A
5251680 Miezawa et al. Oct 1993 A
5252354 Cronin et al. Oct 1993 A
5253050 Karasundai Oct 1993 A
5253109 O'Farrell et al. Oct 1993 A
5255442 Schierbeek et al. Oct 1993 A
5265172 Markandey et al. Nov 1993 A
5266873 Arditi et al. Nov 1993 A
5267160 Ito et al. Nov 1993 A
5276389 Levers Jan 1994 A
5277986 Cronin et al. Jan 1994 A
5285060 Larson et al. Feb 1994 A
5289182 Brillard et al. Feb 1994 A
5289321 Secor Feb 1994 A
5291424 Asayama et al. Mar 1994 A
5293162 Bachalo Mar 1994 A
5296924 Blancard et al. Mar 1994 A
5298732 Chen Mar 1994 A
5301115 Nouso et al. Apr 1994 A
5302956 Asbury et al. Apr 1994 A
5303205 Gauthier et al. Apr 1994 A
5304980 Maekawa Apr 1994 A
5305012 Faris Apr 1994 A
5307136 Saneyoshi Apr 1994 A
5307419 Tsujino et al. Apr 1994 A
5309137 Kajiwara May 1994 A
5313072 Vachss May 1994 A
5318143 Parker et al. Jun 1994 A
5325096 Pakett Jun 1994 A
5325386 Jewell et al. Jun 1994 A
5327288 Wellington et al. Jul 1994 A
5329206 Slotkowski et al. Jul 1994 A
5330149 Haan et al. Jul 1994 A
5331312 Kudoh Jul 1994 A
5336980 Levers Aug 1994 A
5341437 Nakayama Aug 1994 A
5343206 Ansaldi et al. Aug 1994 A
5347456 Zhang et al. Sep 1994 A
5351044 Mathur et al. Sep 1994 A
D351370 Lawlor et al. Oct 1994 S
5355118 Fukuhara Oct 1994 A
5355284 Roberts Oct 1994 A
5359666 Nakayama et al. Oct 1994 A
5361190 Roberts et al. Nov 1994 A
5367457 Ishida Nov 1994 A
5369590 Karasudani et al. Nov 1994 A
5371535 Takizawa Dec 1994 A
5371659 Pastrick et al. Dec 1994 A
5373911 Yasui Dec 1994 A
5374852 Parkes Dec 1994 A
5379196 Kobayashi et al. Jan 1995 A
5379353 Hasegawa et al. Jan 1995 A
5386285 Asayama Jan 1995 A
5388048 Yavnayi et al. Feb 1995 A
5394333 Kao Feb 1995 A
5398041 Hyatt Mar 1995 A
5406395 Wilson et al. Apr 1995 A
5406414 O'Farrell et al. Apr 1995 A
5408346 Trissel et al. Apr 1995 A
5408357 Buekema Apr 1995 A
5410346 Saneyoshi et al. Apr 1995 A
5414257 Stanton May 1995 A
5414439 Groves et al. May 1995 A
5414461 Kishi et al. May 1995 A
5414625 Hattori May 1995 A
5416313 Larson et al. May 1995 A
5416318 Hegyi May 1995 A
5416478 Morinaga May 1995 A
5416711 Gran et al. May 1995 A
5418610 Fischer May 1995 A
5422756 Weber Jun 1995 A
5424865 Lynam Jun 1995 A
5424952 Asayama Jun 1995 A
5426294 Kobayashi et al. Jun 1995 A
5430431 Nelson Jul 1995 A
5430450 Holmes Jul 1995 A
5432496 Lin Jul 1995 A
5434407 Bauer et al. Jul 1995 A
5434927 Brady et al. Jul 1995 A
5436839 Dausch et al. Jul 1995 A
5440428 Hegg et al. Aug 1995 A
5444478 LeLong et al. Aug 1995 A
5446576 Lynam et al. Aug 1995 A
5448180 Kienzler et al. Sep 1995 A
5450057 Watanabe Sep 1995 A
5451822 Bechtel et al. Sep 1995 A
5457493 Leddy et al. Oct 1995 A
5461357 Yoshioka et al. Oct 1995 A
5461361 Moore Oct 1995 A
5465079 Bouchard et al. Nov 1995 A
5467284 Yoshioka et al. Nov 1995 A
5469298 Suman et al. Nov 1995 A
5471515 Fossum et al. Nov 1995 A
5473515 Liu Dec 1995 A
5475366 Van Lente et al. Dec 1995 A
5475494 Nishida et al. Dec 1995 A
5481409 Roberts Jan 1996 A
5482133 Iwata et al. Jan 1996 A
5483060 Sugiura et al. Jan 1996 A
5483168 Reid Jan 1996 A
5483453 Uemura et al. Jan 1996 A
5485378 Franke et al. Jan 1996 A
5487116 Nakano et al. Jan 1996 A
5487522 Hook Jan 1996 A
5488496 Pine Jan 1996 A
5493269 Durley et al. Feb 1996 A
5493392 Blackmon et al. Feb 1996 A
5497305 Pastrick et al. Mar 1996 A
5497306 Pastrick Mar 1996 A
5498866 Bendicks et al. Mar 1996 A
5500760 Varaprasad et al. Mar 1996 A
5500766 Stonecypher Mar 1996 A
5508592 Lapatovich et al. Apr 1996 A
5510983 Iino Apr 1996 A
5515448 Nishitani May 1996 A
5521633 Nakajima et al. May 1996 A
5521760 DeYoung et al. May 1996 A
5523811 Wada et al. Jun 1996 A
5525264 Cronin et al. Jun 1996 A
5528474 Roney et al. Jun 1996 A
5528698 Kamei et al. Jun 1996 A
5529138 Shaw et al. Jun 1996 A
5530240 Larson et al. Jun 1996 A
5530420 Tsuchiya et al. Jun 1996 A
5530421 Marshall et al. Jun 1996 A
5530771 Maekawa Jun 1996 A
5535056 Caskey et al. Jul 1996 A
5535144 Kise Jul 1996 A
5535314 Alves et al. Jul 1996 A
5537003 Bechtel et al. Jul 1996 A
5539397 Asanuma et al. Jul 1996 A
5541590 Nishio Jul 1996 A
5545960 Ishikawa Aug 1996 A
5550677 Schofield et al. Aug 1996 A
5555136 Waldmann et al. Sep 1996 A
5555312 Shima et al. Sep 1996 A
5555503 Kyrtsos Sep 1996 A
5555555 Sato et al. Sep 1996 A
5558123 Castel et al. Sep 1996 A
5559695 Daily Sep 1996 A
5566224 ul Azam et al. Oct 1996 A
5567360 Varaprasad et al. Oct 1996 A
5568027 Teder Oct 1996 A
5568316 Schrenck et al. Oct 1996 A
5570127 Schmidt Oct 1996 A
5572315 Krell Nov 1996 A
5572354 Desmond et al. Nov 1996 A
5574443 Hsieh Nov 1996 A
5576687 Blank et al. Nov 1996 A
5576854 Schmidt et al. Nov 1996 A
5576975 Sasaki et al. Nov 1996 A
5581464 Woll et al. Dec 1996 A
5582383 Mertens et al. Dec 1996 A
5587236 Agrawal et al. Dec 1996 A
5588123 Loibl Dec 1996 A
5594222 Caldwell Jan 1997 A
5598164 Reppas et al. Jan 1997 A
5602457 Anderson et al. Feb 1997 A
5602542 Widmann et al. Feb 1997 A
5608550 Epstein et al. Mar 1997 A
5610756 Lynam et al. Mar 1997 A
5611966 Varaprasad et al. Mar 1997 A
5612686 Takano et al. Mar 1997 A
5612883 Shaffer et al. Mar 1997 A
5614788 Mullins et al. Mar 1997 A
5614885 Van Lente et al. Mar 1997 A
5615023 Yang Mar 1997 A
5615857 Hook Apr 1997 A
5617085 Tsutsumi et al. Apr 1997 A
5619370 Guinosso Apr 1997 A
5626800 Williams et al. May 1997 A
5627586 Yamasaki May 1997 A
5631089 Center, Jr. et al. May 1997 A
5632092 Blank et al. May 1997 A
5632551 Roney et al. May 1997 A
5633944 Guibert et al. May 1997 A
5634709 Iwama Jun 1997 A
5638116 Shimoura et al. Jun 1997 A
5642238 Sala Jun 1997 A
5642299 Hardin et al. Jun 1997 A
5644851 Blank et al. Jul 1997 A
5646612 Byon Jul 1997 A
5646614 Abersfelder et al. Jul 1997 A
5648835 Uzawa Jul 1997 A
5649756 Adams et al. Jul 1997 A
5650765 Park Jul 1997 A
5650944 Kise Jul 1997 A
5660454 Mori et al. Aug 1997 A
5661303 Teder Aug 1997 A
5662375 Adams et al. Sep 1997 A
5666028 Bechtel et al. Sep 1997 A
5666157 Aviv Sep 1997 A
5668663 Varaprasad et al. Sep 1997 A
5669698 Veldman et al. Sep 1997 A
5669699 Pastrick et al. Sep 1997 A
5669704 Pastrick Sep 1997 A
5669705 Pastrick et al. Sep 1997 A
5670935 Schofield et al. Sep 1997 A
5671996 Bos et al. Sep 1997 A
5673019 Dantoni Sep 1997 A
5673994 Fant, Jr. et al. Oct 1997 A
5673999 Koenck Oct 1997 A
5675489 Pomerleau Oct 1997 A
5676484 Chamberlin et al. Oct 1997 A
5677851 Kingdon et al. Oct 1997 A
5680123 Lee Oct 1997 A
5680245 Lynam Oct 1997 A
5680263 Zimmermann et al. Oct 1997 A
5686975 Lipton Nov 1997 A
5689241 Clark, Sr. et al. Nov 1997 A
D388107 Huckins Dec 1997 S
5699044 Van Lente et al. Dec 1997 A
5699057 Ikeda et al. Dec 1997 A
5699149 Kuroda et al. Dec 1997 A
5706355 Raboisson et al. Jan 1998 A
5707129 Kobayashi Jan 1998 A
5708410 Blank et al. Jan 1998 A
5708857 Ishibashi Jan 1998 A
5710633 Klappenbach et al. Jan 1998 A
5715093 Schierbeek et al. Feb 1998 A
5719551 Flick Feb 1998 A
5724187 Varaprasad et al. Mar 1998 A
5724316 Brunts Mar 1998 A
5729194 Spears et al. Mar 1998 A
5737226 Olson et al. Apr 1998 A
5741966 Handfield et al. Apr 1998 A
5745050 Nakagawa Apr 1998 A
5745266 Smith Apr 1998 A
5748287 Takahashi et al. May 1998 A
5751211 Shirai et al. May 1998 A
5751390 Crawford et al. May 1998 A
5751489 Caskey et al. May 1998 A
5754099 Nishimura et al. May 1998 A
5757949 Kinoshita et al. May 1998 A
5760826 Nayer Jun 1998 A
5760828 Cortes Jun 1998 A
5760931 Saburi et al. Jun 1998 A
5760962 Schofield et al. Jun 1998 A
5761094 Olson et al. Jun 1998 A
5762823 Hikmet Jun 1998 A
5764139 Nojima et al. Jun 1998 A
5765116 Wilson-Jones et al. Jun 1998 A
5765940 Levy et al. Jun 1998 A
5767793 Agravante et al. Jun 1998 A
5775762 Vitito Jul 1998 A
5781437 Wiemer et al. Jul 1998 A
5786772 Schofield et al. Jul 1998 A
5788357 Muth et al. Aug 1998 A
5790403 Nakayama Aug 1998 A
5790973 Blaker et al. Aug 1998 A
5793308 Rosinski et al. Aug 1998 A
5793420 Schmidt Aug 1998 A
5796094 Schofield et al. Aug 1998 A
5796176 Kramer et al. Aug 1998 A
5798057 Hikmet Aug 1998 A
5798575 O'Farrell et al. Aug 1998 A
5798688 Schofield Aug 1998 A
5802727 Blank et al. Sep 1998 A
5803579 Turnbull et al. Sep 1998 A
5804719 Didelot et al. Sep 1998 A
5805367 Kanazawa Sep 1998 A
5806965 Deese Sep 1998 A
5808589 Fergason Sep 1998 A
5808713 Broer et al. Sep 1998 A
5808777 Lynam et al. Sep 1998 A
5808778 Bauer et al. Sep 1998 A
5811888 Hsieh Sep 1998 A
5812321 Schierbeek et al. Sep 1998 A
5813745 Fant, Jr. et al. Sep 1998 A
5818625 Bauer et al. Oct 1998 A
5820097 Spooner Oct 1998 A
5820245 Desmond et al. Oct 1998 A
5823654 Pastrick et al. Oct 1998 A
5825527 Forgette et al. Oct 1998 A
5835255 Miles Nov 1998 A
5835613 Breed et al. Nov 1998 A
5835614 Aoyama et al. Nov 1998 A
5837994 Stam et al. Nov 1998 A
5841126 Fossum et al. Nov 1998 A
5844505 Van Ryzin Dec 1998 A
5844682 Kiyomoto et al. Dec 1998 A
5845000 Breed et al. Dec 1998 A
5847755 Wixson et al. Dec 1998 A
5848802 Breed et al. Dec 1998 A
5850176 Kinoshita et al. Dec 1998 A
5850254 Takano et al. Dec 1998 A
5863116 Pastrick et al. Jan 1999 A
5867591 Onda Feb 1999 A
5871275 O'Farrell et al. Feb 1999 A
5877707 Kowalick Mar 1999 A
5877897 Schofield et al. Mar 1999 A
5878370 Olson Mar 1999 A
5879074 Pastrick Mar 1999 A
5883193 Karim Mar 1999 A
5883605 Knapp Mar 1999 A
5883739 Ashihara et al. Mar 1999 A
5884212 Lion Mar 1999 A
5888431 Tonar et al. Mar 1999 A
5890021 Onoda Mar 1999 A
5890083 Franke et al. Mar 1999 A
5896085 Mori et al. Apr 1999 A
5899551 Neijzen et al. May 1999 A
5899956 Chan May 1999 A
5904725 Iisaka et al. May 1999 A
5904729 Ruzicka May 1999 A
5910854 Varaprasad et al. Jun 1999 A
5912534 Benedict Jun 1999 A
5914815 Bos Jun 1999 A
5917664 O'Neill et al. Jun 1999 A
5923027 Stam et al. Jul 1999 A
5923457 Byker et al. Jul 1999 A
5924212 Domanski Jul 1999 A
5928572 Tonar et al. Jul 1999 A
5929784 Kawaziri et al. Jul 1999 A
5929786 Schofield et al. Jul 1999 A
5938320 Crandall Aug 1999 A
5938321 Bos et al. Aug 1999 A
5940011 Agravante et al. Aug 1999 A
5940120 Frankhouse et al. Aug 1999 A
5940201 Ash et al. Aug 1999 A
5942853 Piscart Aug 1999 A
5949331 Schofield et al. Sep 1999 A
5956079 Ridgley Sep 1999 A
5956181 Lin Sep 1999 A
5959367 O'Farrell et al. Sep 1999 A
5959555 Furuta Sep 1999 A
5963247 Banitt Oct 1999 A
5964822 Alland et al. Oct 1999 A
5965247 Jonza et al. Oct 1999 A
5971552 O'Farrell et al. Oct 1999 A
5973760 Dehmlow Oct 1999 A
5975715 Bauder Nov 1999 A
5982288 Sawatari et al. Nov 1999 A
5984482 Rumsey et al. Nov 1999 A
5986730 Hansen et al. Nov 1999 A
5986796 Miles Nov 1999 A
5990469 Bechtel et al. Nov 1999 A
5990649 Nagao et al. Nov 1999 A
5998617 Srinivasa et al. Dec 1999 A
5998929 Bechtel et al. Dec 1999 A
6000823 Desmond et al. Dec 1999 A
6001486 Varaprasad et al. Dec 1999 A
6002511 Varaprasad et al. Dec 1999 A
6002544 Yatsu Dec 1999 A
6007222 Thau Dec 1999 A
6008486 Stam et al. Dec 1999 A
6008871 Okumura Dec 1999 A
6009336 Harris et al. Dec 1999 A
6009359 El-Hakim et al. Dec 1999 A
6016035 Eberspacher et al. Jan 2000 A
6016215 Byker Jan 2000 A
6019411 Carter et al. Feb 2000 A
6019475 Lynam et al. Feb 2000 A
6020704 Buschur Feb 2000 A
6023229 Bugno et al. Feb 2000 A
6025872 Ozaki et al. Feb 2000 A
6028537 Suman et al. Feb 2000 A
6031484 Bullinger Feb 2000 A
6037689 Bingle et al. Mar 2000 A
6037860 Zander et al. Mar 2000 A
6037975 Aoyama Mar 2000 A
6042253 Fant, Jr. et al. Mar 2000 A
6045243 Muth et al. Apr 2000 A
6045643 Byker et al. Apr 2000 A
6046766 Sakata Apr 2000 A
6046837 Yamamoto Apr 2000 A
6049171 Stam et al. Apr 2000 A
6052124 Stein et al. Apr 2000 A
6057754 Kinoshita et al. May 2000 A
6060989 Gehlot May 2000 A
6061002 Weber et al. May 2000 A
6064508 Forgette et al. May 2000 A
6065840 Caskey et al. May 2000 A
6066933 Ponziana May 2000 A
6067111 Hahn et al. May 2000 A
6072391 Suzuki et al. Jun 2000 A
6074777 Reimers et al. Jun 2000 A
6078355 Zengel Jun 2000 A
6078865 Koyanagi Jun 2000 A
6082881 Hicks Jul 2000 A
6084519 Coulling et al. Jul 2000 A
6084700 Knapp et al. Jul 2000 A
6086131 Bingle et al. Jul 2000 A
6086229 Pastrick Jul 2000 A
6087012 Varaprasad et al. Jul 2000 A
6087953 DeLine et al. Jul 2000 A
6091833 Yasui et al. Jul 2000 A
6097023 Schofield et al. Aug 2000 A
6097024 Stam et al. Aug 2000 A
6099131 Fletcher et al. Aug 2000 A
6099155 Pastrick et al. Aug 2000 A
6102559 Nold et al. Aug 2000 A
6104552 Thau et al. Aug 2000 A
6106121 Buckley et al. Aug 2000 A
6107939 Sorden Aug 2000 A
6111498 Jobes et al. Aug 2000 A
6111683 Cammenga et al. Aug 2000 A
6111684 Forgette et al. Aug 2000 A
6111685 Tench et al. Aug 2000 A
6111696 Allen et al. Aug 2000 A
6115651 Cruz Sep 2000 A
6116743 Hoek Sep 2000 A
6122597 Saneyoshi et al. Sep 2000 A
6124647 Marcus et al. Sep 2000 A
6124886 DeLine et al. Sep 2000 A
6128576 Nishimoto et al. Oct 2000 A
6130421 Bechtel et al. Oct 2000 A
6130448 Bauer et al. Oct 2000 A
6132072 Turnbull et al. Oct 2000 A
6139171 Waldmann Oct 2000 A
6139172 Bos et al. Oct 2000 A
6140933 Bugno et al. Oct 2000 A
6144022 Tenenbaum et al. Nov 2000 A
6144158 Beam Nov 2000 A
6146003 Thau Nov 2000 A
6149287 Pastrick et al. Nov 2000 A
6150014 Chu et al. Nov 2000 A
6150930 Cooper Nov 2000 A
6151065 Steed et al. Nov 2000 A
6151539 Bergholz et al. Nov 2000 A
6152590 Furst et al. Nov 2000 A
6154149 Tychkowski et al. Nov 2000 A
6154306 Varaprasad et al. Nov 2000 A
6157294 Urai et al. Dec 2000 A
6158655 DeVries, Jr. et al. Dec 2000 A
6166625 Teowee et al. Dec 2000 A
6166628 Andreas Dec 2000 A
6166847 Tench et al. Dec 2000 A
6166848 Cammenga et al. Dec 2000 A
6167755 Damson et al. Jan 2001 B1
6169955 Fultz Jan 2001 B1
6170956 Rumsey et al. Jan 2001 B1
6172600 Kakinami et al. Jan 2001 B1
6172601 Wada et al. Jan 2001 B1
6172613 DeLine et al. Jan 2001 B1
6173501 Blank et al. Jan 2001 B1
6175164 O'Farrell et al. Jan 2001 B1
6175300 Kendrick Jan 2001 B1
6176602 Pastrick et al. Jan 2001 B1
6178034 Allemand et al. Jan 2001 B1
6183119 Desmond et al. Feb 2001 B1
6184781 Ramakesavan Feb 2001 B1
6185492 Kagawa et al. Feb 2001 B1
6188505 Lomprey et al. Feb 2001 B1
6191704 Takenaga et al. Feb 2001 B1
6196688 Caskey et al. Mar 2001 B1
6198409 Schofield et al. Mar 2001 B1
6199014 Walker et al. Mar 2001 B1
6200010 Anders Mar 2001 B1
6201642 Bos Mar 2001 B1
6210008 Hoekstra et al. Apr 2001 B1
6210012 Broer Apr 2001 B1
6217181 Lynam et al. Apr 2001 B1
6218934 Regan Apr 2001 B1
6222447 Schofield et al. Apr 2001 B1
6222460 DeLine et al. Apr 2001 B1
6227689 Miller May 2001 B1
6229319 Johnson May 2001 B1
6232937 Jacobsen et al. May 2001 B1
6239898 Byker et al. May 2001 B1
6243003 DeLine et al. Jun 2001 B1
6245262 Varaprasad et al. Jun 2001 B1
6249214 Kashiwazaki Jun 2001 B1
6250148 Lynam Jun 2001 B1
6250766 Strumolo et al. Jun 2001 B1
6250783 Stidham et al. Jun 2001 B1
6255639 Stam et al. Jul 2001 B1
6257746 Todd et al. Jul 2001 B1
6259412 Duroux Jul 2001 B1
6259475 Ramachandran et al. Jul 2001 B1
6265968 Betzitza et al. Jul 2001 B1
6266082 Yonezawa et al. Jul 2001 B1
6266442 Laumeyer et al. Jul 2001 B1
6268803 Gunderson et al. Jul 2001 B1
6269308 Kodaka et al. Jul 2001 B1
6274221 Smith et al. Aug 2001 B2
6276821 Pastrick et al. Aug 2001 B1
6276822 Bedrosian et al. Aug 2001 B1
6277471 Tang Aug 2001 B1
6278377 DeLine et al. Aug 2001 B1
6278941 Yokoyama Aug 2001 B1
6280068 Mertens et al. Aug 2001 B1
6280069 Pastrick et al. Aug 2001 B1
6281804 Haller et al. Aug 2001 B1
6285393 Shimoura et al. Sep 2001 B1
6285778 Nakajima et al. Sep 2001 B1
6286965 Caskey et al. Sep 2001 B1
6286984 Berg Sep 2001 B1
6289332 Menig et al. Sep 2001 B2
6290378 Buchalla et al. Sep 2001 B1
6291906 Marcus et al. Sep 2001 B1
6292752 Franke et al. Sep 2001 B1
6294989 Schofield et al. Sep 2001 B1
6296379 Pastrick Oct 2001 B1
6297781 Turnbull et al. Oct 2001 B1
6299333 Pastrick et al. Oct 2001 B1
6300879 Regan et al. Oct 2001 B1
6302545 Schofield et al. Oct 2001 B1
6304173 Pala et al. Oct 2001 B2
6305807 Schierbeek Oct 2001 B1
6310611 Caldwell Oct 2001 B1
6310714 Lomprey et al. Oct 2001 B1
6310738 Chu Oct 2001 B1
6311119 Sawamoto et al. Oct 2001 B2
6313454 Bos et al. Nov 2001 B1
6315421 Apfelbeck et al. Nov 2001 B1
6317057 Lee Nov 2001 B1
6317248 Agrawal et al. Nov 2001 B1
6318870 Spooner et al. Nov 2001 B1
6320176 Schofield et al. Nov 2001 B1
6320282 Caldwell Nov 2001 B1
6320612 Young Nov 2001 B1
6324295 Valery et al. Nov 2001 B1
6324450 Iwama Nov 2001 B1
6326613 Heslin et al. Dec 2001 B1
6326900 DeLine et al. Dec 2001 B2
6329925 Skiver et al. Dec 2001 B1
6330511 Ogura et al. Dec 2001 B2
6331066 Desmond et al. Dec 2001 B1
6333759 Mazzilli Dec 2001 B1
6335680 Matsuoka Jan 2002 B1
6336737 Thau Jan 2002 B1
6341523 Lynam Jan 2002 B2
6344805 Yasui et al. Feb 2002 B1
6346698 Turnbull Feb 2002 B1
6347880 Fürst et al. Feb 2002 B1
6348858 Weis et al. Feb 2002 B2
6353392 Schofield et al. Mar 2002 B1
6356206 Takenaga et al. Mar 2002 B1
6356376 Tonar et al. Mar 2002 B1
6357883 Strumolo et al. Mar 2002 B1
6362548 Bingle et al. Mar 2002 B1
6362729 Hellmann et al. Mar 2002 B1
6363326 Scully Mar 2002 B1
6366213 DeLine et al. Apr 2002 B2
6366236 Farmer et al. Apr 2002 B1
6370329 Teuchert Apr 2002 B1
6371636 Wesson Apr 2002 B1
6386742 DeLine et al. May 2002 B1
6388565 Bernhard et al. May 2002 B1
6388580 Graham May 2002 B1
6389340 Rayner May 2002 B1
6390529 Bingle et al. May 2002 B1
6390635 Whitehead et al. May 2002 B2
6396397 Bos et al. May 2002 B1
6396637 Roest et al. May 2002 B2
6407847 Poll et al. Jun 2002 B1
6408247 Ichikawa et al. Jun 2002 B1
6411204 Bloomfield et al. Jun 2002 B1
6411328 Franke et al. Jun 2002 B1
6412959 Tseng Jul 2002 B1
6412973 Bos et al. Jul 2002 B1
6415230 Maruko et al. Jul 2002 B1
6416208 Pastrick et al. Jul 2002 B2
6418376 Olson Jul 2002 B1
6419300 Pavao et al. Jul 2002 B1
6420975 DeLine et al. Jul 2002 B1
6421081 Markus Jul 2002 B1
6424272 Gutta et al. Jul 2002 B1
6424273 Gutta et al. Jul 2002 B1
6424892 Matsuoka Jul 2002 B1
6426492 Bos et al. Jul 2002 B1
6427349 Blank et al. Aug 2002 B1
6428172 Hutzel et al. Aug 2002 B1
6429594 Stam et al. Aug 2002 B1
6430303 Naoi et al. Aug 2002 B1
6433676 DeLine et al. Aug 2002 B2
6433680 Ho Aug 2002 B1
6433817 Guerra Aug 2002 B1
6433914 Lomprey et al. Aug 2002 B1
6437688 Kobayashi Aug 2002 B1
6438491 Farmer Aug 2002 B1
6439755 Fant et al. Aug 2002 B1
6441748 Takagi et al. Aug 2002 B1
6441872 Ho Aug 2002 B1
6442465 Breed et al. Aug 2002 B2
6445287 Schofield et al. Sep 2002 B1
6447128 Lang et al. Sep 2002 B1
6452533 Yamabuchi et al. Sep 2002 B1
6463369 Sadano et al. Oct 2002 B2
6466701 Ejiri et al. Oct 2002 B1
6469739 Bechtel et al. Oct 2002 B1
6472977 Poechmueller Oct 2002 B1
6472979 Schofield et al. Oct 2002 B2
6473001 Blum Oct 2002 B1
6474853 Pastrick et al. Nov 2002 B2
6476731 Miki et al. Nov 2002 B1
6477460 Kepler Nov 2002 B2
6477464 McCarthy et al. Nov 2002 B2
6483429 Yasui et al. Nov 2002 B1
6483438 DeLine et al. Nov 2002 B2
6485155 Duroux et al. Nov 2002 B1
6487500 Lemelson et al. Nov 2002 B2
6494602 Pastrick et al. Dec 2002 B2
6497503 Dassanayake et al. Dec 2002 B1
6498620 Schofield et al. Dec 2002 B2
6501387 Skiver et al. Dec 2002 B2
6509832 Bauer et al. Jan 2003 B1
6512624 Tonar et al. Jan 2003 B2
6513252 Schierbeek et al. Feb 2003 B1
6515581 Ho Feb 2003 B1
6515597 Wada et al. Feb 2003 B1
6516272 Lin Feb 2003 B2
6516664 Lynam Feb 2003 B2
6520667 Mousseau Feb 2003 B1
6522451 Lynam Feb 2003 B1
6522969 Kannonji Feb 2003 B2
6523964 Schofield et al. Feb 2003 B2
6534884 Marcus et al. Mar 2003 B2
6535242 Strumolo et al. Mar 2003 B1
6539306 Turnbull Mar 2003 B2
6540193 DeLine Apr 2003 B1
6542085 Yang Apr 2003 B1
6542182 Chutorash Apr 2003 B1
6545598 de Villeroche Apr 2003 B1
6547133 DeVries, Jr. et al. Apr 2003 B1
6549335 Trapani et al. Apr 2003 B1
6550949 Bauer et al. Apr 2003 B1
6553130 Lemelson et al. Apr 2003 B1
6553308 Uhlmann et al. Apr 2003 B1
6559435 Schofield et al. May 2003 B2
6567708 Bechtel et al. May 2003 B1
6568839 Pastrick et al. May 2003 B1
6572233 Northman et al. Jun 2003 B1
6574033 Chui et al. Jun 2003 B1
6575643 Takahashi Jun 2003 B2
6578017 Ebersole et al. Jun 2003 B1
6580373 Ohashi Jun 2003 B1
6581007 Hasegawa et al. Jun 2003 B2
6583730 Lang et al. Jun 2003 B2
6587573 Stam et al. Jul 2003 B1
6587968 Leyva Jul 2003 B1
6589625 Kothari et al. Jul 2003 B1
6591192 Okamura et al. Jul 2003 B2
6593565 Heslin et al. Jul 2003 B2
6593698 Stam et al. Jul 2003 B2
6593960 Sugimoto et al. Jul 2003 B1
6594065 Byker et al. Jul 2003 B2
6594067 Poll et al. Jul 2003 B2
6594090 Kruschwitz et al. Jul 2003 B2
6594583 Ogura et al. Jul 2003 B2
6594614 Studt et al. Jul 2003 B2
6597489 Guarr et al. Jul 2003 B1
6611202 Schofield et al. Aug 2003 B2
6611227 Nebiyeloul-Kifle Aug 2003 B1
6611610 Stam et al. Aug 2003 B1
6611759 Brosche Aug 2003 B2
6614387 Deadman Sep 2003 B1
6614579 Roberts et al. Sep 2003 B2
6616313 Fürst et al. Sep 2003 B2
6616764 Krämer Sep 2003 B2
6617564 Ockerse et al. Sep 2003 B2
6618672 Sasaki et al. Sep 2003 B2
6627918 Getz et al. Sep 2003 B2
6630888 Lang et al. Oct 2003 B2
6631316 Stam et al. Oct 2003 B2
6631994 Suzuki et al. Oct 2003 B2
6636258 Strumolo Oct 2003 B2
6642840 Lang et al. Nov 2003 B2
6642851 DeLine et al. Nov 2003 B2
6648477 Hutzel et al. Nov 2003 B2
6650233 DeLine et al. Nov 2003 B2
6650455 Miles Nov 2003 B2
6653614 Stam et al. Nov 2003 B2
6661830 Reed et al. Dec 2003 B1
6665592 Kodama Dec 2003 B2
6670207 Roberts Dec 2003 B1
6670910 Delcheccolo et al. Dec 2003 B2
6671080 Poll et al. Dec 2003 B2
6672731 Schnell et al. Jan 2004 B2
6672734 Lammers Jan 2004 B2
6672744 DeLine et al. Jan 2004 B2
6672745 Bauer et al. Jan 2004 B1
6674370 Rodewald et al. Jan 2004 B2
6674562 Miles Jan 2004 B1
6674878 Retterath et al. Jan 2004 B2
6675075 Engelsburg et al. Jan 2004 B1
6678056 Downs Jan 2004 B2
6678083 Anstee Jan 2004 B1
6678614 McCarthy et al. Jan 2004 B2
6680792 Miles Jan 2004 B2
6681163 Stam et al. Jan 2004 B2
6683539 Trajkovic et al. Jan 2004 B2
6683969 Nishigaki et al. Jan 2004 B1
6685348 Pastrick et al. Feb 2004 B2
6690268 Schofield et al. Feb 2004 B2
6690413 Moore Feb 2004 B1
6693517 McCarthy et al. Feb 2004 B2
6693518 Kumata Feb 2004 B2
6693519 Keirstead Feb 2004 B2
6693524 Payne Feb 2004 B1
6700605 Toyoda et al. Mar 2004 B1
6700692 Tonar et al. Mar 2004 B2
6703925 Steffel Mar 2004 B2
6704621 Stein et al. Mar 2004 B1
6709136 Pastrick et al. Mar 2004 B2
6710908 Miles et al. Mar 2004 B2
6711474 Treyz et al. Mar 2004 B1
6714331 Lewis et al. Mar 2004 B2
6717610 Bos et al. Apr 2004 B1
6717712 Lynam et al. Apr 2004 B2
6726337 Whitehead et al. Apr 2004 B2
6727808 Uselmann et al. Apr 2004 B1
6727844 Zimmermann et al. Apr 2004 B1
6728393 Stam et al. Apr 2004 B2
6728623 Takenaga et al. Apr 2004 B2
6731332 Yasui et al. May 2004 B1
6734807 King May 2004 B2
6735506 Breed et al. May 2004 B2
6737964 Samman et al. May 2004 B2
6738088 Uskolovsky et al. May 2004 B1
6741186 Ross May 2004 B2
6741377 Miles May 2004 B2
6744353 Sjönell Jun 2004 B2
6748211 Isaac et al. Jun 2004 B1
6754367 Ito et al. Jun 2004 B1
6756912 Skiver et al. Jun 2004 B2
6757109 Bos Jun 2004 B2
6759113 Tang Jul 2004 B1
6760157 Allen et al. Jul 2004 B1
6762867 Lippert et al. Jul 2004 B2
6764210 Akiyama Jul 2004 B2
6765480 Tseng Jul 2004 B2
6774810 DeLine et al. Aug 2004 B2
6774988 Stam et al. Aug 2004 B2
6778904 Iwami et al. Aug 2004 B2
6784828 Delcheccolo et al. Aug 2004 B2
6794119 Miles Sep 2004 B2
6795221 Urey Sep 2004 B1
6797396 Liu et al. Sep 2004 B1
6800871 Matsuda et al. Oct 2004 B2
6801127 Mizusawa et al. Oct 2004 B2
6801244 Takeda et al. Oct 2004 B2
6802617 Schofield et al. Oct 2004 B2
6806452 Bos et al. Oct 2004 B2
6807287 Hermans Oct 2004 B1
6810323 Bullock et al. Oct 2004 B1
6811330 Tozawa Nov 2004 B1
6812463 Okada Nov 2004 B2
6813545 Stromme Nov 2004 B2
6819231 Berberich et al. Nov 2004 B2
6819779 Nichani Nov 2004 B1
6822563 Bos et al. Nov 2004 B2
6823241 Shirato et al. Nov 2004 B2
6823261 Sekiguchi Nov 2004 B2
6824281 Schofield et al. Nov 2004 B2
6831261 Schofield et al. Dec 2004 B2
6832848 Pastrick Dec 2004 B2
6836725 Millington et al. Dec 2004 B2
6838980 Gloger et al. Jan 2005 B2
6842189 Park Jan 2005 B2
6842276 Poll et al. Jan 2005 B2
6847487 Burgner Jan 2005 B2
6848817 Bos et al. Feb 2005 B2
6850629 Jeon Feb 2005 B2
6853491 Ruhle et al. Feb 2005 B1
6853738 Nishigaki et al. Feb 2005 B1
6859148 Miller et al. Feb 2005 B2
6861809 Stam Mar 2005 B2
6870655 Northman et al. Mar 2005 B1
6870656 Tonar et al. Mar 2005 B2
6871982 Holman et al. Mar 2005 B2
6873253 Veziris Mar 2005 B2
6877888 DeLine et al. Apr 2005 B2
6882287 Schofield Apr 2005 B2
6888447 Hori et al. May 2005 B2
6889161 Winner et al. May 2005 B2
6891563 Schofield et al. May 2005 B2
6898518 Padmanabhan May 2005 B2
6902284 Hutzel et al. Jun 2005 B2
6906620 Nakai et al. Jun 2005 B2
6906639 Lemelson et al. Jun 2005 B2
6909753 Meehan et al. Jun 2005 B2
6912396 Sziraki et al. Jun 2005 B2
6914521 Rothkop Jul 2005 B2
6928180 Stam et al. Aug 2005 B2
6930737 Weindorf et al. Aug 2005 B2
6932669 Lee et al. Aug 2005 B2
6933837 Gunderson et al. Aug 2005 B2
6934067 Ash et al. Aug 2005 B2
6940423 Takagi et al. Sep 2005 B2
6946978 Schofield Sep 2005 B2
6947576 Stam et al. Sep 2005 B2
6947577 Stam et al. Sep 2005 B2
6950035 Tanaka et al. Sep 2005 B2
6951410 Parsons Oct 2005 B2
6951681 Hartley et al. Oct 2005 B2
6952312 Weber et al. Oct 2005 B2
6953253 Schofield et al. Oct 2005 B2
6956469 Hirvonen et al. Oct 2005 B2
6959994 Fujikawa et al. Nov 2005 B2
6961178 Sugino et al. Nov 2005 B2
6961661 Sekiguchi et al. Nov 2005 B2
6967569 Weber et al. Nov 2005 B2
6968736 Lynam Nov 2005 B2
6972888 Poll et al. Dec 2005 B2
6975215 Schofield et al. Dec 2005 B2
6975775 Rykowski et al. Dec 2005 B2
6977702 Wu Dec 2005 B2
6980092 Turnbull et al. Dec 2005 B2
6985291 Watson et al. Jan 2006 B2
6989736 Berberich et al. Jan 2006 B2
6990397 Albou et al. Jan 2006 B2
6992718 Takahara Jan 2006 B1
6995687 Lang et al. Feb 2006 B2
7001058 Inditsky Feb 2006 B2
7004592 Varaprasad et al. Feb 2006 B2
7004593 Weller et al. Feb 2006 B2
7004606 Schofield Feb 2006 B2
7005974 McMahon et al. Feb 2006 B2
7006173 Hiyama et al. Feb 2006 B1
7009751 Tonar et al. Mar 2006 B2
7012727 Hutzel et al. Mar 2006 B2
7023331 Kodama Apr 2006 B2
7027387 Reinold et al. Apr 2006 B2
7027615 Chen Apr 2006 B2
7030738 Ishii Apr 2006 B2
7030775 Sekiguchi Apr 2006 B2
7030778 Ra Apr 2006 B2
7038577 Pawlicki et al. May 2006 B2
7042616 Tonar et al. May 2006 B2
7046418 Lin et al. May 2006 B2
7046448 Burgner May 2006 B2
7057505 Iwamoto Jun 2006 B2
7057681 Hinata et al. Jun 2006 B2
7062300 Kim Jun 2006 B1
7065432 Moisel et al. Jun 2006 B2
7068289 Satoh et al. Jun 2006 B2
7068844 Javidi et al. Jun 2006 B1
7085633 Nishira et al. Aug 2006 B2
7085637 Breed et al. Aug 2006 B2
7091837 Nakai et al. Aug 2006 B2
7092548 Laumeyer et al. Aug 2006 B2
7095432 Nakayama et al. Aug 2006 B2
7106213 White Sep 2006 B2
7108409 DeLine et al. Sep 2006 B2
7110021 Nobori et al. Sep 2006 B2
7110156 Lawlor et al. Sep 2006 B2
7113867 Stein Sep 2006 B1
7116246 Winter et al. Oct 2006 B2
7121028 Shoen et al. Oct 2006 B2
7123168 Schofield Oct 2006 B2
7125131 Olczak Oct 2006 B2
7132064 Li et al. Nov 2006 B2
7133661 Hatae et al. Nov 2006 B2
7149613 Stam et al. Dec 2006 B2
7151996 Stein Dec 2006 B2
7151997 Uhlmann et al. Dec 2006 B2
7154657 Poll et al. Dec 2006 B2
7160017 Lee et al. Jan 2007 B2
7167796 Taylor et al. Jan 2007 B2
7171027 Satoh Jan 2007 B2
7175291 Li Feb 2007 B1
7184190 McCabe et al. Feb 2007 B2
7184585 Hamza et al. Feb 2007 B2
7187498 Bengoechea et al. Mar 2007 B2
7188963 Schofield et al. Mar 2007 B2
7193764 Lin et al. Mar 2007 B2
7195381 Lynam et al. Mar 2007 B2
7202776 Breed Apr 2007 B2
7202987 Varaprasad et al. Apr 2007 B2
7205904 Schofield Apr 2007 B2
7209277 Tonar et al. Apr 2007 B2
7215473 Fleming May 2007 B2
7221363 Roberts et al. May 2007 B2
7224324 Quist et al. May 2007 B2
7227459 Bos et al. Jun 2007 B2
7227611 Hull et al. Jun 2007 B2
7235918 McCullough et al. Jun 2007 B2
7245336 Hiyama et al. Jul 2007 B2
7248283 Takagi et al. Jul 2007 B2
7248344 Morcom Jul 2007 B2
7249860 Kulas et al. Jul 2007 B2
7251079 Capaldo et al. Jul 2007 B2
7253723 Lindahl et al. Aug 2007 B2
7255451 McCabe et al. Aug 2007 B2
7255465 DeLine et al. Aug 2007 B2
7262916 Kao et al. Aug 2007 B2
7269327 Tang Sep 2007 B2
7269328 Tang Sep 2007 B2
7271951 Weber et al. Sep 2007 B2
7274501 McCabe et al. Sep 2007 B2
7286280 Whitehead et al. Oct 2007 B2
7290919 Pan et al. Nov 2007 B2
7304661 Ishikura Dec 2007 B2
7310177 McCabe et al. Dec 2007 B2
7311406 Schofield et al. Dec 2007 B2
7311428 DeLine et al. Dec 2007 B2
7323819 Hong et al. Jan 2008 B2
7324261 Tonar et al. Jan 2008 B2
7325934 Schofield et al. Feb 2008 B2
7325935 Schofield et al. Feb 2008 B2
7328103 McCarthy et al. Feb 2008 B2
7337055 Matsumoto et al. Feb 2008 B2
7338177 Lynam Mar 2008 B2
7339149 Schofield et al. Mar 2008 B1
7344261 Schofield et al. Mar 2008 B2
7344284 Lynam et al. Mar 2008 B2
7349143 Tonar et al. Mar 2008 B2
7355524 Schofield Apr 2008 B2
7360932 Uken et al. Apr 2008 B2
7362505 Hikmet et al. Apr 2008 B2
7362883 Otsuka et al. Apr 2008 B2
7370983 DeWind et al. May 2008 B2
7372611 Tonar et al. May 2008 B2
7375803 Bamji May 2008 B1
7379224 Tonar et al. May 2008 B2
7379225 Tonar et al. May 2008 B2
7379243 Horsten et al. May 2008 B2
7380948 Schofield et al. Jun 2008 B2
7388182 Schofield et al. Jun 2008 B2
7402786 Schofield et al. Jul 2008 B2
7403659 Das et al. Jul 2008 B2
7411732 Kao et al. Aug 2008 B2
7412328 Uhlmann et al. Aug 2008 B2
7417781 Tonar et al. Aug 2008 B2
7420756 Lynam Sep 2008 B2
7423248 Schofield et al. Sep 2008 B2
7423821 Bechtel et al. Sep 2008 B2
7425076 Schofield et al. Sep 2008 B2
7429998 Kawauchi et al. Sep 2008 B2
7432248 Roberts et al. Oct 2008 B2
7432967 Bechtel et al. Oct 2008 B2
7446462 Lim et al. Nov 2008 B2
7446650 Schofield et al. Nov 2008 B2
7446924 Schofield et al. Nov 2008 B2
7448776 Tang Nov 2008 B2
7459664 Schofield et al. Dec 2008 B2
7460007 Schofield et al. Dec 2008 B2
7467883 DeLine et al. Dec 2008 B2
7471438 McCabe et al. Dec 2008 B2
7474963 Taylor et al. Jan 2009 B2
7477439 Tonar et al. Jan 2009 B2
7480149 DeWard et al. Jan 2009 B2
7489374 Utsumi et al. Feb 2009 B2
7490007 Taylor et al. Feb 2009 B2
7494231 Varaprasad et al. Feb 2009 B2
7495719 Adachi et al. Feb 2009 B2
7502156 Tonar et al. Mar 2009 B2
7511872 Tonar et al. Mar 2009 B2
7525604 Xue Apr 2009 B2
7526103 Schofield et al. Apr 2009 B2
7540620 Weller et al. Jun 2009 B2
7541743 Salmeen et al. Jun 2009 B2
7543946 Ockerse et al. Jun 2009 B2
7545429 Travis Jun 2009 B2
7547467 Olson et al. Jun 2009 B2
7548291 Lee et al. Jun 2009 B2
7551103 Schofield Jun 2009 B2
7551354 Horsten et al. Jun 2009 B2
7561181 Schofield et al. Jul 2009 B2
7565006 Stam et al. Jul 2009 B2
7566851 Stein et al. Jul 2009 B2
7567291 Bechtel et al. Jul 2009 B2
7572490 Park et al. Aug 2009 B2
7581867 Lee et al. Sep 2009 B2
7586566 Nelson et al. Sep 2009 B2
7586666 McCabe et al. Sep 2009 B2
7613327 Stam et al. Nov 2009 B2
7616781 Schofield et al. Nov 2009 B2
7619508 Lynam et al. Nov 2009 B2
7629996 Rademacher et al. Dec 2009 B2
7633383 Dunsmoir et al. Dec 2009 B2
7639149 Katoh Dec 2009 B2
7650030 Shan et al. Jan 2010 B2
7653215 Stam Jan 2010 B2
7655894 Schofield et al. Feb 2010 B2
7663798 Tonar et al. Feb 2010 B2
7676087 Dhua et al. Mar 2010 B2
7683326 Stam et al. Mar 2010 B2
7702133 Muramatsu et al. Apr 2010 B2
7720580 Higgins-Luthman May 2010 B2
7724434 Cross et al. May 2010 B2
7731403 Lynam et al. Jun 2010 B2
7742864 Sekiguchi Jun 2010 B2
7786898 Stein et al. Aug 2010 B2
7791694 Molsen et al. Sep 2010 B2
7792329 Schofield et al. Sep 2010 B2
7825600 Stam et al. Nov 2010 B2
7842154 Lynam Nov 2010 B2
7843451 Lafon Nov 2010 B2
7854514 Conner et al. Dec 2010 B2
7855755 Weller et al. Dec 2010 B2
7855778 Yung et al. Dec 2010 B2
7859565 Schofield et al. Dec 2010 B2
7877175 Higgins-Luthman Jan 2011 B2
7881496 Camilleri Feb 2011 B2
7903324 Kobayashi et al. Mar 2011 B2
7903335 Nieuwkerk et al. Mar 2011 B2
7914187 Higgins-Luthman et al. Mar 2011 B2
7930160 Hosagrahara et al. Apr 2011 B1
7949152 Schofield et al. May 2011 B2
7965357 Van De Witte et al. Jun 2011 B2
7991522 Higgins-Luthman Aug 2011 B2
7994462 Schofield et al. Aug 2011 B2
8017898 Lu et al. Sep 2011 B2
8027691 Bernas et al. Sep 2011 B2
8045760 Stam et al. Oct 2011 B2
8064643 Stein et al. Nov 2011 B2
8082101 Stein et al. Dec 2011 B2
8090153 Schofield et al. Jan 2012 B2
8095310 Taylor et al. Jan 2012 B2
8098142 Schofield et al. Jan 2012 B2
8120652 Bechtel et al. Feb 2012 B2
8164628 Stein et al. Apr 2012 B2
8184159 Luo May 2012 B2
8203440 Schofield et al. Jun 2012 B2
8222588 Schofield et al. Jul 2012 B2
8224031 Saito Jul 2012 B2
8233045 Luo et al. Jul 2012 B2
8254635 Stein et al. Aug 2012 B2
8288711 Heslin et al. Oct 2012 B2
8289430 Bechtel et al. Oct 2012 B2
8305471 Bechtel et al. Nov 2012 B2
8308325 Takayanazi et al. Nov 2012 B2
8314689 Schofield et al. Nov 2012 B2
8324552 Schofield et al. Dec 2012 B2
8325986 Schofield et al. Dec 2012 B2
8339526 Minikey, Jr. et al. Dec 2012 B2
8378851 Stein et al. Feb 2013 B2
8386114 Higgins-Luthman Feb 2013 B2
8405726 Schofield et al. Mar 2013 B2
8452055 Stein et al. May 2013 B2
8531278 DeWard et al. Sep 2013 B2
8553088 Stein et al. Oct 2013 B2
20010019356 Takeda et al. Sep 2001 A1
20010022616 Rademacher et al. Sep 2001 A1
20010026316 Senatore Oct 2001 A1
20010045981 Gloger et al. Nov 2001 A1
20020003571 Schofield et al. Jan 2002 A1
20020005778 Breed Jan 2002 A1
20020029103 Breed et al. Mar 2002 A1
20020049535 Rigo et al. Apr 2002 A1
20020060522 Stam et al. May 2002 A1
20020072026 Lynam et al. Jun 2002 A1
20020080235 Jeon Jun 2002 A1
20020085155 Arikawa Jul 2002 A1
20020093826 Bos et al. Jul 2002 A1
20020113873 Williams Aug 2002 A1
20020116106 Breed et al. Aug 2002 A1
20020126497 Pastrick Sep 2002 A1
20020154007 Yang Oct 2002 A1
20020159270 Lynam et al. Oct 2002 A1
20020172053 Pastrick et al. Nov 2002 A1
20020191409 DeLine et al. Dec 2002 A1
20030002165 Mathias et al. Jan 2003 A1
20030007261 Hutzel et al. Jan 2003 A1
20030016125 Lang et al. Jan 2003 A1
20030016287 Nakayama et al. Jan 2003 A1
20030016542 Pastrick et al. Jan 2003 A1
20030020603 DeLine et al. Jan 2003 A1
20030025596 Lang et al. Feb 2003 A1
20030025597 Schofield Feb 2003 A1
20030030546 Tseng Feb 2003 A1
20030030551 Ho Feb 2003 A1
20030030724 Okamoto Feb 2003 A1
20030035050 Mizusawa Feb 2003 A1
20030043269 Park Mar 2003 A1
20030052969 Satoh et al. Mar 2003 A1
20030058338 Kawauchi et al. Mar 2003 A1
20030067383 Yang Apr 2003 A1
20030069690 Correia et al. Apr 2003 A1
20030070741 Rosenberg et al. Apr 2003 A1
20030076415 Strumolo Apr 2003 A1
20030080877 Takagi et al. May 2003 A1
20030085806 Samman et al. May 2003 A1
20030088361 Sekiguchi May 2003 A1
20030090568 Pico May 2003 A1
20030090569 Poechmuller May 2003 A1
20030090570 Takagi et al. May 2003 A1
20030095331 Bengoechea et al. May 2003 A1
20030098908 Misaiji et al. May 2003 A1
20030103141 Bechtel et al. Jun 2003 A1
20030103142 Hitomi et al. Jun 2003 A1
20030117522 Okada Jun 2003 A1
20030122929 Minaudo et al. Jul 2003 A1
20030122930 Schofield et al. Jul 2003 A1
20030125855 Breed et al. Jul 2003 A1
20030133014 Mendoza Jul 2003 A1
20030137586 Lewellen Jul 2003 A1
20030141965 Gunderson et al. Jul 2003 A1
20030146831 Berberich et al. Aug 2003 A1
20030169158 Paul, Jr. Sep 2003 A1
20030169522 Schofield et al. Sep 2003 A1
20030179293 Oizumi Sep 2003 A1
20030189754 Sugino et al. Oct 2003 A1
20030202096 Kim Oct 2003 A1
20030209893 Breed Nov 2003 A1
20030210369 Wu Nov 2003 A1
20030214576 Koga Nov 2003 A1
20030214584 Ross, Jr. Nov 2003 A1
20030214733 Fujikawa et al. Nov 2003 A1
20030222793 Tanaka et al. Dec 2003 A1
20030222982 Hamdan et al. Dec 2003 A1
20030222983 Nobori et al. Dec 2003 A1
20030227546 Hilborn et al. Dec 2003 A1
20040004541 Hong Jan 2004 A1
20040016870 Pawlicki et al. Jan 2004 A1
20040027695 Lin Feb 2004 A1
20040032321 McMahon et al. Feb 2004 A1
20040032675 Weller et al. Feb 2004 A1
20040036768 Green Feb 2004 A1
20040046870 Leigh Travis Mar 2004 A1
20040051634 Schofield et al. Mar 2004 A1
20040056955 Berberich et al. Mar 2004 A1
20040057131 Hutzel et al. Mar 2004 A1
20040064241 Sekiguchi Apr 2004 A1
20040066285 Sekiguchi Apr 2004 A1
20040075603 Kodama Apr 2004 A1
20040077359 Bernas et al. Apr 2004 A1
20040080404 White Apr 2004 A1
20040080431 White Apr 2004 A1
20040085196 Miller et al. May 2004 A1
20040086153 Tsai et al. May 2004 A1
20040090314 Iwamoto May 2004 A1
20040090317 Rothkop May 2004 A1
20040096082 Nakai et al. May 2004 A1
20040098196 Sekiguchi May 2004 A1
20040105614 Kobayashi et al. Jun 2004 A1
20040107030 Nishira et al. Jun 2004 A1
20040107617 Shoen et al. Jun 2004 A1
20040109060 Ishii Jun 2004 A1
20040114039 Ishikura Jun 2004 A1
20040128065 Taylor et al. Jul 2004 A1
20040145457 Schofield et al. Jul 2004 A1
20040164228 Fogg et al. Aug 2004 A1
20040200948 Bos et al. Oct 2004 A1
20040202001 Roberts et al. Oct 2004 A1
20040243303 Padmanabhan Dec 2004 A1
20040251804 McCullough et al. Dec 2004 A1
20050024591 Lian et al. Feb 2005 A1
20050024729 Ockerse et al. Feb 2005 A1
20050073853 Stam Apr 2005 A1
20050078347 Lin et al. Apr 2005 A1
20050078389 Kulas et al. Apr 2005 A1
20050079326 Varaprasad et al. Apr 2005 A1
20050083577 Varaprasad et al. Apr 2005 A1
20050099559 Lee et al. May 2005 A1
20050111070 Lin et al. May 2005 A1
20050131607 Breed Jun 2005 A1
20050140855 Utsumi et al. Jun 2005 A1
20050185278 Horsten et al. Aug 2005 A1
20050219852 Stam et al. Oct 2005 A1
20050226490 Phillips et al. Oct 2005 A1
20050237385 Kosaka et al. Oct 2005 A1
20050259326 Weber et al. Nov 2005 A1
20060007550 Tonar et al. Jan 2006 A1
20060018511 Stam et al. Jan 2006 A1
20060018512 Stam et al. Jan 2006 A1
20060028730 Varaprasad et al. Feb 2006 A1
20060050018 Hutzel et al. Mar 2006 A1
20060091813 Stam et al. May 2006 A1
20060103727 Tseng May 2006 A1
20060164230 DeWind et al. Jul 2006 A1
20060164725 Horsten et al. Jul 2006 A1
20060250224 Steffel et al. Nov 2006 A1
20060250501 Wildmann et al. Nov 2006 A1
20060274218 Xue Dec 2006 A1
20070024724 Stein et al. Feb 2007 A1
20070041096 Nieuwkerk et al. Feb 2007 A1
20070104476 Yasutomi et al. May 2007 A1
20070109406 Schofield et al. May 2007 A1
20070118287 Taylor et al. May 2007 A1
20070120657 Schofield et al. May 2007 A1
20070132567 Schofield et al. Jun 2007 A1
20070154063 Breed Jul 2007 A1
20070162229 McCarthy et al. Jul 2007 A1
20070171037 Schofield et al. Jul 2007 A1
20070183066 Varaprasad et al. Aug 2007 A1
20070184284 Varaprasad et al. Aug 2007 A1
20070193811 Breed et al. Aug 2007 A1
20070242339 Bradley Oct 2007 A1
20080002106 Van De Witte et al. Jan 2008 A1
20080013153 McCabe et al. Jan 2008 A1
20080043099 Stein et al. Feb 2008 A1
20080068520 Minikey, Jr. et al. Mar 2008 A1
20080094684 Varaprasad et al. Apr 2008 A1
20080094685 Varaprasad et al. Apr 2008 A1
20080147321 Howard et al. Jun 2008 A1
20080180529 Taylor et al. Jul 2008 A1
20080180781 Varaprasad et al. Jul 2008 A1
20080183355 Taylor et al. Jul 2008 A1
20080186724 Lynam et al. Aug 2008 A1
20080201075 Taylor et al. Aug 2008 A1
20080212189 Baur et al. Sep 2008 A1
20080212215 Schofield et al. Sep 2008 A1
20080225538 Lynam et al. Sep 2008 A1
20080234899 Breed et al. Sep 2008 A1
20080266389 DeWind et al. Oct 2008 A1
20080266396 Stein Oct 2008 A1
20080291522 Varaprasad et al. Nov 2008 A1
20080308219 Lynam Dec 2008 A1
20090010494 Bechtel et al. Jan 2009 A1
20090015736 Weller et al. Jan 2009 A1
20090033837 Molsen et al. Feb 2009 A1
20090040465 Conner et al. Feb 2009 A1
20090040588 Tonar et al. Feb 2009 A1
20090040778 Takayanazi et al. Feb 2009 A1
20090052003 Schofield et al. Feb 2009 A1
20090066065 Breed et al. Mar 2009 A1
20090113509 Tseng et al. Apr 2009 A1
20090174776 McCarthy et al. Jul 2009 A1
20090190015 Bechtel et al. Jul 2009 A1
20090201137 Weller et al. Aug 2009 A1
20090243824 Hook et al. Oct 2009 A1
20090256938 Bechtel et al. Oct 2009 A1
20090262422 Cross et al. Oct 2009 A1
20120045112 Lundblad et al. Feb 2012 A1
20120069185 Stein Mar 2012 A1
20120200707 Stein et al. Aug 2012 A1
20120314071 Rosenbaum et al. Dec 2012 A1
20130141580 Stein et al. Jun 2013 A1
20130147957 Stein Jun 2013 A1
Foreign Referenced Citations (465)
Number Date Country
A-4031795 Feb 1995 AU
1008142 Jan 1996 BE
6446 Jul 1984 CH
2074262 Apr 1991 CN
2185701 Dec 1994 CN
1104741 Jul 1995 CN
2204254 Aug 1995 CN
941408 Apr 1956 DE
944531 Jul 1956 DE
1152627 Aug 1963 DE
1182971 Dec 1964 DE
1190413 Apr 1965 DE
1196598 Jul 1965 DE
1214174 Apr 1966 DE
2064839 Dec 1970 DE
7323996 Nov 1973 DE
2631713 Feb 1977 DE
2946561 May 1981 DE
3004247 Aug 1981 DE
3040555 May 1982 DE
3101855 Aug 1982 DE
3240498 Nov 1982 DE
3248511 Jul 1984 DE
3248511 Jul 1984 DE
3301945 Jul 1984 DE
3301945 Jul 1984 DE
3433671 Mar 1985 DE
3515116 Oct 1986 DE
3528220 Feb 1987 DE
3601388 Jul 1987 DE
3614882 Nov 1987 DE
3637165 May 1988 DE
3636946 Jun 1988 DE
3642196 Jun 1988 DE
4015959 Jun 1988 DE
3734066 Apr 1989 DE
3737395 May 1989 DE
3833022 Apr 1990 DE
3839512 May 1990 DE
3839513 May 1990 DE
3840425 Jun 1990 DE
3844364 Jul 1990 DE
9010196 Sep 1990 DE
3932216 Apr 1991 DE
4007646 Sep 1991 DE
4107965 Sep 1991 DE
4111993 Oct 1991 DE
4116255 Dec 1991 DE
4023952 Feb 1992 DE
4130010 Mar 1992 DE
4032927 Apr 1992 DE
4133882 Apr 1992 DE
4035956 May 1992 DE
4122531 Jan 1993 DE
4124654 Jan 1993 DE
4300941 Jul 1993 DE
9306989.8 Jul 1993 DE
4206142 Sep 1993 DE
4137551 Nov 1993 DE
4214223 Nov 1993 DE
4231137 Feb 1994 DE
4328304 Mar 1994 DE
4328902 Mar 1994 DE
4332612 Apr 1994 DE
4238599 Jun 1994 DE
4337756 Jun 1994 DE
4332836 Sep 1994 DE
4407082 Sep 1994 DE
4407757 Sep 1994 DE
4411179 Oct 1994 DE
4412669 Oct 1994 DE
4418122 Dec 1994 DE
4423966 Jan 1995 DE
4336288 Mar 1995 DE
4428069 Mar 1995 DE
4434698 Mar 1995 DE
4341409 Jun 1995 DE
69107283 Jul 1995 DE
4415885 Nov 1995 DE
19518978 Nov 1995 DE
4480341 Mar 1996 DE
29703084 Apr 1997 DE
29703084 Jun 1997 DE
29805142 May 1998 DE
19755008 Jul 1999 DE
29902344 Jul 1999 DE
19530617 Feb 2009 DE
0049722 Oct 1980 EP
0048492 Sep 1981 EP
0072406 Feb 1983 EP
0165817 Dec 1985 EP
0202460 Nov 1986 EP
0348691 Jun 1989 EP
0354561 Aug 1989 EP
169734 Oct 1989 EP
0340735 Nov 1989 EP
0353200 Jan 1990 EP
0360880 Apr 1990 EP
361914 Apr 1990 EP
0426503 May 1991 EP
0433538 Jun 1991 EP
0450553 Oct 1991 EP
454516 Oct 1991 EP
0479271 Oct 1991 EP
455524 Nov 1991 EP
0387817 Dec 1991 EP
477986 Apr 1992 EP
0254435 May 1992 EP
487100 May 1992 EP
0487465 May 1992 EP
0492591 Jul 1992 EP
495508 Jul 1992 EP
496411 Jul 1992 EP
0501345 Sep 1992 EP
505237 Sep 1992 EP
0513476 Nov 1992 EP
0513476 Nov 1992 EP
514343 Nov 1992 EP
527665 Feb 1993 EP
0532379 Mar 1993 EP
0533508 Mar 1993 EP
0550397 Jul 1993 EP
0558027 Sep 1993 EP
0564858 Oct 1993 EP
0567059 Oct 1993 EP
0606586 Dec 1993 EP
582236 Feb 1994 EP
0586857 Mar 1994 EP
0588815 Mar 1994 EP
0590588 Apr 1994 EP
0591743 Apr 1994 EP
0299509 Jun 1994 EP
0605045 Jul 1994 EP
617296 Sep 1994 EP
C 0615882 Sep 1994 EP
626654 Nov 1994 EP
640903 Mar 1995 EP
0640903 Mar 1995 EP
642950 Mar 1995 EP
654392 May 1995 EP
0341985 Jul 1995 EP
0667254 Aug 1995 EP
0667708 Aug 1995 EP
686865 Dec 1995 EP
0729864 Dec 1995 EP
0697641 Feb 1996 EP
0728618 Aug 1996 EP
756968 Feb 1997 EP
0769419 Apr 1997 EP
0788947 Aug 1997 EP
0487332 Oct 1997 EP
677428 Jun 1998 EP
0677428 Jun 1998 EP
1058220 Dec 2000 EP
1074430 Feb 2001 EP
1152285 Nov 2001 EP
0830267 Dec 2001 EP
0975709 Feb 2002 EP
860325 Nov 2002 EP
1359557 May 2003 EP
1376207 Jan 2004 EP
0889801 Jul 2008 EP
1766469 May 2009 EP
1021987 Feb 1953 FR
1461419 Dec 1966 FR
2585991 Feb 1987 FR
2610401 Aug 1988 FR
2641237 Jul 1990 FR
2646383 Nov 1990 FR
2672857 Aug 1992 FR
2673499 Sep 1992 FR
2674201 Sep 1992 FR
2674354 Sep 1992 FR
2687000 Aug 1993 FR
2706211 Dec 1994 FR
2721872 Jan 1996 FR
2726144 Apr 1996 FR
2 759 045 Aug 1998 FR
810010 Mar 1959 GB
914827 Jan 1963 GB
934037 Aug 1963 GB
1000265 Aug 1965 GB
1008411 Oct 1965 GB
1054064 Jan 1967 GB
1098608 Jan 1968 GB
1106339 Mar 1968 GB
1136134 Dec 1968 GB
1178416 Jan 1970 GB
1197710 Jul 1970 GB
2137573 Oct 1984 GB
2210835 Jun 1989 GB
2210836 Jun 1989 GB
224187 Nov 1991 GB
2255539 Nov 1992 GB
2255649 Nov 1992 GB
2261339 May 1993 GB
2262829 Jun 1993 GB
227139 Apr 1994 GB
2280810 Feb 1995 GB
2289332 Nov 1995 GB
2292857 Mar 1996 GB
2297632 Aug 1996 GB
2351055 Dec 2000 GB
970014 Jul 1998 IE
55039843 Mar 1980 JP
S5539843 Mar 1980 JP
5730639 Feb 1982 JP
57208530 Dec 1982 JP
5830729 Feb 1983 JP
S5871230 Apr 1983 JP
58110334 Jun 1983 JP
S58110334 Jun 1983 JP
S58122421 Jul 1983 JP
58209635 Dec 1983 JP
59-114139 Jul 1984 JP
S59127200 Jul 1984 JP
S6047737 Mar 1985 JP
6079889 May 1985 JP
6080953 May 1985 JP
S6078312 May 1985 JP
S60206746 Oct 1985 JP
60240545 Nov 1985 JP
S60219133 Nov 1985 JP
60261275 Dec 1985 JP
S60255537 Dec 1985 JP
S6141929 Feb 1986 JP
S6185238 Apr 1986 JP
S61105245 May 1986 JP
S61191937 Aug 1986 JP
S61285151 Dec 1986 JP
S61285152 Dec 1986 JP
S6221010 Jan 1987 JP
S6226141 Feb 1987 JP
S6280143 Apr 1987 JP
6272245 May 1987 JP
S62115600 May 1987 JP
2122844 Jun 1987 JP
62122487 Jun 1987 JP
62131837 Jun 1987 JP
S62-131837 Jun 1987 JP
S62253543 Nov 1987 JP
S62253546 Nov 1987 JP
S62287164 Dec 1987 JP
63106730 May 1988 JP
63106731 May 1988 JP
63258236 Oct 1988 JP
S63192788 Dec 1988 JP
64-14700 Jan 1989 JP
01123587 May 1989 JP
H01123587 May 1989 JP
H01233129 Sep 1989 JP
H01265400 Oct 1989 JP
H01275237 Nov 1989 JP
H0268237 Mar 1990 JP
02190978 Jul 1990 JP
H02212232 Aug 1990 JP
H0314739 Jan 1991 JP
30061192 Mar 1991 JP
H0374231 Mar 1991 JP
03099952 Apr 1991 JP
03266739 Nov 1991 JP
04239400 Nov 1991 JP
H03246413 Nov 1991 JP
H03273953 Dec 1991 JP
H03282707 Dec 1991 JP
H03282709 Dec 1991 JP
H042909 Jan 1992 JP
H0410200 Jan 1992 JP
4-114587 Apr 1992 JP
04127280 Apr 1992 JP
H04-127280 Apr 1992 JP
H04238219 Aug 1992 JP
H04242391 Aug 1992 JP
40245886 Sep 1992 JP
H04250786 Sep 1992 JP
4291405 Oct 1992 JP
H04303047 Oct 1992 JP
50000638 Jan 1993 JP
H0516722 Jan 1993 JP
050883 Mar 1993 JP
0577657 Mar 1993 JP
H05155287 Jun 1993 JP
5189694 Jul 1993 JP
H05172638 Jul 1993 JP
5-213113 Aug 1993 JP
H05201298 Aug 1993 JP
05244596 Sep 1993 JP
H05229383 Sep 1993 JP
H05297141 Nov 1993 JP
06000927 Jan 1994 JP
H0640286 Feb 1994 JP
H0648247 Feb 1994 JP
H0672234 Mar 1994 JP
H0676200 Mar 1994 JP
6107035 Apr 1994 JP
06113215 Apr 1994 JP
06117924 Apr 1994 JP
06150198 May 1994 JP
H06162398 Jun 1994 JP
H06174845 Jun 1994 JP
H06191344 Jul 1994 JP
6227318 Aug 1994 JP
06230115 Aug 1994 JP
H06215291 Aug 1994 JP
H06229739 Aug 1994 JP
06247246 Sep 1994 JP
06267304 Sep 1994 JP
06270733 Sep 1994 JP
06274626 Sep 1994 JP
06276524 Sep 1994 JP
H06262963 Sep 1994 JP
H06267303 Sep 1994 JP
H06275104 Sep 1994 JP
06295601 Oct 1994 JP
H06289138 Oct 1994 JP
H06293236 Oct 1994 JP
05093981 Nov 1994 JP
06321007 Nov 1994 JP
H06321010 Nov 1994 JP
H06324144 Nov 1994 JP
H06337938 Dec 1994 JP
H06341821 Dec 1994 JP
074170 Jan 1995 JP
07025286 Jan 1995 JP
H072021 Jan 1995 JP
H072022 Jan 1995 JP
732936 Feb 1995 JP
7032935 Feb 1995 JP
7047878 Feb 1995 JP
7052706 Feb 1995 JP
H0737180 Feb 1995 JP
H0740782 Feb 1995 JP
H0746460 Feb 1995 JP
7069125 Mar 1995 JP
07078240 Mar 1995 JP
H0757200 Mar 1995 JP
H0764632 Mar 1995 JP
H0771916 Mar 1995 JP
07105496 Apr 1995 JP
H07101291 Apr 1995 JP
H07105487 Apr 1995 JP
H07108873 Apr 1995 JP
H07108874 Apr 1995 JP
3395289 May 1995 JP
7125571 May 1995 JP
7137574 May 1995 JP
H07125570 May 1995 JP
H07141588 Jun 1995 JP
H07144577 Jun 1995 JP
7186818 Jul 1995 JP
H07249128 Sep 1995 JP
H08320997 Sep 1995 JP
7277072 Oct 1995 JP
H07280563 Oct 1995 JP
06113215 Dec 1995 JP
H07315122 Dec 1995 JP
H0840138 Feb 1996 JP
H0840140 Feb 1996 JP
H0843082 Feb 1996 JP
H0844999 Feb 1996 JP
H0850697 Feb 1996 JP
08-166221 Jun 1996 JP
08166221 Jun 1996 JP
06321007 Jul 1996 JP
07078240 Sep 1996 JP
H07239714 Dec 1996 JP
2630604 Apr 1997 JP
H0991596 Apr 1997 JP
O9330415 Dec 1997 JP
10134183 May 1998 JP
H10222792 Aug 1998 JP
11069211 Mar 1999 JP
11078693 Mar 1999 JP
H1178737 Mar 1999 JP
11250228 Sep 1999 JP
H11259634 Sep 1999 JP
2000127849 May 2000 JP
2001001832 Jan 2001 JP
2001092970 Apr 2001 JP
2001180401 Jul 2001 JP
2002022439 Jan 2002 JP
2002046506 Feb 2002 JP
200274339 Mar 2002 JP
2002079895 Mar 2002 JP
2002099908 Apr 2002 JP
2002109699 Apr 2002 JP
2002211428 Jul 2002 JP
2003083742 Mar 2003 JP
2003123058 Apr 2003 JP
2003150938 May 2003 JP
2003168197 Jun 2003 JP
2003178397 Jun 2003 JP
2003217099 Jul 2003 JP
2003259361 Sep 2003 JP
2003281700 Oct 2003 JP
20041658 Jan 2004 JP
2004032460 Jan 2004 JP
2004146904 May 2004 JP
336535 Apr 1970 SE
WO8605147 Sep 1986 WO
WO8809023 Nov 1988 WO
WO9004528 May 1990 WO
WO9300647 Jan 1993 WO
WO9310550 May 1993 WO
WO9321596 Oct 1993 WO
WO9419212 Feb 1994 WO
WO 9412368 Jun 1994 WO
WO 9419212 Sep 1994 WO
WO 9427262 Nov 1994 WO
WO9518979 Jul 1995 WO
WO9523082 Aug 1995 WO
WO9602817 Feb 1996 WO
WO 9603475 Feb 1996 WO
WO9621581 Jul 1996 WO
WO 9621581 Jul 1996 WO
9638319 Dec 1996 WO
WO9701246 Jan 1997 WO
WO9721127 Jun 1997 WO
WO9729926 Aug 1997 WO
WO 9734186 Sep 1997 WO
9735743 Oct 1997 WO
WO9748134 Dec 1997 WO
WO 9748134 Dec 1997 WO
WO 9838547 Sep 1998 WO
WO9843850 Oct 1998 WO
WO 9844384 Oct 1998 WO
WO 9844385 Oct 1998 WO
WO 9844386 Oct 1998 WO
WO 9914943 Mar 1999 WO
WO9943242 Sep 1999 WO
WO 9945081 Sep 1999 WO
WO9947396 Sep 1999 WO
WO 0011723 Mar 2000 WO
WO 0015462 Mar 2000 WO
WO 0017009 Mar 2000 WO
WO 0017702 Mar 2000 WO
WO0015462 Mar 2000 WO
WO 0022471 Apr 2000 WO
WO 0018612 Apr 2000 WO
WO0022881 Apr 2000 WO
WO 0033134 Jun 2000 WO
WO 0055685 Sep 2000 WO
WO 0055685 Sep 2000 WO
WO 0066679 Nov 2000 WO
WO0164481 Feb 2001 WO
WO 0164464 Sep 2001 WO
WO 0164481 Sep 2001 WO
WO0126332 Dec 2001 WO
WO 02062623 Aug 2002 WO
WO02071487 Sep 2002 WO
WO 03065084 Aug 2003 WO
WO 03079318 Sep 2003 WO
WO03093857 Nov 2003 WO
WO2004005073 Jan 2004 WO
WO 2004058540 Jul 2004 WO
2005019873 Mar 2005 WO
WO 2005024500 Mar 2005 WO
WO 2005045481 May 2005 WO
WO 2005050267 Jun 2005 WO
WO 2005071646 Aug 2005 WO
WO 2005082015 Sep 2005 WO
2007081407 Jul 2007 WO
WO 2007103573 Sep 2007 WO
2007111984 Oct 2007 WO
2009073054 Jun 2009 WO
Non-Patent Literature Citations (213)
Entry
Stewart, James W.; HP SnapLED: LED Assemblies for Automotive Signal Applications; Nov. 1, 1998; Hewlett-Packard Journal; vol. 50, No. 1, www.hpl.hp.com/hpjournal/98nov/nov98al.pdf.
Edgar, Julian; Goodbye 12 Volts . . . Hello 42 Volts!; Oct. 5, 1999; Autospeed 50; Issue 50; www.autospeed.co.nz/cms/A—0319/article.html.
Kobe, Gerry; 42 Volts Goes Underhood; Mar. 2000; Automotive Industries; Cahners Publishing Company; www.findarticles.com/p/articles/mi—m3012/is—3—180/ai—61361677.
Jewett, Dale; Aug. 2000; Automotive Industries; Cahners Publising Company; www.findarticles.com/p/articles/mi—m3012/is—8—180ai—64341779.
Dana H. Ballard and Christopher M. Brown, Computer Vision, article, 4 pages Prentice-Hall, Englewood Cliffs, New Jersey, believed to be published more than one year prior to the filed of the present application.
G. Wang, D. Renshaw, P.B. Denyer and M. Lu, CMOS Video Cameras, article, 1991, 4 pages, University of Edinburgh, UK.
National Semiconductor, LM78S40, Universal Switching Regulator Subsystem, National Semiconductor Corporation, Apr. 1996, p. 6.
N.R. Lynam, “Electrochromic Automotive Day/Night Mirror,” SAE Technical Paper Series, 870636 (1987).
N.R. Lynam, “Smart Windows for Automobiles,” SAE Technical Paper Series, 900419 (1990).
N.R. Lynam and A. Agrawal, “Automotive Applications of Chromogenic Materials,” from Large Area Chromogenics: Materials and Devices for Transmittance Control, C.M. Lampert and C.G. Granquist, EDS, Optical Engineering Press, Washington (1990).
Kobe, Gerry, “Hypnotic Wizardry! (interior electronics),” Automotive Industries, vol. 169, No. 5, p. 60, published May 1989. Relevant section is entitled “Instrumentation.”
SAE Information Report, “Vision Factors Considerations in Rear View Mirror Design—SAE J985 Oct. 88,” approved Oct. 1988, and located in 1995 SAE Handbook, vol. 3.
Denyer, Peter B.; Renshaw, David; Wang, Gouyu; Lu, Ming Ying; and Anderson, Stuart, On-Chip CMOS Sensors for VLSI Imaging Systems, pp. 4b1.1-4b1.5.
Improved Obstacle Detection by Sensor Fusion, p. 1.
Article entitled “Generation of Vision Technology,” published by VLSI Vision Limited, publication date unknown.
R.H. Tribe, Intelligent Autonomous Systems for Cars, Advanced Robotics and Intelligent Machines, Peter Peregrinus (Nov. 1994).
R.H. Tribe et al., Collision Warning, Autotech '93, Seminar 9, NEC Birmingham, UK (Nov. 1993).
M.A. Arain et al., Action planning for the collision avoidance system using neural networks, Intelligent Vehicle Symposium, Tokyo, Japan (Jul. 1993).
P.E. An et al., Aspects of Neural Networks in Intelligent Collision Avoidance Systems for Prometheus, JFIT 93, pp. 129-135 (Mar. 1993).
R. Page et al., Advanced technologies for collision avoidance, Eureka on Campus (Summer 1992).
S. Kidd et al., Speed Over Ground Measurement, SAE Technical Paper Series, No. 910272, pp. 29-36 (Feb.-Mar. 1991).
Andreas Kuehnle, Symmetry-based recognition of vehicle rears, Pattern Recognition Letters 12, pp. 249-258 (1991).
M.A. Arain et al., Application of Neural Networks for Traffic Scenario Identification, 4th Prometheus Workshop, University of Compiegne, Paris, France, pp. 102-111 (Sep. 1990).
R.H. Tribe, Collision Avoidance, Advances, Issue No. 4, May 1990.
R.H. Tribe et al., Collision Avoidance, Lucas International Symposium, Paris, France (1989).
S.J. Pacaud et al., Ground Speed Sensing, Lucas International Symposium, Paris, France (1989).
Bow, Sing T., “Pattern Recognition and Image Preprocessing (Signal Processing and Communications)”, CRC Press, Jan. 15, 2002, pp. 557-559.
J. Borenstein et al., “Where am I? Sensors and Method for Mobile Robot Positioning”, University of Michigan, Apr. 1996, pp. 2, 125-128.
Pratt, “Digital Image Processing, Passage—ED.3”, John Wiley & Sons, US, Jan. 1, 2001, pp. 657-659, XP002529771.
Reexamination Control No. 90/007,519, dated Jun. 9, 2005, Reexamination of U.S. Patent No. 6,222,447, issued to Schofield et al.
Reexamination Control No. 90/007,520, dated Jun. 9, 2005, Reexamination of U.S. Patent No. 5,949,331, issued to Schofield et al.
Reexamination Control No. 90/011,477, dated Mar. 14, 2011, Reexamination of U.S. Patent No. 5,949,331, issued to Schofield et al.
Reexamination Control No. 90/011,478, dated Mar. 28, 2011, Reexamination of U.S. Patent No. 6,222,447, issued to Schofield et al.
Tokimaru et al., “CMOS Rear-View TV System with CCD Camera”, National Technical Report vol. 34, No. 3, pp. 329-336, Jun. 1988 (Japan).
Van Leeuwen et al., “Motion Estimation with a Mobile Camera for Traffic Applications”, IEEE, US, vol. 1, Oct. 3, 2000, pp. 58-63.
Van Leeuwen et al., “Motion Interpretation for In-Car Vision Systems”, IEEE, US, vol. 1, Sep. 30, 2002, p. 135-140.
Van Leeuwen et al., “Requirements for Motion Estimation in Image Sequences for Traffic Applications”, IEEE, US, vol. 1, May 24, 1999, pp. 145-150, XP010340272.
Van Leuven et al., “Real-Time Vehicle Tracking in Image Sequences”, IEEE, US, vol. 3, May 21, 2001, pp. 2049-2054, XP010547308.
Vlacic et al., (Eds), “Intelligent Vehicle Tecnologies, Theory and Applications”, Society of Automotive Engineers Inc., edited by SAE International, 2001.
Jul. 1, 1992; Schwarzinger et al.; “Vision-based car-following: detection, tracking, and identification”.
Sep. 26, 1995; Jochem et al.; “PANS: a portable navigation platform”.
Nov. 16, 1995; Taktak et al.; “Vehicle detection at night using image processing and pattern recognition”.
Oct. 1995; Ritter et al.;“Traffic sign recognition using colour information”.
Kuehnle; Apr. 1991; Symmetry-based recognition of vehicle rears.
Mar. 1996; Dubuisson-Jolly ;“Vehicle segmentation and classification using deformable templates”.
Dec. 8, 1993; Mendis et al.; “A 128×128 CMOS active pixel image sensor for highly integrated imaging systems”.
Oct. 1994; Thomanek et al.; “Multiple object recognition and scene interpretation for autonomous road vehicle guidance”.
1995; Vosselman et al.; “Road tracking by profile matching and Kalman filtering”.
Feb. 1992; Dickmanns et al.; “Recursive 3-D road and relative ego-state recognition”.
1991; Dickmanns; “4-D dynamic vision for intelligent motion control”.
Dec. 1990; Dickmanns et al.; “An integrated spatio-temporal approach to automatic visual guidance of autonomous vehicles”.
Oct. 11, 1995; Fossum; “Low power camera-on-a-chip using CMOS active pixel sensor technology”.
Mar. 14, 1995; Auty et al.; “Image acquisition system for traffic monitoring applications”.
Mar. 1996; Barron et al.;“The role of electronic controls for future automotive mechatronic systems”.
Feb. 14, 1992; Cassiano et al.; “Review of filtering methods in mobile vision from ground vehicles in low light conditions”.
1992; Huijsing; “Integrated smart sensors”.
Nov. 16, 1994 ; Lee et al.; “Automatic recognition of a car license plate using color image processing”.
Sep. 1995; Malik et al.; “A Machine Vision Based System for Guiding Lane-change Maneuvers”.
Jul. 1, 1992; Nguyen et al.; “Obstacle detection using bi-spectrum CCD camera and image processing”.
1995; Nixon et al.; “128×128 CMOS Photodiode-Type Active Pixel Sensor With On-Chip Timing, Control and Signal Chain Electronics”.
Raglan Tribe Video—1994 ; Raglan Tribe; “Robot Car Raglan Tribe” http://www.youtube.com/watch?v=AlLZhcnpXYI.
Eureka Prometheus Project, 1987-1995, and any corresponding publications and public demonstrations.
Lucas Demonstrates Intelligent Cruise Control, Detroit Feb. 27, 1995 available at http://www.thefreelibrary.com/LUCAS+DEMONSTRATES+INTELLIGENT+CUISE+CONTR OL=a016602459.
Siegle, G. “Autonomous Driving on a Road Network,” Proceedings of the Intelligent Vehicles '92 Symposium Detroit, Michigan, ISBN 0-7803-0747-X; Jun. 29-Jul. 1, 1992.
May 11, 1992; Smith et al.; “Optical sensors for automotive applications”.
Oct. 5, 1994; Smith et al.; “Vision sensing for intelligent vehicle and highway systems”.
May 19, 1994; Vellacott; “CMOS in camera”.
1995; Yu et al.; “Vehicles Recognition by Video Camera”.
Oct. 1992; Sukthankar; “RACCOON: A Real-time Autonomous Car Chaser Operating Optimally at Night”.
IEEE Paper 1988; Thorpe, at al, Vision and Navigation for the Carnegie-Mellon Navlab.
IEEE Paper 1988; Turk, et al., “VITS—A Vision System for Autonomous Land Vehicle Navigation”.
Paper from Society of Automotive Engineers, Inc. 1988; Metzler, “Computer Vision Applied to Vehicle Operation”.
IEE Paper 1991; Crisman and Thorpe; “UNSCARF, A Color Vision System for the Detection of Unstructured Roads”.
John Oldenburg; “Comments on the Autronic Eye”.
Technical Report AIP-77 Department of Psychology, Carnegie Mellon University; Mar. 13, 1990; Pomerleau; “Alvinn: An Autonomous Land Vehicle in a Neural Network”.
Roberts, J.M., “Attentive Visual Tracking and Trajectory Estimation for Dynamic Scene Segmentation,” Dec. 1994, University of Southampton, PhD submission.
Matthews, N.D., “Visual Collision Avoidance,” Oct. 1994, University of Southampton, PhD submission.
Manigel, J. “Vehicle control by computer vision,” Industrial Electronics, IEEE Transactions on, Jun. 1992 v. 39 issue 3, 181-188.
Nguyen, HG et al., “Obstacle detection using bi-spectrum CCD camera and image processing”, Proceedings of the Intelligent Vehicles '92 Symposium, Jun. 29-Jul. 1, 1992, p. 42-50.
Industrial Electronics, Control, Instrumentation, and Automation, 1992. Power Electronics and Motion Control., Proceedings of the 1992 International Conference on Date of Conference: Nov. 9-13, 1992.
Ichiro Masaki , Vision-based vehicle guidance, Springer-Verlag, New York, 1992.
J Manigel, W Leonhard, “Computer control of an autonomous road vehicle by computer vision”—Industrial Electronics, Control and Instrumentation, 1991. Proceedings. IECON '91., 1991 International Conference on, p. 19-24 vol. 1.
Linkwitz, K., High Precision Navigation: Integration of Navigational and Geodetic Methods, Springer-Verlag, Jul. 5, 1989.
Yu, X., Road tracking, lane segmentation and obstacle recognition by mathematical morphology, Intelligent Vehicles '92 Symposium, Proceedings of the, p. 166-172.
Donnelly Panoramic Vision™ on Renault Talisman Concept Car at Frankfort Motor Show, PR Newswire, Frankfort, Germany Sep. 10, 2001.
Achler et al., “Vehicle Wheel Detector using 2D Filter Banks,” IEEE Intelligent Vehicles Symposium of Jun. 2004.
Behringer et al., “Simultaneous Estimation of Pitch Angle and Lane Width from the Video Image of a Marked Road,” pp. 966-973, Sep. 12-16, 1994.
Broggi et al., “Automatic Vehicle Guidance: The Experience of the ARGO Vehicle”, World Scientific Publishing Co., 1999.
Broggi et al., “Multi-Resolution Vehicle Detection using Artificial Vision,” IEEE Intelligent Vehicles Symposium of Jun. 2004.
Broggi, Robust Real-time Lane and Road Detection in Critical Shadow Conditions, International Symposium on Computer Vision, 21-23.
Brown, A Survey of Image Registration Techniques, vol. 24, ACM Computing Surveys, pp. 325-376, 1992.
Burger et al., “Estimating 3-D Egomotion from Perspective Image Sequences”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 12, No. 11, pp. 1040-1058, Nov. 1990.
Burt et al., A Multiresolution Spline with Application to Image Mosaics, ACM Transactions on Graphics, vol. 2. No. 4, pp. 217-236, Oct. 1983.
Chien, et. al. “Efficient moving object segmentation algorithm using background registration technique”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 12.
Crisman et al., “UNSCARF, A Color Vision System for the Detection of Unstructured Roads” IEEE Paper 1991.
Cucchiara et al., Vehicle Detection under Day and Night Illumination, 1999.
Cucchiara, et al., “Detecting moving objects, ghosts, and shadows in video streams”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, No. 10.
Cucchiara, et al., “Improving Shadow Suppression in Moving Object Detection with HSV Color Information”, Proceeding of IEEE International Conference on Intelligent Transportation Systems.
De la Escalera et al, 1994 “Neural traffic sign recognition for autonomous vehicles”.
DeFauw, A System for Small Target Detection, Tracking, and Classification, Intelligent Transportation System.
Dickmanns et al., “A Curvature-based Scheme for Improving Road Vehicle Guidance by Computer Vision,” University of Bundeswehr München, 1986.
Franke et al., “Autonomous driving approaches downtown”, Intelligent Systems and Their Applications, IEEE 13 (6), 40-48, Nov./Dec. 1999.
Fung. et al. “Effective moving cast shadow detection for monocular color traffic image sequences”, Optical Engineering, vol. 41, No. 6.
Greene et al., Creating Raster Omnimax Images from Multiple Perspective Views Using the Elliptical Weighted Average Filter, IEEE Computer Graphics and Applications, vol. 6, No. 6, pp. 21-27, Jun. 1986.
Haritaoglu, et al., “W4: Real-Time Surveillance of People and Their Activities”, IEEE Transactions Patter Analysis and Machine Intelligence, vol. 22, No. 8.
Honda Worldwide, “Honda Announces a Full Model Change for the Inspire.” Jun. 18, 2003.
Horprasert, et al. “A Statistical Approach for Real-Time Robust Background Subtraction and Shadow Detection”, Proceeding of IEEE International Conference on Computer vision FRAME—RATE Workshop.
Hsieh, et al. “A shadow elimination method for vehicle analysis”, Proceeding of IEEE International Conference on Pattern Recognition, vol. 4.
Hsieh, et al. “Shadow elimination for effective moving object detection by Gaussian shadow modeling”, Image and Vision Computing, vol. 21.
IEEE Paper 1987; McTamaney, “Mobile Robots Real-Time Intelligent Control”.
IEEE Paper 1991; Thorpe, at al, “Toward autonomous driving: the CMU Navlab. I. Perception”.
IEEE Paper 1993, Charkari et al., “A new approach for real time moving vehicle detection”.
Kan et al., “Model-based vehicle tracking from image sequences with an application to road surveillance,” Purdue University, XP000630885, vol. 35, No. 6, Jun. 1996.
Kang et al., High Dynamic Range Video, ACM Transactions on Graphics, vol. 22.
Kastrinaki et al., “A survey of video processing techniques for traffic applications”.
Kehtarnavaz et al, 1995, “Traffic sign recognition in noisy outdoor scenes”.
Kluge et al., “Representation and Recovery of Road Geometry in YARF,” Carnegie Mellon University, pp. 114-119.
Koller et al., “Binocular Stereopsis and Lane Marker Flow for Vehicle Navigation: Lateral and Longitudinal Control,” University of California, Mar. 24, 1994.
Kuhnert, “A vision system for real time road and object recognition for vehicle guidance,” in Proc. SPIE Mobile Robot Conf, Cambridge, MA, Oct. 1986, pp. 267-272.
Martel-Brisson and Zaccarin, “Moving cast shadow detection from a Gaussian mixture shadow model”, Proceeding of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2.
McKenna, et al., “Tracking Groups of People”, Computer Vision and Image Understanding, vol. 80.
Mei Chen et al., AURORA: A Vision-Based Roadway Departure Warning System, The Robotics Institute, Carnegie Mellon University, published Aug. 9, 1995.
Mikic, et al., “Moving shadow and object detection in traffic scenes”, Proceeding of IEEE International Conference on Pattern Recognition, vol. 1.
Morgan et al., “Road edge tracking for robot road following: a real-time implementation,” vol. 8, No. 3, Aug. 1990.
Nadimi and Bhanu,. “Physical models for moving shadow and object detection in video”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26.
Nathan, Digital Video Data Handling, NASA JPL Tech Report 32-877, Pasadena, CA, Jan. 5, 1966.
Parker (ed.), McGraw-Hill Dictionary of Scientific and Technical Terms Fifth Edition (1993).
Philomin et al., “Pedestrain Tracking from a Moving Vehicle”.
Piccioli et al, 1994 “Robust road sign detection and recognition from image sequences”.
Porter et al., “Compositing Digital Images,” Computer Graphics (Proc. Siggraph), vol. 18, No. 3, pp. 253-259, Jul. 1984.
Prati, et al. “Detecting moving shadows: algorithms and evaluation”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25.
Proceedings of the Intelligent Vehicles Symposium, 1992-present.
Sahli et al., “A Kalman Filter-Based Update Scheme for Road Following,” IAPR Workshop on Machine Vision Applications, pp. 5-9, Nov. 12-14, 1996.
Salvador, et al., “Cast shadow segmentation using invariant color features”, Computer Vision and Image Understanding, vol. 95.
Shirai, 1985, “Future Generation Computer Systems”.
Siala, et al., “Moving shadow detection with support vector domain description in the color ratios space”, Proceeding of IEEE International Conference on Pattern Recognition. vol. 4.
Stander, et a 1., “Detection of moving cast shadows for object segmentation”, IEEE Transactions on Multimedia, vol. 1.
Stengel, et al., 1989, “Intelligent Guidance for Headway and Lane Control”.
Sun et al., “On-road vehicle detection using optical sensors: a review”.
Szeliski, Image Mosaicing for Tele-Reality Applications, DEC Cambridge Research Laboratory, CRL 94/2, May 1994.
Thongkamwitoon, and, Chalidabhongse, “An adaptive real-time background subtraction and moving shadows detection”, Proceeding of IEEE International Conference on Multimedia and Expo. vol. 2.
Thorpe, Vision and Navigation: The Carnegie Mellon Navlab (The Springer International Series in Engineering and Computer Science).
Toth, et al., “Detection of moving shadows using mean shift clustering and a significance test”, Proceeding of IEEE International Conference on Pattern Recognition, vol. 4.
Toyota Motor Corporation, “Present and future of safety technology development at Toyota.” 2004.
Tsugawa et al., “An automobile with artificial intelligence,” in Proc. Sixth IJCAI, 1979.
Tzomakas and von Seelen, Vehicle Detection in Traffic Scenes Using Shadows, Internal report, Institut Für Neuroinformatik Bochum, Internal Report 98-06.
Van Leeuwen et al., “Requirements for Motion Estimation in Image Sequences for Traffic Applications”, IEEE, 2000, pp. 354-359, XP002529773.
Vellacott, Oliver, “CMOS in Camera,” IEE Review, pp. 111-114 (May 1994).
Wang, et al., “A probabilistic method for foreground and shadow segmentation”, Proceeding of IEEE International Conference on Image Processing, vol. 3.
Wolberg, “A Two-Pass Mesh Warping Implementation of Morphing,” Dr. Dobb's Journal, No. 202, Jul. 1993.
Wolberg, Digital Image Warping, IEEE Computer Society Press, 1990.
Xu, et al., “Cast shadow detection in video segmentation”, Pattern Recognition Letters, vol. 26.
Yeh et al., 1994, “Image-Based Dynamic Measurement for Vehicle Steering Control”.
Yoneyama, et al., “Moving cast shadow elimination for robust vehicle extraction based on 2D joint vehicle/shadow models”, Proceeding of IEEE International Conference on Advanced Video and Signal Based Surveillance.
Yoneyama, et al., “Robust vehicle and traffic information extraction for highway surveillance”, EURASIP Journal on Applied Signal Processing. vol. 2005.
Young et al., Cantata: Visual Programming Environment for the Khoros System, ACM SIGGRAPH Computer Graphics-Special focus: modular visualization environments (MVEs), vol. 29, issue 2.
Zheng et al., “An Adaptive System for Traffic Sign Recognition,” IEEE Proceedings of the Intelligent Vehicles '94 Symposium, pp. 165-170 (Oct. 1994).
Amidi, O., “Integrated Mobile Robot Control”, M.S. Thesis, Carnegie Mellon University, May 1990.
Begault, D.R., “Head-Up Auditory Displays for Traffic Collision Avoidance System Advisories: A Preliminary Investigation,” Human Factors, 35(4), Dec. 1993, pp. 707-717.
Beucher et al.; “Road Segmentation and Obstacle Detection by a Fast Watershed Transformation”.
Brauckmann et al.; “Towards All Around Automatic Visual Obstacle Sensing for Cars”.
Carnegie Mellon NAVLAB (including public use, demonstrations, and publications related to NAVLAB vehicles NAVLAB 1-NAVLAB 11).
Chrysler Automatic Beam control, including public use and sales of vehicles including such systems, and associated advertisements and service manuals., 1960-1970.
Crisman, J., et al.; Carnegie Mellon University; “Vision and Navigation—The Carnegie Mellon Navlab” edited by Charles E. Thorpe, 1990.
Davis, L. S., Kushner, T. R., Le Moigne, J. J. and Waxman, A. M., “Road Boundary Detection for Autonomous Vehicle Navigation,” Optical Engineering, vol. 25, No. 3, Mar. 1986, pp. 409-414.
Dérutin et al.; “Real-time collision avoidance at road-crossings on board the Prometheus—ProLab 2 vehicle”.
Dickmanns, E. et al.; “The seeing passenger car ‘VaMoRs-P’”, Oct. 24, 1994.
“Final Report of the Working Group on Advanced Vehicle Control Systems (AVCS) Mobility 2000,” Mar. 1990.
General Motors Autotronic Eye, including public use and sales of vehicles including such systems, and associated advertisements and service manuals., 1953-1988.
Hess, R.A. and Modjtahedzadeh, A., “A Control Theoretic Model of Driver Steering Behavior,” IEEE Control Systems Magazine, vol. 10, No. 5, Aug. 1990, pp. 3-8.
Hessburg, T., Peng, H., Masayoshi, T., Zhang, W. B., “An Experimental Study on Lateral Control of a Vehicle,” California Partners for Advanced Transit and Highways (PATH), Jan. 1, 1991.
IEEE Paper 1993; Schönfeld et al.; Compact Hardware Realization for Hough Based Extraction of Line Segments in Image Sequences for Vehicle Guidance, 1993.
IEEE Paper 1993; Xie et al., Active and Intelligent Sensing of Road Obstacles: Application to the European Eureka—PROMETHEUS Project, 1993.
IEEE Paper 1994; Rombaut et al.; “Dynamic data temporal multisensory fusion in the Prometheus ProLab2 demonstrator”, 1994.
Juberts, M., Murphy, K., Nashman, M., Scheiderman, H., Scott, H., and Szabo, S., “Development and Test Results for a Vision-Based Approach to AVCS.” in Proceedings of the 26th International Symposium on Automotive Technology and Automation, Aachen, Germany, Sep. 1993, pp. 1-9.
Kenue, S.K, “Lanelok: Detection of Lane Boundaries and Vehicle Tracking Using Image-Processing Techniques,” SPIE Conference on Mobile Robots IV, 1989.
Knipling, R., “IVHS Technologies Applied to Collision Avoidance: Perspectives on Six Target Crash Types and Countermeasures,” Technical Paper presented at Safety & Human Factors session of 1993 IVHS America Annual Meeting, Apr. 14-17, 1993, pp. 1-22.
Knipling, R.R., Wierwille, W.W, “Vehicle-Based Drowsy Driver Detection: Current Status and Future Prospects,” IVHS America Fourth Annual Meeting, Atlanta, GA, Apr. 17-20, 1994, pp. 1-24.
Litkouhi, B.B., Lee, A.Y. and Craig, D.B., “Estimator and Controller Design for LaneTrak, a Vision-Based Automatic Vehicle Steering System,” Proceedings of the 32nd Conference on Decision and Control, San Antonio, Texas, Dec. 1993, pp. 1868-1873.
Lumia, J.; “Mobile system for measuring retroreflectance of traffic signs”, Mar. 1, 1991.
Lutz et al.; “New Results on Traffic Sign Recognition”.
Manigel, J. and Leonhard, W., “Computer Control for an Autonomous Road Vehicle by Computer Vision,” 1991 International Conference on Industrial Electronics, Control and Instrumentation, vol. 1, Oct. 28-Nov. 1, 1991, pp. 19-24.
Manigel, J. and Leonhard, W., “Vehicle Control by Computer Vision,” IEEE Transactions on Industrial Electronics, vol. 39, No. 3, Jun. 1992, pp. 181-188.
Mironer, M., Hendricks, D., “Examination of Single Vehicle Roadway Departure Crashes and Potential IVHS Countermeasures,” U.S. Department of Transportation, Aug. 1994.
NAHSC TV Coverage Ree http://youtu.be/6sDmA-sK3aE.
Najm, W.; “Comparison of alternative crash-avoidance sensor technologies”, Jan. 6, 1995.
Nashman, M., and Schneiderman, H., “Real-time Visual Processing for Autonomous Driving,” in Proceedings of the IEEE Intelligent Vehicles, vol. 93, Jun. 1993, pp. 14-16.
NavLab 1984-1994 http://youtu.be/5-acCtyKf7E.
Navlab 5—1997—Driving on highways around Pittsburgh http://youtu.be/xkJVV1—4I8E.
Navlab 90 http://youtu.be/0GXuqw3cgwU.
Navlab 96 http://youtu.be/Tat70DqpKw8.
Navlab 97 http://youtu.be/rV0H7u6tmlk.
Navlab on CNN http://youtu.be/bdQ5rsVgPuk.
Navlab on KDKA http://youtu.be/laolqVMd6tc.
Navlab on Scientific American Frontiers http://youtu.be/507-gmGsq1o.
Peng, H., “Vehicle Lateral Control for Highway Automation,” Ph.D. Thesis—University of California Berkeley, 1992.
Peng, H., Zhang, W. B., Arai, A., Lin, Y., Hessburg, T., Devlin, P., Masayoshi, T., Shladover, S., “Experimental Automatic Lateral Control System for an Automobile,” California Partners for Advanced Transit and Highways (PATH), Jan. 1, 1992.
Raboisson et al.; “Obstacle Detection in Highway Environment by Colour CCD Camera and Image Processing Prototype Installed in a Vehicle”.
Regensburger, Uwe and Volker Graef; “Visual Recognition of Obstacles on Roads”, 1994.
Run-Off-Road Project Final Video http://youtu.be/nzxfMzVmnhM.
SAE Technical Paper 2001-01-0072—LIN.
SafeTRAC Lane Departure Warning http://youtu.be/yPHkQfyO15g.
Schneiderman, H. and Nashman, M., “Visual Processing for Autonomous Driving,” IEEE Workshop on Applications of Computer Vision, Palm Springs, CA., Nov. 30-Dec. 2, 1992.
Schumann, J., Godthelp, J. and Hoekstra, W.H., “An Exploratory Simulator Study on the Use of Active Control Devices in Car Driving,” No. IZF-1992-B-2. Institute for Perception RVO-TNO Soesterberg (Netherlands), May 1992.
Shimizu, S. et al.; “A moving image processing system for personal vehicle system”, Nov. 9, 1992.
Schladover, S.E., “Highway Electrification and Automation,” California Partners for Advanced Transit and Highways (PATH), Jan. 1, 1992.
Shladover, S.E., “Research and Development Needs for Advanced Vehicle Control Systems,” Micro, IEEE , vol. 13, No. 1, Feb. 1993, pp. 11-19.
Shladover, S.E., Desoer, C.A., Hendrick, J.K., Tomizuka, M., Walrand, J., Zhang, W., McMahon, D.H., Peng, H., Sheikholeslam, S., McKeown, N., “Automatic Vehicle Control Developments in the Path Program,” IEEE Transaction on Vehicular Technology, vol. 40, No. 1, Feb. 1991, pp. 114-130.
Tsugawa, S. et al.; “Vision-based vehicles in japan; machine vision systems and driving control systems”, Aug. 1994.
Twilight Sentinel headlamp control, including public use and sales of vehicles including such systems, and associated advertisements and service manuals., 1960-1998.
Wallace, R., Matsuzaki, K., Goto, Y., Crisman, J., Webb, J., Kanade, T., “Progress in Robot Road-Following,” Proceedings of the 1986 IEEE International Conference on Robotics and Automation, vol. 3, 1986, pp. 1615-1621.
Wan et al.; “A New Edge Detector for Obstacle Detection with a Linear Stereo Vision System”.
Wiles, C.; “Science: Next, the cars that can see in the dark Charles Wiles on a battlefield camera that works with a pilot's aid to help reduce road accidents”, Oct. 20, 1993.
Related Publications (1)
Number Date Country
20100118146 A1 May 2010 US
Continuations (4)
Number Date Country
Parent 12496357 Jul 2009 US
Child 12688146 US
Parent 11122880 May 2005 US
Child 12496357 US
Parent 10324679 Dec 2002 US
Child 11122880 US
Parent 08952026 US
Child 10324679 US