Vehicle imaging system and method for distinguishing between vehicle tail lights and flashing red stop lights

Information

  • Patent Grant
  • 9619720
  • Patent Number
    9,619,720
  • Date Filed
    Tuesday, August 19, 2014
    11 years ago
  • Date Issued
    Tuesday, April 11, 2017
    8 years ago
Abstract
An imaging system for a vehicle is provided for distinguishing between tail lights of another vehicle and a flashing red stop light. The system includes an imager configured to image a forward external scene and to generate image data corresponding to the acquired images; and a processor configured to receive and analyze the image data to identify red light sources and to further analyze each red light source to determine if the red light source is detected for a predetermined time period. If the red light source is not detected within a predetermined time period after it is detected, the processor determines that the red light source is a flashing red stop light. Otherwise, if the red light source is detected for a predetermined time period, the processor determines that the red light source may be a tail light of another vehicle.
Description
FIELD OF THE INVENTION

The present invention generally relates to a system and method for imaging and detecting objects in a scene forward of a vehicle.


SUMMARY OF THE INVENTION

According to one aspect of the present invention, an imaging system is provided for a vehicle. The imaging system comprises: an imager configured to image a scene external and forward of the vehicle and to generate image data corresponding to the acquired images; and a processor configured to receive and analyze the image data to identify red light sources and further to analyze each red light source to determine if the red light source is detected for a predetermined time period, wherein if the red light source is not detected within a predetermined time period after it is detected, the processor determines that the red light source is a flashing red stop light, otherwise if the red light source is detected for a predetermined time period, the processor determines that the red light source may be a tail light of another vehicle.


According to another aspect of the present invention, a method is provided for distinguishing between tail lights of another vehicle and a flashing red stop light. The method comprises the steps of: providing an imager for use in a vehicle; imaging a scene external and forward of the vehicle and generating image data corresponding to the acquired images; receiving and analyzing the image data in a processor to identify red light sources and further analyze each red light source to determine if the red light source is detected for a predetermined time period; if the red light source is not detected within a predetermined time period after it is detected, determining that the red light source is a flashing red stop light; and otherwise if the red light source is detected for a predetermined time period, determining that the red light source may be a tail light of another vehicle.


According to another aspect of the present invention, a non-transitory tangible computer readable medium is provided having stored thereon software instructions that, when executed by a processor within a vehicle, cause the processor to distinguish between a tail light of another vehicle and a flashing red stop light by executing the steps comprising: controlling an imager to image a scene external and forward of the vehicle and generating image data corresponding to the acquired images; receiving and analyzing the image data in the processor to identify red light sources and further analyze each red light source to determine if the red light source is detected for a predetermined time period; if the red light source is not detected within a predetermined time period after it is detected, determining that the red light source is a flashing red stop light; and otherwise if the red light source is detected for a predetermined time period, determining that the red light source may be a tail light of another vehicle.


These and other features, advantages, and objects of the present invention will be further understood and appreciated by those skilled in the art by reference to the following specification, claims, and appended drawings.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will be more fully understood from the detailed description and the accompanying drawings, wherein:



FIG. 1 is a block diagram of a system constructed according to one embodiment;



FIG. 2 is a partial cross section of a rearview assembly incorporating the system of FIG. 1;



FIG. 3 is a flow chart illustrating the steps of a method for distinguishing between vehicle tail lights of another vehicle and a flashing red stop lamp;



FIG. 4A is a graphic illustration of an example of red light intensity that may be detected over a number of frames as the controlled vehicle approaches a flashing red stop light; and



FIG. 4B is a graphic illustration of the state of the high beam headlamps as changed in response to the detected red light intensity shown in the example in FIG. 4A.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Reference will now be made in detail to the present preferred embodiments, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numerals will be used throughout the drawings to refer to the same or like parts. In the drawings, the depicted structural elements are not to scale and certain components are enlarged relative to the other components for purposes of emphasis and understanding.


The embodiments described herein relate to an imaging system for a vehicle and a method of detecting and/or categorizing objects in a scene forward of the vehicle. To assist in the understanding of an application of these embodiments, examples are provided that pertain to the use of the imaging system in an exterior light control system for controlling exterior lights of a controlled vehicle in response to image data acquired from an image sensor, which captures images forward of the vehicle. Prior systems are known for controlling exterior vehicle lights in response to images captured forward of the vehicle. In these prior systems, a controller would analyze the captured images and determine if any preceding or oncoming vehicles were present in a glare area in front of the vehicle employing the system. This “glare area” was the area in which the exterior lights would cause excessive glare to a driver if the exterior lights were in a high-beam state (or some state other than a low-beam state). If a vehicle was present in the glare area, the controller would respond by changing the state of the exterior lights so as to not cause glare for the other driver(s). Examples of such systems are described in U.S. Pat. Nos. 5,837,994, 5,990,469, 6,008,486, 6,049,171, 6,130,421, 6,130,448, 6,166,698, 6,379,013, 6,403,942, 6,587,573, 6,611,610, 6,631,316, 6,774,988, 6,861,809, 7,321,112, 7,417,221, 7,565,006, 7,567,291, 7,653,215, 7,683,326, 7,881,839, 8,045,760, 8,120,652, and 8,543,254, the entire disclosures of which are incorporated herein by reference.


One of the challenges for such systems is to distinguish lights from other vehicles and nonvehicle light sources. If the system were unable to accurately make such distinctions, it would respond to detection of a nonvehicle light source in the same way it would respond to an oncoming or leading vehicle and would reduce its lighting level when it does not need to be reduced. Such unnecessary reduction in lighting levels can be a distraction and a nuisance to the driver. Current systems analyze a number of different characteristics of each light source including brightness, size, shape, location, motion, AC modulation, etc. to compute a probability that a light source is a headlight or tail light of another vehicle.


One particular nonvehicle light source that current systems sometimes do not readily distinguish from vehicle tail lights is a flashing red stop light. As a result, the system will often turn off high beams or otherwise reduce illumination levels upon detecting a flashing red stop light at a considerably greater distance than is necessary. Accordingly, the embodiments described herein are provided to better distinguish a flashing red stoplight from vehicle tail lights such that the system will not reduce the illumination level of the exterior vehicle lights until the vehicle is within 100 meters of the stop light or even within 50 meters of the stop light.


To detect that a light is a red flashing light, the system analyzes the image data and identifies a red light source, which at this point may be a tail light or a red flashing light. Because the frequency at which red flashing stop lights flash is governed by law and therefore known, the system may then determines whether the identified red light source is flashing at the known frequency. The specific manner by which the system may perform this function is described further below following a description of an example of the basic structure of the system.


A first embodiment of an imaging system 10 is shown in FIG. 1. Imaging system 10 may be provided for controlling exterior lights 80 and, optionally, other equipment (50, 62) of a vehicle. System 10 includes an imager 20 and a processor 30. Imager 20 includes an image sensor (201, FIG. 2) that is configured to image a scene external and forward of the vehicle and to generate image data corresponding to the acquired images. Processor 30 receives and analyzes the image data to detect objects in the forward scene, and wherein the analysis of the image data performed by processor 30 may further include categorizing objects detected in the image data. If used in an exterior light control system, processor 30 may generate an exterior light control signal that may be used to control exterior lights 80 and may generate control signals to control any additional equipment (50, 62). These control signals are generated in response to analysis of the image data.


If imaging system 10 is used in a vehicle equipment control system, processor 30 may be configured to directly connect to the equipment (50) being controlled such that the generated control signals directly control the equipment. Alternatively, processor 30 may be configured to connect to an equipment control (60 and 70), which, in turn, is connected to the equipment being controlled (62 and 80) such that the control signals generated by processor 30 only indirectly control the equipment. For example, in the case of the equipment being exterior lights 80, processor 30 may analyze the image data from imager 20 so as to generate control signals that are more of a recommendation for an exterior light control 70 to use when controlling exterior lights 80. Thus, it can be said that the control signals are used to control the equipment. The control signals may further include not just a recommendation, but also a code representing a reason for the recommendation so that equipment controls 60 and 70 may determine whether or not to override a recommendation.


As shown in FIG. 1, various inputs (such as inputs 21-24) may be provided to processor 30 that may be taken into account in analyzing the image data or forming a recommendation or direct control signal. In some cases, such inputs may instead be provided to equipment control (60 and 70). For example, input from manual switches may be provided to equipment control (60 and 70), which may allow equipment control (60 and 70) to override a recommendation from processor 30. It will be appreciated that various levels of interaction and cooperation between processor 30 and equipment controls (60 and 70) may exist. One reason for separating control functions is to allow imager 20 to be located in the best location in the vehicle for obtaining images, which may be a distance from the equipment to be controlled and to allow communication over the vehicle bus 25.


According to one embodiment, the equipment that imaging system 10 can control may include one or more exterior lights 80 and the control signal generated by processor 30 may be an exterior light control signal. In this embodiment, exterior lights 80 may be controlled directly by processor 30 or by an exterior light control 70, which receives a control signal from processor 30. As used herein, the “exterior lights” broadly include any exterior lighting on the vehicle. Such exterior lights may include headlamps (both low and high beam if separate from one another), tail lights, foul weather lights (such as fog lights), brake lights, center-mounted stop lights (CHMSLs), turn signals, and back-up lights, etc. The exterior lights may be operated in several different modes including conventional low-beam and high-beam states. They may also be operated as daytime running lights, and additionally as super-bright high beams in those countries where they are permitted.


The brightness of the exterior lights may also be continuously varied between the low, high, and super-high states. Separate lights may be provided for obtaining each of these exterior lighting states or the actual brightness of the exterior lights may be varied to provide these different exterior lighting states. In either case, the “perceived brightness” or illumination pattern of the exterior lights is varied. As used herein, the term “perceived brightness” means the brightness of the exterior lights as perceived by an observer outside the vehicle. Most typically, such observers will be drivers or passengers in a preceding vehicle or in a vehicle traveling along the same street in the opposite direction. Ideally, the exterior lights are controlled such that if an observer is located in a vehicle within a “glare area” relative to the vehicle (i.e., the area in which the observer would perceive the brightness of the exterior lights as causing excessive glare), the beam illumination pattern is varied such that the observer is no longer in the glare area. The perceived brightness and/or glare area of the exterior lights may be varied by changing the illumination output of one or more exterior lights, by steering one or more lights to change the aim of one or more of the exterior lights, selectively blocking or otherwise activating or deactivating some or all of the exterior lights, altering the illumination pattern forward of the vehicle, or a combination of the above.


Imager 20 may be any conventional imager. Examples of suitable imagers are disclosed in published United States Patent Application Publication Nos. US 20080192132 A1 and US 20120072080 A1, and in U.S. Provisional Application Nos. 61/500,418 entitled “MEDIAN FILTER” filed on Jun. 23, 2011, by Jon H. Bechtel et al.; 61/544,315 entitled “MEDIAN FILTER” and filed on Oct. 7, 2011, by Jon H. Bechtel et al.; and 61/556,864 entitled “HIGH DYNAMIC RANGE CAMERA LOW LIGHT LEVEL FILTERING” filed on Nov. 8, 2011, by Jon H. Bechtel et al., the entire disclosures of which are incorporated herein by reference.


The imager includes an image sensor (or camera) to capture images that may then be displayed and/or analyzed in order to detect and optionally categorize objects or to optionally control vehicle equipment such as exterior lights. For example, such imagers have been used for lane departure warning systems, forward collision warning systems, adaptive cruise control systems, pedestrian detection systems, night vision systems, terrain detection systems, parking assist systems, traffic sign recognition systems, and reverse camera display systems. Examples of systems using imagers for such purposes are disclosed in U.S. Pat. Nos. 5,837,994, 5,990,469, 6,008,486, 6,049,171, 6,130,421, 6,130,448, 6,166,698, 6,379,013, 6,403,942, 6,587,573, 6,611,610, 6,631,316, 6,774,988, 6,861,809, 7,321,112, 7,417,221, 7,565,006, 7,567,291, 7,653,215, 7,683,326, 7,881,839, 8,045,760, and 8,120,652, and in U.S. Provisional Application Nos. 61/512,213 entitled “RAISED LANE MARKER DETECTION SYSTEM AND METHOD THEREOF” and filed on Jul. 27, 2011, by Brock R. Rycenga et al., and 61/512,158 entitled “COLLISION WARNING SYSTEM AND METHOD THEREOF” and filed on Jul. 27, 2011, by Brock R. Rycenga et al., which together correspond to published United States Patent Application Publication No. US 2013/0028473 A1, the entire disclosures of which are incorporated herein by reference.


In the example shown in FIG. 1, imager 20 may be controlled by processor 30. Communication of imager parameters as well as image data occurs over communication bus 40, which may be a bi-directional serial bus, parallel bus, a combination of both, or other suitable means. Processor 30 may perform equipment control functions by analyzing images from imager 20, determining an equipment (or exterior light) state based upon information detected within those images, and communicating the determined equipment (or exterior light) state to the equipment 50, equipment control 60, or exterior light control 70 through bus 42, which may be the vehicle bus 25, a CAN bus, a LIN bus or any other suitable communication link. Processor 30 may control the imager to be activated in several different modes with different exposure times and different readout windows. Processor 30 may be used to both perform the equipment or exterior light control function and control the parameters of imager 20.


Processor 30 can also take advantage of the availability of signals (such as vehicle speed and yaw) communicated via discrete connections or over the vehicle bus 25 in making decisions regarding the operation of the exterior lights 80. In particular, speed input 21 provides vehicle speed information to the processor 30 from which speed can be a factor in determining the control state for the exterior lights 80 or other equipment. The reverse signal 22 informs processor 30 that the vehicle is in reverse, responsive to which the processor 30 may clear an electrochromic mirror element regardless of signals output from light sensors. Auto ON/OFF switch input 23 is connected to a switch having two states to dictate to processor 30 whether the vehicle exterior lights 80 should be automatically or manually controlled. The auto ON/OFF switch (not shown) connected to the ON/OFF switch input 23 may be incorporated with the headlamp switches that are traditionally mounted on the vehicle dashboard or incorporated into steering wheel column levels. Manual dimmer switch input 24 is connected to a manually actuated switch (not shown) to provide a manual override signal for an exterior light control state. Some or all of the inputs 21, 22, 23, 24 and outputs 42a, 42b, and 42c, as well as any other possible inputs or outputs, such as a steering wheel input, can optionally be provided through vehicle bus 25 shown in FIG. 1. Alternatively, these inputs 21-24 may be provided to equipment control 60 or exterior light control 70.


Processor 30 can control or interact with, at least in part, other equipment 50 within the vehicle which is connected to processor 30 via vehicle bus 42. Specifically, the following are some examples of one or more equipment 50 that may be controlled by processor 30: exterior lights 80, a rain sensor, a compass, information displays, windshield wipers, a heater, a defroster, a defogger, an air conditioning system, a telephone system, a navigation system, a security system, a tire pressure monitoring system, a garage door opening transmitter, a remote keyless entry system, a telematics system, a voice recognition system such as a digital signal processor-based voice actuation system, a vehicle speed control, interior lights, rearview mirrors, an audio system, an engine control system, and various other switches and other display devices that may be located throughout the vehicle.


In addition, processor 30 may be, at least in part, located within a rearview assembly of a vehicle or located elsewhere within the vehicle. The processor 30 may also use a second processor (or processors) such as equipment control 60, which may be located in a rearview assembly or elsewhere in the vehicle in order to control certain kinds of equipment 62. Equipment control 60 can be connected to receive via vehicle bus 42 control signals generated by processor 30. Equipment control 60 subsequently communicates and controls equipment 62 via bus 61. For example, equipment control 60 may be a windshield wiper control unit which controls windshield wiper equipment, turning this equipment ON or OFF. Equipment control may also be an electrochromic mirror control unit where processor 30 is programmed to communicate with the electrochromic control unit in order for the electrochromic control unit to change the reflectivity of the electrochromic mirror(s) in response to information obtained from an ambient light sensor, a glare sensor, as well as any other components coupled to the processor. Specifically, equipment control unit 60 in communication with processor 30 may control the following equipment: exterior lights, a rain sensor, a compass, information displays, windshield wipers, a heater, a defroster, a defogger, air conditioning, a telephone system, a navigation system, a security system, a tire pressure monitoring system, a garage door opening transmitter, a remote keyless entry, a telemetry system, a voice recognition system such as a digital signal processor-based voice actuation system, a vehicle speed warning, interior lights, rearview mirrors, an audio system, a climate control, an engine control, and various other switches and other display devices that may be located throughout the vehicle.


Portions of imaging system 10 can be advantageously integrated into a rearview assembly 200 as illustrated in FIG. 2, wherein imager 20 is integrated into a mount 203 of rearview assembly 200. This location provides an unobstructed forward view through a region of the windshield 202 of the vehicle that is typically cleaned by the vehicle's windshield wipers (not shown). Additionally, mounting the image sensor 201 of imager 20 in the rearview assembly permits sharing of circuitry such as the power supply, microcontroller and light sensors.


Referring to FIG. 2, image sensor 201 is mounted within rearview mount 203, which is mounted to vehicle windshield 202. The rearview mount 203 provides an opaque enclosure for the image sensor with the exception of an aperture through which light is received from a forward external scene.


Processor 30 of FIG. 1 may be provided on a main circuit board 215 and mounted in rearview housing 204 as shown in FIG. 2. As discussed above, processor 30 may be connected to imager 20 by a bus 40 or other means. The main circuit board 215 may be mounted within rearview housing 204 by conventional means. Power and a communication link 42 with the vehicle electrical system, including the exterior lights 80 (FIG. 1), are provided via a vehicle wiring harness 217 (FIG. 2).


Rearview assembly 200 may include a mirror element or a display that displays a rearward view. The mirror element may be a prismatic element or an electro-optic element, such as an electrochromic element.


Additional details of the manner by which imaging system 10 may be integrated into a rearview mirror assembly 200 are described in U.S. Pat. No. 6,611,610, the entire disclosure of which is incorporated herein by reference. Alternative rearview mirror assembly constructions used to implement exterior light control systems are disclosed in U.S. Pat. No. 6,587,573, the entire disclosure of which is incorporated herein by reference.


The method for distinguishing between vehicle tail lights and flashing red stop lights will now be described with reference to FIGS. 3A, 3B, and 4. This method is described below as being implemented by processor 30 using image data received from imager 20. This method may be a subroutine executed by any processor, and thus this method may be embodied in a non-transitory computer readable medium having stored thereon software instructions that, when executed by a processor, cause the processor to control the equipment of the controlled vehicle, by executing the steps of the method described below. In other words, aspects of the inventive method may be achieved by software stored on a non-transitory tangible computer readable medium or software modifications or updates to existing software residing in a non-transitory computer readable medium. Such software or software updates may be downloaded into a first non-transitory computer readable media 32 of processor 30 (or locally associated with processor 30 or some other processor) typically prior to being installed in a vehicle, from a second non-transitory computer readable media 90 located remote from first non-transitory computer readable media 32. Second non-transitory computer readable media 90 may be in communication with first non-transitory computer readable media 32 by any suitable means, which may at least partially include the Internet or a local or wide area wired or wireless network.



FIG. 3 shows a flow chart of the steps of the method for distinguishing between vehicle tail lights and flashing red stop lights that may be executed by processor 30 when interacting with imager 20. The method begins with imaging a scene external and forward of the vehicle and generating image data corresponding to the acquired images (step 300). Next, processor 30 receives and analyzes the image data to identify red light sources (step 302) and further analyzes each red light source to determine if the red light source is detected for a predetermined time period (step 304). If the red light source is not detected within a predetermined time period after it is detected (about one second), processor 30 determines that the red light source is a flashing red stop light (step 306). Otherwise, if the red light source is detected for a predetermined time period, processor 30 determines that the red light source may be a tail light of another vehicle (step 308). If the imaging system is used to control vehicle equipment, such as exterior lights, processor 30 may execute the optional step of generating a signal indicating the presence of a vehicle tail light if other characteristics of the red light source indicate that the red light source is a vehicle tail light. Such other characteristics may include any known characteristics such as motion and brightness, for example. Optionally, processor 30 may be configured to determine that the red light source is a flashing red stop light when it is detected periodically and is not detected within the predetermined time period following a period of detection. The frames of image data may be captured at a predetermined image capture rate and the step of determining whether the red light source is a flashing red stop light includes determining whether the red light source is not detected within a predetermined number of frames acquired after a frame in which the red light source is detected, wherein the predetermined number of frames corresponds to the predetermined time period multiplied by the predetermined image capture rate. Processor 30 may further be configured to generate a signal indicating that an illumination level of the exterior lights should be reduced when a brightness of a red flashing stop light exceeds a brightness threshold. An example of this method is further described below with respect to FIGS. 4A and 4B.



FIG. 4A shows a graphic illustration of the intensity of red light detected as the vehicle approaches a red flashing stop light. The horizontal scale is the number of seconds of image data that has been captured. FIG. 4B shows the state of the high beam headlamps for the same time period as shown in FIG. 4A. As illustrated in FIG. 4B, the high beams are initially turned on, but turned off upon detecting a bright red reflector. Because the detected light from the reflector is significantly reduced upon turning off the high beams, the processor determines that the bright red reflected light was that from a reflector and the processor returns the high beams to an ON state.


As shown through the graph in FIG. 4A, a pulsing red light is detected of increasing intensity. Between each pulse, however, the detected brightness returns to zero on a periodic basis. The frequency at which flashing red stop lights flash is typically within a certain legally specified limit. Thus, processor 30 can detect that a pulsing red light is in fact a stop light upon determining that the pulsing red light is pulsing at a frequency corresponding to that of a stop light.


Frames of image data are captured at a predetermined image capture rate of, for example, five frames per second (one frame captured every 200 milliseconds). Thus, if a red light is initially detected and then not detected within a predetermined number of frames (e.g., about four frames) thereafter, processor 30 can determine that the pulsing red light is a flashing red stop light. Given the known image capture rate, the predetermined number of frames is selected to correspond to a predetermined time period in which one would expect a stop light to turn on and then off based on the known frequency at which such stop lights flash.


Accordingly, referring back to FIG. 4A, processor 30 would determine that the pulsing red light appearing throughout the window of second 3 to second 23 is a stop light and therefore would not generate a control signal to have the high beams turned off. Thus, the high beams remain on as shown in FIG. 4B. However, at about second 23, a tail light of another vehicle is detected where the brightness of which is superimposed with that of the stop lamp in the graph of FIG. 4A where it is labeled “Tail Light detected.” Because the value of red light detected does not drop to zero within a predetermined number of frames or seconds, processor 30 may determine that a tail light of a vehicle may be present and therefore generate a control signal indicating such presence which causes the high beams to be turned off as shown in FIG. 4B.


Next, around frame 31, the tail light is no longer present and the value of red light detected again periodically drops to zero. Therefore, processor 30 determines that no vehicles are currently present and that the red light is from a stop light, and thus generates a control signal to turn the high beams back on. As shown in FIG. 4B, the high beams then remain on until the brightness of the stop light exceeds a brightness threshold (at about 70 seconds, at which point, processor 30 determines that the controlled vehicle has neared the intersection where the stop light is located and then generates a control signal to turn the high beams off). The control signal may include an indication that the basis for the command is an approaching intersection. Preferably, processor 30 causes the high beams to turn off within 100 meters of the red flashing stop light, more preferably within 50 meters. Based on simulations, the above described embodiment is capable of turning off the high beams within 50 meters of the red flashing stop light.


It should be noted that processor 30 is capable of determining that a flashing red light is a stop light as opposed to just some nonvehicle light. If processor 30 determined that such a flashing red light was just a nonvehicle light source, processor 30 would not necessarily know to turn off the high beams as the brightness increased. Nevertheless, when supplied with vehicle speed data, processor 30 may turn off the high beams when the vehicle speed falls below a threshold.


The above description is considered that of the preferred embodiments only. Modifications of the invention will occur to those skilled in the art and to those who make or use the invention. Therefore, it is understood that the embodiments shown in the drawings and described above are merely for illustrative purposes and not intended to limit the scope of the invention, which is defined by the claims as interpreted according to the principles of patent law, including the doctrine of equivalents.

Claims
  • 1. An imaging system for a vehicle, comprising: an imager configured to acquire one or more images of a scene external and forward of the vehicle and to generate image data corresponding to the one or more acquired images; anda processor configured to receive and analyze the image data to detect one or more red light sources and to further analyze each detected red light source to determine if each respective red light source is detected for a predetermined time period,based on a determination that the respective red light source is not detected for the predetermined time period, the processor is further configured to: (i) determine that the respective red light source is a flashing red stop light, and (ii) generate a signal indicating that an illumination level of exterior lights of the vehicle should be reduced when a brightness of the flashing red stop light exceeds a brightness threshold regardless of an absence of any other vehicles in the imaged scene.
  • 2. The imaging system of claim 1, wherein the processor is further configured to, based on the determination that the respective red light source may be a tail light of another vehicle, generate a signal indicating the presence of another vehicle.
  • 3. The imaging system of claim 1, wherein the predetermined time period is about one second.
  • 4. The imaging system of claim 1, wherein the processor is further configured to determine that the respective red light source is a flashing red stop light based on detecting the respective red light source periodically and not detecting the respective red light source for the predetermined time period.
  • 5. The imaging system of claim 1, wherein the imager is configured to acquire one or more images by capturing frames of image data at a predetermined image frame capture rate; and wherein the processor is configured to determine that the respective red light source is a flashing red stop light based on not detecting the respective red light source with a predetermined number of frames acquired after a frame in which the respective red light source is detected, the predetermined number of frames corresponding to the predetermined time period multiplied by the predetermined image capture rate.
  • 6. A method for distinguishing between tail lights of another vehicle and a flashing red stop light, comprising: providing an imager for use in a vehicle;imaging a scene external and forward of the vehicle and generating image data corresponding to the acquired images;receiving and analyzing the image data in a processor to identify red light sources and further analyze each red light source to determine if the red light source is detected for a predetermined time period;if the red light source is not detected within the predetermined time period, determining that the red light source is a flashing red stop light and generating a signal indicating that an illumination level of the exterior lights should be reduced when a brightness of the flashing red stop light exceeds a brightness threshold regardless of an absence of any other vehicles in the imaged scene; andotherwise if the red light source is detected for the predetermined time period, determining that the red light source may be a tail light of another vehicle.
  • 7. The method of claim 6, and further comprising generating a control signal that is used to control exterior lights of the vehicle in which the imaging system is used in response to analysis of the image data, wherein, when generating a control signal, indicating the presence of another vehicle if a tail light of the other vehicle is detected.
  • 8. The method of claim 6, wherein the predetermined time period is about one second.
  • 9. The method of claim 6, wherein the step of determining that the red light source is a flashing red stop light includes determining whether the red light source is detected periodically and is not detected within the predetermined time period.
  • 10. The method of claim 6, wherein frames of image data are captured at a predetermined image capture rate and the step of determining whether the red light source is a flashing red stop light includes determining whether the red light source is not detected within a predetermined number of frames acquired after a frame in which the red light source is detected, wherein the predetermined number of frames corresponds to the predetermined time period multiplied by the predetermined image capture rate.
  • 11. A non-transitory tangible computer readable medium having stored thereon software instructions that, when executed by a processor within a vehicle, cause the processor to distinguish between tail lights of another vehicle and a flashing red stop light, by executing the steps comprising: controlling an imager to image a scene external and forward of the vehicle and generating image data corresponding to the acquired images;receiving and analyzing the image data in the processor to identify red light sources and further analyzing each red light source to determine if the red light source is detected for a predetermined time period;if the red light source is not detected within the predetermined time period, determining that the red light source is a flashing red stop light and generating a signal that indicates that an illumination level of the exterior lights should be reduced when a brightness of the flashing red stop light exceeds a brightness threshold regardless of an absence of any other vehicles in the imaged scene; andotherwise if the red light source is detected for the predetermined time period, determining that the red light source may be a tail light of another vehicle.
  • 12. The non-transitory tangible computer readable medium of claim 11, wherein the software instructions cause the processor to generate a control signal that is used to control exterior lights of the vehicle in which the imaging system is used, in response to analysis of the image data, and when generating a control signal, to indicate the presence of another vehicle if a tail light of the other vehicle is detected.
  • 13. The non-transitory tangible computer readable medium of claim 11, wherein the predetermined time period is about one second.
  • 14. The non-transitory tangible computer readable medium of claim 11, wherein the step of determining that the red light source is a flashing red stop light includes determining whether the red light source is detected periodically and is not detected within the predetermined time period.
  • 15. The non-transitory tangible computer readable medium of claim 11, wherein frames of image data are captured at a predetermined image capture rate and the step of determining whether the red light source is a flashing red stop light includes determining whether the red light source is not detected within a predetermined number of frames acquired after a frame in which the red light source is detected, wherein the predetermined number of frames corresponds to the predetermined time period multiplied by the predetermined image capture rate.
  • 16. The non-transitory tangible computer readable medium of claim 11, wherein the brightness threshold is selected such that the signal is generated when a distance to the flashing red stop light is within at least about 50 meters.
  • 17. The imaging system of claim 1, wherein the brightness threshold is selected such that the signal is generated when a distance to the flashing red stop light is within at least about 50 meters.
  • 18. The method of claim 6, wherein the brightness threshold is selected such that the signal is generated when a distance to the flashing red stop light is within at least about 50 meters.
  • 19. The imaging system of claim 1, wherein, based on a determination that the respective red light source is detected for the predetermined time period, the processor is further configured to determine that the respective red light source may be a tail light of another vehicle.
CROSS-REFERENCE TO RELATED APPLICATION

This application claims priority to and the benefit under 35 U.S.C. §119(e) of U.S. Provisional Patent Application No. 61/867,216, filed on Aug. 19, 2013, entitled “SYSTEM AND METHOD FOR CONTROLLING EXTERIOR VEHICLE LIGHTS IN RESPONSE TO FLASHING RED STOP LIGHTS,” the entire disclosure of which is hereby incorporated herein by reference.

US Referenced Citations (210)
Number Name Date Kind
2131888 Harris Oct 1938 A
2632040 Rabinow Mar 1953 A
2827594 Rabinow Mar 1958 A
3179845 Kulwiec Apr 1965 A
3581276 Newman May 1971 A
3663819 Hicks et al. May 1972 A
4139801 Linares Feb 1979 A
4151526 Hinachi et al. Apr 1979 A
4236099 Rosenblum Nov 1980 A
4258979 Mahin Mar 1981 A
4286308 Wolff Aug 1981 A
4357558 Massoni et al. Nov 1982 A
4376909 Tagami et al. Mar 1983 A
4479173 Rumpakis Oct 1984 A
4599544 Martin Jul 1986 A
4645975 Meitzler et al. Feb 1987 A
4665321 Chang et al. May 1987 A
4692798 Seko et al. Sep 1987 A
4716298 Etoh Dec 1987 A
4727290 Smith et al. Feb 1988 A
4768135 Kretschmer et al. Aug 1988 A
4862037 Farber et al. Aug 1989 A
4891559 Matsumoto et al. Jan 1990 A
4930742 Schofield et al. Jun 1990 A
4934273 Endriz Jun 1990 A
4967319 Seko Oct 1990 A
5008946 Ando Apr 1991 A
5036437 Macks Jul 1991 A
5072154 Chen Dec 1991 A
5086253 Lawler Feb 1992 A
5096287 Kakinami et al. Mar 1992 A
5124549 Michaels et al. Jun 1992 A
5166681 Bottesch et al. Nov 1992 A
5182502 Slotkowski et al. Jan 1993 A
5187383 Taccetta et al. Feb 1993 A
5235178 Hegyi Aug 1993 A
5329206 Slotkowski et al. Jul 1994 A
5347261 Adell Sep 1994 A
5347459 Greenspan et al. Sep 1994 A
5355146 Chiu et al. Oct 1994 A
5379104 Takao Jan 1995 A
5396054 Krichever et al. Mar 1995 A
5402170 Parulski et al. Mar 1995 A
5416318 Hegyi May 1995 A
5426294 Kobayashi et al. Jun 1995 A
5428464 Silverbrook Jun 1995 A
5430450 Holmes Jul 1995 A
5434407 Bauer et al. Jul 1995 A
5451822 Bechtel et al. Sep 1995 A
5452004 Roberts Sep 1995 A
5471515 Fossum et al. Nov 1995 A
5475441 Parulski et al. Dec 1995 A
5481268 Higgins Jan 1996 A
5483346 Butzer Jan 1996 A
5485155 Hibino Jan 1996 A
5508592 Lapatovich et al. Apr 1996 A
5537003 Bechtel et al. Jul 1996 A
5541724 Hoashi Jul 1996 A
5550677 Schofield et al. Aug 1996 A
5554912 Thayer et al. Sep 1996 A
5574463 Shirai et al. Nov 1996 A
5587929 League et al. Dec 1996 A
5592146 Kover, Jr. Jan 1997 A
5614788 Mullins et al. Mar 1997 A
5621460 Hatlestad et al. Apr 1997 A
5660454 Mori et al. Aug 1997 A
5666028 Bechtel et al. Sep 1997 A
5684473 Hibino et al. Nov 1997 A
5707129 Kobayashi Jan 1998 A
5710565 Shirai et al. Jan 1998 A
5714751 Chen Feb 1998 A
5715093 Schierbeek et al. Feb 1998 A
5736816 Strenke et al. Apr 1998 A
5751832 Panter et al. May 1998 A
5781105 Bitar et al. Jul 1998 A
5786787 Eriksson et al. Jul 1998 A
5796094 Schofield et al. Aug 1998 A
5798727 Shirai et al. Aug 1998 A
5811888 Hsieh Sep 1998 A
5812321 Schierbeek et al. Sep 1998 A
5835613 Breed et al. Nov 1998 A
5837994 Stam et al. Nov 1998 A
5841126 Fossum et al. Nov 1998 A
5845000 Breed et al. Dec 1998 A
5867214 Anderson et al. Feb 1999 A
5877897 Schofield et al. Mar 1999 A
5905457 Rashid May 1999 A
5912534 Benedict Jun 1999 A
5923027 Stam et al. Jul 1999 A
5942853 Piscart Aug 1999 A
5990469 Bechtel et al. Nov 1999 A
6008486 Stam et al. Dec 1999 A
6018308 Shirai Jan 2000 A
6049171 Stam et al. Apr 2000 A
6097023 Schofield et al. Aug 2000 A
6102546 Carter Aug 2000 A
6130421 Bechtel et al. Oct 2000 A
6130448 Bauer et al. Oct 2000 A
6140933 Bugno et al. Oct 2000 A
6144158 Beam Nov 2000 A
6166698 Turnbull et al. Dec 2000 A
6184781 Ramakesavan Feb 2001 B1
6255639 Stam et al. Jul 2001 B1
6281632 Stam et al. Aug 2001 B1
6329925 Skiver et al. Dec 2001 B1
6349782 Sekiya et al. Feb 2002 B1
6356376 Tonar et al. Mar 2002 B1
6379013 Bechtel et al. Apr 2002 B1
6396040 Hill May 2002 B1
6396397 Bos et al. May 2002 B1
6403942 Stam Jun 2002 B1
6429594 Stam et al. Aug 2002 B1
6442465 Breed et al. Aug 2002 B2
6443602 Tanabe et al. Sep 2002 B1
6465962 Fu et al. Oct 2002 B1
6469739 Bechtel et al. Oct 2002 B1
6483438 DeLine et al. Nov 2002 B2
6491416 Strazzanti Dec 2002 B1
6507779 Breed et al. Jan 2003 B2
6550943 Strazzanti Apr 2003 B2
6553130 Lemelson et al. Apr 2003 B1
6558026 Strazzanti May 2003 B2
6559435 Schofield et al. May 2003 B2
6587573 Stam et al. Jul 2003 B1
6593698 Stam et al. Jul 2003 B2
6611610 Stam et al. Aug 2003 B1
6617564 Ockerse et al. Sep 2003 B2
6631316 Stam et al. Oct 2003 B2
6677986 Pöchmüller Jan 2004 B1
6728393 Stam et al. Apr 2004 B2
6772057 Breed et al. Aug 2004 B2
6774988 Stam et al. Aug 2004 B2
6824281 Schofield et al. Nov 2004 B2
6861809 Stam Mar 2005 B2
6902307 Strazzanti Jun 2005 B2
6913375 Strazzanti Jul 2005 B2
6928180 Stam et al. Aug 2005 B2
6946978 Schofield Sep 2005 B2
6999004 Comaniciu et al. Feb 2006 B2
7012543 Deline et al. Mar 2006 B2
7038577 Pawlicki et al. May 2006 B2
7049945 Breed et al. May 2006 B2
7255465 Deline et al. Aug 2007 B2
7262406 Heslin et al. Aug 2007 B2
7265342 Heslin et al. Sep 2007 B2
7311428 Deline et al. Dec 2007 B2
7321112 Stam et al. Jan 2008 B2
7365769 Mager Apr 2008 B1
7417221 Creswick et al. Aug 2008 B2
7446650 Schofield et al. Nov 2008 B2
7467883 Deline et al. Dec 2008 B2
7468651 Deline et al. Dec 2008 B2
7526103 Schofield et al. Apr 2009 B2
7533998 Schofield et al. May 2009 B2
7565006 Stam et al. Jul 2009 B2
7567291 Bechtel et al. Jul 2009 B2
7575348 Hasegawa Aug 2009 B2
7653215 Stam Jan 2010 B2
7658521 Deline et al. Feb 2010 B2
7683326 Stam et al. Mar 2010 B2
7708434 Moizard et al. May 2010 B2
7719408 Deward et al. May 2010 B2
7720580 Higgins-Luthman May 2010 B2
7815326 Blank et al. Oct 2010 B2
7877175 Higgins-Luthman Jan 2011 B2
7881839 Stam et al. Feb 2011 B2
7888629 Heslin et al. Feb 2011 B2
7914188 Deline et al. Mar 2011 B2
7972045 Schofield Jul 2011 B2
7994471 Heslin et al. Aug 2011 B2
8045760 Stam et al. Oct 2011 B2
8063753 Deline et al. Nov 2011 B2
8090153 Schofield et al. Jan 2012 B2
8100568 Deline et al. Jan 2012 B2
8116929 Higgins-Luthman Feb 2012 B2
8120652 Bechtel et al. Feb 2012 B2
8142059 Higgins-Luthman et al. Mar 2012 B2
8162518 Schofield Apr 2012 B2
8203443 Bos et al. Jun 2012 B2
8217830 Lynam Jul 2012 B2
8222588 Schofield et al. Jul 2012 B2
8258433 Byers et al. Sep 2012 B2
8289142 Pawlicki et al. Oct 2012 B2
8289430 Bechtel et al. Oct 2012 B2
8325028 Schofield et al. Dec 2012 B2
8492698 Schofield et al. Jul 2013 B2
8543254 Schut et al. Sep 2013 B1
8718899 Schwindt et al. May 2014 B2
20020040962 Schofield Apr 2002 A1
20030202357 Strazzanti Oct 2003 A1
20040143380 Stam Jul 2004 A1
20040145905 Strazzanti Jul 2004 A1
20080044062 Stam Feb 2008 A1
20080165028 Fechner et al. Jul 2008 A1
20080192132 Bechtel et al. Aug 2008 A1
20080252488 Bos Oct 2008 A1
20080294315 Breed Nov 2008 A1
20090010494 Bechtel Jan 2009 A1
20110135155 Kudo Jun 2011 A1
20120072080 Jeromin et al. Mar 2012 A1
20120176499 Winter et al. Jul 2012 A1
20120249795 Ito Oct 2012 A1
20130028473 Hilldore et al. Jan 2013 A1
20130101161 Faber Apr 2013 A1
20130229520 Aimura Sep 2013 A1
20130253754 Ferguson Sep 2013 A1
20140036080 Schut Feb 2014 A1
20140132769 Kido May 2014 A1
20140192190 Takahashi Jul 2014 A1
20140247351 Murillo Amaya et al. Sep 2014 A1
Foreign Referenced Citations (29)
Number Date Country
2946561 May 1981 DE
1504957 Sep 2005 EP
2437233 Apr 2012 EP
2641237 Jul 1990 FR
2726144 Apr 1996 FR
2313973 Dec 1997 GB
5744541 Mar 1982 JP
6015237 Jan 1985 JP
62131837 Jun 1987 JP
01233129 Sep 1989 JP
5139203 Jun 1993 JP
5342901 Dec 1993 JP
06151067 May 1994 JP
06267304 Sep 1994 JP
6276524 Sep 1994 JP
6295601 Oct 1994 JP
6321007 Nov 1994 JP
732936 Feb 1995 JP
747878 Feb 1995 JP
7052706 Feb 1995 JP
769125 Mar 1995 JP
8166221 Jun 1996 JP
08221700 Aug 1996 JP
2010020483 Jan 2010 JP
8605147 Sep 1986 WO
9735743 Oct 1997 WO
9843850 Oct 1998 WO
9947396 Oct 1999 WO
0022881 Apr 2000 WO
Non-Patent Literature Citations (14)
Entry
Christopher M. Kormanyos, “SAE Paper No. 980003 entitled “HID System with Adaptive Vertical AIM Control”,” p. 13-18.
J.P. Lowenau et al, “SAE Paper No. 980007 entitled “Adaptive Light Control—A New Light Concept Controlled by Vehicle Dynamics and Navigation”,” p. 33-38.
Franz-Josef Kalze, “SAE Paper No. 980005 entitled “Xenon Light for Main and Dipped Beam”,” p. 23-26.
Tohru Shimizu et al, (SAE Paper No. 980322 entitled “Development of PWM DRL with Low RF Emissions and Low Heat”, p. 113-117.
International Searching Authority, Patent Cooperation Treaty, International Search Report and Written Opinion, Sep. 12, 2013, 6 pages.
International Searching Authority, Patent Cooperation Treaty, International Search Report and Written Opinion, Dec. 25, 2014, 6 pages.
JP Abstract of Patent No. 60-015237, “Headlight Device,” (Jan. 25, 1985).
JP Abstract of Patent No. 01233129 A, “Optical Axis Changing Device for Illuminating Lamp,” (Sep. 18, 1989).
JP Abstract of Patent No. 5342901, “Projection headlamp for vehicles,” p. 1.
JP Abstract for Patent No. 6321007, “Vehicular Dazzle Sensor,” p. 1.
JP Abstract of Patent No. 08221700 A, “Stop Lamp Recognition Device,” p. 1.
JP Abstract of Publication No. 2010-020483, “Vehicle Detection Device, Vehicle Detection Program and Light Controller,” p. 1.
International Searching Authority, Patent Cooperation Treaty, International Search Report and Written Opinion, Nov. 20, 2014, 6 pages.
Patent Cooperation Treaty, International Searching Authority, International Search Report, Written Opinion of the International Searching Authority and Notification of Transmittal, Nov. 20, 2014, 6 Pages.
Related Publications (1)
Number Date Country
20150049197 A1 Feb 2015 US
Provisional Applications (1)
Number Date Country
61867216 Aug 2013 US