The present invention relates to vision systems for vehicles and, more particularly, to an imaging sensor or camera for a vision or imaging system of a vehicle.
Traditional color imagers or image sensors or pixelated imaging arrays use a Bayer pattern of pixels and pixel filters, which has a red-green-green-blue (R-G-G-B) pixel/filter configuration (such as shown in
The three color filters (R, G, and B) not only pass ranges of wavelengths or spectral bands that are corresponding to red, green and blue colors, they also pass through a significant amount of wavelengths in the infrared (IR) or near infrared (NIR) region or band of the spectrum. Therefore, the color imager sensitivity or quantum efficiency spectrum typically has a rich IR or NIR response even with the R, G, and B color pixels. For example, a typical silicon CMOS color sensor's spectrum response is shown in
On the other hand, for IR imaging applications that have the main spectrum of interest in the NIR region of the spectrum, where silicon CCD or CMOS can provide good quantum efficiency, one may need to cut the visible spectrum off from the light reaching the imager. A long pass filter that is a multilayer coating on a glass or plastic flat plate is typically used to let through NIR light only. Security cameras and some special industrial machine vision systems are among the applications of such a filter-imager configuration. For cameras that operate in both day and night for such applications, mechanical switches of IR cutoff filters and IR pass filters are used on silicon CMOS or CCD cameras. The additional cost of two filters and mechanical switches, as well as the reliability of the moving mechanical switch, make such systems undesirable, and such a design is not suitable for automotive or consumer digital cameras or cell phone cameras.
The present invention provides an imaging sensor or camera having a pixel configuration that combines red (R), green (G), and blue (B) sensing pixels with an infrared (IR) sensing pixel. The present invention provides de-mosaicing techniques for processing the R-G-B-IR imager pixels. The imaging sensor or camera of the present invention is suitable for use in automotive camera products, such as for vision-based systems, such as headlamp control systems, reverse aid systems, lane departure warning systems, object detection systems, rain sensor systems, front collision warning, night vision and/or the like. The imaging sensor or camera of the present invention may also be suitable for use in security cameras, consumer digital cameras, cell phone cameras and industrial machine vision systems and/or the like.
Therefore, the imaging sensor of the present invention can provide both visible and IR imaging capability without external filters and shuttering/switching mechanism. Such a pixel configuration of R-G-B-IR pixels is applicable to various pixelated imaging array type sensors, such as CMOS, CCD and/or other image sensing devices. The imaging sensor of the present invention is suitable for applications in security cameras, automotive cameras, consumer digital cameras, cell phone cameras, and industrial machine vision cameras, as well as other markets and/or applications.
These and other objects, advantages, purposes and features of the present invention will become more apparent upon review of the following specification in conjunction with the drawings.
Imaging Sensor Description
The imaging sensor of the present invention comprises a two-dimensional pixelated imaging array having a plurality of photo-sensing pixels arranged or disposed or established on a semiconductor substrate. For example, the imaging sensor may comprise a complementary-metal-oxide-semiconductor (CMOS) or a CCD imaging sensor or device or the like, and may utilize aspects of the imaging sensors described in U.S. Pat. Nos. 5,796,094; 6,097,023; 6,320,176; 6,313,454; 6,559,435; 6,831,261; 6,396,397; 5,877,897; 6,498,620; 5,670,935; 6,806,452; 6,946,978; 7,123,168; 7,004,606; 7,005,974 and/or 5,550,677, and/or PCT Application No. PCT/US07/75702, filed Aug. 10, 2007 and published Feb. 28, 2008 as International Publication No. WO 2008/024639, and/or U.S. patent application Ser. No. 11/239,980, filed Sep. 30, 2005, now U.S. Pat. No. 7,881,496; Ser. No. 11/105,757, filed Apr. 14, 2005, now U.S. Pat. No. 7,526,103; and/or Ser. No. 10/534,632, filed May 11, 2005, now U.S. Pat. No. 7,965,336; and/or PCT Application No. PCT/US03/36177, filed Nov. 14, 2003 and published Jun. 3, 2004 as International Publication No. WO 2004/047421; and/or PCT Application No. PCT/US06/41709, filed Oct. 27, 2006 and published May 10, 2007 as International Publication No. WO 07/053404, which are hereby incorporated herein by reference in their entireties.
The arrangement of pixels has a pixel pattern that is derived from the Bayer pattern, but has one of the green filters or pixels replaced by an infrared (IR) filter or pixel 12d, such as shown in
Optionally, the R-G-B-IR filter arrangement may be any suitable arrangement of pixels/filters in a repeating pattern, such as the 2×2 arrangements of pixels shown in
Since the red, green and blue pixels typically may have a similar IR response as the IR pixel, the establishment of the IR pixel on each sub-array of the pixelated array allows for a determination of the IR component present in the light being imaged. Thus, the IR component (the IR or NIR energy or light sensed by the IR pixel) can be subtracted from the red, green and blue pixel response signal to obtain a more accurate or true color response for the pixels and to limit or avoid IR color wash-out. Further, with the IR cut-off in the sensor pixel level, there is no need to add IR cut-off filter in a lens. This simplifies the lens design and results in significant cost savings in manufacturing of the optical lens for the imager. Because the sensitivities in the IR region of the red, green and blue pixels may be different, factors or ratios may be applied to the IR signal when subtracting the IR value or component from the color signals. For example, the color values of the pixel signals may be determined via the following equations:
Red Color Value=Red Pixel Value−AR*IRPixel Value;
Green Color Value=Green Pixel Value−AG*IRPixel Value;
Blue Color Value=Blue Pixel Value−AB*IRPixel Value; and
IR Value=IR Pixel Value,
where AR, AG, and AB are the selected factors for the red pixels, green pixels and blue pixels, respectively.
Optionally, for example, these factors can be selected to be 1 to simplify the de-mosaicing circuitry or software. However, the spectral responses of the R, G and B pixels in the NIR region may not match well with that of the IR pixel. As a result, it may be difficult to totally cancel out the IR signal from the R, G and B pixels that sensing light in the visible range of the spectrum. Optionally, the factors of each R, G and B color can be determined by the corresponding spectral sensitivities of the sensor. The spectrum contribution of the light source should also be considered. For example, for applications where the scene is illuminated by a light source or light sources emitting illumination within a known spectral band or region of the spectrum, a combination of the light source and the imager's spectral sensitivity may determine the factors. For example, if an imaged object or illumination source contains narrow-band IR illumination around 750 nm, the imager with spectral characteristics such as shown in
Optionally, for illumination sources that generate a light or energy output with changing or un-controlled spectrum or spectral characteristics, such as sun light at different times of the day, more sophistical image processing can be applied to determine the factors. For example, an iterative approach can be used in which multiple passes with different factors determine the best factors. As another example, a color conversion matrix can be used in a manner that is similar to an auto white balance technique that is commonly used in visible light cameras, wherein a statistical analysis of the scene image is performed to determine the contribution of individual visible colors and the IR spectrum and to determine the values of three factors.
With the pixel arrangements discussed above and the de-mosaicing techniques described below, an imager or imaging system of the present invention can produce a visible color image and a NIR image at the same time. The following sections describe the different de-mosaicing techniques of the present invention.
De-Mosaicing Techniques
A pixelated imaging array of the present invention may have an R-G-B-IR pixel pattern, such as described above and shown in
2×2 Window De-Mosaicing
For an R-G-B-IR pixel pattern, each pixel of the array has its associated primary color (or is sensitive to a particular band of the spectrum and is less sensitive to other bands of the spectrum), which comes from its filter type. For example, and with reference to
R(r0c0)=R1−AR*IR1,G(r0c0)=G1−AG*IR1,B(r0c0)=B1−AB*IR1, and IR(r0c0)=IR1
R(r0c1)=R1−AR*IR1,G(r0c1)=G1−AG*IR1,B(r0c1)=B1−AB*IR1, and IR(r0c1)=IR1;
R(r1c0)=R1−AR*IR1,G(r1c0)=G1−AG*IR1,B(r1c0)=B1−AB*IR1, and IR(r1c0)=IR1; and
R(r1c1)=R1−AR*IR1,G(r1c1)=G1−AG*IR1,B(r1c1)=B1−AB*IR1, and IR(r1c1)=IR1,
where AR, AG, and AB are the selected factors for the red pixels, green pixels and blue pixels, respectively, and where R(rx,cx) is the red color value for the respective pixel location, G(rx,cx) is the green color value for the respective pixel location, B(rx,cx) is the blue color value for the respective pixel location, and IR(rx,cx) is the infrared value for the respective pixel location, and where R1 is an output value of the red pixel of the window or sub-array, G1 is an output value of the green pixel of the window or sub-array, B1 is an output value of the blue pixel of the window or sub-array, and IR1 is an output value of the IR pixel of the window or sub-array.
Thus, the value of the IR component for that pixel location (as determined by multiplying the value of the IR pixel by a factoring value for each color) is subtracted from the value of the respective color pixel (and the color for the adjacent pixels also has the factored IR value subtracted therefrom) to determine the color value of each pixel location. This is repeated for each 2×2 window of the array.
2×2 Rolling Window De-Mosaicing
Unlike the 2×2 Window De-mosaicing method above, the 2×2 Rolling Window De-mosaicing process shifts the 2×2 window by one pixel at a time in both the row and column directions, such as can be seen with reference to
R(r0c0)=R1−AR*IR1,G(r0c0)=G1−AG*IR1,B(r0c0)=B1−AB*IR1, and IR(r0c0)=IR1
R(r0c1)=R2−AR*IR1,G(r0c1)=G1−AG*IR1,B(r0c1)=B2−AB*IR1, and IR(r0c1)=IR1
R(r1c0)=R3−AR*IR1,G(r1c0)=G3−AG*IR1,B(r1c0)=B1−AB*IR1, and IR(r1c0)=IR1; and
R(r1c1)=R4−AR*IR1,G(r1c1)=G3−AG*IR1,B(r1c1)=B2−AB*IR1, and IR(r1c1)=IR1,
where AR, AG, and AB are the selected factors for the red pixels, green pixels and blue pixels, respectively, and where R(rx,cx) is the red color value for the respective pixel location, G(rx,cx) is the green color value for the respective pixel location, B(rx,cx) is the blue color value for the respective pixel location, and IR(rx,cx) is the infrared value for the respective pixel location, and where Rx is an output value of the red pixel of a respective sub-array, Gx is an output value of the green pixel of a respective sub-array, Bx is an output value of the blue pixel of a respective sub-array and IRx is an output value of the IR pixel of a respective sub-array.
Thus, the color values for the individual pixel locations are based on the color values for the respective color pixel within the particular 2×2 window, as reduced by the factored IR value of the IR pixel within the particular 2×2 window. The 2×2 window moves or shifts along the rows and columns during the processing of the pixelated array.
3×3 Rolling Window De-Mosaicing
Optionally, the imaging array may be de-mosaiced via a 3×3 rolling window de-mosaicing technique. In this method, each pixel will have its red, green, blue and IR information from averaging the adjacent pixels except for the primary color itself, such as can be seen with reference to
R(r1c1)=(R1+R2+R3+R4)/4−AR*IR1,G(r1c1)=(G1+G3)/2−AG*IR1,B(r1c1)=(B1+B2)/2−AB*IR1, and IR(r0c0)=IR1;
R(r1c2)=(R2+R4)/2−AR*(IR1+IR2)/2,G(r1c2)=(G1+G2+G3+G4)/4−AG*(IR1+IR2)/2,B(r0c1)=B2−AB*(IR1+IR2)/2, and IR(r0c1)=(IR1+IR2)/2;
R(r2c1)=(R3+R4)/2−AR*(IR1+IR3)/2,G(r2c1)=G3−AG*(IR1+IR3)/2,B(r2c1)=(B1+B2+B3+B4)/4−AB*(IR1+IR3)/2, and IR(r2c1)=(IR1+IR3)/2; and
R(r2c2)=R4−AR*(IR1+IR2+IR3+IR4)/4,G(r1c1)=(G3+G4)/2−AG*(IR1+IR2+IR3+IR4)/4,B(r2c2)=(B2+B4)/2−AB*(IR1+IR2+IR3+IR4)/4, and IR(r2c2)=(IR1+IR2+IR3+IR4)/4,
where AR, AG, and AB are the selected factors for the red pixels, green pixels and blue pixels, respectively, and where R(rx,cx) is the red color value for the respective pixel location, G(rx,cx) is the green color value for the respective pixel location, B(rx,cx) is the blue color value for the respective pixel location, and IR(rx,cx) is the infrared value for the respective pixel location, and where Rx is an output value of the red pixel of a respective sub-array, Gx is an output value of the green pixel of a respective sub-array, Bx is an output value of the blue pixel of a respective sub-array and IRx is an output value of the IR pixel of a respective sub-array.
Thus, the values for each color component of each row/column location is determined by averaging the respective colored pixels that are adjacent to the particular pixel location or coordinate. For example, the red color component for the pixel location r1,c1 is the average of the four red pixels R1, R2, R3, R4 (at r0,c0; r0,c2; r2,c0 and r2,c2) that are adjacent to the pixel location r1,c1, while the green color component for the pixel location r1,c1 is the average of the two green pixels G1, G3 (at r0,c1 and r2,c1) that are adjacent to the pixel location r1,c1, and the blue color component for the pixel location r1,c1 is the average of the two blue pixels B1, B2 (at r1,c0 and r1,c2) that are adjacent to the pixel location r1,c1. For the infrared value at the pixel location r1,c1, the infrared value is determined by the value of the IR1 pixel at that particular location. The color or IR values for the other coordinates or pixel locations are determined in a similar manner.
Implementation Methods
It is envisioned that any of the above de-mosaicing techniques may be implemented in conjunction with any suitable imaging array sensor. For example, the above de-mosaicing techniques can be implemented on or in conjunction with an imager chip, which is often called a system-on-chip (SOC). Optionally, such processing techniques may be implemented outside the imager chip by means of a FPGA, DSP or microprocessor or the like, while remaining within the spirit and scope of the present invention.
Applications
IHC Reflective Object Identification
In Intelligent Headlamp Control (IHC) applications, spectral information, such as R, G, B color information may be used to assist identifying tail lights, head lamps and traffic signs, such as described in U.S. Pat. Nos. 5,796,094; 6,097,023; 6,320,176; 6,313,454; 6,559,435; 6,831,261; 6,396,397; 5,877,897; 6,498,620; 5,670,935; 6,946,978; 7,123,168; 7,004,606; 7,005,974 and/or 5,550,677, and/or PCT Application No. PCT/US07/75702, filed Aug. 10, 2007 and published Feb. 28, 2008 as International Publication No. WO 2008/024639, and/or U.S. patent application Ser. No. 11/239,980, filed Sep. 30, 2005, now U.S. Pat. No. 7,881,496; and/or Ser. No. 10/534,632, filed May 11, 2005, now U.S. Pat. No. 7,965,336; and/or PCT Application No. PCT/US03/36177, filed Nov. 14, 2003 and published Jun. 3, 2004 as International Publication No. WO 2004/047421, which are hereby incorporated herein by reference in their entireties. It is usually challenging to identify and separate reflective objects, like traffic signs in night scene, from tail lights, head lamps and street lights. The techniques that have been often used are identifying color (traffic signs are often yellow or white) and shapes (traffic signs are often square or rectangle). However, these techniques are not always effective and accurate.
The present invention provides a processing technique that uses IR information to increase the identification accuracy. In addition to three visible colors, IR adds one more dimension to the color space. All tail lights, head lamps and traffic signs have their IR signatures based on how much they emit or reflect IR light. For instance, a traffic sign reflects the light from the head lamps of an approaching vehicle, with the head lamp illumination carrying a known IR signature. Quantitative IR signature of objects helps to determine the types of objects, such as for example, whether an object is a light or a road side traffic sign or other object. The IR signatures include:
In the signatures of the types 2 and 3 above, the vision algorithm can calibrate and define an area, range or segment of these color spaces to classify the object types located inside these regions. Thus, the system may recognize or identify an object based on whether the signature of the detected object falls within a particular signature range that is previously categorized or identified as being associated with or indicative of or representative of a particular object or type or class of object.
The system thus may utilize the CIE color space, as is known in the color art, such as by utilizing aspects of the systems described in U.S. patent application Ser. No. 11/239,980, filed Sep. 30, 2005, now U.S. Pat. No. 7,881,496, which is hereby incorporated herein by reference in its entirety. The Centre Internationale de l'Eclairage (CIE) is an international organization that establishes methods for measuring color and that recommends standards and procedures for light and lighting, including colorimetry. The color standards for colormetric measurements are internationally accepted specifications that define color values mathematically. CIE defines color as a combination of three axes: x, y, and z. CIE color models are considered device independent because the colors should not differ, theoretically, from one output device to another if properly calibrated.
In addition to the spectral domain, one can apply IR processing techniques to identify an object in the time domain. For example, a vehicle can include an IR light source, such as an IR LED or the like, at or near or in the headlamp assembly. The headlamp's IR output can be modulated with a pulse width modulation (PWM) sequence. Since only traffic signs, not the head lamps of other oncoming vehicles or tail lights of other leading vehicles reflect the IR light, the traffic signs appear in the image frames that the headlamp IR source is turned on, and is diminished or reduced in the image frames that the headlamp IR source is turned off. For the objects seen in the IR image that possibly come from oncoming vehicle headlamps, proceeding vehicle tail lights or street lights, the captured images of these objects or illumination sources will not have such an on and off sequence, which is synchronized with the IR light source PWM sequence. Such a processing technique can also help identify and separate the reflective objects in the forward field of view of the imager, such as retro-reflective marks/signs that are placed on the mailbox, front yard, building side or reflective stripes on trucks. The modulated IR light is not visible or discernible to humans, so it does not impede drivers of other vehicles or the driver of the subject or host vehicle and does not violate related regulations. The PWM sequence of each vehicle can be encoded or controlled with unique “signature” to avoid the interference from other vehicles equipped with the same type of PWM IR lights. Optionally, for example, this signature can be one of the following or a combination of the following: (1) unique modulation sequence, (2) unique modulation frequency.
NIR Night Vision Capability in Forward Viewing Camera
A NIR night vision system of the present invention may include a NIR light source (such as a LED, laser, or other NIR light sources) that illuminates the scene in the dark environment, a NIR video camera that captures image/video of the scene, and a display that shows the captured NIR images to a driver. For example, and with reference to
The NIR night vision system allows a driver to see a farther distance ahead of the vehicle and to see and discern more objects than he or she can normally see with typical head lamp illumination. Such a vision system thus gives the driver more time to react to different driving conditions. It is desirable to have a NIR camera mounted inside the vehicle compartment and looking forward. To avoid blocking the viewing angle of driver, it is best to mount the camera behind the rear view mirror, and preferably with a forward field of view through the windshield of the vehicle and through an area that is wiped or cleaned by the windshield wipers of the vehicle. The NIR illumination source can be either mounted inside one or both headlamps or inside the driver compartment, such as at or in or near or behind the interior rear view mirror or accessory module or windshield electronics module or the like.
The forward viewing camera thus may be mounted behind the rear view mirror and inside the vehicle, and the system may provide driver assistance and safety functions, such as intelligent headlamp control, lane departure warning, traffic sign recognition, adaptive cruise control, and/or the like. Optionally, the forward viewing camera may provide or be associated with a rain sensing function, or there may be other sensing devices, such as a rain sensor or the like, mounted at or near or behind the rear view mirror. Adding a NIR camera to an already crowded space behind the rear view mirror can be challenging and costly. It is thus desirable to combine the NIR camera to a visible forward viewing camera to save space and cost. The R-G-B-IR imager of the present invention provides a solution to combine the visible forward viewing camera and the NIR camera into a single imager and camera body.
Rear Viewing Camera With NIR Night Vision
A typical automotive rear viewing camera includes a color imager and a lens with IR cut-off filter. The IR cut-off filter in the lens prevents the camera from IR contamination and provides good visible color images for the camera in normal lighting conditions. However, in a night condition when no or minimum ambient light is available to illuminate the scene behind the vehicle, the only illumination comes from the vehicle tail lights. The tail lights, which include running lights, brake lights and back up lights, typically produce illumination of less than a few Lux. Thus, in a rear viewing camera image with an IR cut-off filter, the night scene image is mostly noisy, dark and low-contrast. One can thus only view a very short distance behind the vehicle. This is partially because the IR cut-off filter cuts off a significant amount of the existing IR light in the tail light illumination which otherwise assists the imaging sensitivity.
To provide enhanced safety and enhanced convenience to the drivers, it is desirable to increase the low light sensitivity of the rear viewing camera. It may be desirable to take off the IR cutoff filter from the lens, so that the scene illuminated by the near IR light component from the tail lights, street light, garage light, and/or the like can be better imaged by the imager, since incandescent light bulbs, which are typically used in vehicle tail lights, street lights, garage lights, and the like, emit a relatively high amount of near IR light. It may also be desired to include an additional invisible near IR light source (such as, for example, a near infrared LED or the like) at or near or in the tail light assembly for illuminating the rearward scene of the vehicle. The high sensitivity of silicon imager in the near IR region helps increase the image sensitivity of the night scene.
The R-G-B-IR imager described above provides a possibility of using NIR to enhance low light sensitivity of rear viewing cameras. A rear viewing camera, such as shown in
For example, and with reference to
To further increase the signal strength and sensitivity of the low light images, it is desirable to use all the available photons that the imager can detect in the NIR and visible wavelength ranges. The red and IR pixels are more sensitive to the light sources in low light night conditions, which includes incandescent bulb tail lights, an added NIR illuminator, garage lamp lights, street lights, and the like, than green and blue pixels, since most of these lights have stronger red and NIR spectrum range components than green and blue spectrum range components. To take advantage of this feature, a de-mosaicing method of the present invention may be implemented to extract the desired image color components. As can be seen with reference to
The switching between normal color image and low light monochrome image may be done in an imager de-mosaicing unit. For example, and with reference to
Optionally, the imaging sensor may include a plurality of sub-arrays, such as a plurality of 2×2 sub-arrays or the like, that may each comprise a principally red sensing pixel, a principally blue sensing pixel, a principally green sensing pixel and a principally IR or near IR sensing pixel, as discussed above, or may include other types of sub-arrays having different arrangements of pixels such as one or more principally IR or NIR sensing pixels and one or more color sensing pixels. For example, an imaging sensor may include a plurality of sub-arrays where each sub-array may comprise a color sensing pixel (such as a pixel primarily sensing red light from the red region of the spectrum, blue light from the blue region of the spectrum or green light from the green region of the spectrum, such as may be achieved by placing a color filter in front of or over the respective pixel) and a clear pixel (that has a clear spectral filter or no spectral filter and is receptive to all or substantially all incoming visible and infrared light or energy incident thereon) and a principally IR or NIR sensing pixel (such as a principally IR or NIR sensing pixel as described above). Such an imaging sensor may function in a similar manner as described above, where the IR value (as sensed by the principally IR or NIR sensing pixel) may be subtracted by the accompanying image processor from the color value sensed by the color pixel and the clear value sensed by the clear pixel to obtain a more accurate or true color response for the color and clear pixels and to limit or avoid infrared color wash-out. Clearly, other arrangements of pixels and filters, such as an imaging sensor having a sub-array having one or more color sensing pixels and/or one or more clear pixels and one or more principally IR or NIR sensing pixels, such as an imaging sensor having three pixel sub-arrays or 2×2 sub-arrays or 3×3 sub-arrays or the like, may be implemented with one or more IR or NIR sensing pixels while remaining within the spirit and scope of the present invention.
A benefit of the present invention is that a principally IR or NIR sensing pixel is in close proximity and optionally contiguous with the associated cluster of spectrally selective and/or clear pixels (such as red, green and/or blue pixels and/or clear pixels), so that the image data associated with a particular object in the field of view being imaged (such as, for example, a headlamp or taillight in a forward field of view or a child or object in a rearward field of view or a vehicle or object in a forward field of view or a child or seat occupant in an interior cabin field of view or the like) can be processed to subtract out or account for the IR or NIR content of light or energy in the field of view by subtracting out the IR or NIR values (as sensed by the principally IR or NIR sensing pixel) from the other pixels of a particular sub-array of pixels (that images the headlamp or taillight or object or child or occupant that is in the field of view of the imaging sensor) to enhance the image output and/or to obtain a more accurate or true color response for the pixels and to limit or avoid infrared color wash-out. Thus, the IR or NIR values local to or associated with an illumination source or object being sensed by a respective sub-array of the imaging sensor may be subtracted from the color values or clear spectral values sensed by the color and/or clear pixels of that respective sub-array (with the color and/or clear pixels of the sub-array being proximate to or associated with the principally IR or NIR sensing pixel or pixels of that sub-array), so that the system may provide enhanced detection of and/or identification of the illumination sources or objects exterior or interior of the vehicle that are being imaged by the imaging sensor and system.
Optionally, the imaging sensor may be suitable for various forward imaging systems, such as for headlamp control systems, rain sensor systems, lane departure warning systems, object detection systems, and/or the like. For example, the imaging sensor of the present invention may be associated with one or more systems of the types described in U.S. Pat. Nos. 5,796,094; 6,396,397; 6,806,452; 6,690,268; 7,005,974; 7,123,168; 7,004,606; 6,946,978; 7,038,577; 6,353,392; 6,320,176; 6,313,454 and 6,824,281, which are all hereby incorporated herein by reference in their entireties.
Optionally, the imaging sensor may be suitable for various in-cabinet imaging systems, such as driver drowsiness monitoring, passenger identification for intelligent airbag deployment, back seat passenger monitoring, and/or the like.
Optionally, the imaging sensor may be suitable for use in connection with other vehicle imaging systems, such as, for example, a blind spot detection system, where a blind spot indicator may be operable to provide an indication to the driver of the host vehicle that an object or other vehicle has been detected in the lane or area adjacent to the side of the host vehicle. In such a blind spot detector/indicator system, the blind spot detection system may include an imaging sensor or sensors, or ultrasonic sensor or sensors, or sonar sensor or sensors or the like. For example, the blind spot detection system may utilize aspects of the blind spot detection and/or imaging and/or indicating systems described in U.S. Pat. Nos. 7,038,577; 6,882,287; 6,198,409; 5,929,786 and/or 5,786,772, and/or U.S. pat. application Ser. No. 11/315,675, filed Dec. 22, 2005, now U.S. Pat. No. 7,720,580; Ser. No. 11/239,980, filed Sep. 30, 2005, now U.S. Pat. No. 7,881,496; and/or Ser. No. 11/933,697, filed Nov. 1, 2007, now U.S. Pat. No. 7,777,611, and/or PCT Application No. PCT/US06/26148, filed Jul. 5, 2006 and published Jan. 11, 2007 as International Publication No. WO 2007/005942, and/or PCT Application No. PCT/US07/082099, filed Oct. 22, 2007 and published May 2, 2008 as International Publication No. WO 2008/051910, and/or U.S. provisional applications, Ser. No. 60/618,686, filed Oct. 14, 2004; Ser. No. 60/853,850, filed Oct. 24, 2006; Ser. No. 60/918,089, filed Mar. 15, 2007; Ser. No. 60/970,687, filed Sep. 7, 2007; and/or Ser. No. 60/857,025, filed Nov. 6, 2006, and/or of the reverse or backup aid systems, such as the rearwardly directed vehicle vision systems described in U.S. Pat. Nos. 5,550,677; 5,760,962; 5,670,935; 6,201,642; 6,396,397; 6,498,620; 6,717,610; 6,757,109 and/or 7,005,974, and/or of the rain sensors described in U.S. Pat. Nos. 6,250,148 and 6,341,523, and/or of other imaging systems, such as the types described in U.S. Pat. Nos. 7,123,168; 6,353,392 and 6,313,454, with all of the above referenced U.S. patents, patent applications and provisional applications and PCT applications being commonly assigned and being hereby incorporated herein by reference in their entireties.
Optionally, the mirror assembly and/or accessory module or windshield electronics module may include one or more displays, such as for displaying the captured images or video images captured by the imaging sensor or sensors of the vehicle, such as the displays of the types disclosed in U.S. Pat. Nos. 7,004,593; 5,530,240 and/or 6,329,925, which are hereby incorporated herein by reference, and/or display-on-demand or transflective type displays, such as the types disclosed in U.S. Pat. Nos. 7,274,501; 7,255,451; 7,195,381; 7,184,190; 6,690,268; 5,668,663 and/or 5,724,187, and/or in U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US 2006-0050018; Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US 2006-0061008; Ser. No. 10/993,302, filed Nov. 19, 2004, now U.S. Pat. No. 7,338,177; and/or Ser. No. 11/284,543, filed Nov. 22, 2005, now U.S. Pat. No. 7,370,983, and/or PCT Application No. PCT/US03/29776, filed Sep. 9, 2003 and published Apr. 1, 2004 as International Publication No. WO 2004/026633; and/or PCT Application No. PCT/US03/35381, filed Nov. 5, 2003 and published May 21, 2004 as International Publication No. WO 2004/042457, and/or PCT Application No. PCT/US03/40611, filed Dec. 19, 2003 and published Jul. 15, 2004 as International Publication No. WO 2004/058540, and/or PCT Application No. PCT/US06/18567, filed May 15, 2006 and published Nov. 23, 2006 as International Publication No. WO 2006/124682, which are all hereby incorporated herein by reference in their entireties, or may include or incorporate video displays or the like, such as the types described in PCT Application No. PCT/US03/40611, filed Dec. 19, 2003 and published Jul. 15, 2004 as International Publication No. WO 2004/058540, and/or PCT Application No. PCT/US07/042718, filed Oct. 31, 2006 and published May 10, 2007 as International Publication No. WO 07/053710, and/or U.S. Pat. application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US 2006-0050018; and/or Ser. No. 11/284,543, filed Nov. 22, 2005, now U.S. Pat. No. 7,370,983, and/or U.S. provisional applications, Ser. No. 60/732,245, filed Nov. 1, 2005; Ser. No. 60/759,992, filed Jan. 18, 2006; and/or Ser. No. 60/836,219, filed Aug. 8, 2006, which are hereby incorporated herein by reference.
The imaging sensor may be incorporated at or in an accessory module or windshield electronics module (such as described above), or may be incorporated at or in an interior rearview mirror assembly of the vehicle, while remaining within the spirit and scope of the present invention. Optionally, the accessory module may utilize aspects of other accessory modules or windshield electronics modules or the like, such as the types described in U.S. patent application Ser. No. 11/201,661, filed Aug. 11, 2005, and/or U.S. Pat. Nos. 7,188,963; 7,004,593; 6,824,281; 6,690,268; 6,250,148; 6,341,523; 6,593,565; 6,428,172; 6,501,387; 6,329,925 and 6,326,613, and/or in PCT Application No. PCT/US03/40611, filed Dec. 19, 2003 and published Jul. 15, 2004 as International Publication No. WO 2004/058540, and/or Ireland pat. applications, Ser. No. S2004/0614, filed Sep. 15, 2004; Ser. No. S2004/0838, filed Dec. 14, 2004; and Ser. No. S2004/0840, filed Dec. 15, 2004, which are all hereby incorporated herein by reference.
Optionally, the mirror assembly and/or module may support one or more other accessories or features, such as one or more electrical or electronic devices or accessories. For example, illumination sources or lights, such as map reading lights or one or more other lights or illumination sources, such as illumination sources of the types disclosed in U.S. Pat. Nos. 6,690,268; 5,938,321; 5,813,745; 5,820,245; 5,673,994; 5,649,756; 5,178,448; 5,671,996; 4,646,210; 4,733,336; 4,807,096; 6,042,253; 6,971,775; 7,249,860; 7,195,381 and/or 5,669,698, which are hereby incorporated herein by reference in their entireties, may be included in the mirror assembly or module. The illumination sources and/or the circuit board may be connected to one or more buttons or inputs for activating and deactivating the illumination sources. Optionally, the mirror assembly or module may also or otherwise include other accessories, such as microphones, such as analog microphones or digital microphones or the like, such as microphones of the types disclosed in U.S. Pat. Nos. 6,243,003; 6,278,377 and/or 6,420,975, and/or in U.S. patent application Ser. No. 10/529,715, filed Mar. 30, 2005. Optionally, the mirror assembly may also or otherwise include other accessories, such as a telematics system, speakers, antennas, including global positioning system (GPS) or cellular phone antennas, such as disclosed in U.S. Pat. No. 5,971,552, a communication module, such as disclosed in U.S. Pat. No. 5,798,688, a voice recorder, transmitters and/or receivers, such as for a garage door opener or a vehicle door unlocking system or the like (such as a remote keyless entry system), a digital network, such as described in U.S. Pat. No. 5,798,575, a memory mirror system, such as disclosed in U.S. Pat. No. 5,796,176, a hands-free phone attachment, a video device for internal cabin surveillance (such as for sleep detection or driver drowsiness detection or the like) and/or video telephone function, such as disclosed in U.S. Pat. Nos. 5,760,962 and/or 5,877,897, a remote keyless entry receiver, a seat occupancy detector, a remote starter control, a yaw sensor, a clock, a carbon monoxide detector, status displays, such as displays that display a status of a door of the vehicle, a transmission selection (4wd/2wd or traction control (TCS) or the like), an antilock braking system, a road condition (that may warn the driver of icy road conditions) and/or the like, a trip computer, a tire pressure monitoring system (TPMS) receiver (such as described in U.S. Pat. Nos. 6,124,647; 6,294,989; 6,445,287; 6,472,979 and/or 6,731,205; and/or U.S. patent application Ser. No. 11/232,324, filed Sep. 21, 2005, now U.S. Pat. No. 7,423,522, and/or an ONSTAR® system and/or any other accessory or circuitry or the like (with all of the above-referenced patents and PCT and U.S. patent applications being commonly assigned to Donnelly Corporation, and with the disclosures of the referenced patents and patent applications being hereby incorporated herein by reference in their entireties).
Changes and modifications to the specifically described embodiments may be carried out without departing from the principles of the present invention, which is intended to be limited by the scope of the appended claims, as interpreted in accordance with the principles of patent law.
The present application is a continuation of U.S. patent application Ser. No. 16/010,581, filed Jun. 18, 2018, now U.S. Pat. No. 10,616,507, which is a continuation of U.S. patent application Ser. No. 14/562,826, filed Dec. 8, 2014, now U.S. Pat. No. 10,003,755, which is a continuation of U.S. patent application Ser. No. 13/896,845, filed May 17, 2013, now U.S. Pat. No. 8,908,040, which is a continuation of U.S. patent application Ser. No. 12/681,204, filed May 28, 2010, now U.S. Pat. No. 8,446,470, which is a 371 national phase filing of PCT Application No. PCT/US08/78700, filed Oct. 3, 2008, which claims the benefit of U.S. provisional application Ser. No. 60/977,430, filed Oct. 4, 2007, which is hereby incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
2632040 | Rabinow | Mar 1953 | A |
2827594 | Rabinow | Mar 1958 | A |
3349394 | Carver | Oct 1967 | A |
3601614 | Platzer | Aug 1971 | A |
3612666 | Rabinow | Oct 1971 | A |
3665224 | Kelsey | May 1972 | A |
3680951 | Jordan et al. | Aug 1972 | A |
3689695 | Rosenfield et al. | Sep 1972 | A |
3708231 | Walters | Jan 1973 | A |
3746430 | Brean et al. | Jul 1973 | A |
3807832 | Castellion | Apr 1974 | A |
3811046 | Levick | May 1974 | A |
3813540 | Albrecht | May 1974 | A |
3862798 | Hopkins | Jan 1975 | A |
3947095 | Moultrie | Mar 1976 | A |
3962600 | Pittman | Jun 1976 | A |
3985424 | Steinacher | Oct 1976 | A |
3986022 | Hyatt | Oct 1976 | A |
4037134 | Loper | Jul 1977 | A |
4052712 | Ohama et al. | Oct 1977 | A |
4093364 | Miller | Jun 1978 | A |
4111720 | Michel et al. | Sep 1978 | A |
4161653 | Bedini et al. | Jul 1979 | A |
4200361 | Malvano et al. | Apr 1980 | A |
4214266 | Myers | Jul 1980 | A |
4218698 | Bart et al. | Aug 1980 | A |
4236099 | Rosenblum | Nov 1980 | A |
4238760 | Carr | Dec 1980 | A |
4247870 | Gabel et al. | Jan 1981 | A |
4249160 | Chilvers | Feb 1981 | A |
4266856 | Wainwright | May 1981 | A |
4277804 | Robison | Jul 1981 | A |
4281898 | Ochiai et al. | Aug 1981 | A |
4288814 | Talley et al. | Sep 1981 | A |
4355271 | Noack | Oct 1982 | A |
4357558 | Massoni et al. | Nov 1982 | A |
4381888 | Momiyama | May 1983 | A |
4388701 | Aichelmann, Jr. et al. | Jun 1983 | A |
4420238 | Felix | Dec 1983 | A |
4431896 | Lodetti | Feb 1984 | A |
4443057 | Bauer et al. | Apr 1984 | A |
4460831 | Oettinger et al. | Jul 1984 | A |
4481450 | Watanabe et al. | Nov 1984 | A |
4491390 | Tong-Shen | Jan 1985 | A |
4512637 | Ballmer | Apr 1985 | A |
4529275 | Ballmer | Jul 1985 | A |
4529873 | Ballmer et al. | Jul 1985 | A |
4546551 | Franks | Oct 1985 | A |
4549208 | Kamejima et al. | Oct 1985 | A |
4571082 | Downs | Feb 1986 | A |
4572619 | Reininger et al. | Feb 1986 | A |
4580875 | Bechtel et al. | Apr 1986 | A |
4600913 | Caine | Jul 1986 | A |
4603946 | Kato et al. | Aug 1986 | A |
4614415 | Hyatt | Sep 1986 | A |
4620141 | McCumber et al. | Oct 1986 | A |
4623222 | Itoh et al. | Nov 1986 | A |
4626850 | Chey | Dec 1986 | A |
4629941 | Ellis et al. | Dec 1986 | A |
4630109 | Barton | Dec 1986 | A |
4632509 | Ohmi et al. | Dec 1986 | A |
4638287 | Umebayashi et al. | Jan 1987 | A |
4647161 | Muller | Mar 1987 | A |
4653316 | Fukuhara | Mar 1987 | A |
4669825 | Itoh et al. | Jun 1987 | A |
4669826 | Itoh et al. | Jun 1987 | A |
4671615 | Fukada et al. | Jun 1987 | A |
4672457 | Hyatt | Jun 1987 | A |
4676601 | Itoh et al. | Jun 1987 | A |
4690508 | Jacob | Sep 1987 | A |
4692798 | Seko et al. | Sep 1987 | A |
4697883 | Suzuki et al. | Oct 1987 | A |
4701022 | Jacob | Oct 1987 | A |
4713685 | Nishimura et al. | Dec 1987 | A |
4717830 | Botts | Jan 1988 | A |
4727290 | Smith et al. | Feb 1988 | A |
4731669 | Hayashi et al. | Mar 1988 | A |
4741603 | Miyagi et al. | May 1988 | A |
4768135 | Kretschmer et al. | Aug 1988 | A |
4772942 | Tuck | Sep 1988 | A |
4789904 | Peterson | Dec 1988 | A |
4793690 | Gahan et al. | Dec 1988 | A |
4817948 | Simonelli | Apr 1989 | A |
4820933 | Hong et al. | Apr 1989 | A |
4825232 | Howdle | Apr 1989 | A |
4838650 | Stewart et al. | Jun 1989 | A |
4847772 | Michalopoulos et al. | Jul 1989 | A |
4855822 | Narendra et al. | Aug 1989 | A |
4862037 | Farber et al. | Aug 1989 | A |
4867561 | Fujii et al. | Sep 1989 | A |
4871917 | O'Farrell et al. | Oct 1989 | A |
4872051 | Dye | Oct 1989 | A |
4881019 | Shiraishi et al. | Nov 1989 | A |
4882565 | Gallmeyer | Nov 1989 | A |
4886960 | Molyneux et al. | Dec 1989 | A |
4891559 | Matsumoto et al. | Jan 1990 | A |
4892345 | Rachael, III | Jan 1990 | A |
4895790 | Swanson et al. | Jan 1990 | A |
4896030 | Miyaji | Jan 1990 | A |
4907870 | Brucker | Mar 1990 | A |
4910591 | Petrossian et al. | Mar 1990 | A |
4916374 | Schierbeek et al. | Apr 1990 | A |
4917477 | Bechtel et al. | Apr 1990 | A |
4937796 | Tendler | Jun 1990 | A |
4953305 | Van Lente et al. | Sep 1990 | A |
4956591 | Schierbeek et al. | Sep 1990 | A |
4961625 | Wood et al. | Oct 1990 | A |
4967319 | Seko | Oct 1990 | A |
4970653 | Kenue | Nov 1990 | A |
4971430 | Lynas | Nov 1990 | A |
4974078 | Tsai | Nov 1990 | A |
4987357 | Masaki | Jan 1991 | A |
4991054 | Walters | Feb 1991 | A |
5001558 | Burley | Mar 1991 | A |
5003288 | Wilhelm | Mar 1991 | A |
5012082 | Watanabe | Apr 1991 | A |
5016977 | Baude et al. | May 1991 | A |
5027001 | Torbert | Jun 1991 | A |
5027200 | Petrossian et al. | Jun 1991 | A |
5044706 | Chen | Sep 1991 | A |
5055668 | French | Oct 1991 | A |
5059877 | Teder | Oct 1991 | A |
5064274 | Alten | Nov 1991 | A |
5072154 | Chen | Dec 1991 | A |
5086253 | Lawler | Feb 1992 | A |
5096287 | Kakinami et al. | Mar 1992 | A |
5097362 | Lynas | Mar 1992 | A |
5121200 | Choi | Jun 1992 | A |
5124549 | Michaels et al. | Jun 1992 | A |
5130709 | Toyama et al. | Jul 1992 | A |
5140465 | Yasui et al. | Aug 1992 | A |
5148014 | Lynam et al. | Sep 1992 | A |
5168378 | Black | Dec 1992 | A |
5170374 | Shimohigashi et al. | Dec 1992 | A |
5172235 | Wilm et al. | Dec 1992 | A |
5177685 | Davis et al. | Jan 1993 | A |
5182502 | Slotkowski et al. | Jan 1993 | A |
5184956 | Langlais et al. | Feb 1993 | A |
5189561 | Hong | Feb 1993 | A |
5193000 | Lipton et al. | Mar 1993 | A |
5193029 | Schofield et al. | Mar 1993 | A |
5204778 | Bechtel | Apr 1993 | A |
5208701 | Maeda | May 1993 | A |
5245422 | Borcherts et al. | Sep 1993 | A |
5253109 | O'Farrell et al. | Oct 1993 | A |
5276389 | Levers | Jan 1994 | A |
5285060 | Larson et al. | Feb 1994 | A |
5289182 | Brillard et al. | Feb 1994 | A |
5289321 | Secor | Feb 1994 | A |
5305012 | Faris | Apr 1994 | A |
5307136 | Saneyoshi | Apr 1994 | A |
5309137 | Kajiwara | May 1994 | A |
5313072 | Vachss | May 1994 | A |
5325096 | Pakett | Jun 1994 | A |
5325386 | Jewell et al. | Jun 1994 | A |
5329206 | Slotkowski et al. | Jul 1994 | A |
5331312 | Kudoh | Jul 1994 | A |
5336980 | Levers | Aug 1994 | A |
5341437 | Nakayama | Aug 1994 | A |
5351044 | Mathur et al. | Sep 1994 | A |
5355118 | Fukuhara | Oct 1994 | A |
5374852 | Parkes | Dec 1994 | A |
5386285 | Asayama | Jan 1995 | A |
5394333 | Kao | Feb 1995 | A |
5406395 | Wilson et al. | Apr 1995 | A |
5410346 | Saneyoshi et al. | Apr 1995 | A |
5414257 | Stanton | May 1995 | A |
5414461 | Kishi et al. | May 1995 | A |
5416313 | Larson et al. | May 1995 | A |
5416318 | Hegyi | May 1995 | A |
5416478 | Morinaga | May 1995 | A |
5424952 | Asayama | Jun 1995 | A |
5426294 | Kobayashi et al. | Jun 1995 | A |
5430431 | Nelson | Jul 1995 | A |
5434407 | Bauer et al. | Jul 1995 | A |
5440428 | Klegg et al. | Aug 1995 | A |
5444478 | Lelong et al. | Aug 1995 | A |
5451822 | Bechtel et al. | Sep 1995 | A |
5457493 | Leddy et al. | Oct 1995 | A |
5461357 | Yoshioka et al. | Oct 1995 | A |
5461361 | Moore | Oct 1995 | A |
5469298 | Suman et al. | Nov 1995 | A |
5471515 | Fossum et al. | Nov 1995 | A |
5475494 | Nishida et al. | Dec 1995 | A |
5487116 | Nakano et al. | Jan 1996 | A |
5498866 | Bendicks et al. | Mar 1996 | A |
5500766 | Stonecypher | Mar 1996 | A |
5510983 | Lino | Apr 1996 | A |
5515448 | Nishitani | May 1996 | A |
5521633 | Nakajima et al. | May 1996 | A |
5528698 | Kamei et al. | Jun 1996 | A |
5529138 | Shaw et al. | Jun 1996 | A |
5530240 | Larson et al. | Jun 1996 | A |
5530420 | Tsuchiya et al. | Jun 1996 | A |
5535314 | Alves et al. | Jul 1996 | A |
5537003 | Bechtel et al. | Jul 1996 | A |
5539397 | Asanuma et al. | Jul 1996 | A |
5541590 | Nishio | Jul 1996 | A |
5550677 | Schofield et al. | Aug 1996 | A |
5555312 | Shima et al. | Sep 1996 | A |
5555555 | Sato et al. | Sep 1996 | A |
5568027 | Teder | Oct 1996 | A |
5574443 | Hsieh | Nov 1996 | A |
5581464 | Woll et al. | Dec 1996 | A |
5594222 | Caldwell | Jan 1997 | A |
5614788 | Mullins | Mar 1997 | A |
5619370 | Guinosso | Apr 1997 | A |
5634709 | Iwama | Jun 1997 | A |
5642299 | Hardin et al. | Jun 1997 | A |
5648835 | Uzawa | Jul 1997 | A |
5650944 | Kise | Jul 1997 | A |
5660454 | Mori et al. | Aug 1997 | A |
5661303 | Teder | Aug 1997 | A |
5666028 | Bechtel et al. | Sep 1997 | A |
5668663 | Varaprasad et al. | Sep 1997 | A |
5670935 | Schofield et al. | Sep 1997 | A |
5675489 | Pomerleau | Oct 1997 | A |
5677851 | Kingdon et al. | Oct 1997 | A |
5699044 | Van Lente et al. | Dec 1997 | A |
5703568 | Hegyi | Dec 1997 | A |
5724187 | Varaprasad et al. | Mar 1998 | A |
5724316 | Brunts | Mar 1998 | A |
5737226 | Olson et al. | Apr 1998 | A |
5757949 | Kinoshita et al. | May 1998 | A |
5760826 | Nayar | Jun 1998 | A |
5760828 | Cortes | Jun 1998 | A |
5760931 | Saburi et al. | Jun 1998 | A |
5760962 | Schofield et al. | Jun 1998 | A |
5761094 | Olson et al. | Jun 1998 | A |
5765116 | Wilson-Jones et al. | Jun 1998 | A |
5781437 | Wiemer et al. | Jul 1998 | A |
5786772 | Schofield et al. | Jul 1998 | A |
5790403 | Nakayama | Aug 1998 | A |
5790973 | Blaker et al. | Aug 1998 | A |
5793308 | Rosinski et al. | Aug 1998 | A |
5793420 | Schmidt | Aug 1998 | A |
5796094 | Schofield et al. | Aug 1998 | A |
5798575 | O'Farrell et al. | Aug 1998 | A |
5835255 | Miles | Nov 1998 | A |
5837994 | Stam et al. | Nov 1998 | A |
5844505 | Van Ryzin | Dec 1998 | A |
5844682 | Kiyomoto et al. | Dec 1998 | A |
5845000 | Breed et al. | Dec 1998 | A |
5848802 | Breed et al. | Dec 1998 | A |
5850176 | Kinoshita et al. | Dec 1998 | A |
5850254 | Takano et al. | Dec 1998 | A |
5867591 | Onda | Feb 1999 | A |
5877707 | Kowalick | Mar 1999 | A |
5877897 | Schofield et al. | Mar 1999 | A |
5878370 | Olson | Mar 1999 | A |
5883739 | Ashihara et al. | Mar 1999 | A |
5884212 | Lion | Mar 1999 | A |
5890021 | Onoda | Mar 1999 | A |
5896085 | Mori et al. | Apr 1999 | A |
5899956 | Chan | May 1999 | A |
5914815 | Bos | Jun 1999 | A |
5923027 | Stam et al. | Jul 1999 | A |
5929786 | Schofield et al. | Jul 1999 | A |
5940120 | Frankhouse et al. | Aug 1999 | A |
5949331 | Schofield et al. | Sep 1999 | A |
5956181 | Lin | Sep 1999 | A |
5959367 | O'Farrell et al. | Sep 1999 | A |
5959555 | Furuta | Sep 1999 | A |
5963247 | Banitt | Oct 1999 | A |
5964822 | Alland et al. | Oct 1999 | A |
5971552 | O'Farrell et al. | Oct 1999 | A |
5986796 | Miles | Nov 1999 | A |
5990469 | Bechtel et al. | Nov 1999 | A |
5990649 | Nagao et al. | Nov 1999 | A |
6001486 | Varaprasad et al. | Dec 1999 | A |
6009336 | Harris et al. | Dec 1999 | A |
6020704 | Buschur | Feb 2000 | A |
6049171 | Stam et al. | Apr 2000 | A |
6066933 | Ponziana | May 2000 | A |
6084519 | Coulling et al. | Jul 2000 | A |
6087953 | DeLine et al. | Jul 2000 | A |
6097023 | Schofield et al. | Aug 2000 | A |
6097024 | Stam et al. | Aug 2000 | A |
6116743 | Hoek | Sep 2000 | A |
6118383 | Hegyi | Sep 2000 | A |
6124647 | Marcus et al. | Sep 2000 | A |
6124886 | DeLine et al. | Sep 2000 | A |
6139172 | Bos et al. | Oct 2000 | A |
6144022 | Tenenbaum et al. | Nov 2000 | A |
6172613 | DeLine et al. | Jan 2001 | B1 |
6175164 | O'Farrell et al. | Jan 2001 | B1 |
6175300 | Kendrick | Jan 2001 | B1 |
6198409 | Schofield et al. | Mar 2001 | B1 |
6201642 | Bos | Mar 2001 | B1 |
6222447 | Schofield et al. | Apr 2001 | B1 |
6222460 | DeLine et al. | Apr 2001 | B1 |
6243003 | DeLine et al. | Jun 2001 | B1 |
6250148 | Lynam | Jun 2001 | B1 |
6259412 | Duroux | Jul 2001 | B1 |
6266082 | Yonezawa et al. | Jul 2001 | B1 |
6266442 | Laumeyer et al. | Jul 2001 | B1 |
6285393 | Shimoura et al. | Sep 2001 | B1 |
6291906 | Marcus et al. | Sep 2001 | B1 |
6294989 | Schofield et al. | Sep 2001 | B1 |
6297781 | Turnbull et al. | Oct 2001 | B1 |
6302545 | Schofield et al. | Oct 2001 | B1 |
6310611 | Caldwell | Oct 2001 | B1 |
6313454 | Bos et al. | Nov 2001 | B1 |
6317057 | Lee | Nov 2001 | B1 |
6320176 | Schofield et al. | Nov 2001 | B1 |
6320282 | Caldwell | Nov 2001 | B1 |
6326613 | Heslin et al. | Dec 2001 | B1 |
6329925 | Skiver et al. | Dec 2001 | B1 |
6333759 | Mazzilli | Dec 2001 | B1 |
6341523 | Lynam | Jan 2002 | B2 |
6353392 | Schofield et al. | Mar 2002 | B1 |
6366213 | DeLine et al. | Apr 2002 | B2 |
6370329 | Teuchert | Apr 2002 | B1 |
6396397 | Bos et al. | May 2002 | B1 |
6411204 | Bloomfield et al. | Jun 2002 | B1 |
6411328 | Franke et al. | Jun 2002 | B1 |
6420975 | DeLine et al. | Jul 2002 | B1 |
6424273 | Gutta et al. | Jul 2002 | B1 |
6428172 | Hutzel et al. | Aug 2002 | B1 |
6430303 | Naoi et al. | Aug 2002 | B1 |
6433676 | DeLine et al. | Aug 2002 | B2 |
6433817 | Guerra | Aug 2002 | B1 |
6442465 | Breed et al. | Aug 2002 | B2 |
6477464 | McCarthy et al. | Nov 2002 | B2 |
6485155 | Duroux et al. | Nov 2002 | B1 |
6497503 | Dassanayake et al. | Dec 2002 | B1 |
6498620 | Schofield et al. | Dec 2002 | B2 |
6513252 | Schierbeek | Feb 2003 | B1 |
6516664 | Lynam | Feb 2003 | B2 |
6523964 | Schofield et al. | Feb 2003 | B2 |
6534884 | Marcus et al. | Mar 2003 | B2 |
6539306 | Turnbull | Mar 2003 | B2 |
6547133 | Devries, Jr. et al. | Apr 2003 | B1 |
6553130 | Lemelson et al. | Apr 2003 | B1 |
6559435 | Schofield et al. | May 2003 | B2 |
6574033 | Chui et al. | Jun 2003 | B1 |
6587573 | Stam et al. | Jul 2003 | B1 |
6589625 | Kothari et al. | Jul 2003 | B1 |
6593565 | Heslin et al. | Jul 2003 | B2 |
6594583 | Ogura et al. | Jul 2003 | B2 |
6611202 | Schofield et al. | Aug 2003 | B2 |
6611610 | Stam et al. | Aug 2003 | B1 |
6625587 | Erten et al. | Sep 2003 | B1 |
6627918 | Getz et al. | Sep 2003 | B2 |
6636258 | Strumolo | Oct 2003 | B2 |
6648477 | Hutzel et al. | Nov 2003 | B2 |
6650233 | DeLine et al. | Nov 2003 | B2 |
6650455 | Miles | Nov 2003 | B2 |
6672731 | Schnell et al. | Jan 2004 | B2 |
6674562 | Miles | Jan 2004 | B1 |
6678056 | Downs | Jan 2004 | B2 |
6678614 | McCarthy et al. | Jan 2004 | B2 |
6680792 | Miles | Jan 2004 | B2 |
6690268 | Schofield et al. | Feb 2004 | B2 |
6700605 | Toyoda et al. | Mar 2004 | B1 |
6703925 | Steffel | Mar 2004 | B2 |
6704621 | Stein et al. | Mar 2004 | B1 |
6710908 | Miles et al. | Mar 2004 | B2 |
6711474 | Treyz et al. | Mar 2004 | B1 |
6714331 | Lewis et al. | Mar 2004 | B2 |
6717610 | Bos et al. | Apr 2004 | B1 |
6735506 | Breed et al. | May 2004 | B2 |
6741377 | Miles | May 2004 | B2 |
6744353 | Sjonell | Jun 2004 | B2 |
6757109 | Bos | Jun 2004 | B2 |
6759646 | Acharya | Jul 2004 | B1 |
6762867 | Lippert et al. | Jul 2004 | B2 |
6794119 | Miles | Sep 2004 | B2 |
6795221 | Urey | Sep 2004 | B1 |
6802617 | Schofield et al. | Oct 2004 | B2 |
6806452 | Bos et al. | Oct 2004 | B2 |
6822563 | Bos et al. | Nov 2004 | B2 |
6823241 | Shirato et al. | Nov 2004 | B2 |
6824281 | Schofield et al. | Nov 2004 | B2 |
6831261 | Schofield et al. | Dec 2004 | B2 |
6847487 | Burgner | Jan 2005 | B2 |
6882287 | Schofield | Apr 2005 | B2 |
6889161 | Winner et al. | May 2005 | B2 |
6891563 | Schofield et al. | May 2005 | B2 |
6909753 | Meehan et al. | Jun 2005 | B2 |
6946978 | Schofield | Sep 2005 | B2 |
6953253 | Schofield et al. | Oct 2005 | B2 |
6968736 | Lynam | Nov 2005 | B2 |
6975775 | Rykowski et al. | Dec 2005 | B2 |
7004593 | Weller et al. | Feb 2006 | B2 |
7004606 | Schofield | Feb 2006 | B2 |
7005974 | McMahon et al. | Feb 2006 | B2 |
7038577 | Pawlicki et al. | May 2006 | B2 |
7046448 | Burgner | May 2006 | B2 |
7062300 | Kim | Jun 2006 | B1 |
7065432 | Moisel et al. | Jun 2006 | B2 |
7085637 | Breed et al. | Aug 2006 | B2 |
7092548 | Laumeyer et al. | Aug 2006 | B2 |
7116246 | Winter et al. | Oct 2006 | B2 |
7123168 | Schofield | Oct 2006 | B2 |
7133661 | Hatae et al. | Nov 2006 | B2 |
7149613 | Stam et al. | Dec 2006 | B2 |
7167796 | Taylor et al. | Jan 2007 | B2 |
7195381 | Lynam et al. | Mar 2007 | B2 |
7202776 | Breed | Apr 2007 | B2 |
7224324 | Quist et al. | May 2007 | B2 |
7227459 | Bos et al. | Jun 2007 | B2 |
7227611 | Hull et al. | Jun 2007 | B2 |
7249860 | Kulas et al. | Jul 2007 | B2 |
7253723 | Lindahl et al. | Aug 2007 | B2 |
7255451 | McCabe et al. | Aug 2007 | B2 |
7311406 | Schofield et al. | Dec 2007 | B2 |
7325934 | Schofield et al. | Feb 2008 | B2 |
7325935 | Schofield et al. | Feb 2008 | B2 |
7338177 | Lynam | Mar 2008 | B2 |
7339149 | Schofield et al. | Mar 2008 | B1 |
7344261 | Schofield et al. | Mar 2008 | B2 |
7360932 | Uken et al. | Apr 2008 | B2 |
7370983 | DeWind et al. | May 2008 | B2 |
7375803 | Bamji | May 2008 | B1 |
7380948 | Schofield et al. | Jun 2008 | B2 |
7388182 | Schofield et al. | Jun 2008 | B2 |
7402786 | Schofield et al. | Jul 2008 | B2 |
7423248 | Schofield et al. | Sep 2008 | B2 |
7423821 | Bechtel et al. | Sep 2008 | B2 |
7425076 | Schofield et al. | Sep 2008 | B2 |
7459664 | Schofield et al. | Dec 2008 | B2 |
7526103 | Schofield et al. | Apr 2009 | B2 |
7541743 | Salmeen et al. | Jun 2009 | B2 |
7546026 | Pertsel et al. | Jun 2009 | B2 |
7561181 | Schofield et al. | Jul 2009 | B2 |
7565006 | Stam et al. | Jul 2009 | B2 |
7616781 | Schofield et al. | Nov 2009 | B2 |
7619508 | Lynam et al. | Nov 2009 | B2 |
7633383 | Dunsmoir et al. | Dec 2009 | B2 |
7639149 | Katoh | Dec 2009 | B2 |
7676087 | Dhua et al. | Mar 2010 | B2 |
7720580 | Higgins-Luthman | May 2010 | B2 |
7746396 | Mikkonen | Jun 2010 | B2 |
7755016 | Toda et al. | Jul 2010 | B2 |
7792329 | Schofield et al. | Sep 2010 | B2 |
7859565 | Schofield et al. | Dec 2010 | B2 |
7881496 | Camilleri et al. | Feb 2011 | B2 |
7914187 | Higgins-Luthman et al. | Mar 2011 | B2 |
8017898 | Lu et al. | Sep 2011 | B2 |
8095310 | Taylor et al. | Jan 2012 | B2 |
8098142 | Schofield et al. | Jan 2012 | B2 |
8224031 | Saito | Jul 2012 | B2 |
8237794 | Moritz et al. | Aug 2012 | B2 |
8446470 | Lu | May 2013 | B2 |
8908040 | Lu | Dec 2014 | B2 |
10003755 | Lu | Jun 2018 | B2 |
10616507 | Lu et al. | Apr 2020 | B2 |
20020015153 | Downs | Feb 2002 | A1 |
20020044065 | Quist et al. | Apr 2002 | A1 |
20020113873 | Williams | Aug 2002 | A1 |
20020159270 | Lynam et al. | Oct 2002 | A1 |
20030137586 | Lewellen | Jul 2003 | A1 |
20030222982 | Hamdan et al. | Dec 2003 | A1 |
20030227777 | Schofield | Dec 2003 | A1 |
20040012488 | Schofield | Jan 2004 | A1 |
20040016870 | Pawlicki et al. | Jan 2004 | A1 |
20040032321 | McMahon et al. | Feb 2004 | A1 |
20040051634 | Schofield et al. | Mar 2004 | A1 |
20040114381 | Salmeen et al. | Jun 2004 | A1 |
20040128065 | Taylor et al. | Jul 2004 | A1 |
20040200948 | Bos et al. | Oct 2004 | A1 |
20050029456 | Eggers | Feb 2005 | A1 |
20050078389 | Kulas et al. | Apr 2005 | A1 |
20050134966 | Burgner | Jun 2005 | A1 |
20050134983 | Lynam | Jun 2005 | A1 |
20050146792 | Schofield et al. | Jul 2005 | A1 |
20050169003 | Lindahl et al. | Aug 2005 | A1 |
20050195488 | McCabe et al. | Sep 2005 | A1 |
20050200700 | Schofield et al. | Sep 2005 | A1 |
20050219852 | Stam et al. | Oct 2005 | A1 |
20050232469 | Schofield et al. | Oct 2005 | A1 |
20050237385 | Kosaka et al. | Oct 2005 | A1 |
20050264891 | Uken et al. | Dec 2005 | A1 |
20050270784 | Hahn | Dec 2005 | A1 |
20060018511 | Stam et al. | Jan 2006 | A1 |
20060018512 | Stam et al. | Jan 2006 | A1 |
20060028731 | Schofield et al. | Feb 2006 | A1 |
20060050018 | Hutzel et al. | Mar 2006 | A1 |
20060091813 | Stam et al. | May 2006 | A1 |
20060103727 | Tseng | May 2006 | A1 |
20060164230 | DeWind et al. | Jul 2006 | A1 |
20060171704 | Bingle et al. | Aug 2006 | A1 |
20060250501 | Wildmann et al. | Nov 2006 | A1 |
20060250711 | Noguchi | Nov 2006 | A1 |
20070023613 | Schofield et al. | Feb 2007 | A1 |
20070104476 | Yasutomi et al. | May 2007 | A1 |
20070109406 | Schofield et al. | May 2007 | A1 |
20070109651 | Schofield et al. | May 2007 | A1 |
20070109652 | Schofield et al. | May 2007 | A1 |
20070109653 | Schofield et al. | May 2007 | A1 |
20070109654 | Schofield et al. | May 2007 | A1 |
20070120657 | Schofield et al. | May 2007 | A1 |
20070145273 | Chang | Jun 2007 | A1 |
20070153335 | Hosaka | Jul 2007 | A1 |
20070176080 | Schofield et al. | Aug 2007 | A1 |
20070201738 | Toda et al. | Aug 2007 | A1 |
20070216768 | Smith | Sep 2007 | A1 |
20070242339 | Bradley | Oct 2007 | A1 |
20080029701 | Onozawa | Feb 2008 | A1 |
20080036576 | Stein | Feb 2008 | A1 |
20080063237 | Rubenstein | Mar 2008 | A1 |
20080080028 | Bakin et al. | Apr 2008 | A1 |
20080147321 | Howard et al. | Jun 2008 | A1 |
20080180529 | Taylor et al. | Jul 2008 | A1 |
20080192132 | Bechtel et al. | Aug 2008 | A1 |
20080218597 | Cho | Sep 2008 | A1 |
20080278591 | Barna et al. | Nov 2008 | A1 |
20080298708 | Ovsiannikov et al. | Dec 2008 | A1 |
20090113509 | Tseng et al. | Apr 2009 | A1 |
20090141128 | Seger | Jun 2009 | A1 |
20090160987 | Bechtel et al. | Jun 2009 | A1 |
20090190015 | Bechtel et al. | Jul 2009 | A1 |
20090208058 | Schofield | Aug 2009 | A1 |
20090256938 | Bechtel et al. | Oct 2009 | A1 |
20100045797 | Schofield et al. | Feb 2010 | A1 |
20100214791 | Schofield | Aug 2010 | A1 |
20120045112 | Lundblad et al. | Feb 2012 | A1 |
20130329989 | Palum et al. | Dec 2013 | A1 |
Number | Date | Country |
---|---|---|
4107965 | Sep 1991 | DE |
4118208 | Nov 1991 | DE |
4139515 | Jun 1992 | DE |
4123641 | Jan 1993 | DE |
0416222 | Mar 1991 | EP |
0426503 | May 1991 | EP |
0513476 | Nov 1992 | EP |
2641237 | Jul 1990 | FR |
2672857 | Aug 1992 | FR |
2673499 | Sep 1992 | FR |
2726144 | Apr 1996 | FR |
2244187 | Nov 1991 | GB |
2255539 | Nov 1992 | GB |
2327823 | Feb 1999 | GB |
59114139 | Jul 1984 | JP |
6079889 | Oct 1986 | JP |
6272245 | May 1987 | JP |
64-014700 | Jan 1989 | JP |
1123587 | May 1989 | JP |
01-141137 | Jun 1989 | JP |
03-061192 | Mar 1991 | JP |
03099952 | Apr 1991 | JP |
03284413 | Dec 1991 | JP |
04114587 | Apr 1992 | JP |
04-245866 | Sep 1992 | JP |
05-000638 | Jan 1993 | JP |
0550883 | Mar 1993 | JP |
05-213113 | Aug 1993 | JP |
61-056638 | Jun 1994 | JP |
62-027318 | Aug 1994 | JP |
06267304 | Sep 1994 | JP |
06276524 | Sep 1994 | JP |
06295601 | Oct 1994 | JP |
07004170 | Jan 1995 | JP |
07-032936 | Feb 1995 | JP |
07047878 | Feb 1995 | JP |
07052706 | Feb 1995 | JP |
07069125 | Mar 1995 | JP |
07015496 | Apr 1995 | JP |
08166221 | Jun 1996 | JP |
05-077657 | Jul 1997 | JP |
2630604 | Jul 1997 | JP |
2003083742 | Mar 2003 | JP |
199621581 | Jul 1996 | WO |
1986005147 | Sep 1996 | WO |
1997035743 | Oct 1997 | WO |
1998014974 | Apr 1998 | WO |
Entry |
---|
Article entitled “Generation of Vision Technology,” published by VLSI Vision Limited, publication date unknown. |
Article entitled On-Chip CMOS Sensors for VLSI Imaging Systems,: published by VLSI Vision Limited, 1991. |
Ballard et al., Computer Vision, Prentice-Hall, Englewood Cliffs, New Jersey, 5 pages, 1982. |
Borenstein et al., “Where am I? Sensors and Method for Mobile Robot Positioning”, University of Michigan, Apr. 1996, pp. 2, 125-128. |
Bow, Sing T., “Pattern Recognition and Image Preprocessing (Signal Processing and Communications)”, CRC Press, Jan. 15, 2002, pp. 557-559. |
Decision—Motions—Bd. R. 125(a), dated Aug. 29, 2006 in connection with Interference No. 105,325, which involved U.S. Appl. No. 09/441,341, filed Nov. 16, 1999 by Schofield et al. and U.S. Pat. No. 5,837,994, issued to Stam et al. cited byapplicant. |
Hamit, Francis “360-Degree Interactivity: New Video and Still Cameras Provide a Global Roaming Viewpoint,” Advanced Imaging, Mar. 1997, p. 50. |
Johannes, Laura “A New Microchip Ushers in Cheaper Digital Cameras,” The Wall Street Journal, Aug. 21, 1998, p. B1. |
Pratt, “Digital Image Processing, Passage—ED.3”, John Wiley & Sons, US, Jan. 1, 2001, pp. 657-659, XP002529771. |
Reexamination Control No. 90/007,519, Reexamination of U.S. Pat. No. 6,222,447, issued to Schofield et al. |
Reexamination Control No. 90/007,520, Reexamination of U.S. Pat. No. 5,949,331, issued to Schofield et al. |
Reexamination Control No. 90/011,477, Reexamination of U.S. Pat. No. 5,949,331, issued to Schofield et al. |
Reexamination Control No. 90/011,478, Reexamination of U.S. Pat. No. 6,222,447, issued to Schofield et al. |
Tokimaru et al., “CMOS Rear-View Tv System with CCD Camera”, National Technical Report vol. 34, No. 3, pp. 329-336, Jun. 1988 (Japan). |
Van Leeuwan et al., “Motion Estimation in Image Sequences for Traffic Applications”, vol. 1, May 1, 2000, pp. 354-359, XP002529773. |
Van Leeuwan et al., “Motion Estimation with a Mobile Camera for Traffic Applications”, IEEE, US, vol. 1, Oct. 3, 2000, pp. 58-63. |
Van Leeuwan et al., “Motion Interpretation for In-Car Vision Systems”, IEEE, US, vol. 1, Sep. 30, 2002, p. 135-140. |
Van Leeuwan et al., “Real-Time Vehicle Tracking in Image Sequences”, IEEE, US, vol. 3, May 21, 2001, pp. 2049-2054, XP010547308. |
Van Leeuwan et al., “Requirements for Motion Estimation in Image Sequences for Traffic Applications”, IEEE, US, vol. 1, May 24, 1999, pp. 145-150, XP010340272. |
Vlacic et al. (Eds), “Intelligent Vehicle Tecnologies, Theory and Applications”, Society of Automotive Engineers Inc., edited by SAE International, 2001. |
Wang et al., CMOS Video Cameras, article, 1991, 4 pages, University of Edinburgh, UK. |
Number | Date | Country | |
---|---|---|---|
20200236306 A1 | Jul 2020 | US |
Number | Date | Country | |
---|---|---|---|
60977430 | Oct 2007 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16010581 | Jun 2018 | US |
Child | 16840586 | US | |
Parent | 14562826 | Dec 2014 | US |
Child | 16010581 | US | |
Parent | 13896845 | May 2013 | US |
Child | 14562826 | US | |
Parent | 12681204 | US | |
Child | 13896845 | US |