Vehicular driver monitoring system with camera view optimization

Information

  • Patent Grant
  • 12131511
  • Patent Number
    12,131,511
  • Date Filed
    Monday, April 29, 2024
    8 months ago
  • Date Issued
    Tuesday, October 29, 2024
    2 months ago
Abstract
A vehicular driver monitoring system includes a camera disposed at a mirror head of an interior rearview mirror of a vehicle and viewing at least the head of a driver of the vehicle. The camera moves together and in tandem with the mirror head when the mirror head is adjusted by the driver of the vehicle to adjust a rearward view of the driver. Responsive to the driver of the vehicle adjusting the orientation of the mirror head, and responsive to processing at the ECU of image data captured by the camera, the vehicular driver monitoring system maintains the camera's view of at least the head of the driver as it was before the driver adjusted the orientation of the mirror head. The vehicular driver monitoring system, at least in part via processing at the ECU of image data captured by the camera, monitors the driver of the vehicle.
Description
FIELD OF THE INVENTION

The present invention relates generally to a vehicle vision system for a vehicle and, more particularly, to a vehicle vision system that utilizes one or more cameras at a vehicle.


BACKGROUND OF THE INVENTION

Use of imaging sensors in vehicle imaging systems is common and known. Examples of such known systems are described in U.S. Pat. Nos. 5,949,331; 5,670,935 and/or 5,550,677, which are hereby incorporated herein by reference in their entireties.


SUMMARY OF THE INVENTION

The present invention provides a driving assistance system or vision system or imaging system for a vehicle that includes a camera disposed at an interior rearview mirror of a vehicle equipped with the vehicular driver assist system and viewing interior of the vehicle. The camera captures image data. The system includes an electronic control unit (ECU) that has electronic circuitry and associated software. The electronic circuitry of the ECU includes an image processor for processing image data captured by the camera to detect presence of objects (e.g., a vehicle driver and/or passenger) in the field of view of the camera. The camera includes an imaging array sensor and a lens. The lens is disposed at the imaging array sensor with a center axis of the lens vertically offset from a center region of the imaging array sensor such that a center region of the lens is disposed at and images at a lower region of the imaging array sensor when the camera is disposed at the interior rearview mirror, and an upper peripheral region of the lens images at the center region and a lower region of the imaging array sensor when the camera is disposed at the interior rearview mirror. The ECU, responsive to processing by the image processor of image data captured by the camera, determines presence of an occupant of the vehicle.


These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a plan view of a vehicle with a vision system that incorporates cameras in accordance with the present invention;



FIG. 2 is a side view of an interior of a vehicle with an interior camera;



FIG. 3 is a plan view of an outline of a lens over an outline of an imager of the camera of the vision system of FIG. 1; and



FIG. 4 is a side view of an interior of a vehicle with the vision system of FIG. 1.





DESCRIPTION OF THE PREFERRED EMBODIMENTS

A vehicle vision system and/or driver or driving assist system and/or object detection system and/or alert system operates to capture data of an interior of the vehicle and may process the data to detect objects within the vehicle. The system includes a processor or processing system that is operable to receive data from one or more sensors.


Referring now to the drawings and the illustrative embodiments depicted therein, a vehicle 10 includes a sensing system or driver assist system 12 that includes at least one interior viewing imaging sensor or camera, such as a rearview mirror imaging sensor or camera 16 (FIG. 1). Optionally, an interior viewing camera may be disposed at the windshield of the vehicle. The vision system 12 includes a control or electronic control unit (ECU) 18 having electronic circuitry and associated software, with the electronic circuitry including a data processor or image processor that is operable to process image data captured by the sensor or camera or cameras, whereby the ECU may detect or determine presence of objects or the like (although shown in FIG. 1 as being part of or incorporated in or at an interior rearview mirror assembly 20 of the vehicle, the control and/or the display device may be disposed elsewhere at or in the vehicle). The data transfer or signal communication from the sensor or camera to the ECU may comprise any suitable data or communication link, such as a vehicle network bus or the like of the equipped vehicle.


Referring now to FIG. 2, in some driver monitoring systems or driver assist systems, an interior camera is attached to or is mounted within the interior rearview mirror to, for example, monitor the interior of the vehicle. Thus, when the driver adjusts the rearview mirror (i.e., rotates, tilts, or otherwise moves the interior rearview mirror to adjust the driver's rearward view), the field of view of the interior camera is also changed. The change in the field of view of the camera may lead to blind spots in the surveillance of the cabin or otherwise decrease the effectiveness of the interior camera. For example, some areas of the cabin (e.g., the driver or other occupant) may be captured by a lower resolution portion of the image sensor of the camera as the resolution may decrease in outer regions of the field of view of the camera (such as due to distortion in the image data imaged at the imager by a wide angle or curved or spherical lens). As shown in FIG. 2, the field of view of the camera is suboptimal, as a large portion of the field of view of the camera is of the ceiling or roofliner or headliner of the vehicle.


Implementations herein include a vehicular driver monitoring system that may align a lens of the camera vertically and/or horizontally offset from a center of the image sensor (FIGS. 3 and 4). For example, and such as shown in FIG. 3, the lens may be disposed at the imaging sensor with a center axis of the lens vertically offset from a center region of the imaging array sensor and a center region of the lens is disposed at and images at a lower region of the imaging array sensor when the camera is disposed at the interior rearview mirror. An upper peripheral region of the lens may image at the center region and a lower region of the imaging array sensor when the camera is disposed at the interior rearview mirror (FIG. 4). By doing this, regions of low interest (e.g., the roofliner or ceiling of the vehicle) may be imaged by a lower resolution or more distorted portion of the lens/image sensor (e.g., edges of the lens or edges of the image sensor) while regions of high interest (e.g., portions where occupants heads are likely to be) are imaged with a higher resolution or with less distortion as these regions such as areas that are more centrally located with respect to the lens. The lens and imager construction may utilize aspects of the cameras and systems described in U.S. Pat. No. 10,525,883 and/or U.S. Publication No. US-2016-0137126, which are both hereby incorporated herein by reference in their entireties.


Optionally, the system or camera may adjust the location of the lens respective to the image sensor based on adjustments to the interior rearview mirror (i.e., by the driver). That is, when the driver of the vehicle adjusts the interior rearview mirror, the system may, in response, adjust the position of the lens relative to the image sensor to maintain an optimal field of view within the cabin of the vehicle using, for example, actuators, motors, or the like. For example, the field of view of the camera is adjusted such that the field of view includes less of low priority areas of the cabin (e.g., the roof) and the field of view includes more of high priority areas (e.g., the occupants).


Optionally, the vehicular driver monitoring system may include a gimbal or gear system or other pivoted support or mounting system for the camera. Such a mounting system decouples the field of view of the camera with the orientation of the mirror. For example, when the driver adjusts the mirror, the mounting system (e.g., the gimbal, ball and socket joint, etc.) adjusts the camera so that the field of view of the camera remains the same. The mounting system may allow adjustment of the camera about a single axis or about two or three axes. Thus, even with a very tall or very short driver who makes extreme adjustments to the mirror, the camera can maintain an optimal or ideal field of view. For example, when the driver moves the field of view off the mirror higher, the field of view of the camera may remain relatively lower of the cabin of the vehicle.


Optionally, the camera may alternatively or additionally (i.e., in addition to functioning as an interior driver or occupant monitoring camera) function as a light detector for an auto dimmable mirror. For example, the camera may measure ambient light and/or glare light (e.g., from headlights of vehicles behind the equipped vehicle) and the rear-view mirror and/or side mirrors may dim based on the measurements. Thus, other redundant light detectors or sensors may be removed to reduce costs. Such a camera and ambient light sensing system may utilize aspects of the cameras and systems described in U.S. Publication Nos. US-2021-0316663; US-2021-0245662; US-2019-0146297 and/or US-2019-0047475, and/or International PCT Application No. PCT/US22/70062, filed Jan. 6, 2022, and published Jul. 14, 2022 as International Publication No. WO 2022/150826, which are all hereby incorporated herein by reference in their entireties.


Optionally, the ECU, based on image data captured by the interior camera, calculate areas of vehicle mirrors (e.g., rearview mirror, side mirrors, etc.) that should be dimmed when the driver or other occupants are at risk of glare from a light source from behind the vehicle. That is, based on the captured image data, only portions of the mirror may be dimmed (i.e., the portions that reflect the detected glare light). The ECU may perform the calculations via, for example, triangulation of the mirror(s) and the camera positions. Thus, using the captured image data, the mirror may function as a high-dynamic-range (HDR) mirror. The ECU may use the resolution of the camera to create coordinates (e.g., Cartesian coordinates) of the dimmable area of the mirror. For example, the ECU may define the mirror surface as an array or foil or pixelated structure where each element or portion or region or sub-array of pixels may be selectively or individually dimmed.


The system thus may determine a location of the driver's head and/or eyes, and may determine when glare light (emanating from a headlight of a vehicle rearward of the equipped vehicle) may reflect off the interior mirror reflective element (or exterior mirror reflective element at either side of the vehicle) toward the driver's eyes. Based on the determined location of the driver's head/eyes and the location of the rearward light source relative to the vehicle and mirrors, the system determines which area of each mirror reflective element to dim to reduce or limit or avoid glare light to the driver of the vehicle. The system may utilize aspects of the adaptive masks and systems described in U.S. Publication Nos. US-2021-0114439 and/or US-2020-0296273, which are hereby incorporated herein by reference in their entireties.


The camera or sensor may comprise any suitable camera or sensor. Optionally, the camera may comprise a “smart camera” that includes the imaging sensor array and associated circuitry and image processing circuitry and electrical connectors and the like as part of a camera module, such as by utilizing aspects of the vision systems described in U.S. Pat. Nos. 10,099,614 and/or 10,071,687, which are hereby incorporated herein by reference in their entireties.


The system includes an image processor operable to process image data captured by the camera or cameras, such as for detecting objects or other vehicles or pedestrians or the like in the field of view of one or more of the cameras. For example, the image processor may comprise an image processing chip selected from the EYEQ family of image processing chips available from Mobileye Vision Technologies Ltd. of Jerusalem, Israel, and may include object detection software (such as the types described in U.S. Pat. Nos. 7,855,755; 7,720,580 and/or 7,038,577, which are hereby incorporated herein by reference in their entireties), and may analyze image data to detect vehicles and/or other objects. Responsive to such image processing, and when an object or other vehicle is detected, the system may generate an alert to the driver of the vehicle and/or may generate an overlay at the displayed image to highlight or enhance display of the detected object or vehicle, in order to enhance the driver's awareness of the detected object or vehicle or hazardous condition during a driving maneuver of the equipped vehicle.


The vehicle may include any type of sensor or sensors, such as imaging sensors or radar sensors or lidar sensors or ultrasonic sensors or the like. The imaging sensor or camera may capture image data for image processing and may comprise any suitable camera or sensing device, such as, for example, a two dimensional array of a plurality of photosensor elements arranged in at least 640 columns and 480 rows (at least a 640×480 imaging array, such as a megapixel imaging array or the like), with a respective lens focusing images onto respective portions of the array. The photosensor array may comprise a plurality of photosensor elements arranged in a photosensor array having rows and columns. Preferably, the imaging array has at least 300,000 photosensor elements or pixels, more preferably at least 500,000 photosensor elements or pixels and more preferably at least 1 million photosensor elements or pixels. The imaging array may capture color image data, such as via spectral filtering at the array, such as via an RGB (red, green and blue) filter or via a red/red complement filter or such as via an RCC (red, clear, clear) filter or the like. The logic and control circuit of the imaging sensor may function in any known manner, and the image processing and algorithmic processing may comprise any suitable means for processing the images and/or image data.


For example, the vision system and/or processing and/or camera and/or circuitry may utilize aspects described in U.S. Pat. Nos. 9,233,641; 9,146,898; 9,174,574; 9,090,234; 9,077,098; 8,818,042; 8,886,401; 9,077,962; 9,068,390; 9,140,789; 9,092,986; 9,205,776; 8,917,169; 8,694,224; 7,005,974; 5,760,962; 5,877,897; 5,796,094; 5,949,331; 6,222,447; 6,302,545; 6,396,397; 6,498,620; 6,523,964; 6,611,202; 6,201,642; 6,690,268; 6,717,610; 6,757,109; 6,802,617; 6,806,452; 6,822,563; 6,891,563; 6,946,978; 7,859,565; 5,550,677; 5,670,935; 6,636,258; 7,145,519; 7,161,616; 7,230,640; 7,248,283; 7,295,229; 7,301,466; 7,592,928; 7,881,496; 7,720,580; 7,038,577; 6,882,287; 5,929,786 and/or 5,786,772, and/or U.S. Publication Nos. US-2014-0340510; US-2014-0313339; US-2014-0347486; US-2014-0320658; US-2014-0336876; US-2014-0307095; US-2014-0327774; US-2014-0327772; US-2014-0320636; US-2014-0293057; US-2014-0309884; US-2014-0226012; US-2014-0293042; US-2014-0218535; US-2014-0218535; US-2014-0247354; US-2014-0247355; US-2014-0247352; US-2014-0232869; US-2014-0211009; US-2014-0160276; US-2014-0168437; US-2014-0168415; US-2014-0160291; US-2014-0152825; US-2014-0139676; US-2014-0138140; US-2014-0104426; US-2014-0098229; US-2014-0085472; US-2014-0067206; US-2014-0049646; US-2014-0052340; US-2014-0025240; US-2014-0028852; US-2014-005907; US-2013-0314503; US-2013-0298866; US-2013-0222593; US-2013-0300869; US-2013-0278769; US-2013-0258077; US-2013-0258077; US-2013-0242099; US-2013-0215271; US-2013-0141578 and/or US-2013-0002873, which are all hereby incorporated herein by reference in their entireties. The system may communicate with other communication systems via any suitable means, such as by utilizing aspects of the systems described in U.S. Pat. Nos. 10,071,687; 9,900,490; 9,126,525 and/or 9,036,026, which are hereby incorporated herein by reference in their entireties.


Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.

Claims
  • 1. A vehicular driver monitoring system, the vehicular driver monitoring system comprising: a camera disposed at a mirror head of an interior rearview mirror of a vehicle equipped with the vehicular driver monitoring system and viewing interior of the vehicle, the camera capturing image data;wherein the mirror head accommodates a mirror reflective element;wherein orientation of the mirror head is adjustable by a driver of the vehicle to adjust a rearward view of the driver;wherein the camera moves together and in tandem with the mirror head when the mirror head is adjusted by the driver of the vehicle;wherein the camera comprises a CMOS imaging array, and wherein the CMOS imaging array comprises at least one million photosensors arranged in rows and columns;an electronic control unit (ECU) comprising electronic circuitry and associated software;wherein the electronic circuitry of the ECU comprises an image processor for processing image data captured by the camera;wherein the camera views at least the head of the driver of the vehicle;wherein the camera comprises a mounting system that adjustably mounts the camera to the mirror head of the interior rearview mirror;wherein, responsive to the driver of the vehicle adjusting the orientation of the mirror head, and responsive to processing at the ECU of image data captured by the camera, the vehicular driver monitoring system adjusts the mounting system so that the camera adjusts relative to the mirror head to maintain the view of at least the head of the driver of the vehicle as it was before the driver adjusted the orientation of the mirror head; andwherein the vehicular driver monitoring system, at least in part via processing at the ECU of image data captured by the camera, monitors the driver of the vehicle.
  • 2. The vehicular driver monitoring system of claim 1, wherein the mounting system comprises a gimbal.
  • 3. The vehicular driver monitoring system of claim 2, wherein the mounting system comprises a ball-and-socket joint.
  • 4. The vehicular driver monitoring system of claim 1, wherein the mounting system adjusts the orientation of the camera relative to the mirror head about a single axis.
  • 5. The vehicular driver monitoring system of claim 1, wherein the mounting system adjusts the orientation of the camera relative to the mirror head about a plurality of axes.
  • 6. The vehicular driver monitoring system of claim 5, wherein the plurality of axes comprises three axes.
  • 7. The vehicular driver monitoring system of claim 1, wherein the camera comprises a lens disposed at the CMOS imaging array with a centerline of the lens that does not pass through a center of the CMOS imaging array, and wherein, with the camera disposed at the interior rearview mirror, a center region of the lens is disposed at and images at a lower region of the CMOS imaging array and an upper region of the lens images at the center region of the CMOS imaging array.
  • 8. The vehicular driver monitoring system of claim 7, wherein a vertical offset of the lens relative to the CMOS imaging array increases resolution of image data captured by the camera.
  • 9. The vehicular driver monitoring system of claim 7, wherein a vertical offset of the lens relative to the CMOS imaging array decreases distortion in image data captured by the camera.
  • 10. The vehicular driver monitoring system of claim 1, wherein the mirror reflective element comprises a dimmable mirror reflective element, and wherein the vehicular driver monitoring system, responsive to processing by the image processor of image data captured by the camera, determines intensity of glare light, and wherein the vehicular driver monitoring system, responsive to determining that the intensity of glare light is greater than a threshold level, at least partially dims the dimmable mirror reflective element.
  • 11. The vehicular driver monitoring system of claim 10, wherein, based at least in part on processing of image data captured by the camera, the vehicular driver monitoring system determines a portion of the dimmable mirror reflective element that is reflecting glare light toward eyes of the driver of the vehicle, and wherein the vehicular driver monitoring system, responsive to determining the portion of the dimmable mirror reflective element that is reflecting the glare light, dims the portion of the dimmable mirror reflective element that is reflecting the glare light while not dimming other portions of the dimmable mirror reflective element that are not reflecting the glare light.
  • 12. The vehicular driver monitoring system of claim 11, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element that is reflecting the glare light based on triangulation and a position of the dimmable mirror reflective element and the camera.
  • 13. The vehicular driver monitoring system of claim 12, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element based on an array of locations that represent a surface of the dimmable mirror reflective element.
  • 14. A vehicular driver monitoring system, the vehicular driver monitoring system comprising: a camera disposed at a mirror head of an interior rearview mirror of a vehicle equipped with the vehicular driver monitoring system and viewing interior of the vehicle, the camera capturing image data;wherein the mirror head accommodates a mirror reflective element;wherein orientation of the mirror head is adjustable by a driver of the vehicle to adjust a rearward view of the driver;wherein the camera moves together and in tandem with the mirror head when the mirror head is adjusted by the driver of the vehicle;wherein the camera comprises a CMOS imaging array and a lens, and wherein the CMOS imaging array comprises at least one million photosensors arranged in rows and columns;wherein the lens is disposed at the CMOS imaging array with a centerline of the lens that does not pass through a center of the CMOS imaging array, and wherein, with the camera disposed at the interior rearview mirror, a center region of the lens is disposed at and images at a lower region of the CMOS imaging array and an upper region of the lens images at the center region of the CMOS imaging array;wherein a vertical offset of the lens relative to the CMOS imaging array decreases distortion of images imaged at the lower region of the CMOS imaging array;an electronic control unit (ECU) comprising electronic circuitry and associated software;wherein the electronic circuitry of the ECU comprises an image processor for processing image data captured by the camera;wherein the camera views at least the head of the driver of the vehicle such that the head of the driver is imaged at the lower region of the CMOS imaging array;wherein, responsive to the driver of the vehicle adjusting the orientation of the mirror head, and responsive to processing at the ECU of image data captured by the camera, the vehicular driver monitoring system maintains the camera's view of at least the head of the driver of the vehicle as it was before the driver adjusted the orientation of the mirror head; andwherein the vehicular driver monitoring system, at least in part via processing at the ECU of image data captured by the camera, monitors the driver of the vehicle.
  • 15. The vehicular driver monitoring system of claim 14, wherein the vehicular driver monitoring system adjusts orientation of the camera relative to the mirror head about a single axis to maintain the camera's view of at least the head of the driver of the vehicle.
  • 16. The vehicular driver monitoring system of claim 14, wherein the vehicular driver monitoring system adjusts orientation of the camera relative to the mirror head about a plurality of axes to maintain the camera's view of at least the head of the driver of the vehicle.
  • 17. The vehicular driver monitoring system of claim 16, wherein the plurality of axes comprises three axes.
  • 18. The vehicular driver monitoring system of claim 14, wherein the vertical offset of the lens relative to the CMOS imaging array increases resolution of image data captured by the camera at the lower region of the CMOS imaging array.
  • 19. The vehicular driver monitoring system of claim 14, wherein the mirror reflective element comprises a dimmable mirror reflective element, and wherein the vehicular driver monitoring system, responsive to processing by the image processor of image data captured by the camera, determines intensity of glare light, and wherein the vehicular driver monitoring system, responsive to determining that the intensity of glare light is greater than a threshold level, at least partially dims the dimmable mirror reflective element.
  • 20. The vehicular driver monitoring system of claim 19, wherein, based at least in part on processing of image data captured by the camera, the vehicular driver monitoring system determines a portion of the dimmable mirror reflective element that is reflecting glare light toward eyes of the driver of the vehicle, and wherein the vehicular driver monitoring system, responsive to determining the portion of the dimmable mirror reflective element that is reflecting the glare light, dims the portion of the dimmable mirror reflective element that is reflecting the glare light while not dimming other portions of the dimmable mirror reflective element that are not reflecting the glare light.
  • 21. The vehicular driver monitoring system of claim 20, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element that is reflecting the glare light based on triangulation and a position of the dimmable mirror reflective element and the camera.
  • 22. The vehicular driver monitoring system of claim 21, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element based on an array of locations that represent a surface of the dimmable mirror reflective element.
  • 23. A vehicular driver monitoring system, the vehicular driver monitoring system comprising: a camera disposed at a mirror head of an interior rearview mirror of a vehicle equipped with the vehicular driver monitoring system and viewing interior of the vehicle, the camera capturing image data;wherein the mirror head accommodates a dimmable mirror reflective element;wherein orientation of the mirror head is adjustable by a driver of the vehicle to adjust a rearward view of the driver;wherein the camera moves together and in tandem with the mirror head when the mirror head is adjusted by the driver of the vehicle;wherein the camera comprises a CMOS imaging array, and wherein the CMOS imaging array comprises at least one million photosensors arranged in rows and columns;an electronic control unit (ECU) comprising electronic circuitry and associated software;wherein the electronic circuitry of the ECU comprises an image processor for processing image data captured by the camera;wherein the camera views at least the head of the driver of the vehicle;wherein, based at least in part on processing of image data captured by the camera, the vehicular driver monitoring system determines a portion of the dimmable mirror reflective element that is reflecting glare light toward eyes of the driver of the vehicle, and wherein the vehicular driver monitoring system, responsive to determining the portion of the dimmable mirror reflective element that is reflecting the glare light, dims the portion of the dimmable mirror reflective element that is reflecting the glare light while not dimming other portions of the dimmable mirror reflective element that are not reflecting the glare light;wherein, responsive to the driver of the vehicle adjusting the orientation of the mirror head, and responsive to processing at the ECU of image data captured by the camera, the vehicular driver monitoring system maintains the camera's view of at least the head of the driver of the vehicle as it was before the driver adjusted the orientation of the mirror head; andwherein the vehicular driver monitoring system, at least in part via processing at the ECU of image data captured by the camera, monitors the driver of the vehicle.
  • 24. The vehicular driver monitoring system of claim 23, wherein the vehicular driver monitoring system adjusts orientation of the camera relative to the mirror head about a single axis to maintain the camera's view of at least the head of the driver of the vehicle.
  • 25. The vehicular driver monitoring system of claim 23, wherein the vehicular driver monitoring system adjusts orientation of the camera relative to the mirror head about a plurality of axes to maintain the camera's view of at least the head of the driver of the vehicle.
  • 26. The vehicular driver monitoring system of claim 25, wherein the plurality of axes comprises three axes.
  • 27. The vehicular driver monitoring system of claim 23, wherein the vehicular driver monitoring system, responsive to processing by the image processor of image data captured by the camera, determines intensity of the glare light, and wherein the vehicular driver monitoring system, responsive to determining that the intensity of the glare light is greater than a threshold level, at least partially dims the portion of the dimmable mirror reflective element.
  • 28. The vehicular driver monitoring system of claim 23, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element that is reflecting the glare light based on triangulation and a position of the dimmable mirror reflective element and the camera.
  • 29. The vehicular driver monitoring system of claim 23, wherein the vehicular driver monitoring system determines the portion of the dimmable mirror reflective element based on an array of locations that represent a surface of the dimmable mirror reflective element.
CROSS REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. patent application Ser. No. 17/650,255, filed Feb. 8, 2022, now U.S. Pat. No. 11,972,597, which claims the filing benefits of U.S. provisional application Ser. No. 63/200,003, filed Feb. 9, 2021, which is hereby incorporated herein by reference in its entirety.

US Referenced Citations (174)
Number Name Date Kind
5550677 Schofield et al. Aug 1996 A
5567360 Varaprasad et al. Oct 1996 A
5570127 Schmidt Oct 1996 A
5670935 Schofield et al. Sep 1997 A
5760962 Schofield et al. Jun 1998 A
5796094 Schofield et al. Aug 1998 A
5877897 Schofield et al. Mar 1999 A
6097023 Schofield et al. Aug 2000 A
6158655 DeVries, Jr. et al. Dec 2000 A
6483438 DeLine et al. Nov 2002 B2
6552342 Holz et al. Apr 2003 B2
6593565 Heslin et al. Jul 2003 B2
6627918 Getz et al. Sep 2003 B2
6690268 Schofield et al. Feb 2004 B2
6703925 Steffel Mar 2004 B2
6824281 Schofield et al. Nov 2004 B2
7038577 Pawlicki et al. May 2006 B2
7184190 McCabe et al. Feb 2007 B2
7195381 Lynam et al. Mar 2007 B2
7224324 Quist et al. May 2007 B2
7249860 Kulas et al. Jul 2007 B2
7253723 Lindahl et al. Aug 2007 B2
7255451 McCabe et al. Aug 2007 B2
7274501 McCabe et al. Sep 2007 B2
7289037 Uken et al. Oct 2007 B2
7338177 Lynam Mar 2008 B2
7360932 Uken et al. Apr 2008 B2
7420756 Lynam Sep 2008 B2
7477758 Piirainen et al. Jan 2009 B2
7480149 DeWard et al. Jan 2009 B2
7626749 Baur et al. Dec 2009 B2
7720580 Higgins-Luthman May 2010 B2
7855755 Weller et al. Dec 2010 B2
7914187 Higgins-Luthman et al. Mar 2011 B2
8049640 Uken et al. Nov 2011 B2
8258932 Wahlstrom Sep 2012 B2
8446470 Lu et al. May 2013 B2
8451107 Lu et al. May 2013 B2
8508831 De Wind et al. Aug 2013 B2
8529108 Uken et al. Sep 2013 B2
8730553 De Wind et al. May 2014 B2
8743203 Karner et al. Jun 2014 B2
8876342 Wimbert et al. Nov 2014 B2
8922422 Klar et al. Dec 2014 B2
9090213 Lawlor et al. Jul 2015 B2
9126525 Lynam et al. Sep 2015 B2
9174578 Uken et al. Nov 2015 B2
9280202 Gieseke et al. Mar 2016 B2
9346403 Uken et al. May 2016 B2
9405120 Graf et al. Aug 2016 B2
9487159 Achenbach Nov 2016 B2
9493122 Krebs Nov 2016 B2
9598016 Blank et al. Mar 2017 B2
9609757 Steigerwald Mar 2017 B2
9701258 Tiryaki Jul 2017 B2
9827913 De Wind et al. Nov 2017 B2
9878669 Kendall Jan 2018 B2
9900490 Ihlenburg et al. Feb 2018 B2
10017114 Bongwald Jul 2018 B2
10029614 Larson Jul 2018 B2
10046706 Larson et al. Aug 2018 B2
10065574 Tiryaki Sep 2018 B2
10166924 Baur Jan 2019 B2
10166926 Krebs et al. Jan 2019 B2
10247941 Fursich Apr 2019 B2
10261648 Uken et al. Apr 2019 B2
10315573 Bongwald Jun 2019 B2
10421404 Larson et al. Sep 2019 B2
10442360 LaCross et al. Oct 2019 B2
10466563 Kendall et al. Nov 2019 B2
10525883 Goseberg et al. Jan 2020 B2
10567633 Ihlenburg et al. Feb 2020 B2
10567705 Ziegenspeck et al. Feb 2020 B2
10703204 Hassan et al. Jul 2020 B2
10769434 Weller et al. Sep 2020 B2
10906463 Pflug et al. Feb 2021 B2
10908417 Fürsich Feb 2021 B2
10922563 Nix et al. Feb 2021 B2
10946798 Fürsich et al. Mar 2021 B2
10948798 Lynam et al. Mar 2021 B2
10958830 Koravadi Mar 2021 B2
10967796 Uken et al. Apr 2021 B2
11167771 Caron et al. Nov 2021 B2
11205083 Lynam Dec 2021 B2
11214199 LaCross et al. Jan 2022 B2
11240427 Koravadi Feb 2022 B2
11242008 Blank et al. Feb 2022 B2
11247609 Fürsich et al. Feb 2022 B2
11252376 Ihlenburg Feb 2022 B2
11341671 Lu et al. May 2022 B2
11348374 Kramer et al. May 2022 B2
11433906 Lu Sep 2022 B2
11465561 Peterson et al. Oct 2022 B2
11488399 Wacquant Nov 2022 B2
11493918 Singh Nov 2022 B2
11518401 Kulkarni Dec 2022 B2
11582425 Liu Feb 2023 B2
11972597 Rother et al. Apr 2024 B2
20010022550 Steffel Sep 2001 A1
20020005999 Hutzel et al. Jan 2002 A1
20020024713 Roberts et al. Feb 2002 A1
20040252993 Sato Dec 2004 A1
20060050018 Hutzel et al. Mar 2006 A1
20070182528 Breed et al. Aug 2007 A1
20080310005 Tonar et al. Dec 2008 A1
20090040778 Takayanagi et al. Feb 2009 A1
20090135493 Takayanagi et al. May 2009 A1
20100085653 Uken et al. Apr 2010 A1
20110080481 Bellingham Apr 2011 A1
20110188122 Habibi et al. Aug 2011 A1
20120236136 Boddy Sep 2012 A1
20140022390 Blank et al. Jan 2014 A1
20140285666 O'Connell et al. Sep 2014 A1
20140293169 Uken et al. Oct 2014 A1
20140313563 Uken et al. Oct 2014 A1
20140336876 Gieseke et al. Nov 2014 A1
20150009010 Biemer Jan 2015 A1
20150015710 Tiryaki Jan 2015 A1
20150022664 Pflug et al. Jan 2015 A1
20150092042 Fursich Apr 2015 A1
20150097955 De Wind et al. Apr 2015 A1
20150232030 Bongwald Aug 2015 A1
20150294169 Zhou et al. Oct 2015 A1
20150296135 Wacquant et al. Oct 2015 A1
20150352953 Koravadi Dec 2015 A1
20160044284 Goseberg et al. Feb 2016 A1
20160137126 Fursich et al. May 2016 A1
20160209647 Fursich Jul 2016 A1
20160221509 Takada et al. Aug 2016 A1
20170217367 Pflug et al. Aug 2017 A1
20170237946 Schofield et al. Aug 2017 A1
20170274906 Hassan et al. Sep 2017 A1
20170355312 Habibi et al. Dec 2017 A1
20170355314 May Dec 2017 A1
20180134217 Peterson et al. May 2018 A1
20180222414 Ihlenburg et al. Aug 2018 A1
20180231976 Singh Aug 2018 A1
20180304727 Choi et al. Oct 2018 A1
20190009786 Liu Jan 2019 A1
20190054899 Hoyos et al. Feb 2019 A1
20190118717 Blank et al. Apr 2019 A1
20190146297 Lynam et al. May 2019 A1
20190168669 Lintz et al. Jun 2019 A1
20190188504 Aihara et al. Jun 2019 A1
20190258131 Lynam et al. Aug 2019 A9
20190364199 Koravadi Nov 2019 A1
20190381938 Hopkins Dec 2019 A1
20200133390 Patton Apr 2020 A1
20200143560 Lu et al. May 2020 A1
20200148120 Englander et al. May 2020 A1
20200202151 Wacquant Jun 2020 A1
20200320320 Lynam Oct 2020 A1
20200327323 Noble Oct 2020 A1
20200377022 LaCross et al. Dec 2020 A1
20210056306 Hu et al. Feb 2021 A1
20210122404 Lisseman et al. Apr 2021 A1
20210155167 Lynam et al. May 2021 A1
20210162926 Lu Jun 2021 A1
20210245662 Blank et al. Aug 2021 A1
20210291739 Kasarla et al. Sep 2021 A1
20210306538 Solar Sep 2021 A1
20210316663 Buerkle et al. Oct 2021 A1
20210323473 Peterson et al. Oct 2021 A1
20210368082 Solar Nov 2021 A1
20220111857 Kulkarni Apr 2022 A1
20220229309 Laskin et al. Jul 2022 A1
20220242438 Sobecki et al. Aug 2022 A1
20220254132 Rother Aug 2022 A1
20220377219 Conger et al. Nov 2022 A1
20230131471 Sobecki et al. Apr 2023 A1
20230137004 Huizen et al. May 2023 A1
20230302994 Miller et al. Sep 2023 A1
20240017610 Ravichandran et al. Jan 2024 A1
20240064274 Blank et al. Feb 2024 A1
Foreign Referenced Citations (2)
Number Date Country
2527091 Dec 2015 GB
2022150826 Jul 2022 WO
Related Publications (1)
Number Date Country
20240282069 A1 Aug 2024 US
Provisional Applications (1)
Number Date Country
63200003 Feb 2021 US
Continuations (1)
Number Date Country
Parent 17650255 Feb 2022 US
Child 18648562 US