The present invention relates generally to a vision system for a vehicle and, more particularly, to a vehicular vision system that utilizes one or more cameras to monitor attentiveness of the driver of the vehicle.
Driver attentiveness is an important element of safe driving, with driver distraction and/or driver fatigue being significant contributors to road accidents and the like. Driver monitoring systems utilizing an in-cabin mounted camera that images the head/face/eyes of the driver to determine driver attentiveness are known. However, such known driver monitoring systems have disadvantages, especially when the driver is driving the vehicle during daylight and on sunny days.
The present invention provides a driver monitoring system for a vehicle that utilizes one or more cameras to capture image data representative of images interior of the vehicle, and provides an illumination source that emits light that illuminates at least a portion of a driver of the vehicle. The illumination source modulates or codes the emitted light so that the system (via processing of image data captured by the camera) can process image data in accordance with the modulated light to avoid image washout that may otherwise occur in high ambient lighting conditions, such as a sunny day. A control includes an image processor operable to process image data captured by the camera. The camera captures image data representative of the light that is emitted by the illumination source and that reflects off the illuminated portion of the driver of the vehicle. The control, responsive to processing of image data captured by the camera, monitors the illuminated portion of the driver to determine the driver's attentiveness.
The light emitted by the illumination source may be in a near infrared (near IR) spectral region that is invisible to the human eye, and that is outside the visible spectral region present in insolation (solar radiation at a given area) and that is at a wavelength within the sensitivity of the camera. The camera may comprise a CMOS imaging array sensor (comprising multiple rows and columns of photosensing elements) that is sensitive within the near infrared spectral region.
Because the intensity of light emitted by the illumination source that is incident at the imaging sensor of the camera is modulated at the frequency of light emitted modulation of the light emitter, electronic analysis/filtering of the signal output of the imaging sensor of the camera can be used to distinguish the information of interest (i.e., the episodically illuminated driver's head region) from ambient lighting/sunlight directly impinging on the imaging sensor (that is not modulated as is the light emitted by the illumination source). The present invention is particularly advantageous in distinguishing the information of interest from ambient lighting when sunlight is streaming through side windows of the vehicle or through a rear window of the vehicle so as to be directly incident on the imaging array that views into the vehicle cabin towards the driver's head region.
These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.
Referring now to the drawings and the illustrative embodiments depicted therein, a vehicle 10 includes an imaging system or vision system 12 that includes at least one interior or driver monitoring camera 14 and a light source or illumination source 16 (
The camera has an imaging array sensor (comprising multiple rows and columns of photosensing elements) that is sensitive to at least the spectral region of the light emitted by the illumination source. The ECU may electronically or digitally filter the output signal of the imaging array sensor of the camera to distinguish the component of that output arising from the light-intensity modulated illuminated driver's head (as illuminated by the illumination source when modulated) from the component of the output signal of the imaging array sensor due to illumination of the driver's head by ambient lighting (typically sunlight streaming into the interior cabin such as though the windows of the quipped vehicle, and especially through side windows and/or the rear window). By syncing the camera frame capture rate with the intensity modulation rate of the illumination source, the system can compare illuminated frames of image data with non-illuminated frames of image data to provide enhanced processing and to avoid image washout issues in bright sunlight conditions.
For example, the camera may capture frames of image data at 60 frames per second and the illumination source may be modulated so that it is at its high intensity 30 times per second (and thus at its low intensity 30 times per second), whereby every other frame of image data captured by the camera is of the illuminated driver's head region. The system compares the frames of image data captured when the illumination source is activated to the frames of image data captured when the illumination source is not activated to determine the component of the captured image data that is due to the illumination source and the component of the captured image data that is due to ambient light, and the system may then filter the output signal or adjust the camera settings to reduce or remove the component of the captured image data that is due to ambient light, so as to avoid washout during bright lighting conditions.
The driver monitoring camera 14 captures image data representative of the driver's head. The interior cabin-viewing camera may be disposed at any suitable location in interior cabin of the vehicle, such as at a dashboard or instrument panel of the vehicle or within the movable head of an interior rearview mirror or part of a windshield electronics module (that includes a forward-viewing camera for advanced driver assistance functions such as described in U.S. Pat. Nos. 9,896,039; 9,871,971; 9,596,387; 8,256,821; 7,480,149; 6,824,281 and/or 6,690,268, which are all hereby incorporated herein by reference in their entireties) with the field of view of the cabin-viewing camera encompassing the driver head region in the vehicle (either directly or via a reflector). The light source 16 is operable to illuminate the driver front head region in a manner that enhances capturing and processing of image data in high ambient lighting conditions, as discussed below.
Driver monitoring systems may be used to track the head and eyes of the driver of a vehicle. To accurately track the eyes of the driver, the pupils of the eyes must be clearly viewed by a camera or other imaging device. Because of this, a preferred camera position is directly in front of the driver and at or below a line-of-sight to the road. This allows the camera to view both pupils in most situations and also avoid having the driver's eyelashes in the way (which is a problem that occurs when cameras are mounted above the line-of-sight to the road). In high ambient lighting conditions, image washout is a concern and the system may not be as capable of detecting the slight changes and differences in the driver's gaze or attentiveness.
The driver monitoring system of the present invention uses an additional light source 16 that modulates (e.g., modulates the intensity of or encodes a modulation signal on) the light that is emitted by the light source to illuminate the driver's face. The camera senses that light and the image processor processes image data captured by the camera when sensing that light and when not sensing that light and/or to distinguish light having a characteristic light modulation signature to provide enhanced sensing capabilities to monitor the driver's head and eyes. The system then can subtract out or filter out the images to remove the high ambient light component of the captured image data so that the image data is not washed out on a sunny day or other high ambient lighting conditions.
The light source 16 may emit visible and/or non-visible (e.g., infrared or near-infrared) light, and may comprise a light emitting diode (LED) or other light source disposed in the vehicle that emits the light toward the driver's head region, whereby the camera images the illuminated driver's head. The LED and camera may operate together when the driver monitoring system is operating. The LED may be disposed at the camera or the LED may be disposed elsewhere in the vehicle and emit light generally directly toward the driver's head region.
The light source may emit visible and/or non-visible light. During daytime lighting conditions, it does not matter if the driver's head is illuminated with extra visible light, since the driver will not typically notice that due to the sunlight present. During nighttime lighting conditions, the potential image washout due to the sun is less of a concern.
Sunlight comprises substantial amount of visible, UV and IR/near IR radiation. On average, most of the solar energy is between 300 nm and 2,100 nm, while Ultraviolet (UV) radiation constitutes around 3 percent of solar radiation (up to 400 nm), visible light or radiation constitutes around 48 percent (between 400 nm and 700 nm), and infrared (IR) light or radiation constitutes around 49 percent. NIR LEDs with wavelengths 770 nm or 870 nm or 880 nm or 940 nm or 950 nm are available from the likes of Excelitas Technologies with a place of business in Waltham, Massachusetts, and pulsed semiconductor lasers in the near IR (NIR) are also available from the likes of Excelitas Technologies, which offers a wide range of pulsed 905 nm lasers including monolithic layered structures with up to 4 active areas per chip, resulting in up to 100 Watts of peak output power.
The system of the present invention modulates the light emitted by the light source, such as by modulating the intensity of the light emitted by the light source. For example, the light source may emit a particular spectral band of light (such as IR or near IR light) or may emit visible light covering a wide spectral band, and the light source may be pulsed or modulated to turn on/off the emitted light and/or increase/decrease the intensity of the emitted light. The light source may comprise a white light-emitting LED or a red light-emitting LED or the like, and may be modulated during daytime lighting conditions, whereby the driver's head is illuminated with the additional light. In bright sunlight, the white light-emitted LED may be dominated by the sunlight, so the system may reduce the light sensitivity of the camera to a level where the image washout is reduced. For example, when the ambient light level is high (such as greater than, for example, 10,000 lux), then the system may reduce the sensitivity of the camera and/or may increase the intensity of the peak illumination levels of the white light-emitting LED.
Optionally, the light emitted by the illumination source may be in a near infrared (near IR) spectral region that is invisible to the human eye, and that is outside the visible spectral region present in insolation (solar radiation at a given area) and that is at a wavelength within the spectral sensitivity of the camera. Because the intensity of light emitted by the illumination source that is incident at the imaging sensor of the camera is modulated, electronic analysis/processing/filtering of the signal output of the imaging sensor of the camera can be used to distinguish the information of interest (i.e., the light-modulated illuminated driver's head region) from ambient lighting/sunlight that is or may be directly impinging on the imaging sensor (and that is not modulated by the control or system). Thus, the system can process and compare the frames of image data (illuminated and non-illuminated) to determine differences and to filter out the direct sunlight effects to provide enhanced sensing and monitoring of the driver's head and eyes. The present invention is particularly advantageous in distinguishing the information of interest from ambient lighting when sunlight is streaming through side windows of the vehicle or through a rear window of the vehicle so as to be directly incident on the imaging array that views into the vehicle cabin towards the driver's head region.
Thus, the system uses a light emitting illumination source (that modulates the light emitted thereby) and a camera to monitor, for example, the head and/or eyes of a driver of a vehicle. Using the selected modulation of the light source, the system processes image data captured by the camera to detect and monitor the driver's head illuminated by the modulated light. The system may, for example, determine the general viewing direction of the driver, the focal distance of the driver's gaze, or alertness of the driver.
The driver monitoring system may utilize aspects of head and face direction and position tracking systems and/or eye tracking systems and/or gesture recognition systems. Such head and face direction and/or position tracking systems and/or eye tracking systems and/or gesture recognition systems may utilize aspects of the systems described in U.S. Pat. No. 7,311,406 and/or 6,523,964, and/or U.S. Publication Nos. US-2017-0274906; US-2016-0137126; US-2015-0352953; US-2015-0296135; US-2015-0294169; US-2015-0232030; US-2015-0022664; US-2015-0015710; US-2015-0009010 and/or US-2014-0336876, and/or U.S. patent application Ser. No. 16/718,860, filed Dec. 18, 2019 and published Jun. 25, 2020 as U.S. Publication No. US-2020-0202151, and/or Ser. No. 16/668,035, filed Oct. 30, 2019 and published May 7, 2020 as U.S. Publication No. US-2020-0143560, which are all hereby incorporated herein by reference in their entireties.
The vehicle or vision system may include at least one exterior viewing imaging sensor or camera, such as a rearward viewing imaging sensor or camera 20a (and the system may optionally include multiple exterior viewing imaging sensors or cameras, such as a forward viewing camera 20b at the front (or at the windshield) of the vehicle, and a sideward/rearward viewing camera 20c, 20d at respective sides of the vehicle), which captures images exterior of the vehicle, with the camera having a lens for focusing images at or onto an imaging array or imaging plane or imager of the camera (
The imaging device and control and image processor and any associated illumination source, if applicable, may comprise any suitable components, and may utilize aspects of the cameras (such as various imaging sensors or imaging array sensors or cameras or the like, such as a CMOS imaging array sensor, a CCD sensor or other sensors or the like) and vision systems described in U.S. Pat. Nos. 5,760,962; 5,715,093; 6,922,292; 6,757,109; 6,717,610; 6,590,719; 6,201,642; 5,796,094; 6,559,435; 6,831,261; 6,822,563; 6,946,978; 7,720,580; 8,542,451; 7,965,336; 7,480,149; 5,877,897; 6,498,620; 5,670,935; 5,796,094; 6,396,397; 6,806,452; 6,690,268; 7,005,974; 7,937,667; 7,123,168; 7,004,606; 6,946,978; 7,038,577; 6,353,392; 6,320,176; 6,313,454 and/or 6,824,281, and/or International Publication Nos. WO 2009/036176; WO 2009/046268; WO 2010/099416; WO 2011/028686 and/or WO 2013/016409, and/or U.S. Publication Nos. US-2010-0020170 and/or US-2009-0244361, which are all hereby incorporated herein by reference in their entireties.
Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.
The present application is a continuation of U.S. patent application Ser. No. 17/645,066, filed Dec. 20, 2021, now U.S. Pat. No. 11,645,856, which is a continuation of U.S. patent application Ser. No. 16/836,968, filed Apr. 1, 2020, now U.S. Pat. No. 11,205,083, which claims the filing benefits of U.S. provisional application Ser. No. 62/828,128, filed Apr. 2, 2019, which is hereby incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5550677 | Schofield et al. | Aug 1996 | A |
5567360 | Varaprasad et al. | Oct 1996 | A |
5570127 | Schmidt | Oct 1996 | A |
5670935 | Schofield et al. | Sep 1997 | A |
5715093 | Schierbeek et al. | Feb 1998 | A |
5760962 | Schofield et al. | Jun 1998 | A |
5796094 | Schofield et al. | Aug 1998 | A |
5877897 | Schofield et al. | Mar 1999 | A |
6097023 | Schofield et al. | Aug 2000 | A |
6158655 | DeVries, Jr. et al. | Dec 2000 | A |
6201642 | Bos | Mar 2001 | B1 |
6313454 | Bos et al. | Nov 2001 | B1 |
6320176 | Schofield et al. | Nov 2001 | B1 |
6353392 | Schofield et al. | Mar 2002 | B1 |
6396397 | Bos et al. | May 2002 | B1 |
6483438 | DeLine et al. | Nov 2002 | B2 |
6498620 | Schofield et al. | Dec 2002 | B2 |
6523964 | Schofield et al. | Feb 2003 | B2 |
6559435 | Schofield et al. | May 2003 | B2 |
6590719 | Bos | Jul 2003 | B2 |
6593565 | Heslin et al. | Jul 2003 | B2 |
6627918 | Getz et al. | Sep 2003 | B2 |
6690268 | Schofield et al. | Feb 2004 | B2 |
6703925 | Steffel | Mar 2004 | B2 |
6717610 | Bos et al. | Apr 2004 | B1 |
6757109 | Bos | Jun 2004 | B2 |
6806452 | Bos et al. | Oct 2004 | B2 |
6822563 | Bos et al. | Nov 2004 | B2 |
6824281 | Schofield et al. | Nov 2004 | B2 |
6831261 | Schofield et al. | Dec 2004 | B2 |
6922292 | Bos | Jul 2005 | B2 |
6946978 | Schofield | Sep 2005 | B2 |
7004606 | Schofield | Feb 2006 | B2 |
7005974 | McMahon et al. | Feb 2006 | B2 |
7038577 | Pawlicki et al. | May 2006 | B2 |
7123168 | Schofield | Oct 2006 | B2 |
7184190 | McCabe et al. | Feb 2007 | B2 |
7195381 | Lynam et al. | Mar 2007 | B2 |
7224324 | Quist et al. | May 2007 | B2 |
7249860 | Kulas et al. | Jul 2007 | B2 |
7253723 | Lindahl et al. | Aug 2007 | B2 |
7255451 | McCabe et al. | Aug 2007 | B2 |
7274501 | McCabe et al. | Sep 2007 | B2 |
7289037 | Uken et al. | Oct 2007 | B2 |
7311406 | Schofield et al. | Dec 2007 | B2 |
7338177 | Lynam | Mar 2008 | B2 |
7360932 | Uken et al. | Apr 2008 | B2 |
7420756 | Lynam | Sep 2008 | B2 |
7480149 | DeWard et al. | Jan 2009 | B2 |
7626749 | Baur et al. | Dec 2009 | B2 |
7720580 | Higgins-Luthman | May 2010 | B2 |
7855755 | Weller et al. | Dec 2010 | B2 |
7914187 | Higgins-Luthman et al. | Mar 2011 | B2 |
7937667 | Kramer et al. | May 2011 | B2 |
7965336 | Bingle et al. | Jun 2011 | B2 |
8049640 | Uken et al. | Nov 2011 | B2 |
8256821 | Lawlor et al. | Sep 2012 | B2 |
8258932 | Wahlstrom | Sep 2012 | B2 |
8446470 | Lu et al. | May 2013 | B2 |
8451107 | Lu et al. | May 2013 | B2 |
8508831 | De Wind et al. | Aug 2013 | B2 |
8529108 | Uken et al. | Sep 2013 | B2 |
8542451 | Lu et al. | Sep 2013 | B2 |
8725311 | Breed | May 2014 | B1 |
8730553 | De Wind et al. | May 2014 | B2 |
8743203 | Karner et al. | Jun 2014 | B2 |
8876342 | Wimbert et al. | Nov 2014 | B2 |
9041806 | Baur et al. | May 2015 | B2 |
9090213 | Lawlor et al. | Jul 2015 | B2 |
9126525 | Lynam et al. | Sep 2015 | B2 |
9174578 | Uken et al. | Nov 2015 | B2 |
9346403 | Uken et al. | May 2016 | B2 |
9405120 | Graf et al. | Aug 2016 | B2 |
9487159 | Achenbach | Nov 2016 | B2 |
9493122 | Krebs | Nov 2016 | B2 |
9596387 | Achenbach et al. | Mar 2017 | B2 |
9598016 | Blank et al. | Mar 2017 | B2 |
9609757 | Steigerwald | Mar 2017 | B2 |
9827913 | De Wind et al. | Nov 2017 | B2 |
9871971 | Wang et al. | Jan 2018 | B2 |
9878669 | Kendall | Jan 2018 | B2 |
9896039 | Achenbach et al. | Feb 2018 | B2 |
9900490 | Ihlenburg et al. | Feb 2018 | B2 |
10017114 | Bongwald | Jul 2018 | B2 |
10029614 | Larson | Jul 2018 | B2 |
10046706 | Larson et al. | Aug 2018 | B2 |
10065574 | Tiryaki | Sep 2018 | B2 |
10166924 | Baur | Jan 2019 | B2 |
10166926 | Krebs et al. | Jan 2019 | B2 |
10261648 | Uken et al. | Apr 2019 | B2 |
10315573 | Bongwald | Jun 2019 | B2 |
10421404 | Larson et al. | Sep 2019 | B2 |
10442360 | LaCross et al. | Oct 2019 | B2 |
10466563 | Kendall et al. | Nov 2019 | B2 |
10567633 | Ihlenburg et al. | Feb 2020 | B2 |
10567705 | Ziegenspeck et al. | Feb 2020 | B2 |
10703204 | Hassan et al. | Jul 2020 | B2 |
10793067 | Ihlenburg et al. | Oct 2020 | B2 |
10922563 | Nix et al. | Feb 2021 | B2 |
10958830 | Koravadi | Mar 2021 | B2 |
11167771 | Caron et al. | Nov 2021 | B2 |
11205083 | Lynam | Dec 2021 | B2 |
11214199 | LaCross et al. | Jan 2022 | B2 |
11240427 | Koravadi | Feb 2022 | B2 |
11242008 | Blank et al. | Feb 2022 | B2 |
11252376 | Ihlenburg | Feb 2022 | B2 |
11341671 | Lu et al. | May 2022 | B2 |
11348374 | Kramer et al. | May 2022 | B2 |
11433906 | Lu | Sep 2022 | B2 |
11465561 | Peterson et al. | Oct 2022 | B2 |
11488399 | Wacquant | Nov 2022 | B2 |
11493918 | Singh | Nov 2022 | B2 |
11518401 | Kulkarni | Dec 2022 | B2 |
11582425 | Liu | Feb 2023 | B2 |
11645856 | Lynam | May 2023 | B2 |
20020003571 | Schofield | Jan 2002 | A1 |
20020005999 | Hutzel et al. | Jan 2002 | A1 |
20030209893 | Breed | Nov 2003 | A1 |
20040129478 | Breed | Jul 2004 | A1 |
20050259309 | Blank | Nov 2005 | A1 |
20060050018 | Tutzel et al. | Mar 2006 | A1 |
20070182528 | Breed et al. | Aug 2007 | A1 |
20090040778 | Takayanagi et al. | Feb 2009 | A1 |
20090244361 | Gebauer et al. | Oct 2009 | A1 |
20100020170 | Higgins-Luthman et al. | Jan 2010 | A1 |
20100085653 | Uken et al. | Apr 2010 | A1 |
20110080481 | Bellingham | Apr 2011 | A1 |
20110122049 | Lvovskiy | May 2011 | A1 |
20120293658 | Higgins-Luthman | Nov 2012 | A1 |
20130016409 | Kurtz et al. | Jan 2013 | A1 |
20130044138 | Koga | Feb 2013 | A1 |
20140022390 | Blank et al. | Jan 2014 | A1 |
20140118131 | Chiu | May 2014 | A1 |
20140285666 | O'Connell et al. | Sep 2014 | A1 |
20140293169 | Uken et al. | Oct 2014 | A1 |
20140313563 | Uken et al. | Oct 2014 | A1 |
20140336876 | Gieseke et al. | Nov 2014 | A1 |
20150009010 | Biemer | Jan 2015 | A1 |
20150015710 | Tiryaki | Jan 2015 | A1 |
20150022664 | Pflug et al. | Jan 2015 | A1 |
20150092042 | Fursich | Apr 2015 | A1 |
20150097955 | De Wind et al. | Apr 2015 | A1 |
20150232030 | Bongwald | Aug 2015 | A1 |
20150294169 | Zhou et al. | Oct 2015 | A1 |
20150296135 | Wacquant et al. | Oct 2015 | A1 |
20150352953 | Koravadi | Dec 2015 | A1 |
20160137126 | Fursich et al. | May 2016 | A1 |
20160209647 | Fursich | Jul 2016 | A1 |
20170217367 | Pflug et al. | Aug 2017 | A1 |
20170237946 | Schofield et al. | Aug 2017 | A1 |
20170274906 | Hassan et al. | Sep 2017 | A1 |
20170355312 | Habibi et al. | Dec 2017 | A1 |
20170372143 | Barcus | Dec 2017 | A1 |
20180134217 | Peterson et al. | May 2018 | A1 |
20180222414 | Ihlenburg et al. | Aug 2018 | A1 |
20180231976 | Singh | Aug 2018 | A1 |
20180324367 | Siddiqui | Nov 2018 | A1 |
20190001930 | Dellock | Jan 2019 | A1 |
20190052792 | Baba | Feb 2019 | A1 |
20190054899 | Hoyos et al. | Feb 2019 | A1 |
20190118717 | Blank et al. | Apr 2019 | A1 |
20190146297 | Lynam et al. | May 2019 | A1 |
20190156134 | Krishnan | May 2019 | A1 |
20190168669 | Lintz et al. | Jun 2019 | A1 |
20190258131 | Lynam et al. | Aug 2019 | A9 |
20190275945 | Chang | Sep 2019 | A1 |
20190364199 | Koravadi | Nov 2019 | A1 |
20190377933 | Hara | Dec 2019 | A1 |
20190381938 | Hopkins | Dec 2019 | A1 |
20200062168 | Yuchi | Feb 2020 | A1 |
20200104571 | Osuga | Apr 2020 | A1 |
20200143560 | Lu et al. | May 2020 | A1 |
20200148098 | Salter | May 2020 | A1 |
20200148120 | Englander et al. | May 2020 | A1 |
20200202151 | Wacquant | Jun 2020 | A1 |
20200247315 | Foutis | Aug 2020 | A1 |
20200320320 | Lynam | Oct 2020 | A1 |
20200327323 | Noble | Oct 2020 | A1 |
20200377022 | LaCross et al. | Dec 2020 | A1 |
20210056306 | Hu et al. | Feb 2021 | A1 |
20210122404 | Lisseman et al. | Apr 2021 | A1 |
20210155167 | Lynam et al. | May 2021 | A1 |
20210162926 | Lu | Jun 2021 | A1 |
20210199954 | Kawasaki | Jul 2021 | A1 |
20210245662 | Blank et al. | Aug 2021 | A1 |
20210291739 | Kasarla et al. | Sep 2021 | A1 |
20210306538 | Solar | Sep 2021 | A1 |
20210323473 | Peterson et al. | Oct 2021 | A1 |
20210368082 | Solar | Nov 2021 | A1 |
20210383104 | Kramer et al. | Dec 2021 | A1 |
20220111857 | Kulkarni | Apr 2022 | A1 |
20220242438 | Sobecki et al. | Aug 2022 | A1 |
20220254132 | Rother | Aug 2022 | A1 |
20220377219 | Conger et al. | Nov 2022 | A1 |
20230131471 | Sobecki et al. | Apr 2023 | A1 |
20230137004 | Huizen et al. | May 2023 | A1 |
Number | Date | Country |
---|---|---|
205523966 | Aug 2016 | CN |
209534893 | Oct 2019 | CN |
10325646 | Jan 2004 | DE |
102005000650 | Jul 2006 | DE |
3071788 | Apr 2019 | FR |
2022150826 | Jul 2022 | WO |
2023034956 | Mar 2023 | WO |
Entry |
---|
Seeing Machines' FOVIO Driver Monitoring (FDM) processor [description of which is available at https://china.xilinx.com/publications/presentations/c_D3_03-Driver-Monitoring-Systems.pdf], Jan. 14, 2021. |
Number | Date | Country | |
---|---|---|---|
20230274558 A1 | Aug 2023 | US |
Number | Date | Country | |
---|---|---|---|
62828128 | Apr 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17645066 | Dec 2021 | US |
Child | 18313510 | US | |
Parent | 16836968 | Apr 2020 | US |
Child | 17645066 | US |