The present invention relates to vehicular vision systems that use one or more cameras of a vehicle to monitor a driver of the vehicle and/or display video images derived from image data captured by the cameras of the vehicle.
It is known to provide a video display at the exterior rearview mirror assembly, such as described in U.S. Pat. No. 7,777,611, which is hereby incorporated herein by reference in its entirety, or to provide a video display at an interior rearview mirror assembly to display sideward and/or rearward images, such as described in U.S. Pat. No. 5,670,935, which is hereby incorporated herein by reference in its entirety. A variety of interior and exterior mirror assemblies with indicators are known in the art, such as U.S. Pat. Nos. 5,668,663; 5,355,284; 5,788,357; 6,257,746; 6,005,724; 5,481,409; 6,111,683; 6,045,243; 6,264,353; 6,512,624; 6,356,376; 2,263,382; 2,580,014; 3,266,016; 4,499,451; 4,588,267; 4,630,904; 4,623,222; 4,721,364; 4,906,085; 5,313,335; 5,587,699; 5,575,552; 5,436,741; 5,587,699; 5,938,320; 6,700,692 and 5,786,772, which are all hereby incorporated herein by reference in their entireties.
A driver monitoring system (DMS) includes a camera that views a driver's head and body region in the vehicle cabin and an image processor that processes image data captured by the camera to determine posture of the driver sitting in the driver seat of the vehicle. The system may also or otherwise include a camera that views a passenger's head and body region at a passenger seat in the vehicle cabin, whereby the image processor processes image data captured by the passenger-viewing camera to determine posture of the passenger sitting in the passenger seat of the vehicle. The system, responsive to determination (via processing by the image processor of image data captured by the camera) that the driver's posture is poor (outside of a threshold range of acceptable postures or positions), generates an alert to the driver (or passenger) so the driver (or passenger) can improve his or her posture.
The system includes an electronic control unit (ECU) of the vehicle that receives image data captured by the camera and that may include the image processor for processing the captured image data. The ECU may output (such as via a coaxial cable) a signal to a video display of the interior rearview mirror assembly for displaying video images at the display screen based on the captured image data or for displaying a visible alert at the mirror reflective element of the interior rearview mirror assembly.
These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.
A vehicle vision system and/or driver or driving assist system and/or driver monitoring system and/or alert system operates to capture images interior of the vehicle and may process the captured image data to monitor the driver of the vehicle (and/or one or more passengers of the vehicle). The system includes an image processor or image processing system that is operable to receive image data from one or more cameras and to provide an output to one or more alert devices or display devices for alerting the driver and/or passenger.
Referring now to the drawings and the illustrative embodiments depicted therein, a driver monitoring and posture alert system 10 includes a driver monitoring camera 12 that is disposed at an interior rearview mirror assembly 14 of the vehicle. The system 10 also includes a posture indicator 16 (such as one or more visible light-emitting light emitting diodes (LEDs)) and may also include one or more illumination devices 18 (such as one or more infrared LEDs) for illuminating the driver region to enhance image capture of the driver during low ambient lighting conditions.
The system 10 uses one or more driver monitoring system (DMS) cameras 12 (such as one that views the driver seat region of the vehicle and one that views the passenger seat region of the vehicle or a single camera that views both seat regions) to analyze and monitor the driver's and/or passenger's posture. The system 10 may allow for users to set their ideal posture position (such as via pressing a button or otherwise actuating a user input or human machine interface (HMI) to save the user's current posture as a baseline or target posture) or the system 10 may automatically define the ideal posture by analyzing the person's head and spine positions relative to the key features of the vehicle (e.g., relative to the vehicle seat, steering wheel, mirror, etc.).
Referring to
Thus, the driver, when initially operating the vehicle, may actuate a “proper posture” input and the system may capture image data representative of the driver in the proper posture. The proper posture image data is stored in memory for use during that driving maneuver or trip by that driver and optionally for future driving maneuvers or trips by that driver. The system may also include a driver identification feature to determine if the driver in the vehicle is a driver that has previously stored proper posture data. While the driver operates the vehicle, the system may continuously or episodically monitor the driver and compare the determined current driver posture to the stored proper posture data. If the system determines that the driver's determined current posture differs from the stored proper posture by a threshold amount (e.g., the driver's head tilts or the driver slouches by more than a threshold amount compared to the stored posture data), the system generates an alert to the driver so the driver can adjust his or her posture back towards the proper posture/position.
The alert may comprise an audible sound (via one or more speakers in the vehicle), a verbal warning (via one or more speakers in the vehicle), a seat vibration (or steering wheel vibration or other suitable haptic alert), a light indicator 16, a light indicator 16 that changes color/brightness depending on how bad/good the posture is (e.g., a light that is green when the posture is good, and red when the posture is bad). Optionally, the light indicator 16 may comprise a single LED that simply lights up when the system 10 determines bad posture, or the light indicator 16 may comprise two LEDs (or a single LED that can emit different colors of light) such that red light is emitted when posture is determined to be poor and green light is emitted when posture is determined to be good. Optionally, the light indicator 16 may comprise a plurality of LEDs arranged in a row or column, where the number of LEDs (and optionally color too) energized changes with the determined degree of bad posture.
Although shown and described as having the camera 12 and alert device at the interior rearview mirror assembly 14, the monitoring and posture alert system 10 may be located at the instrument panel or at a center console or overhead console of the interior cabin of the vehicle.
As shown in
Optionally, and such as shown in
The system 110 provides a posture indicator 116 at the video display screen, such as via a graphic overlay or iconistic display at the display screen. In the illustrated embodiment, the indicator 116 provides an iconistic representation of a driver with bad posture (and may be displayed as a red color) to alert the driver of his or her bad posture. The color and/or intensity of the iconistic indicator 116 may be adjusted responsive to the degree of bad posture and/or time that the driver (or passenger) maintains the bad posture. Optionally, the iconistic indicator 116 may flash and/or be accompanied by an audible alert based at least in part on the degree of bad posture and/or time that the driver (or passenger) maintains the bad posture. Similarly, the system 110 may provide a passenger side posture indicator 117 (which may be at the passenger side of the mirror) to alert a passenger of his or her bad posture. The icon and/or warning message may be displayed at or on the video mirror display screen or at or on any other of the vehicle's display screens (such as a center console display screen or the like).
Optionally, the system 110 may record the amount of time the driver has good posture versus bad posture and report the information to the vehicle as a matrix that the driver/passage can improve on over time. Thus, as the driver/passenger uses the system 110 over time, he or she can monitor progress and set goals for themselves.
Optionally, the system 110 may process captured image data to determine the driver's skeletal structure mapped from the driver's perceived body position. The system 110 may continually monitor the determined skeletal structure and compare it to a known best posture or a pre-set posture setting. Optionally, the system 110 may monitor the determined skeletal structure to determine a threshold change in posture over time (to determine a time-based posture change), and may alert the driver or passenger when the change exceeds a threshold amount over a given time period.
The interior rearview mirror assembly may comprise any suitable mirror assembly having a video display device disposed behind the mirror reflective element so as to be viewable through a partially reflective and partially visible light transmitting or transflective mirror reflector of the mirror reflective element when the display screen is powered. For example, and such as shown in
The mirror head (including the mirror housing and reflective element and display) is mounted at a stay assembly or mounting structure, which is configured to attach at an interior portion of the vehicle (such as at an overhead console or headliner or at an in-cabin side of the vehicle windshield). The mirror head may be toggled between the mirror mode and the display mode by flipping a toggle tab at the bottom of the mirror head, which causes the mirror head to pivot about a generally horizontal pivot axis relative to the mirror stay.
The system may utilize aspects of the driver monitoring systems described in U.S. Pat. Nos. 10,065,574; 10,017,114; 9,405,120 and/or 7,914,187, and/or U.S. Publication Nos. US-2021-0323473; US-2021-0291739; US-2020-0202151; US-2020-0143560; US-2020-0320320; US-2018-0231976; US-2018-0222414; US-2017-0274906; US-2017-0217367; US-2016-0209647; US-2016-0137126; US-2015-0352953; US-2015-0296135; US-2015-0294169; US-2015-0232030; US-2015-0092042; US-2015-0022664; US-2015-0015710; US-2015-0009010 and/or US-2014-0336876, and/or U.S. patent application Ser. No. 17/450,721, filed Oct. 13, 2021, and/or U.S. provisional applications, Ser. No. 63/260,359, filed Aug. 18, 2021, Ser. No. 63/201,894, filed May 18, 2021, Ser. No. 63/201,757, filed May 12, 2021, Ser. No. 63/201,371, filed Apr. 27, 2021, Ser. No. 63/200,451, filed Mar. 8, 2021, Ser. No. 63/200,315, filed Mar. 1, 2021, Ser. No. 63/200,003, filed Feb. 9, 2021, which are hereby incorporated herein by reference in their entireties.
The cameras may comprise any suitable imaging sensor or camera, such as a pixelated imaging array or the like, such as a CMOS imaging array sensor, a CCD sensor or other sensors or the like, such as a camera or sensor of the types disclosed in commonly assigned, U.S. Pat. Nos. 7,965,336; 5,550,677; 5,760,962; 6,097,023 and 5,796,094, which are hereby incorporated herein by reference in their entireties. Optionally, the cameras may comprise a stereo imaging camera or the like, such as by utilizing aspects of the imaging systems described in U.S. Pat. Nos. 6,396,397 and/or 5,796,094, which are hereby incorporated herein by reference in their entireties. Optionally, the cameras may comprise an infrared or near infrared light sensitive camera and may be suitable for capturing images in low lighting conditions, and/or the camera may include or be associated with an illumination source (such as an infrared or near-infrared light emitting illumination source that, when actuated to emit infrared or near-infrared light at the side of the vehicle, enhances the camera's performance but is not visible or discernible to the driver of the vehicle), such as by utilizing aspects of the cameras described in U.S. Pat. Nos. 7,965,336; 5,550,677; 5,760,962; 6,097,023 and 5,796,094, which are hereby incorporated herein by reference in their entireties.
The mirror assembly may comprise any suitable construction, such as, for example, a mirror assembly with the reflective element being nested in the mirror casing and with a bezel portion that circumscribes a perimeter region of the front surface of the reflective element or with the mirror casing having a curved or beveled perimeter edge around the reflective element and with no overlap onto the front surface of the reflective element (such as by utilizing aspects of the mirror assemblies described in U.S. Pat. Nos. 7,184,190; 7,274,501; 7,255,451; 7,289,037; 7,360,932; 7,626,749; 8,049,640; 8,277,059 and/or 8,529,108, which are hereby incorporated herein by reference in their entireties), or such as a mirror assembly having a rear substrate of an electro-optic or electrochromic reflective element nested in the mirror casing, and with the front substrate having curved or beveled perimeter edges, or such as a mirror assembly having a prismatic reflective element that is disposed at an outer perimeter edge of the mirror casing and with the prismatic substrate having curved or beveled perimeter edges, such as described in U.S. Pat. Nos. 8,508,831; 8,730,553; 9,598,016 and/or 9,346,403, and/or U.S. Publication Nos. US-2014-0313563 and/or US-2015-0097955, which are hereby incorporated herein by reference in their entireties (and with electrochromic and prismatic mirrors of such construction are commercially available from the assignee of this application under the trade name INFINITY™ mirror).
The mirror assembly may include user inputs or actuatable switches or touch sensors or the like for user/driver control of one or more features of the mirror assembly and/or display system. The user inputs or touch sensors may comprise any suitable sensors or inputs, and may utilize aspects of the inputs and sensors described in U.S. Pat. Nos. 9,827,913; 9,598,016; 9,346,403; 8,508,831; 8,730,553; 7,224,324; 7,253,723; 7,255,451 and/or 8,154,418, which are hereby incorporated herein by reference in their entireties.
Changes and modifications to the specifically described embodiments may be carried out without departing from the principles of the present invention, which is intended to be limited only by the scope of the appended claims as interpreted according to the principles of patent law.
The present application claims the filing benefits of U.S. provisional application Ser. No. 63/199,918, filed Feb. 3, 2021, which is hereby incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
2263382 | Gotzinger | Nov 1941 | A |
2580014 | Gazda | Dec 1951 | A |
3266016 | Maru | Aug 1966 | A |
4499451 | Suzuki et al. | Feb 1985 | A |
4588267 | Pastore | May 1986 | A |
4623222 | Itoh et al. | Nov 1986 | A |
4630904 | Pastore | Dec 1986 | A |
4721364 | Itoh et al. | Jan 1988 | A |
4906085 | Sugihara et al. | Mar 1990 | A |
5313335 | Gray et al. | May 1994 | A |
5355284 | Roberts | Oct 1994 | A |
5436741 | Crandall | Jul 1995 | A |
5481409 | Roberts | Jan 1996 | A |
5550677 | Schofield et al. | Aug 1996 | A |
5567360 | Varaprasad et al. | Oct 1996 | A |
5570127 | Schmidt | Oct 1996 | A |
5575552 | Faloon et al. | Nov 1996 | A |
5587699 | Faloon et al. | Dec 1996 | A |
5668663 | Varaprasad et al. | Sep 1997 | A |
5670935 | Schofield et al. | Sep 1997 | A |
5760962 | Schofield et al. | Jun 1998 | A |
5786772 | Schofield et al. | Jul 1998 | A |
5788357 | Muth et al. | Aug 1998 | A |
5796094 | Schofield et al. | Aug 1998 | A |
5877897 | Schofield et al. | Mar 1999 | A |
5938320 | Crandall | Aug 1999 | A |
6005724 | Todd | Dec 1999 | A |
6045243 | Muth et al. | Apr 2000 | A |
6097023 | Schofield et al. | Aug 2000 | A |
6111683 | Cammenga et al. | Aug 2000 | A |
6158655 | DeVries, Jr. et al. | Dec 2000 | A |
6257746 | Todd et al. | Jul 2001 | B1 |
6264353 | Caraher et al. | Jul 2001 | B1 |
6356376 | Tonar et al. | Mar 2002 | B1 |
6483438 | DeLine et al. | Nov 2002 | B2 |
6512624 | Tonar et al. | Jan 2003 | B2 |
6593565 | Heslin et al. | Jul 2003 | B2 |
6627918 | Getz et al. | Sep 2003 | B2 |
6690268 | Schofield et al. | Feb 2004 | B2 |
6700692 | Tonar et al. | Mar 2004 | B2 |
6703925 | Steffel | Mar 2004 | B2 |
6824281 | Schofield et al. | Nov 2004 | B2 |
7038577 | Pawlicki et al. | May 2006 | B2 |
7184190 | McCabe et al. | Feb 2007 | B2 |
7195381 | Lynam et al. | Mar 2007 | B2 |
7224324 | Quist et al. | May 2007 | B2 |
7249860 | Kulas et al. | Jul 2007 | B2 |
7253723 | Lindahl et al. | Aug 2007 | B2 |
7255451 | McCabe et al. | Aug 2007 | B2 |
7274501 | McCabe et al. | Sep 2007 | B2 |
7289037 | Uken et al. | Oct 2007 | B2 |
7338177 | Lynam | Mar 2008 | B2 |
7360932 | Uken et al. | Apr 2008 | B2 |
7420756 | Lynam | Sep 2008 | B2 |
7480149 | DeWard et al. | Jan 2009 | B2 |
7626749 | Baur et al. | Dec 2009 | B2 |
7720580 | Higgins-Luthman | May 2010 | B2 |
7777611 | Desai | Aug 2010 | B2 |
7855755 | Weller et al. | Dec 2010 | B2 |
7914187 | Higgins-Luthman et al. | Mar 2011 | B2 |
8049640 | Uken et al. | Nov 2011 | B2 |
8258932 | Wahlstrom | Sep 2012 | B2 |
8446470 | Lu et al. | May 2013 | B2 |
8451107 | Lu et al. | May 2013 | B2 |
8508831 | De Wind et al. | Aug 2013 | B2 |
8529108 | Uken et al. | Sep 2013 | B2 |
8730553 | De Wind et al. | May 2014 | B2 |
8743203 | Karner et al. | Jun 2014 | B2 |
8876342 | Wimbert et al. | Nov 2014 | B2 |
9090213 | Lawlor et al. | Jul 2015 | B2 |
9126525 | Lynam | Sep 2015 | B2 |
9174578 | Uken et al. | Nov 2015 | B2 |
9346403 | Uken et al. | May 2016 | B2 |
9405120 | Graf et al. | Aug 2016 | B2 |
9487159 | Achenbach | Nov 2016 | B2 |
9493122 | Krebs | Nov 2016 | B2 |
9598016 | Blank et al. | Mar 2017 | B2 |
9609757 | Steigerwald | Mar 2017 | B2 |
9827913 | De Wind et al. | Nov 2017 | B2 |
9878669 | Kendall | Jan 2018 | B2 |
9900490 | Ihlenburg et al. | Feb 2018 | B2 |
10017114 | Bongwald | Jul 2018 | B2 |
10029614 | Larson | Jul 2018 | B2 |
10046706 | Larson et al. | Aug 2018 | B2 |
10065574 | Tiryaki | Sep 2018 | B2 |
10166924 | Baur | Jan 2019 | B2 |
10166926 | Krebs et al. | Jan 2019 | B2 |
10261648 | Uken et al. | Apr 2019 | B2 |
10315573 | Bongwald | Jun 2019 | B2 |
10421404 | Larson et al. | Sep 2019 | B2 |
10442360 | LaCross et al. | Oct 2019 | B2 |
10466563 | Kendall et al. | Nov 2019 | B2 |
10567633 | Ihlenburg et al. | Feb 2020 | B2 |
10567705 | Ziegenspeck et al. | Feb 2020 | B2 |
10703204 | Hassan et al. | Jul 2020 | B2 |
10922563 | Nix et al. | Feb 2021 | B2 |
11205083 | Lynam | Dec 2021 | B2 |
11214199 | LaCross et al. | Jan 2022 | B2 |
11240427 | Koravadi | Feb 2022 | B2 |
11242008 | Blank et al. | Feb 2022 | B2 |
11252376 | Ihlenburg | Feb 2022 | B2 |
11341671 | Lu et al. | May 2022 | B2 |
11348374 | Kramer et al. | May 2022 | B2 |
11433906 | Lu | Sep 2022 | B2 |
11465561 | Peterson et al. | Oct 2022 | B2 |
11488399 | Wacquant | Nov 2022 | B2 |
11493918 | Singh | Nov 2022 | B2 |
11518401 | Kulkarni | Dec 2022 | B2 |
11582425 | Liu | Feb 2023 | B2 |
20020005999 | Hutzel et al. | Jan 2002 | A1 |
20060050018 | Hutzel et al. | Mar 2006 | A1 |
20070182528 | Breed et al. | Aug 2007 | A1 |
20090040778 | Takayanagi et al. | Feb 2009 | A1 |
20100085653 | Uken et al. | Apr 2010 | A1 |
20110080481 | Bellingham | Apr 2011 | A1 |
20140022390 | Blank et al. | Jan 2014 | A1 |
20140285666 | O'Connell et al. | Sep 2014 | A1 |
20140293169 | Uken et al. | Oct 2014 | A1 |
20140313563 | Uken et al. | Oct 2014 | A1 |
20140336876 | Gieseke et al. | Nov 2014 | A1 |
20150009010 | Biemer | Jan 2015 | A1 |
20150015710 | Tiryaki | Jan 2015 | A1 |
20150022664 | Pflug et al. | Jan 2015 | A1 |
20150092042 | Fursich | Apr 2015 | A1 |
20150097955 | De Wind et al. | Apr 2015 | A1 |
20150232030 | Bongwald | Aug 2015 | A1 |
20150294169 | Zhou et al. | Oct 2015 | A1 |
20150296135 | Wacquant et al. | Oct 2015 | A1 |
20150352953 | Koravadi | Dec 2015 | A1 |
20160137126 | Fursich et al. | May 2016 | A1 |
20160209647 | Fursich | Jul 2016 | A1 |
20170158054 | Munaoka | Jun 2017 | A1 |
20170161575 | Banno | Jun 2017 | A1 |
20170197523 | Magaña | Jul 2017 | A1 |
20170210357 | Nagai | Jul 2017 | A1 |
20170217367 | Pflug et al. | Aug 2017 | A1 |
20170237946 | Schofield et al. | Aug 2017 | A1 |
20170274906 | Hassan et al. | Sep 2017 | A1 |
20170355312 | Habibi et al. | Dec 2017 | A1 |
20180134217 | Peterson et al. | May 2018 | A1 |
20180222414 | Ihlenburg et al. | Aug 2018 | A1 |
20180231976 | Singh | Aug 2018 | A1 |
20190047475 | Uken et al. | Feb 2019 | A1 |
20190054899 | Hoyos et al. | Feb 2019 | A1 |
20190118717 | Blank et al. | Apr 2019 | A1 |
20190146297 | Lynam et al. | May 2019 | A1 |
20190168669 | Lintz et al. | Jun 2019 | A1 |
20190258131 | Lynam et al. | Aug 2019 | A9 |
20190364199 | Koravadi | Nov 2019 | A1 |
20190381938 | Hopkins | Dec 2019 | A1 |
20200143560 | Lu et al. | May 2020 | A1 |
20200148120 | Englander et al. | May 2020 | A1 |
20200202151 | Wacquant | Jun 2020 | A1 |
20200320320 | Lynam | Oct 2020 | A1 |
20200327323 | Noble | Oct 2020 | A1 |
20200377022 | LaCross et al. | Dec 2020 | A1 |
20210052206 | Kale | Feb 2021 | A1 |
20210056306 | Hu et al. | Feb 2021 | A1 |
20210122404 | Lisseman et al. | Apr 2021 | A1 |
20210155167 | Lynam et al. | May 2021 | A1 |
20210162926 | Lu | Jun 2021 | A1 |
20210197747 | Yoshizawa | Jul 2021 | A1 |
20210245662 | Blank et al. | Aug 2021 | A1 |
20210291739 | Kasarla et al. | Sep 2021 | A1 |
20210306538 | Solar | Sep 2021 | A1 |
20210323473 | Peterson et al. | Oct 2021 | A1 |
20210368082 | Solar | Nov 2021 | A1 |
20220084316 | Wang | Mar 2022 | A1 |
20220111857 | Kulkarni | Apr 2022 | A1 |
20220242438 | Sobecki et al. | Aug 2022 | A1 |
20220254132 | Rother | Aug 2022 | A1 |
20220377219 | Conger et al. | Nov 2022 | A1 |
Number | Date | Country |
---|---|---|
113384266 | Sep 2021 | CN |
2022187805 | Sep 2022 | WO |
2022241423 | Nov 2022 | WO |
2023034956 | Mar 2023 | WO |
Number | Date | Country | |
---|---|---|---|
20220242438 A1 | Aug 2022 | US |
Number | Date | Country | |
---|---|---|---|
63199918 | Feb 2021 | US |