The present application generally relates to a vehicle warning system, and, more particularly, to a system and method for enhancing a driver's awareness of pedestrians and other objects by showing projected travel paths of the vehicle and pedestrian and/or moving object when the vehicle is making a turn.
Motorized vehicles may be equipped with various kinds of warning systems. These warning systems may be general warning systems that inform the driver of different operating conditions of the vehicle. For example, vehicles may be equipped with warning systems which may be used to warn the driver of low fuel amounts, high and or low engine temperature, a drop in oil pressure, problems in charging the vehicle battery, doors and or trunks that are open and the like.
Motorized vehicles may further be equipped with more advanced warning systems which may be used to warn the driver of the vehicle about potentially dangerous situations involving other traffic participants. For example, warning systems have been designed for vehicles that may be used to provide a driver with back-up collision warning, blind spot detection, lane-departure warnings, as well as driver-alertness monitoring, assisted parking, traffic sign recognition, and the like.
While the above systems do provide the driver with warnings of potential dangers, these systems fail to provide any information about a projected path of the projected danger. For example, while a blind spot detection system is able to monitor if a vehicle or other object is located in a blind spot of a vehicle, the blind spot detection system does not provide any information as to whether the vehicle and or object detected is moving towards or away from the vehicle. Similarly, while back-up warning systems may alert a driver to potential objects located behind the vehicle while the vehicle is in reverse, these systems also fail to provide any information as to whether the object is moving towards or away from the vehicle.
Therefore, it would be desirable to provide a vehicle warning system and method that overcome the above problems.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the DESCRIPTION OF THE APPLICATION. This summary is not intended to identify key features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
A system for warning of potential hazards when a vehicle is turning has a sensor coupled to the vehicle and configured to capture data of objects located around the vehicle. A control unit is coupled to the sensor for processing the data captured by the sensor to generate graphical representations of objects captured by the sensor, graphical representations of projected paths of moving objects captured by the sensor; and a graphical representation of a projected turning path of the vehicle.
A system for warning of potential hazards when a vehicle is turning has image sensors coupled to the vehicle and configured to capture pedestrians located around the vehicle. Monitoring sensors are coupled to the vehicle to measure the speed and turning angle of the vehicle. A control unit is coupled to the image sensors and monitoring sensors for processing data from the image sensors and monitoring sensors and to generate graphical representations of pedestrians captured by the image sensors, graphical representations of virtual crosswalks associated with pedestrians moving near the vehicle; and graphical representations indicating a projected turning path of the vehicle. A display is used to show the graphical representations of pedestrians captured by the sensor, graphical representations of the virtual crosswalks associated with pedestrians moving near the vehicle; and graphical representations indicating the projected turning path of the vehicle.
A system for warning of potential hazards when a vehicle is turning has image sensors coupled to the vehicle and configured to capture pedestrians located around the vehicle. Monitoring sensors are coupled to the vehicle to measure a speed and a turning angle of the vehicle. A control unit is coupled to the image sensors and monitoring sensors for processing data from the image sensors and monitoring sensors, wherein the control unit generates graphical representations of pedestrians captured by the sensor, graphical representations of virtual crosswalks associated with pedestrians moving near the vehicle; and graphical representations indicating a projected turning path of the vehicle when the monitoring sensors indicates that the vehicle is turning, wherein the graphical representation indicating the projected turning path of the vehicle is divided into a plurality of sections, each section having an indicator, wherein each indicator illustrates a probability of impact with the moving pedestrian. A Heads-Up Display HUD system is coupled to the control unit to show the graphical representations of pedestrians captured by the sensor, graphical representations of virtual crosswalks associated with pedestrians moving near the vehicle; and graphical representations indicating the projected turning path of the vehicle.
Embodiments of the disclosure will become more fully understood from the detailed description and the accompanying drawings, wherein:
The description set forth below in connection with the appended drawings is intended as a description of presently preferred embodiments of the disclosure and is not intended to represent the only forms in which the present disclosure can be constructed and/or utilized. The description sets forth the functions and the sequence of steps for constructing and operating the disclosure in connection with the illustrated embodiments. It is to be understood, however, that the same or equivalent functions and sequences can be accomplished by different embodiments that are also intended to be encompassed within the spirit and scope of this disclosure.
Referring to
Referring now to
The warning system 12 may further have one or more monitoring sensors 18. The monitoring sensor 18 may be coupled to one or more operating systems 20 of the vehicle 10. The monitoring sensor 18 may be used to detect operating conditions of the vehicle 10. For example, the monitoring sensor 18 may be used to monitor a speed of the vehicle 10, whether the vehicle 10 is making a turn, or the like. In accordance with one embodiment, the warning system 12 may be configured to not process and display the location and dynamics of pedestrians detected unless the monitoring sensors 18 detects that the vehicle 10 is making a turn. A threshold value may be established so that common road vibrations and slight movement of the a steering mechanism of the vehicle 10 may not accidently trigger the warning system 10 to project the location and dynamics of the pedestrians detected. In one embodiment, the warning system 10 may be used when the user enables their left hand signal.
In accordance with another embodiment, a Global Positioning Satellite (GPS) unit 28 located in the vehicle 10 may be used to determine if the vehicle 10 is moving and or turning. Since the GPS unit 28 may be able to provide speed and direction of travel data on the vehicle 10, the GPS unit 28 may be used to determine if the vehicle 10 is moving and or turning. The warning system 12 may be configured to not process and display the location and dynamics of pedestrians detected unless the GPS unit 28 detects that the vehicle 10 is making a turn and or moving.
The sensors 16 and the monitoring sensors 18 may be coupled to a control unit 22. The control unit 22 may take and process the data captured by the sensors 16. The control unit 22 may process the data in order to detect and identify the different objects detected by the sensors 16. The control unit 22 may identify the position of the different objects as well as the whether the object is moving. If moving, the control unit 22 may be used to calculate the speed and direction of the moving object. The control unit 22 may then take the process data and generate graphical representations of the objects captured by the sensors 16 and provide graphical representations of the projected paths of the moving objects.
The control unit 22 may also process data generated by the monitoring sensors 18 of the vehicle 10. The control unit 22 may receive data from the monitoring sensors 18 in order to determine if the vehicle 10 is making a turn and the speed the vehicle 10 may be traveling. If the control unit 22 determines that the vehicle 10 is turning, the control unit 22 may generate a graphical representation indicating the projected turning path of the vehicle 10.
The warning system 12 is a dynamic system. Thus, the control unit 22 may continuously update graphical representations of the objects captured by the sensors 16 as the vehicle 10 is turning. Thus, graphical representations of projected paths of the objects detected that are moving may be continuously updated. Graphical representations indicating the projected turning path of the vehicle 10 may also be continuously updated as the vehicle 10 progresses into the turn.
The control unit 22 may be coupled to a display 24. The display 24 may be used to show the graphical representations generated by the control unit 22 of the objects captured by the sensors 16, projected paths of the moving objects, as well as the graphical representation indicating the projected turning path of the vehicle 10. Alternatively, the control unit 22 may be coupled to a Heads Up Display (HUD) system 26. The HUD system 26 may be used to display the graphical representations generated by the control unit 22 of the objects captured by the sensors 16, projected paths of the moving objects, as well as the graphical representation indicating the projected turning path of the vehicle 10.
The warning system 12 may be configured so that the display 24 and or HUD system 26 displays the general area viewable in front of the driver when the driver is seated in the vehicle 10 (hereinafter Field of View (FOV) of the driver). However, for objects detected by the sensors 16 but not in the FOV of the driver or towards a far edge of the FOV, for example a pedestrian located behind the vehicle 10, symbols and or warning text and a string indicating a general position of the detected object may be generated and shown on the display 24 and or HUD system 26.
As stated above, a Global Positioning Satellite (GPS) unit 28 may be coupled to the control unit 22. The GPS unit 28 may be used to provide geographical information to the control unit 22. Based on the location indicated by the GPS unit 28, the GPS unit 28 may load and transfer location data about the indicated location. For example, the GPS unit 28 may load and transfer satellite imagery of the current location. This imaging may be sent to the control unit 22 which may generate a graphical representation of the satellite images to be shown on the display 24 and or HUD system 26. Further, as disclosed above, the GPS unit 28 may be used to determine if the vehicle 10 is turning and or moving by monitoring the speed and direction of the vehicle 10.
Referring now to
If an object is detected as moving, the image processing module 30 may be used to calculate the speed and direction of the moving object. Based on the calculated speed and direction of the moving object, a path processing module 32 of the control unit 22 may be used to calculate a projected path of the moving object. The path processing module 32 may further be used to calculate a projected turning path of the vehicle 10. The path processing module 32 may receive data from the monitoring sensor 18 and or GPS unit 28 indicating speed and directional information of the vehicle 10. Based on this information, the path processing module 32 may calculate the projected turning path of the vehicle 10.
An augmented reality processing module 34 of the control unit 22 may be used to generate graphical representations of the different objects detected by the sensors 16, graphical representations of the projected paths of the moving objects and graphical representations indicating the projected turning path of the vehicle 10. The augmented reality processing module 34 may further generate graphical representations of location data provided by the GPS unit 28. The graphical representations generated by the augmented reality processing module 34 may be two dimensional representations or three dimensional representations.
The graphical representations generated by the augmented reality processing module 34 may be shown on a display 24 located within the vehicle 10. Alternatively, the HUD system 26 may be used to display the graphical representations generated by the augmented reality processing module 34.
Referring now to
The control unit 22 may generate one or more graphical representations of pedestrians 40 detected by the sensors 16. In the present embodiment, two pedestrians 40 may be seen as being in the FOV of the driver. To highlight the position of the pedestrians 40, a sign 42 or other symbol may be placed adjacent to each pedestrian 40. In the embodiment shown, the signs 42 may be positioned above each respective pedestrian 40 in the FOV of the driver. The signs 42 may flash and or blink to further draw the driver's attention to the pedestrians 40 in the FOV of the driver.
If a pedestrian 40 is out of the FOV or towards the outer edge of the FOV of the driver, a sign 42 with a pointer 44 may be used to indicate a general position of the pedestrian 40 who is out of or towards the outer edge of the FOV of the driver. The sign 42 and pointer 44 may flash and or blink to further draw the driver's attention to the pedestrians 40 who may be out of or towards the outer edge of the FOV of the driver.
The augmented reality shown may further include a virtual crosswalk 46 for pedestrians 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be used to identify a projected path of the pedestrians 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be dynamic and change position based on the movement of the pedestrians 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be calculated based on the position, speed and direction of the pedestrians 40.
The augmented reality shown may further include a graphical representation of a projected turning path 48 (hereinafter turning path 48) of the vehicle 10. The turning path 48 may be dynamic and adjust in length/size as the vehicle 10 moves during the turn. The turning path 48 may be calculated based on the speed and turning angle of the vehicle 10.
The turning path 48 may be color coded based on the distance between the vehicle 10 intersecting with a particular pedestrian 40. In accordance with one embodiment, the turning path 48 may be color coded to provide a visual indication/warning to the driver. For example, in the turning path 48 shown in
The augmented reality shown may further include a warning sign 50. The warning sign 50 may be positioned along the turning path 48. The warning sign 50 may be used to further provide a visual warning to the driver. For example, based on the current speed and turning angle of the vehicle 10, the warning sign 50 may indicate to the driver to go, yield, stop or the like.
Referring now to
The control unit 22 may generate one or more graphical representations of a pedestrian 40 detected by the sensors 16. In the present embodiment, one pedestrian 40 may be seen as being in the FOV of the driver. To highlight the position of the pedestrian 40, a sign 42 or other symbol may be placed adjacent to each pedestrian 40. In the embodiment shown, the sign 42 may be positioned above the pedestrian 40 in the FOV of the driver. The sign 42 may flash and or blink to further draw the driver's attention to the pedestrian 40 in the FOV of the driver.
If a pedestrian 40 is out of the FOV or towards the outer edge of the FOV of the driver, a sign 42 with a pointer 44 may be used to indicate a general position of the pedestrian 40 who is out of or towards the outer edge of the FOV of the driver. The sign 42 and pointer 44 may flash and or blink to further draw the driver's attention to the pedestrians 40 who may be out of or towards the outer edge of the FOV of the driver.
The augmented reality shown may further include a virtual crosswalk 46 for the pedestrian 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be used to identify a projected path of the pedestrian 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be dynamic and change position based on the movement of the pedestrian 40 who may be seen as being in the FOV of the driver. The virtual crosswalk 46 may be calculated based on the position, speed and direction of the pedestrian 40.
The augmented reality shown may further include a graphical representation of a projected turning path 48 (hereinafter turning path 48) of the vehicle 10. The turning path 48 may be dynamic and adjust in length/size as the vehicle 10 moves during the turn. The turning path 48 may be calculated based on the speed and turning angle of the vehicle 10.
The turning path 48 may be color coded based on the distance between the vehicle 10 intersecting with a particular pedestrian 40. In accordance with one embodiment, the turning path 48 may be color coded to provide a visual indication/warning to the driver. For example, in the turning path 48 shown in
The augmented reality shown may further include a warning sign 50. The warning sign 50 may be positioned along the turning path 48. The warning sign 50 may be used to further provide a visual warning to the driver. For example, based on the current speed and turning angle of the vehicle 10, the warning sign 50 may indicate to the driver to go, yield, stop or the like.
While embodiments of the disclosure have been described in terms of various specific embodiments, those skilled in the art will recognize that the embodiments of the disclosure may be practiced with modifications within the spirit and scope of the claims.
The present application is a Continuation-In-Part of U.S. Patent Application entitle “AUGMENTED REALITY HEADS UP DISPLAY (HUD) FOR LEFT TURN SAFETY CUES,” having Ser. No. 13/800,822, filed Mar. 13, 2013, in the name of Lee Beckwith and Victor Ng-Thow-Hing.
Number | Name | Date | Kind |
---|---|---|---|
5519536 | Hoehn | May 1996 | A |
6727807 | Trajkovic et al. | Apr 2004 | B2 |
6919866 | Kanevsky et al. | Jul 2005 | B2 |
7561966 | Nakamura et al. | Jul 2009 | B2 |
7605773 | Janssen | Oct 2009 | B2 |
7680592 | Ikeda | Mar 2010 | B2 |
7815313 | Ito et al. | Oct 2010 | B2 |
7847678 | Kawamata et al. | Dec 2010 | B2 |
8009024 | Enya et al. | Aug 2011 | B2 |
8035493 | Hioki | Oct 2011 | B2 |
8049609 | Takahashi et al. | Nov 2011 | B2 |
8138990 | Kakizaki | Mar 2012 | B2 |
8144076 | Kakizaki | Mar 2012 | B2 |
9043083 | Kadowaki | May 2015 | B2 |
9073484 | Aimura | Jul 2015 | B2 |
9142132 | Yao | Sep 2015 | B2 |
9294733 | Watanabe | Mar 2016 | B2 |
20050134479 | Isaji et al. | Jun 2005 | A1 |
20050149251 | Donath et al. | Jul 2005 | A1 |
20060055525 | Kubota et al. | Mar 2006 | A1 |
20060103590 | Divon | May 2006 | A1 |
20060151223 | Knoll | Jul 2006 | A1 |
20070182527 | Traylor | Aug 2007 | A1 |
20080015772 | Sanma et al. | Jan 2008 | A1 |
20080133136 | Breed et al. | Jun 2008 | A1 |
20090058761 | Chen et al. | Mar 2009 | A1 |
20090083627 | Onda et al. | Mar 2009 | A1 |
20100098297 | Zhang | Apr 2010 | A1 |
20100157430 | Hotta et al. | Jun 2010 | A1 |
20100253489 | Cui et al. | Oct 2010 | A1 |
20100253494 | Inoue | Oct 2010 | A1 |
20100253540 | Seder et al. | Oct 2010 | A1 |
20100274478 | Takahashi | Oct 2010 | A1 |
20110087433 | Yester | Apr 2011 | A1 |
20110128139 | Tauchi et al. | Jun 2011 | A1 |
20110153198 | Kokkas et al. | Jun 2011 | A1 |
20110181728 | Tieman et al. | Jul 2011 | A1 |
20110199199 | Perkins | Aug 2011 | A1 |
20110288766 | Nagasawa et al. | Nov 2011 | A1 |
20110313653 | Lindner | Dec 2011 | A1 |
20120154441 | Kim | Jun 2012 | A1 |
20120200476 | Kanamori et al. | Aug 2012 | A1 |
20120206483 | Funabashi | Aug 2012 | A1 |
20120218093 | Yoshizawa et al. | Aug 2012 | A1 |
20120224060 | Gurevich et al. | Sep 2012 | A1 |
20120224062 | Lacoste et al. | Sep 2012 | A1 |
20120235805 | Nogami et al. | Sep 2012 | A1 |
20120249589 | Gassner et al. | Oct 2012 | A1 |
20120314071 | Rosenbaum et al. | Dec 2012 | A1 |
20130120850 | Lambert et al. | May 2013 | A1 |
20130293582 | Ng-Thow-Hing | Nov 2013 | A1 |
20130317735 | Mann | Nov 2013 | A1 |
20140063064 | Seo et al. | Mar 2014 | A1 |
20140203925 | Augst | Jul 2014 | A1 |
20140204267 | Akiba | Jul 2014 | A1 |
20140267398 | Beckwith | Sep 2014 | A1 |
20140268353 | Fujimura et al. | Sep 2014 | A1 |
20140362195 | Ng-Thow-Hing et al. | Dec 2014 | A1 |
20150035685 | Strickland | Feb 2015 | A1 |
20150298547 | Inoue | Oct 2015 | A1 |
20150329043 | Skvarce | Nov 2015 | A1 |
20150375680 | Watanabe | Dec 2015 | A1 |
Number | Date | Country |
---|---|---|
2639771 | Sep 2013 | EP |
Entry |
---|
In-Vehicle AR-HUD System to Provide Driving-Safety Information: Park, Hye Sun, Min Woo Park, Kwang Hee Won, Kyong-Ho Kim, and Soon Ki Jung ETRI Journal 35, No. 6 (2013) |
Bergmeier, et al., “Augmented Reality in Vehicle—Technical Realisation of a Contact Analogue Head-up Display Under Automotive Capable Aspects: Usefulness Exemplified Through Night Vision Systems”, AZT online, Edition Sep. 2008. |
Howard, B., “See the Future: Augmented Reality Head-up Displays Beckon”, Extreme Tech, Web, <http://www.extremetech.com/extreme/114635-augmented-reality-head-up-displays-beckon/2>, Jan. 20, 2012. |
Tonnis, et al., “Experimental Evaluation of an Augmented Reality Visualization for Directing a Car Driver's Attention”, Fourth IEEE and ACM International Symposium on Mixed and Augmented Reality (ISMAR'05), pp. 56-59, 2005. |
Tran, C. et al., “A Left-Turn Driving Aid Using Projected Oncoming Vehicle Paths with Augmented Reality”, Proceedings of the 5th International Conference on Automotive USer Interfaces and Interactive Vehicular Applications, p. 300-307, Oct. 28-30, 2013. |
Number | Date | Country | |
---|---|---|---|
20140266656 A1 | Sep 2014 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13800822 | Mar 2013 | US |
Child | 14173506 | US |