The subject matter described herein relates in general to autonomous vehicles and, more particularly, to environmental driver comfort feedback for autonomous vehicles.
Autonomous vehicles react to the driving environment around them. Autonomous vehicles recognize traffic signs, traffic lights, and other surrounding objects. Sometimes the autonomous vehicle may stop, change lanes, swerve, or take other actions based on surrounding objects. The driver and/or occupants of the vehicle may be alarmed by sudden changes in the status of the vehicle if they are unaware of the upcoming change.
In an embodiment herein, a driver notification module is described, the driver notification module comprising a memory, and a processor coupled to the memory. The processor may be configured to: receive first information related to an upcoming event; determine an action for an autonomous vehicle based, at least in part, on the first information; determine one or more characteristics of the action based, at least in part on the first information and a current status of the autonomous vehicle; determine a notification based, at least in part, on the action and the one or more characteristics; and deliver the notification within the autonomous vehicle.
In another embodiment herein, a method for driver notification is described. The method may comprise receiving, by a driver notification module, first information related to an upcoming event; determining, by the driver notification module, an action for an autonomous vehicle based, at least in part, on the first information; determining, by the driver notification module, one or more characteristics of the action based, at least in part on the first information and a current status of the autonomous vehicle; determining, by the driver notification module, a notification based, at least in part, on the action and the one or more characteristics; and delivering, by the driver notification module, the notification within the autonomous vehicle.
In another embodiment herein, an autonomous vehicle is described. The autonomous vehicle may comprise a sensor; and a driver notification module in communication with the sensor. The driver notification module may be configured to receive first information related to an upcoming event; determine an action for the autonomous vehicle based, at least in part, on the first information; determine one or more characteristics of the action based, at least in part on the first information and a current status of the autonomous vehicle; determine a notification based, at least in part, on the action and the one or more characteristics; and deliver the notification within the autonomous vehicle.
Described herein is a system and method for environmental driver comfort feedback for autonomous vehicles. When an autonomous vehicle is operating autonomously, the vehicle may be configured to recognize vehicle traffic, signs, traffic lights, road hazards, and other upcoming events. The vehicle operating autonomously may include any action taken by the vehicle that does not require driver intervention. Upcoming events may include any situation or object that the vehicle may encounter in the future. The autonomous vehicle may be configured to take an action based on the upcoming events. A driver and/or occupants of the vehicle may become uncomfortable if the vehicle takes an action unbeknownst to the driver and/or occupants. For example, if a vehicle is approaching a red light, the vehicle may recognize the red light and abruptly stop. The driver and/or occupants of the vehicle may be uncomfortable because they are unaware of the reason for the stop.
An autonomous vehicle may be configured with a driver notification system. The driver notification system may detect an upcoming event and determine a corresponding action for the upcoming event. The driver notification system may also determine characteristics of the action to be taken. For example, the driver notification may notify the driver that a red light has been detected and that the vehicle will be stopping shortly. The notification may be audible or visual or both. The notification may place the driver at ease when the action is taken.
The event and action database 130 may be a parameterized database that associates events that vehicle 110 may encounter to actions vehicle 110 may take in response to the event. The actions may be based on the type and/or severity of the event. The actions may be associated with a notification to be provided to driver and/or occupants of the vehicle. The notification may notify driver and/or occupants of the upcoming event and the action taken in response to the event. The notification may be a visual notification, an audio notification, or both.
The geography data store 135 may contain information related to geography vehicle 110 may encounter. The geography data store 135 may contain road maps, topography maps, traffic maps, traffic statistics, and other information that may be used by vehicle 110. Geography data store 135 may contain locations of traffic lights and other traffic control devices. Geography data store 135 and sensor 140 may be used alone or in combination to determine upcoming events that a driver may be notified of.
The upcoming event may be detected using data from a sensor, e.g., sensor 140, data from a geography data store, e.g., geography data store 135, other data sources, or any combination thereof. For example, sensor 140 may comprise a camera and may capture an image of a bicyclist on the shoulder of the road ahead of the vehicle. As another example, driver notification module 150, may interact with geography data store 135 to determine a sharp curve in the upcoming road. As yet another example, sensor 140 may detect a pothole in the road and the location of the detected pothole may be correlated with a location of the pothole from geography data store 135.
After an upcoming event is detected, the method may continue at block 320 where an action that corresponds to the event may be detected. The action may be any action that causes the vehicle to change its current operating status. Examples of actions include, but are not limited to deceleration, swerving, stopping lane changing, accelerating, activating a turn signal, etc.
Driver notification module 150 may interact with the event and actions database 130 to determine actions that correspond to the upcoming event. For example, an action associated with the bicyclist on the shoulder of the road ahead of the vehicle may include a lane shift. As another example, an action associated with the sharp curve in the upcoming road may include deceleration of the vehicle. As yet another example, an action associated with the pothole in the road may include the vehicle swerving to avoid the pothole.
After an action is determined, the method may continue at block 330 where characteristics of the action may be determined. The characteristics of the action may be determined based upon the upcoming event and/or the determined action. Examples of characteristics of the action include, but are not limited to intensity of the action, duration of the action, time until the action occurs, etc.
Driver notification module 150 may interact with the event and actions database 130 to determine characteristics of the action. For example, characteristics of a lane shift may include identifying which lane the vehicle will shift to and/or when the shift will occur. As another example, characteristics of deceleration may include identifying how much velocity will decrease as a result of the deceleration, the duration of the deceleration, and/or identifying when the deceleration will begin. As yet another example, characteristics of swerving may include direction of the swerve, intensity of the swerve, and/or when the swerve will occur.
After an action and characteristics of the action have been determined, the method may continue at block 340 where the driver and/or occupants of the vehicle may be notified. The driver notification module 150 may create a notification by retrieving a notification previously stored in a database, e.g., event and actions database 130. The driver notification module 150 may create the notification using a text to speech generator and the previously determined action and characteristics of the action. Creating the notification may include other actions taken by the driver notification module 150 to provide the vehicle occupants with information related to the action and/or characteristics of the action. The notification may include the upcoming event, the action, and/or characteristics of the action. The notification may be provided audibly and/or visually. The notification may be provided via any combination of HUD 220, in-dash display 230, or speaker 240.
Driver notification module 150 may interact with vehicle systems to provide the notification. For example, the driver notification module 150 may cause the notification “bicyclist ahead, the vehicle will change to the left lane in 35 seconds” to be displayed and played to the driver and/or vehicle occupants. As another example, the driver notification module 150 may cause the notification “sharp curve ahead, the vehicle will be decelerating to 25 miles per hour” to be displayed and played to the driver and/or vehicle occupants. As yet another example, the driver notification module 150 may cause the notification “pothole ahead, the vehicle will be swerving to the right in 400 feet” to be displayed and played to the driver and/or vehicle occupants.
After the driver and/or occupants of the vehicle are notified, the method may continue at step 350, where the vehicle takes the action described in the notification. In some embodiments, this step is optional, e.g., the driver may take control of the vehicle or the vehicle may determine autonomous driving is no longer possible based upon the event.
In addition to the processor 410 (which may be referred to as a central processor unit or CPU), the system 400 might include network connectivity devices 420, random access memory (RAM) 430, read only memory (ROM) 440, secondary storage 450, and input/output (I/O) devices 460. These components might communicate with one another via a bus 470. In some cases, some of these components may not be present or may be combined in various combinations with one another or with other components not shown. These components might be located in a single physical entity or in more than one physical entity. Any actions described herein as being taken by the processor 410 might be taken by the processor 410 alone or by the processor 410 in conjunction with one or more components shown or not shown in the drawing, such as a digital signal processor (DSP) 480. Although the DSP 480 is shown as a separate component, the DSP 480 might be incorporated into the processor 410.
The processor 410 executes instructions, codes, computer programs, or scripts that it might access from the network connectivity devices 420, RAM 430, ROM 440, or secondary storage 450 (which might include various disk-based systems such as hard disk, floppy disk, or optical disk). While only one CPU 410 is shown, multiple processors may be present. Thus, while instructions may be discussed as being executed by a processor, the instructions may be executed simultaneously, serially, or otherwise by one or multiple processors. The processor 410 may be implemented as one or more CPU chips and may be a hardware device capable of executing computer instructions. In an embodiment, the processor 410 may be configured to execute instructions that cause a driver notification module to perform the method of
The network connectivity devices 420 may take the form of modems, modem banks, Ethernet devices, universal serial bus (USB) interface devices, serial interfaces, token ring devices, fiber distributed data interface (FDDI) devices, wireless local area network (WLAN) devices, radio transceiver devices such as code division multiple access (CDMA) devices, global system for mobile communications (GSM) radio transceiver devices, universal mobile telecommunications system (UMTS) radio transceiver devices, long term evolution (LTE) radio transceiver devices, worldwide interoperability for microwave access (WiMAX) devices, and/or other well-known devices for connecting to networks. These network connectivity devices 420 may enable the processor 410 to communicate with the Internet or one or more telecommunications networks or other networks from which the processor 410 might receive information or to which the processor 410 might output information. The network connectivity devices 420 might also include one or more transceiver components 425 capable of transmitting and/or receiving data wirelessly.
The RAM 430 might be used to store volatile data and perhaps to store instructions that are executed by the processor 410. The ROM 440 is a non-volatile memory device that typically has a smaller memory capacity than the memory capacity of the secondary storage 450. ROM 440 might be used to store instructions and perhaps data that are read during execution of the instructions. Access to both RAM 430 and ROM 440 is typically faster than to secondary storage 450. The secondary storage 450 is typically comprised of one or more disk drives or tape drives and might be used for non-volatile storage of data or as an over-flow data storage device if RAM 430 is not large enough to hold all working data. Secondary storage 450 may be used to store programs that are loaded into RAM 430 when such programs are selected for execution.
The I/O devices 460 may include liquid crystal displays (LCDs), touch screen displays, keyboards, keypads, switches, dials, mice, trackballs, voice recognizers, card readers, paper tape readers, printers, video monitors, or other well-known input/output devices. Also, the transceiver 425 might be considered to be a component of the I/O devices 460 instead of or in addition to being a component of the network connectivity devices 420.
Detailed embodiments are disclosed herein; however, it is to be understood that the disclosed embodiments are intended only as examples. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for teaching one skilled in the art to variously employ the aspects herein in virtually any appropriately detailed structure. Further, the terms and phrases used herein are not intended to be limiting but rather to provide an understandable description of possible implementations. Various embodiments are shown in
It will be appreciated that for simplicity and clarity of illustration, where appropriate, reference numerals have been repeated among the different figures to indicate corresponding or analogous elements. In addition, numerous specific details are set forth in order to provide a thorough understanding of the embodiments described herein. However, it will be understood by those of ordinary skill in the art that the embodiments described herein can be practiced without these specific details.
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments. In this regard, each block in the flowcharts or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
The systems, components and/or processes described above can be realized in hardware or a combination of hardware and software and can be realized in a centralized fashion in one processing system or in a distributed fashion where different elements are spread across several interconnected processing systems. Any kind of processing system or other apparatus adapted for carrying out the methods described herein is suited. A typical combination of hardware and software can be a processing system with computer-usable program code that, when being loaded and executed, controls the processing system such that it carries out the methods described herein. The systems, components and/or processes also can be embedded in a computer-readable storage, such as a computer program product or other data programs storage device, readable by a machine, tangibly embodying a program of instructions executable by the machine to perform methods and processes described herein. These elements also can be embedded in an application product which comprises all the features enabling the implementation of the methods described herein and, which when loaded in a processing system, is able to carry out these methods.
Furthermore, arrangements described herein may take the form of a computer program product embodied in one or more computer-readable media having computer-readable program code embodied or embedded, e.g., stored, thereon. Any combination of one or more computer-readable media may be utilized. The computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium. The phrase “computer-readable storage medium” means a non-transitory storage medium.
Number | Name | Date | Kind |
---|---|---|---|
4238778 | Ohsumi | Dec 1980 | A |
5983161 | Lemelson | Nov 1999 | A |
6678590 | Burchfiel | Jan 2004 | B1 |
7660436 | Chang | Feb 2010 | B2 |
7667581 | Fujimoto | Feb 2010 | B2 |
7831407 | Huang | Nov 2010 | B2 |
8108147 | Blackburn | Jan 2012 | B1 |
8112225 | Eidehall | Feb 2012 | B2 |
8340883 | Arbitmann | Dec 2012 | B2 |
8504283 | Aso | Aug 2013 | B2 |
8552886 | Bensoussan | Oct 2013 | B2 |
8576055 | Hara | Nov 2013 | B2 |
8589061 | Bengtsson | Nov 2013 | B2 |
8676491 | Taylor | Mar 2014 | B2 |
8818681 | Dolgov | Aug 2014 | B1 |
8874300 | Allard | Oct 2014 | B2 |
8948955 | Zhu | Feb 2015 | B2 |
9452754 | Clarke | Sep 2016 | B2 |
9493118 | Laur | Nov 2016 | B1 |
9511730 | Wu | Dec 2016 | B1 |
9517767 | Kentley | Dec 2016 | B1 |
9633565 | Conrad | Apr 2017 | B2 |
9725036 | Tarte | Aug 2017 | B1 |
9804599 | Kentley-Klay | Oct 2017 | B2 |
9805423 | Konrardy | Oct 2017 | B1 |
9975379 | Gowda | May 2018 | B1 |
20020198632 | Breed | Dec 2002 | A1 |
20030014165 | Baker | Jan 2003 | A1 |
20050060069 | Breed | Mar 2005 | A1 |
20070078600 | Fregene | Apr 2007 | A1 |
20070205875 | De Haan | Sep 2007 | A1 |
20080147253 | Breed | Jun 2008 | A1 |
20090074249 | Moed | Mar 2009 | A1 |
20090076702 | Arbitmann | Mar 2009 | A1 |
20090184862 | Stayton | Jul 2009 | A1 |
20090326796 | Prokhorov | Dec 2009 | A1 |
20100179760 | Petrini | Jul 2010 | A1 |
20100305857 | Byrne | Dec 2010 | A1 |
20110190972 | Timmons | Aug 2011 | A1 |
20110210872 | Molander | Sep 2011 | A1 |
20120029964 | Tengler | Feb 2012 | A1 |
20120092208 | Lemire | Apr 2012 | A1 |
20120126997 | Bensoussan | May 2012 | A1 |
20120143488 | Othmezouri | Jun 2012 | A1 |
20140032049 | Moshchuk | Jan 2014 | A1 |
20140136044 | Conrad | May 2014 | A1 |
20140136045 | Zhu | May 2014 | A1 |
20140139366 | Moses | May 2014 | A1 |
20150003669 | Solgi | Jan 2015 | A1 |
20150134226 | Palmer | May 2015 | A1 |
20150160338 | Bageshwar | Jun 2015 | A1 |
20150166062 | Johnson | Jun 2015 | A1 |
20150235092 | Mei | Aug 2015 | A1 |
20160003636 | Ng-Thow-Hing | Jan 2016 | A1 |
20160125746 | Kunzi | May 2016 | A1 |
20160152180 | Kirsch | Jun 2016 | A1 |
20160167668 | Prokhorov | Jun 2016 | A1 |
20160216130 | Abramson | Jul 2016 | A1 |
20160221500 | Sakai | Aug 2016 | A1 |
20160272201 | Kang | Sep 2016 | A1 |
20160318442 | James | Nov 2016 | A1 |
20160325753 | Stein | Nov 2016 | A1 |
20160334797 | Ross | Nov 2016 | A1 |
20160375767 | Konet | Dec 2016 | A1 |
20160375900 | Laur | Dec 2016 | A1 |
20170028987 | Yamada | Feb 2017 | A1 |
20170057507 | Gordon | Mar 2017 | A1 |
20170120804 | Kentley | May 2017 | A1 |
20170124658 | Gordon | May 2017 | A1 |
20170144596 | Wu | May 2017 | A1 |
20170240096 | Ross | Aug 2017 | A1 |
20170240098 | Sweeney | Aug 2017 | A1 |
20170254880 | Smith | Sep 2017 | A1 |
20170297588 | Doshi | Oct 2017 | A1 |
20170316696 | Bartel | Nov 2017 | A1 |
20180050704 | Tascione | Feb 2018 | A1 |
20180093676 | Emura | Apr 2018 | A1 |
Entry |
---|
Koo J., Kwac J., Ju W., Steinert M., Leifer L., & Nass C., “Why did my car just do that? Explaining semi-autonomous driving actions to improve driver understanding, trust and performance.” in International Journal of Interactive Design and Manufacturing. Springer. Apr. 2014 in 7 pages. |
Number | Date | Country | |
---|---|---|---|
20180052461 A1 | Feb 2018 | US |