The present application generally relates to a system for evaluating performance of a driver.
A system for evaluating performance of a driver of a vehicle with an electronic control unit is provided. The system includes an evaluation processor configured to access driving dynamics data regarding operation of the vehicle, and a driver monitoring sensor in communication with the evaluation processor. The driver monitoring sensor is configured to generate driver status data that relates to a position, orientation, or condition of the driver. The system also includes an external monitoring sensor in communication with the evaluation processor. The external monitoring sensor is configured to generate external interaction data relating to interaction of the driver with an external environment. The evaluation processor is configured to generate a driver rating based upon the driving dynamics data and the driver status data and the external interaction data.
A method for evaluating performance of a driver of a vehicle is also provided. The method comprises the steps of: receiving, by an evaluation processor, driving dynamics data regarding operation of the vehicle; generating, by a driver monitoring sensor in communication with the evaluation processor, driver status data that relates to a position, orientation, or condition of the driver; generating, by an external monitoring sensor in communication with the evaluation processor, external interaction data relating to interaction of the driver with an external environment; and computing, by the evaluation processor, a driver rating based upon the driving dynamics data and the driver status data and the external interaction data.
Further objects, features, and advantages of this application will become readily apparent to persons skilled in the art after a review of the following description, with reference to the drawings and claims that are appended to and form a part of this specification.
Driver feedback may be based either on set driver metrics (attention, engagement, drowsiness, etc) or vehicle dynamics feedback (acceleration, braking, inertial measurements, etc). However, no focus has been put on incorporating actual driver gaze behavior or driver visual monitoring into a system to provide feedback to the driver and/or other entities
Systems may give warnings (haptic, audible, visual alerts) based on set driver metrics, and other systems may compile vehicle dynamic data for driver rating systems like GM's “Teen Driver.” One large gap that has not received consideration includes combining external data (radars, cameras, LiDAR, etc), driver monitoring, and vehicle data for a holistic solution.
The application could be for an improvement on existing Teen Driver/vehicle dynamic based systems, information for insurance agencies to reward low-risk drivers, for driver training to aid in fostering better driving behaviors, or for a general driver “scorecard” that provides feedback on driving metrics.
In one example, a first driver has normal acceleration, braking, keeps within speed limits, is engaged while driving, does not eat or use his cell phone, and always checks for pedestrians both ways at stop signs and cross walks. The first driver may get a great score.
A second driver never uses her seatbelts, speeds, drives while drowsy, does her make-up in the car, angrily yells at other drivers while driving, and does not check blind spots. The second driver may get a bad score with a full breakdown of the items negatively affecting the score.
A third driver may be great at keeping a decent following distance, but does not keep to his lane all the time. He thinks about work sometimes, taking away the focus from driving, but is mostly alert and engaged. The third driver may get a mediocre score, with a breakdown of items that he can improve on.
Referring to
In step 104, the driver status data may be determined for example, through driver monitoring sensors. The driver monitor may be executed by a number of sensors as described elsewhere in this application. The driver status data may include attention zones, engagement of the driver, whether the driver is performing secondary tasks such as eating, drinking, etc., whether the driver has his hands on the steering wheel, impairment of the driver, drowsiness of the driver, the cognitive mode of the driver, emotion of the driver, as well as, alert responsiveness.
In step 106, the environmental interaction is determined. The environmental interaction may include any action by the driver in controlling the vehicle to interact with the outside environment. Specifically, one or more external monitoring sensors is configured to generate external interaction data relating to interaction of the driver with an external, or outside, environment. The external interaction data may be determined by a number of outward looking sensors attached to the vehicle. The environmental interaction may include, for example, the driver's awareness of imminent events, looking for pedestrians on crosswalks, checking blind spots before lane changes, performing full stops at stop signs, or running a red light. In step 108, a driver profile may be generated based on the driver status data and the external interaction data. The driver profile may be uploaded to a network server and may be accessed by various vehicles based on the driver of that vehicle. In some implementations, the driver profile may be loaded onto a mobile device for example, a cell phone, which may then be accessed by the vehicle for example by Wi-Fi when the driver enters the vehicle or when selected by the driver either through the vehicle or through an app on the mobile device. In step 110, a driver baseline is determined. The driver baseline may include the data provided in the driver profile. However, the driver baseline may provide values for each attribute of the driver under a normal (e.g. standardized) conditions. Accordingly, the driver profile over a certain period of time may then be compared to the driver baseline to determine whether the driver is acting substantially different from the norm for that driver in any one of the recorded attributes.
In some embodiments, the evaluation processor may be configure to alert the driver when a currently measured attribute deviates by a first threshold amount from the driver baseline. For example, an audible, haptic, or visual indicator may be presented to the driver in response to determining that the driver is deviating from one or more attributes of the driver baseline by a corresponding first threshold amount. The alert may be tailored to the measured attribute or attributes that deviate from the first threshold. For example, the system may provide one warning message in response to determining that the driver is excessively drowsy, and the system may provide a different warning message in response to determining that the driver is excessively distracted and that the distractions are or could adversely affect their driving ability.
In some embodiments, the evaluation processor may be configure to alert a remote system when a currently measured attribute deviates by a second threshold amount from the driver baseline. In some embodiments, the second threshold may be the same as the first threshold. In some other embodiments, the second threshold may be greater, or farther from the driver baseline than the first threshold. For example, a driver may receive an alert in response to a measured attribute, such as driver drowsiness exceeding the first threshold; and if the driver fails to take corrective action and if the measured attribute continues such that it exceeds the second threshold, the remote system may be alerted. The remote system may be, for example, a server that tracks commercial driving behavior.
For example, a warning message or visual indicator may be presented to the driver in response to determining that the driver is substantially deviating from the driver baseline. Such a warning message or indicator may be tailored to the measured attribute or attributes that deviate from the first threshold. For example, the system may provide one warning message in response to determining that the driver is excessively drowsy, and the system may provide a different warning message in response to determining that the driver is excessively distracted and that the distractions are or could adversely affect their driving ability.
The driver monitor 112 may also be in communication with a driver communication and alert system 118. The driver communication and alert system 118 may include video screens 132, audio system 134, as well as other indicators 136. The screen may be a screen in the console and may be part of the instrument cluster, or a part of a vehicle infotainment system. The audio may be integrated into the vehicle infotainment system or a separate audio feature for example, as part of the navigation or telecommunication systems. The audio may provide noises such as beeps, chirps or chimes or may provide language prompts for example, asking questions or providing statements in an automated or pre-recorded voice. The driver communication and alert system 118 may also include other indicators for example, lamps or LEDs to provide a visual indication or stimulation either on the instrument cluster or elsewhere in the vehicle including for example, on the side view mirrors or rear view mirror.
The driver monitor 112 may also be in communication with an autonomous driving system 150. The autonomous driving system 150 may utilize the driver profile and driver baseline information for making various decisions for example, when and how to provide vehicle control handoff, when making decisions about drivers and objects (e.g. people, vehicles, etc.) around the current vehicle. In one example, a vehicle-to-vehicle communication system may provide information about a driver in a nearby car based on the driver information system and the autonomous driving system 150 may make driving decisions based on the driver profile and/or driver baseline of drivers in surrounding vehicles.
Now referring to
The vehicle 200 may also include biosensors 218. The biosensor 218 may for example, be integrated into a steering wheel of the vehicle. However, other implementations may include integration into seats and/or a seatbelt or within other vehicle controls such as the gear shift or other control knobs. Biosensor 218 may determine a heartbeat, temperature, and/or moisture of the skin of the driver of the vehicle. As such, the condition of the driver may be evaluated by measuring various biosensor readings as provided by the biosensor 218. The system may also have one or more inward or cabin facing cameras 220. The cabin facing cameras 220 may include cameras that operate in the white light spectrum, infrared spectrum, or other available wavelengths. The cameras may be used to determine various gestures of the driver, position or orientation of the driver, or facial expressions of the driver to provide information about the condition of the driver (e.g. emotional state, engagement, drowsiness and impairment of the driver). Further, bioanalysis may be applied to the images from the camera to determine the condition of the driver or if the driver has experienced some symptoms of some medical state. For example, if the driver's eyes are dilated, this may be indicative of a potential medical condition which could be taken into account in controlling the vehicle. As, such, condition of the driver may be determined based on a combination of measurements from one or more sensors. For example, a heart rate in a certain range, a particular facial expression, and skin coloring within a certain range may correspond to a particular emotional state, engagement, drowsiness and/or impairment of the driver.
Cameras 222 may be used to view the external road conditions, such as in front of, behind, or to the side of the vehicle. This may be used to determine the path of the road in front of the vehicle, the lane indications on the road, the condition of the road with regard to road surface, or with regard to the environment external to the vehicle including whether the vehicle is in a rain or snow environment, as well as, lighting conditions external to the vehicle including whether there is glare or glint from the sun or other objects surrounding the vehicle as well as the lack of light due to poor road lighting infrastructure. As discussed previously, the vehicle may include rearward or sideward looking implementations of any of the previously mentioned sensors. As such, a side view mirror sensor 224 may be attached to the side view mirror of the vehicle and may include a radar, Lidar and/or camera sensor for determining external conditions relative to the vehicle including the position of objects such as other vehicles around the instant vehicle. Additionally, rearward facing camera 226 and ultrasonic sensor 228 in the rear bumper of the vehicle provide other exemplary implementations of rearward facing sensors that parallel the functionality of the forward facing sensors described previously.
The vehicle may also include an evaluation processor 230 configured to access driving dynamics data regarding operation of the vehicle. For example, the evaluation processor 230 may be in functional communication with the sensor processer 210. The evaluation processor 230 may be and in functional communication with a driver monitoring sensor configured to generate driver status data that relates to a position, orientation, or condition of the driver. The evaluation processor 230 may also be in functional communication with an external monitoring sensor configured to generate external interaction data relating to interaction of the driver with an external environment. The evaluation processor 230 may be configured to generate a driver rating based upon the driving dynamics data and the driver status data and the external interaction data. In some embodiments, the evaluation processor 230 may be a stand-alone unit. In some other embodiments, the evaluation processor 230 may be implemented integrally with one or more other processors, such as sensor processer 210.
With regard to
Referring to
The driver status data 414 may be generated by a driver monitoring system including sensors configured to monitor the driver. The driver status data 414 may include attention zones, engagement of the driver, whether the driver is performing secondary tasks such as eating, drinking, etc., whether the driver has his hands on the steering wheel, impairment of the driver, drowsiness of the driver, the cognitive mode of the driver, emotion of the driver, as well as alert responsiveness. One or more external monitoring sensors, such as outward looking sensors attached to the vehicle, are configured to generate external interaction data relating to interaction of the driver with an external, or outside, environment. The environmental interaction may include, for example, the driver's awareness of imminent events, looking for pedestrians on crosswalks, checking blind spots before lane changes, performing full stops at stop signs, or running a red light. In some embodiments, the driver rating 418 may be compared to a driver baseline. The driver baseline may include the data provided in a data driver profile. However, the driver baseline may provide values for each attribute of the driver under a normal (e.g. standardized) conditions. Accordingly, the driver profile over a certain period of time may then be compared to the driver baseline to determine whether the driver is acting substantially different from the norm for that driver in any one of the recorded attributes.
A method for evaluating performance of a driver of a vehicle is also provided. The method comprises receiving, by an evaluation processor, driving dynamics data regarding operation of the vehicle. The driving dynamics data may include measured and/or computed data from one or more systems and sensors within the vehicle. The driving dynamics data may include, for example, braking data, acceleration data, maximum speed, stability control events, forward collision alerts, distance from other vehicles, seatbelt usage, lane keeping, and/or data regarding compliance with speed limits.
The method also includes generating, by a driver monitoring sensor in communication with the evaluation processor, driver status data that relates to a position, orientation, or condition of the driver. The driver status data may include, for example, attention zones, engagement of the driver, whether the driver is performing secondary tasks such as eating, drinking, etc., whether the driver has his hands on the steering wheel, impairment of the driver, drowsiness of the driver, the cognitive mode of the driver, emotion of the driver, and/or alert responsiveness.
In some embodiments, the step of generating the driver status data may further include the sub-step of determining a direction that the driver is looking based on data from a camera that is positioned such that the driver is in a field of view of the camera.
In some embodiments, the step of generating the driver status data may further include the sub-step of determining driver contact with a steering wheel of the vehicle. The driver contact may be a binary (yes/no) determination. Alternatively, the driver status data may include specific information regarding the specifics of driver contact with the steering wheel (e.g. how many hands on the wheel, placement of hands on the wheel, etc.).
The method also includes generating, by an external monitoring sensor in communication with the evaluation processor, external interaction data relating to interaction of the driver with an external environment. The external interaction data may describe, for example, the driver's awareness of imminent events, looking for pedestrians on crosswalks, checking blind spots before lane changes, performing full stops at stop signs, or running a red light.
The method also includes the step of computing, by the evaluation processor, a driver rating based upon the driving dynamics data and the driver status data and the external interaction data.
The methods, devices, processing, and logic described above may be implemented in many different ways and in many different combinations of hardware and software. For example, all or parts of the implementations may be circuitry that includes an instruction processor, such as a Central Processing Unit (CPU), microcontroller, or a microprocessor; an Application Specific Integrated Circuit (ASIC), Programmable Logic Device (PLD), or Field Programmable Gate Array (FPGA); or circuitry that includes discrete logic or other circuit components, including analog circuit components, digital circuit components or both; or any combination thereof. The circuitry may include discrete interconnected hardware components and/or may be combined on a single integrated circuit die, distributed among multiple integrated circuit dies, or implemented in a Multiple Chip Module (MCM) of multiple integrated circuit dies in a common package, as examples.
The circuitry may further include or access instructions for execution by the circuitry. The instructions may be stored in a tangible storage medium that is other than a transitory signal, such as a flash memory, a Random Access Memory (RAM), a Read Only Memory (ROM), an Erasable Programmable Read Only Memory (EPROM); or on a magnetic or optical disc, such as a Compact Disc Read Only Memory (CDROM), Hard Disk Drive (HDD), or other magnetic or optical disk; or in or on another machine-readable medium. A product, such as a computer program product, may include a storage medium and instructions stored in or on the medium, and the instructions when executed by the circuitry in a device may cause the device to implement any of the processing described above or illustrated in the drawings.
The implementations may be distributed as circuitry among multiple system components, such as among multiple processors and memories, optionally including multiple distributed processing systems. Parameters, databases, and other data structures may be separately stored and managed, may be incorporated into a single memory or database, may be logically and physically organized in many different ways, and may be implemented in many different ways, including as data structures such as linked lists, hash tables, arrays, records, objects, or implicit storage mechanisms. Programs may be parts (e.g., subroutines) of a single program, separate programs, distributed across several memories and processors, or implemented in many different ways, such as in a library, such as a shared library (e.g., a Dynamic Link Library (DLL)). The DLL, for example, may store instructions that perform any of the processing described above or illustrated in the drawings, when executed by the circuitry.
As a person skilled in the art will readily appreciate, the above description is meant as an illustration of the principles of this application. This description is not intended to limit the scope or application of the claim in that the assembly is susceptible to modification, variation and change, without departing from spirit of this application, as defined in the following claims.
The present application claims the benefit of the filing date of U.S. Provisional Application No. 62/863,130, filed Jun. 18, 2019, the disclosure of which is hereby incorporated herein by reference in its entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2020/038382 | 6/18/2020 | WO |
Number | Date | Country | |
---|---|---|---|
62863130 | Jun 2019 | US |