Reward system related to a vehicle-to-vehicle communication system

Information

  • Patent Grant
  • 10733673
  • Patent Number
    10,733,673
  • Date Filed
    Friday, May 27, 2016
    7 years ago
  • Date Issued
    Tuesday, August 4, 2020
    3 years ago
Abstract
System and methods are disclosed for determining, through vehicle-to-vehicle communication, whether vehicles are involved in autonomous droning. Vehicle driving data and other information may be used to calculate a autonomous droning reward amount. In addition, vehicle involved in a drafting relationship in addition to, or apart from, an autonomous droning relationship may be financially rewarded. Moreover, aspects of the disclosure related to determining ruminative rewards and/or aspects of vehicle insurance procurement/underwriting.
Description
TECHNICAL FIELD

Aspects of the disclosure generally relate to the analysis of vehicle driving data, including vehicle driving data from vehicles involved in aspects of autonomous driving. In particular, various aspects of the disclosure relate to receiving and transmitting driving data, included data related to autonomous driving relationships between vehicles, analyzing driving data, and determining ruminative rewards and/or aspects of vehicle insurance procurement.


BACKGROUND

Many vehicles include sensors and internal computer systems designed to monitor and control vehicle operations, driving conditions, and driving functions. Advanced vehicles systems can perform such tasks as monitoring fuel consumption and optimizing engine operation to achieve higher fuel efficiency, detecting and correcting a loss of traction on an icy road, and detecting a collision and automatically contacting emergency services. Various vehicle-based communication systems allow vehicles to communicate with other devices inside or outside of the vehicle. For example, a Bluetooth system may enable communication between the vehicle and the driver's mobile phone. Telematics systems may be configured to access vehicle computers and sensor data, including on-board diagnostics systems (OBD), and transmit the data to a display within the vehicle, a personal computer or mobile device, or to a centralized data processing system. Data obtained from vehicle sensors and OBD systems has been used for a variety of purposes, including maintenance, diagnosis, and analysis. Additionally, vehicle-to-vehicle (V2V) communication systems can be used to provide drivers with safety warnings and collision alerts based on data received from other nearby vehicles. Additionally, vehicles can include autonomous driving systems that assume all or part of real-time driving functions to operate the vehicle without real-time input from a human operator.


When out on the road, vehicles and drivers may engage in many different types of driving behaviors, including various “social interactions” with other vehicles and drivers. An example social interaction can include vehicle drafting where one vehicle follows another vehicle to reduce the overall effect of drag and improve fuel efficiency. Another example social interaction can include autonomous vehicle droning where a vehicle engages in at least partial autonomous driving based on the driving of a lead or pilot vehicle.


SUMMARY

The following presents a simplified summary in order to provide a basic understanding of some aspects of the disclosure. The summary is not an extensive overview of the disclosure. It is neither intended to identify key or critical elements of the disclosure nor to delineate the scope of the disclosure. The following summary merely presents some concepts of the disclosure in a simplified form as a prelude to the description below.


Aspects of the disclosure relate to methods and computer devices for receiving and transmitting driving data, analyzing driving data, and determining whether a first vehicle is engaged in a drafting relationship with at least a second vehicle and allocating a drafting reward based on the drafting relationship. A drafting characteristic of the drafting relationship can be determined using vehicle operational data. Example drafting characteristics can include a vehicle spacing, vehicle speed, and/or vehicle type. A drafting property associated with the first or second vehicle can be determined using the drafting characteristic. Example drafting properties can include a drafting fuel savings rate, a drafting savings amount, and/or a percentage increase in miles-per-gallon. Example drafting rewards can include a cash payment, a carbon credit, a fuel credit, a tax credit, a rebate, and at least a portion of a drafting fuel savings amount associated with the first vehicle driving analysis computer.


In accordance with further aspects of the present disclosure, a property of an insurance policy may be determined by receiving and analyzing driving data from vehicles engaged in an autonomous droning relationship where a vehicle engages in at least partial autonomous driving based on information from another vehicle. A characteristic of the autonomous droning relationship can be determined from, for example, the received driving data. Example characteristics of an autonomous droning relationship include identification of a lead vehicle and a drone vehicle, an amount of time a vehicle is the lead or drone vehicle, and the amount of driving distance a vehicle is the lead or drone vehicle. Example properties of an insurance policy can include a premium for the first insurance policy, a deductible of the first insurance policy, and a coverage amount of the first insurance policy. Aspects of the disclosure also include determining an autonomous droning insurance factor using vehicle operational data and/or a characteristic of the autonomous droning relationship. The autonomous droning insurance factor can be used to determine the property of the first insurance policy using the autonomous droning insurance factor.


Aspects of the present disclosure provide incentives for beneficial social interactions between vehicles and drivers. Other features and advantages of the disclosure will be apparent from the additional description provided herein.





BRIEF DESCRIPTION OF THE DRAWINGS

A more complete understanding of the present invention and the advantages thereof may be acquired by referring to the following description in consideration of the accompanying drawings, in which like reference numbers indicate like features, and wherein:



FIG. 1 illustrates a network environment and computing systems that may be used to implement aspects of the disclosure.



FIG. 2 is a diagram illustrating various example components of a driving analysis computing device according to one or more aspects of the disclosure.



FIG. 3 is a flow diagram illustrating an example method of analyzing vehicle driving data, including determining whether vehicles are engaged in a drafting relationship, determining a drafting property associated with a vehicle in a drafting relationship, and allocating a drafting reward based on the drafting property.



FIG. 4 is a flow diagram illustrating an example method of analyzing vehicle driving data, including receiving vehicle operational data pertaining to vehicles engaged in an autonomous droning relationship, determining a characteristic of an autonomous droning relationship between vehicles, and determining a property of an insurance policy using the characteristic of the autonomous droning relationship.



FIG. 5 is a flow diagram illustrating an example method used in an autonomous droning reward system.



FIGS. 6A-6B are diagrams illustrating examples of various driving interactions that may be analyzed according to one or more aspects of the disclosure.



FIG. 7 is a diagram illustrating one example of an autonomous droning reward calculator in accordance with various aspects of the disclosure.



FIG. 8 illustrates graphical representations of illustrative formulas for calculating an autonomous droning reward amount based on one or more inputs, in accordance with various aspects of the disclosure.



FIG. 9 is a diagram illustrating one example of an account processor in accordance with various aspects of the disclosure.





DETAILED DESCRIPTION

In the following description of the various embodiments, reference is made to the accompanying drawings, which form a part hereof, and in which is shown by way of illustration, various embodiments of the disclosure that may be practiced. It is to be understood that other embodiments may be utilized.


As will be appreciated by one of skill in the art upon reading the following disclosure, various aspects described herein may be embodied as a method, a computer system, or a computer program product. Accordingly, those aspects may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. In addition, aspects may take the form of a computing device configured to perform specified actions. Furthermore, such aspects may take the form of a computer program product stored by one or more computer-readable storage media having computer-readable program code, or instructions, embodied in or on the storage media. Any suitable computer readable storage media may be utilized, including hard disks, CD-ROMs, optical storage devices, magnetic storage devices, and/or any combination thereof. In addition, various signals representing data or events as described herein may be transferred between a source and a destination in the form of electromagnetic waves traveling through signal-conducting media such as metal wires, optical fibers, and/or wireless transmission media (e.g., air and/or space).



FIG. 1 illustrates a block diagram of a computing device 101 in driving analysis communication system 100 that may be used according to one or more illustrative embodiments of the disclosure. The driving analysis device 101 may have a processor 103 for controlling overall operation of the device 101 and its associated components, including RAM 105, ROM 107, input/output module 109, and memory unit 115. The computing device 101, along with one or more additional devices (e.g., terminals 141, 151) may correspond to any of multiple systems or devices, such as a driving analysis computing devices or systems, configured as described herein for transmitting and receiving vehicle operational data, analyzing vehicle operational data, and determining driving characteristics and various properties related to driver rewards and/or vehicle insurance based on the data. Vehicle operational data can include data collected from vehicle sensors and OBD systems. Vehicle operations can also include data pertaining to the driver of a vehicle. Vehicle operational data can also include data pertaining to other nearby vehicles collected via, for example, V2V communications. As used herein, vehicle operation data is used interchangeably with driving data.


Input/Output (I/O) 109 may include a microphone, keypad, touch screen, and/or stylus through which a user of the computing device 101 may provide input, and may also include one or more of a speaker for providing audio output and a video display device for providing textual, audiovisual and/or graphical output. Software may be stored within memory unit 115 and/or other storage to provide instructions to processor 103 for enabling device 101 to perform various functions. For example, memory unit 115 may store software used by the device 101, such as an operating system 117, application programs 119, and an associated internal database 121. The memory unit 115 includes one or more of volatile and/or non-volatile computer memory to store computer-executable instructions, data, and/or other information. Processor 103 and its associated components may allow the driving analysis system 101 to execute a series of computer-readable instructions to transmit or receive vehicle driving data, analyze driving data, determine driving characteristics from the driving data, and determine properties of, for example, driver rewards or insurance policies based on the driving data.


The driving analysis computing device 101 may operate in a networked environment 100 supporting connections to one or more remote computers, such as terminals/devices 141 and 151. Driving analysis computing device 101, and related terminals/devices 141 and 151, may include devices installed in vehicles, mobile devices that may travel within vehicles, or devices outside of vehicles that are configured to receive and process vehicle and driving data. Thus, the driving analysis computing device 101 and terminals/devices 141 and 151 may each include personal computers (e.g., laptop, desktop, or tablet computers), servers (e.g., web servers, database servers), vehicle-based devices (e.g., on-board vehicle computers, short-range vehicle communication systems, telematics devices), or mobile communication devices (e.g., mobile phones, portable computing devices, and the like), and may include some or all of the elements described above with respect to the driving analysis computing device 101. The network connections depicted in FIG. 1 include a local area network (LAN) 125 and a wide area network (WAN) 129, and a wireless telecommunications network 133, but may also include other networks. When used in a LAN networking environment, the driving analysis computing device 101 may be connected to the LAN 125 through a network interface or adapter 123. When used in a WAN networking environment, the device 101 may include a modem 127 or other means for establishing communications over the WAN 129, such as network 131 (e.g., the Internet). When used in a wireless telecommunications network 133, the device 101 may include one or more transceivers, digital signal processors, and additional circuitry and software for communicating with wireless computing devices 141 (e.g., mobile phones, short-range vehicle communication systems, vehicle telematics devices) via one or more network devices 135 (e.g., base transceiver stations) in the wireless network 133.


It will be appreciated that the network connections shown are illustrative and other means of establishing a communications link between the computers may be used. The existence of any of various network protocols such as TCP/IP, Ethernet, FTP, HTTP and the like, and of various wireless communication technologies such as GSM, CDMA, WiFi, and WiMAX, is presumed, and the various computing devices and driving analysis system components described herein may be configured to communicate using any of these network protocols or technologies.


Additionally, one or more application programs 119 used by the driving analysis computing device 101 may include computer executable instructions (e.g., driving data analysis programs, driving characteristic algorithms, driving and insurance policy properties algorithms, and driver reward algorithms) for transmitting and receiving vehicle driving data, determining driving characteristics, and determining various properties associated with one or more vehicles or drivers, and performing other related functions as described herein.


As used herein, a driving characteristic may refer to one or more actions or events performed by a vehicle and can include aspects of information identified or determined from vehicle operational data collected from a vehicle. A driving characteristic can include, for example, a vehicle speed and/or gas mileage identified from the vehicle operational data. In addition, for example, a driving characteristic may include a vehicle speed or a gas mileage determined from other operational data collected from the vehicle. As discussed below, a driving characteristic may be determined from driving data collected by a vehicle sensors and telematics device, and/or additional data received from other nearby vehicles using vehicle-to-vehicle (V2V) communications. It should be understood that a driving characteristic may be associated with a vehicle, a driver, or a group of vehicles or drivers engaged in social interaction, such as an autonomous droning relationship



FIG. 2 is a diagram of an illustrative driving analysis system 200 including two vehicles 210 and 220, a driving analysis server 250, and additional related components. Each component shown in FIG. 2 may be implemented in hardware, software, or a combination of the two. Additionally, each component of the driving analysis system 200 may include a computing device (or system) having some or all of the structural components described above for computing device 101.


Vehicles 210 and 220 in the driving analysis system 200 may be, for example, automobiles, motorcycles, scooters, buses, recreational vehicles, boats, or other vehicles for which a vehicle driving data may be collected and analyzed. The vehicles 210 and 220 each include vehicle operation sensors 211 and 221 capable of detecting and recording various conditions at the vehicle and operational parameters of the vehicle. For example, sensors 211 and 221 may detect and store data corresponding to the vehicle's location (e.g., GPS coordinates), speed and direction, rates of acceleration or braking, gas mileage, and specific instances of sudden acceleration, braking, and swerving. Sensors 211 and 221 also may detect and store data received from the vehicle's 210 internal systems, such as impact to the body of the vehicle, air bag deployment, headlights usage, brake light operation, door opening and closing, door locking and unlocking, cruise control usage, hazard lights usage, windshield wiper usage, horn usage, turn signal usage, seat belt usage, phone and radio usage within the vehicle, maintenance performed on the vehicle, and other data collected by the vehicle's computer systems, including the vehicle OBD.


Additional sensors 211 and 221 may detect and store the external driving conditions, for example, external temperature, rain, snow, light levels, and sun position for driver visibility. For example, external cameras and proximity sensors 211 and 221 may detect other nearby vehicles, vehicle spacing, traffic levels, road conditions, traffic obstructions, animals, cyclists, pedestrians, and other conditions that may factor into a driving event data analysis. Sensors 211 and 221 also may detect and store data relating to moving violations and the observance of traffic signals and signs by the vehicles 210 and 220. Additional sensors 211 and 221 may detect and store data relating to the maintenance of the vehicles 210 and 220, such as the engine status, oil level, engine coolant temperature, odometer reading, the level of fuel in the fuel tank, engine revolutions per minute (RPMs), and/or tire pressure.


Vehicles sensors 211 and 221 also may include cameras and/or proximity sensors capable of recording additional conditions inside or outside of the vehicles 210 and 220. For example, internal cameras may detect conditions such as the number of the passengers and the types of passengers (e.g. adults, children, teenagers, pets, etc.) in the vehicles, and potential sources of driver distraction within the vehicle (e.g., pets, phone usage, unsecured objects in the vehicle). Sensors 211 and 221 also may be configured to collect data a driver's movements or the condition of a driver. For example, vehicles 210 and 220 may include sensors that monitor a driver's movements, such as the driver's eye position and/or head position, etc. Additional sensors 211 and 221 may collect data regarding the physical or mental state of the driver, such as fatigue or intoxication. The condition of the driver may be determined through the movements of the driver or through other sensors, for example, sensors that detect the content of alcohol in the air or blood alcohol content of the driver, such as a breathalyzer.


Certain vehicle sensors 211 and 221 also may collect information regarding the driver's route choice, whether the driver follows a given route, and to classify the type of trip (e.g. commute, errand, new route, etc.). In certain embodiments, sensors and/or cameras 211 and 221 may determine when and how often the vehicles 210 and 220 stay in a single lane or stray into other lanes. A Global Positioning System (GPS), locational sensors positioned inside the vehicles 210 and 220, and/or locational sensors or devices external to the vehicles 210 and 220 may be used determine the route, lane position, and other vehicle position/location data.


The data collected by vehicle sensors 211 and 221 may be stored and/or analyzed within the respective vehicles 210 and 220, such as for example a driving analysis computer 217, 227 integrated into the vehicle, and/or may be transmitted to one or more external devices. For example, as shown in FIG. 2, sensor data may be transmitted via short-range communication systems 212 and 222 to other nearby vehicles. Additionally, the sensor data may be transmitted via telematics devices 213 and 223 to one or more remote computing devices, such as driving analysis server 250.


Short-range communication systems 212 and 222 are vehicle-based data transmission systems configured to transmit vehicle operational data to other nearby vehicles, and to receive vehicle operational data from other nearby vehicles. In some examples, communication systems 212 and 222 may use the dedicated short-range communications (DSRC) protocols and standards to perform wireless communications between vehicles. In the United States, 75 MHz in the 5.850-5.925 GHz band have been allocated for DSRC systems and applications, and various other DSRC allocations have been defined in other countries and jurisdictions. However, short-range communication systems 212 and 222 need not use DSRC, and may be implemented using other short-range wireless protocols in other examples, such as WLAN communication protocols (e.g., IEEE 802.11), Bluetooth (e.g., IEEE 802.15.1), or one or more of the Communication Access for Land Mobiles (CALM) wireless communication protocols and air interfaces. The vehicle-to-vehicle (V2V) transmissions between the short-range communication systems 212 and 222 may be sent via DSRC, Bluetooth, satellite, GSM infrared, IEEE 802.11, WiMAX, RFID, and/or any suitable wireless communication media, standards, and protocols. In certain systems, short-range communication systems 212 and 222 may include specialized hardware installed in vehicles 210 and 220 (e.g., transceivers, antennas, etc.), while in other examples the communication systems 212 and 222 may be implemented using existing vehicle hardware components (e.g., radio and satellite equipment, navigation computers) or may be implemented by software running on the mobile devices 215 and 225 of drivers and passengers within the vehicles 210 and 220.


The range of V2V communications between vehicle communication systems 212 and 222 may depend on the wireless communication standards and protocols used, the transmission/reception hardware (e.g., transceivers, power sources, antennas), and other factors. Short-range V2V communications may range from just a few feet to many miles, and different types of driving behaviors may be determined depending on the range of the V2V communications. For example, V2V communications ranging only a few feet may be sufficient for a driving analysis computing device 101 in one vehicle to determine that another vehicle is tailgating or cut-off the vehicle, whereas longer communications may allow the device 101 to determine additional types of driving behaviors (e.g., vehicle spacing, yielding, defensive avoidance, proper response to a safety hazard, etc.).


V2V communications also may include vehicle-to-infrastructure (V2I) communications, such as transmissions from vehicles to non-vehicle receiving devices, for example, toll booths, rail road crossings, and road-side traffic monitoring devices. Certain V2V communication systems may periodically broadcast data from a vehicle 210 to any other vehicle, or other infrastructure device capable of receiving the communication, within the range of the vehicle's transmission capabilities. For example, a vehicle 210 may periodically broadcast (e.g., every 0.1 second, every 0.5 seconds, every second, every 5 seconds, etc.) certain vehicle operation data via its short-range communication system 212, regardless of whether or not any other vehicles or reception devices are in range. In other examples, a vehicle communication system 212 may first detect nearby vehicles and receiving devices, and may initialize communication with each by performing a handshaking transaction before beginning to transmit its vehicle operation data to the other vehicles and/or devices.


The types of vehicle operational data, or vehicle driving data, transmitted by vehicles 210 and 220 may depend on the protocols and standards used for the V2V communication, the range of communications, the autonomous driving system, and other factors. In certain examples, vehicles 210 and 220 may periodically broadcast corresponding sets of similar vehicle driving data, such as the location (which may include an absolute location in GPS coordinates or other coordinate systems, and/or a relative location with respect to another vehicle or a fixed point), speed, and direction of travel. In certain examples, the nodes in a V2V communication system (e.g., vehicles and other reception devices) may use internal clocks with synchronized time signals, and may send transmission times within V2V communications, so that the receiver may calculate its distance from the transmitting node based on the difference between the transmission time and the reception time. The state or usage of the vehicle's 210 controls and instruments may also be transmitted, for example, whether the vehicle is accelerating, braking, turning, and by how much, and/or which of the vehicle's instruments are currently activated by the driver (e.g., head lights, turn signals, hazard lights, cruise control, 4-wheel drive, traction control, etc.). Vehicle warnings such as a detection by the vehicle's 210 internal systems that the vehicle is skidding, that an impact has occurred, or that the vehicle's airbags have been deployed, also may be transmitted in V2V communications. In various other examples, any data collected by any vehicle sensors 211 and 221 potentially may be transmitted via V2V communication to other nearby vehicles or infrastructure devices receiving V2V communications from communication systems 212 and 222. Further, additional vehicle driving data not from the vehicle's sensors (e.g., vehicle make/model/year information, driver insurance information, driving route information, vehicle maintenance information, driver scores, etc.) may be collected from other data sources, such as a driver's or passenger's mobile device 215 or 225, driving analysis server 250, and/or another external computer system 230, and transmitted using V2V communications to nearby vehicles and other receiving devices using communication systems 212 and 222.


As shown in FIG. 2, the data collected by vehicle sensors 211 and 221 also may be transmitted to a driving analysis server 250, and one or more additional external servers and devices via telematics devices 213 and 223. Telematics devices 213 and 223 may be computing devices containing many or all of the hardware/software components as the computing device 101 depicted in FIG. 1. As discussed above, the telematics devices 213 and 223 may receive vehicle operation data and driving data from vehicle sensors 211 and 221, and may transmit the data to one or more external computer systems (e.g., driving analysis server 250 of an insurance company, financial institution, or other entity) over a wireless transmission network. Telematics devices 213 and 223 also may be configured to detect or determine additional types of data relating to real-time driving and the condition of the vehicles 210 and 220. In certain embodiments, the telematics devices 213 and 223 may contain or may be integral with one or more of the vehicle sensors 211 and 221. The telematics devices 213 and 223 also may store the type of their respective vehicles 210 and 220, for example, the make, model, trim (or sub-model), year, and/or engine specifications, as well as other information such as vehicle owner or driver information, insurance information, and financing information for the vehicles 210 and 220.


In the example shown in FIG. 2, telematics devices 213 and 223 may receive vehicle driving data from vehicle sensors 211 and 221, and may transmit the data to a driving analysis server 250. However, in other examples, one or more of the vehicle sensors 211 and 221 may be configured to transmit data directly to a driving analysis server 250 without using a telematics device. For instance, telematics devices 213 and 223 may be configured to receive and transmit data from certain vehicle sensors 211 and 221, while other sensors may be configured to directly transmit data to a driving analysis server 250 without using the telematics device 216. Thus, telematics devices 213 and 223 may be optional in certain embodiments.


In certain embodiments, vehicle sensors, vehicle OBD, and/or vehicle communication systems, may collect and/or transmit data pertaining to autonomous driving of the vehicles. In autonomous driving, the vehicle fulfills all or part of the driving without being piloted by a human. An autonomous car can be also referred to as a driverless car, self-driving car, or robot car. For example, in autonomous driving, a vehicle control computer 217, 227 may be configured to operate all or some aspects of the vehicle driving, including but not limited to acceleration, deceleration, steering, and/or route navigation. A vehicle with an autonomous driving capability may sense its surroundings using the vehicle sensors 221, 221 and/or receive inputs regarding control of the vehicle from the vehicle communications systems, including but not limited to short range communication systems 212, Telematics 213, or other vehicle communication system.


In certain embodiments, vehicle operational data can be collected from vehicles engaged in an autonomous droning relationship. As used herein, vehicles engaged in an autonomous droning relationship include where a vehicle engages in at least partial autonomous driving based on following the driving of a lead or pilot vehicle. In other words, in autonomous droning, the driving of the “drone” car is automated based at least in part on the driving of a lead vehicle. The lead vehicle can be a vehicle which the drone is following. For example, referring to FIGS. 6A and 6B, the driving of vehicle 520 may be automated based on the driving of vehicle 510. In such example, vehicle 510 can be referred to as the lead or pilot vehicle. For example, the acceleration, deceleration, braking, steering, and other operations of the drone vehicle 520 can be synchronized with that of the lead vehicle 510. Also, as shown by example in FIG. 6A, autonomous droning can be used to form caravans or platoons of one or more vehicles 530, 520 following the driving of the lead vehicle 510 and/or a vehicle which the drone is following. The drivers of the drone vehicle can, for example, pay attention to matters other than driving when the vehicle is engaged in an autonomous droning relationship. The lead vehicle can be driven manually, autonomously, or partially autonomously. Vehicles can engage in an autonomous droning relationship using systems of the vehicle alone or in cooperation with systems of other vehicles. For example, a drone vehicle may rely on vehicle sensors 211 and the vehicle control computer 217 to automate driving based on a lead vehicle it is following. In addition, a drone vehicle may use information received from vehicle communication systems, including V2V systems, of the lead vehicle to automate its driving. In addition, a drone vehicle may use both vehicle sensors 211 and information from other vehicles, including a lead vehicle, to automate its driving. Vehicle operational data collected from vehicles engaged in an autonomous droning relationship can include all types of information and data described herein.


In certain embodiments, mobile computing devices 215 and 225 within the vehicles 210 and 220 may be used to collect vehicle driving data and/or to receive vehicle driving data from sensors 211 and 221, and then to transmit the vehicle driving data to the driving analysis server 250 and other external computing devices. Mobile computing devices 215 and 225 may be, for example, mobile phones, personal digital assistants (PDAs), or tablet computers of the drivers or passengers of vehicles 210 and 220. Software applications executing on mobile devices 215 and 225 may be configured to detect certain driving data independently and/or may communicate with vehicle sensors 211 and 221 to receive additional driving data. For example, mobile devices 215 and 225 equipped with GPS functionality may determine vehicle location, speed, direction and other basic driving data without needing to communicate with the vehicle sensors 211 or 221, or any vehicle system. In other examples, software on the mobile devices 215 and 225 may be configured to receive some or all of the driving data collected by vehicle sensors 211 and 221. Mobile computing devices 215 and 225 may also be involved with aspects of autonomous driving, including receiving, collecting, and transmitting vehicle operational data regarding autonomous driving and autonomous driving relationship between multiple vehicles.


When mobile computing devices 215 and 225 within the vehicles 210 and 220 are used to detect vehicle driving data and/or to receive vehicle driving data from vehicles 211 and 221, the mobile computing devices 215 and 225 may store, analyze, and/or transmit the vehicle driving data to one or more other devices. For example, mobile computing devices 215 and 225 may transmit vehicle driving data directly to one or more driving analysis servers 250, and thus may be used in conjunction with or instead of telematics devices 213 and 223. Additionally, mobile computing devices 215 and 225 may be configured to perform the V2V communications described above, by establishing connections and transmitting/receiving vehicle driving data to and from other nearby vehicles. Thus, mobile computing devices 215 and 225 may be used in conjunction with or instead of short-range communication systems 212 and 222 in some examples. In addition, mobile computing devices 215 and 225 may be used in conjunction with the vehicle control computers 217 and 227 for purposes of autonomous driving. Moreover, the processing components of the mobile computing devices 215 and 225 may be used to analyze vehicle driving data, determine driving characteristics, determine properties related to rewards and/or aspects of vehicle insurance, and perform other related functions. Therefore, in certain embodiments, mobile computing devices 215 and 225 may be used in conjunction with, or in place of, the driving analysis computers 214 and 224.


Vehicles 210 and 220 may include driving analysis computers 214 and 224, which may be separate computing devices or may be integrated into one or more other components within the vehicles 210 and 220, such as the short-range communication systems 212 and 222, telematics devices 213 and 223, or the internal computing systems of vehicles 210 and 220. As discussed above, driving analysis computers 214 and 224 also may be implemented by computing devices independent from the vehicles 210 and 220, such as mobile computing devices 215 and 225 of the drivers or passengers, or one or more separate computer systems 230 (e.g., a user's home or office computer). In any of these examples, the driving analysis computers 214 and 224 may contain some or all of the hardware/software components as the computing device 101 depicted in FIG. 1. Further, in certain implementations, the functionality of the driving analysis computers, such as storing and analyzing vehicle driving data, determining driving characteristics, and determining rewards and/or aspects of insurance policies, may be performed in a central driving analysis server 250 rather than by individual vehicles 210 and 220. In such implementations, the vehicles 210 and 220 might only collect and transmit vehicle driving data to a driving analysis server 250, and thus the vehicle-based driving analysis computers 214 and 224 may be optional.


Driving analysis computers 214 and 224 may be implemented in hardware and/or software configured to receive vehicle driving data from vehicle sensors 211 and 221, short-range communication systems 212 and 222, telematics devices 213 and 223, vehicle control computer 217 and 227 and/or other driving data sources. Vehicle sensors/OBDs 211 and 221, short-range communication systems 212 and 222, telematics devices 213 and 223, vehicle control computer 217 and 227 and/or other driving data sources can be referred to herein individually or collectively as a vehicle data acquiring component. The driving analysis computer 214, 224 may comprise an electronic receiver to interface with the vehicle data acquiring components to receive the collected data. After receiving, via the electronic receiver, the vehicle driving data from, for example a vehicle data acquiring component, the driving analysis computers 214 and 224 may perform a set of functions to analyze the driving data, determine driving characteristics, and determine properties related to driver rewards and/or vehicle insurance. For example, the driving analysis computers 214 and 224 may include one or more driving characteristic algorithms, which may be executed by software running on generic or specialized hardware within the driving analysis computers. The driving analysis computer 214 in a first vehicle 210 may use the vehicle driving data received from that vehicle's sensors 211, along with vehicle driving data for other nearby vehicles received via the short-range communication system 212, to determine driving characteristics and determine properties related to rewards and/or vehicle insurance applicable to the first vehicle 210 and the other nearby vehicles. Within the driving analysis computer 214, a vehicle insurance property reward function may use the results of the driving analysis performed by the computer 214 to determine/adjust a property of an insurance policy for a driver of a vehicle 210 or other vehicles based on determined driving characteristics. Further descriptions and examples of the algorithms, functions, and analyses that may be executed by the driving analysis computers 214 and 224 are described below in reference to FIGS. 3 and 4.


The system 200 also may include a driving analysis server 250, containing some or all of the hardware/software components as the computing device 101 depicted in FIG. 1. The driving analysis server 250 may include hardware, software, and network components to receive vehicle operational data/driving data from one or more vehicles 210 and 220, and other data sources. The driving analysis server 250 may include a driving data and driver data database 252 and driving analysis computer 251 to respectively store and analyze driving data received from vehicles and other data sources. The driving analysis server 250 may initiate communication with and/or retrieve driving data from vehicles 210 and 220 wirelessly via telematics devices 213 and 223, mobile devices 215 and 225, or by way of separate computing systems (e.g., computer 230) over one or more computer networks (e.g., the Internet). Additionally, the driving analysis server 250 may receive additional data relevant to driving characteristic or other determinations from other non-vehicle data sources, such as external traffic databases containing traffic data (e.g., amounts of traffic, average driving speed, traffic speed distribution, and numbers and types of accidents, etc.) at various times and locations, external weather databases containing weather data (e.g., rain, snow, sleet, and hail amounts, temperatures, wind, road conditions, visibility, etc.) at various times and locations, and other external data sources containing driving hazard data (e.g., road hazards, traffic accidents, downed trees, power outages, road construction zones, school zones, and natural disasters, etc.), and insurance company databases containing insurance data (e.g., driver score, coverage amount, deductible amount, premium amount, insured status) for the vehicle, driver, and/or other nearby vehicles and drivers.


Data stored in the driving data database 252 may be organized in any of several different manners. For example, a table in database 252 may contain all of the vehicle operation data for a specific vehicle 210, similar to a vehicle event log. Other tables in the database 252 may store certain types of data for multiple vehicles. For instance, tables may store specific data sets, including correlations related to the vehicle for use in determining driving characteristics and/or properties related to driver rewards and/or vehicle insurance.


The driving analysis computer 251 within the driving analysis server 250 may be configured to retrieve data from the driving data and driver score database 252, or may receive driving data directly from vehicles 210 and 220 or other data sources, and may perform driving data analyses, driving characteristic determinations, driver reward and/or vehicle insurance determinations, and other related functions. The functions performed by the driving analysis computer 251 may be similar to those of driving analysis computers 214 and 224, and further descriptions and examples of the algorithms, functions, and analyses that may be executed by the driving analysis computer 251 are described below in reference to FIGS. 3 and 4.


In various examples, the driving data analyses, driving characteristic determinations, and driving reward and/or insurance property determinations may be performed entirely in the driving analysis computer 251 of the driving analysis server 250 (in which case driving analysis computers 214 and 224 need not be implemented in vehicles 210 and 220), or may be performed entirely in the vehicle-based driving analysis computers 214 and 224 (in which case the driving analysis computer 251 and/or the driving analysis server 250 need not be implemented). In other examples, certain driving data analyses may be performed by vehicle-based driving analysis computers 214 and 224, while other driving data analyses are performed by the driving analysis computer 251 at the driving analysis server 250. For example, a vehicle-based driving analysis computer 214 may continuously receive and analyze driving data from nearby vehicles to determine certain driving characteristics (e.g., aspects related to vehicle drafting, aspects related to autonomous driving, etc.) so that large amounts of driving data need not be transmitted to the driving analysis server 250. However, for example, after driving characteristic is determined by the vehicle-based driving analysis computer 214, the characteristic may be transmitted to the server 250, and the driving analysis computer 251 may determine if a driver reward and insurance vehicle property should be updated based on the determined driving characteristic.



FIG. 3 is a flow diagram illustrating an example method of determining a property of an insurance policy based on analysis of vehicle operational data of vehicles engaged in an autonomous droning relationship. This example method may be performed by one or more computing devices in a driving analysis system, such as vehicle-based driving analysis computers 214 and 224, a driving analysis computer 251 of a driving analysis server 250, user mobile computing devices 215 and 225, and/or other computer systems.


The steps shown in FIG. 3 describe performing an analysis of vehicle operational data to determine driving characteristics of vehicles engaged in an autonomous droning relationship and determining a property of an insurance policy based on the determined property. In step 301, vehicle operational data may be received from a first vehicle 210. As described above, a driving analysis computer 214 may receive and store vehicle driving data from a vehicle data acquiring component, including but not limited to the vehicle's internal computer systems and any combination of the vehicle's sensors 211 and/or communication systems. The data received in step 301 may include, for example, an identifier that the vehicle is engaged in autonomous driving and/or an autonomous driving relationship with another vehicle. The data received in step 301 may include, for example, the location, speed, and direction of the vehicle 210, object proximity data from the vehicle's external cameras and proximity sensors, and data from the vehicle's various systems used to determine if the vehicle 210 is braking, accelerated, or turning, etc., and to determine the status of the vehicle's user-operated controls (e.g., head lights, turn signals, hazard lights, radio, phone, etc.), along with any other data collected by vehicle sensors 211 or data received from a nearby vehicle.


In step 302, the vehicle operational data is analyzed to determine whether the vehicle is engaged in an autonomous droning relationship with another vehicle. For example, the driving data may include an identifier which indicates that the vehicle is engaged in an autonomous droning relationship. In addition, for example, a driving analysis computer 214 in a first vehicle 210 may compare the driving data (e.g., location, speed, direction) from its own vehicle sensors 211 (received in step 301) with the corresponding driving data (e.g., location, speed, direction) from a nearby vehicle 220. Based on the relative locations, speeds, and directions of travel of vehicles 210 and 220, the driving analysis computer 214 may determine that the vehicles are engaged in an autonomous driving relationship. In an embodiment, the driving data of the nearby vehicle can be collected by the data acquiring component of the first vehicle 210 via, for example, vehicle V2V. In an embodiment, the driving data of the nearby vehicle can be received from the nearby vehicle directly. In an embodiment, the driving data of the nearby vehicle can be determined from vehicle sensors 211 of the first vehicle.


In step 303, the vehicle driving data received in steps 301 may be analyzed, and driving characteristics of the autonomous droning relationship may be determined for the vehicles 210 and 220 based on the driving data. A driving characteristic of the autonomous droning relationship may include any number actions or events performed by the vehicles in the relationship or driving conditions or circumstances impacting the autonomous droning relationship. For example, a characteristic of an autonomous droning relationship can include identification of a lead vehicle, identification of the drone vehicle, an amount of time or distance a vehicle is the lead vehicle, an amount of time or distance a vehicle is a drone vehicle, that the lead vehicle is engaged in manual or autonomous driving, the number of vehicles engaged in the autonomous driving relationship, a weather condition, and/or a driver safety rating. The driving characteristic of the autonomous droning relationship can be determined by, for example, identifying the pertinent information from the received data, including actions or events performed by the vehicle or nearby vehicles. For example, the driving data may include a data point that identifies a vehicle as a lead vehicle in an autonomous driving relationship. In addition, for example the driving data may include time marked data from when a vehicle begins droning to when it completes droning. The driving characteristic can also be calculated using selected driving data. For example, if the driving data includes an identifier for when the vehicle is engaged in autonomous driving as a drone and time marked vehicle speed, the distance in which the vehicle is engaged in the autonomous driving relationship as a drone can be calculated. Example algorithms using time marked driving data are included in US Publications Number 2013/0073112 which is hereby incorporated by reference herein in its entirety. The driving characteristic can also be identified by comparing the driving data associated with a first vehicle with the corresponding driving data of a nearby vehicle. Based on information from, for example, the relative locations, speeds, and directions of travel of vehicles 210 and 220, the driving analysis computer 214 may determine a driving characteristic of the autonomous droning relationship involving the two vehicles.


In step 304, a property of an insurance policy may be determined using the characteristic of the autonomous droning relationship. The property of an insurance policy can include any of a number of aspects of a vehicle insurance policy. For example, a property of an insurance policy can include a premium, a deductible, a coverage term, a coverage amount, or other attribute of an insurance policy. In various embodiments, the property can be determined in accordance with rules set forth by the insurance provider. For example, the property of the vehicle insurance policy may change depending upon the characteristic of the autonomous droning relationship. For example, if the characteristic is that a first vehicle is a lead vehicle in an autonomous droning relationship, the deductible of an insurance policy of the first vehicle may increase in accordance with the number of vehicles engaged as drones. In addition, for example, a premium may decrease if the characteristic is that a first vehicle is engaged as a drone vehicle with a nearby vehicle which is piloted by a driver with a strong safety record and/or good driver safety rating.


In addition, in various embodiments, step 304 can include an additional step of determining an autonomous droning insurance factor using at least one of the vehicle operational data and the characteristic of the autonomous droning relationship and determine a property of an insurance policy using the autonomous droning insurance factor. As used herein, the autonomous droning insurance factor can include a variable used in an algorithm for calculating a property of an insurance policy. For example, an autonomous droning insurance factor can include a ratio of an amount of time a first vehicle is engaged in an autonomous droning relationship versus a total amount of time the first vehicle has driven over a period of time and/or a ratio of an amount of distance the first vehicle is engaged in an autonomous droning relationship versus a total distance the first vehicle has driven over a period of time. The factor can be used in an algorithm for calculating an insurance premium by, for example, assessing different rates for the periods of time or distance which the vehicle is engaged in autonomous driving.


In addition, in various embodiments, information pertaining to an insurance policy of a second vehicle can be used in determining the property of the first insurance policy of the first vehicle. For example, if the driving data includes a coverage amount of a vehicle engaged in the autonomous droning relationship and the coverage amount does not exceed a threshold level, the deductible of the first vehicle can increase.


In step 305, the driving analysis computer can adjust or cause to adjust the insurance policy based on the determined property. In various embodiments, the adjustment can occur during the coverage term and/or prior to the subsequent coverage term. In addition, the policy holder may be notified of the adjustment. Alternatively, the adjustment can come in the form of a reward to be received by the party to which it is allocated. For example, the system may cause the remunerative payment to be debited from a bank account associated the payor (e.g., a vehicle insurance company) and credited to the account of the policy holder of the insurance policy. As such, the paying party (e.g., the policy holder) may enjoy the benefit of the remunerative payment.


As shown in FIG. 3, a single vehicle-based driving analysis computer 214 may receive driving data for a first vehicle 210 (step 301), including driving data received from V2V communications including driving data for one or more other vehicles, may determine from the data whether the vehicle is engaged in an autonomous driving relationship (step 302), and may determine a characteristic of an autonomous droning relationship (step 303), determine a property of an insurance policy based on the characteristic (step 304), and adjust the insurance policy based on the determined property (step 305). However, other driving analysis computers and/or other computing devices may be used to some or all of the steps and functionality described above in reference to FIG. 3. For example, any of steps 301-305 may be performed by a user's mobile device 215 or 225 within the vehicles 210 or 220. These mobile devices 215 or 225, or another computing device 230, may execute software configured to perform similar functionality in place of the driving analysis computers 214 and 224. Additionally, some or all of the driving analysis functionality described in reference to FIG. 3 may be performed by a driving analysis computer 251 at a non-vehicle based driving analysis server 250. For example, vehicles 210 and 220 may be configured to transmit their own vehicle sensor data, and/or the V2V communications data received from other nearby vehicles, to a central driving analysis server 250 via telematics devices 213 and 223. In this example, the driving analysis computer 251 of the server 250 may perform the data analysis of steps 302 through 305.



FIG. 5 is a flow diagram of an illustrative method of awarding (e.g., crediting) one or more user accounts with a reward amount based on involvement in an autonomous droning relationship. The illustrative method steps may be performed by one or more computing devices in an autonomous droning reward system, including, but not limited to, a vehicle-based driving analysis computers 214 and 224, a driving analysis computer 251 of a driving analysis server 250, user mobile computing devices 215 and 225, and/or other computer systems.


Regarding step 502, an autonomous droning reward system may comprise an autonomous droning relationship determinator 226B configured to receive driving data indicating a lead vehicle 220 in an autonomous droning relationship with at least one following vehicle 210. The autonomous droning relationship determinator 226B may determine/identify when a vehicle is in an autonomous droning relationship. In some examples, a vehicle's data acquiring component may collect vehicle operation data and use some of that data to generate driving data. In other examples, vehicle operation data may be used interchangeably with driving data. In any event, the autonomous droning relationship determinator 226B may use the driving data to determine when one or more vehicles are in an autonomous droning relationship.


As explained above in accordance with various aspects of the disclosure, driving characteristic of an autonomous droning relationship can be determined by, for example, identifying pertinent information in the received data, including actions or events performed by vehicles 210, 220 or driving conditions or circumstances impacting the autonomous droning relationship. For example, the driving data may include an express data point/message/packet that identifies a vehicle (e.g., by its VIN number or other identifier) as a lead vehicle 220 in an autonomous driving relationship. In addition, the driving data may include a time-stamped message indicating that a following vehicle has started droning until when it completed droning. From such data, the distance traveled when in an autonomous droning relationship may be calculated. Similarly, the amount of time in the autonomous droning relationship may be calculated.


Furthermore, in some examples, vehicle driving data may be analyzed, and the driving characteristic of the autonomous droning relationship may be determined for the involved vehicles 210 and 220. For example, a characteristic of an autonomous droning relationship may include identification of a lead vehicle, identification of the drone vehicle, an amount of time or distance a vehicle is the lead vehicle, an amount of time or distance a vehicle is a drone vehicle, that the lead vehicle is engaged in manual or autonomous driving, the number of vehicles engaged in the autonomous droning relationship, a weather condition, a driver safety rating, and/or other characteristics. The driving characteristic can also be identified by comparing the driving data of a following vehicle 210 with driving data of a nearby vehicle 220. Based on information from, for example, the relative locations, speeds, and directions of travel of the vehicles 210, 220, autonomous droning relationship determinators 226A, 226B may determine that the two vehicles are in an autonomous droning relationship. Moreover, the autonomous droning relationship determinators 226A, 226B, in coordination with other computing devices 214, 224 at the vehicle, may determine the total number of vehicles engaged in the autonomous droning relationship. For example, three vehicles may be involved in an autonomous droning relationship, as illustrated in FIG. 6A.


While systems already exist for autonomous vehicles, such as the self-driving car by GOOGLE™, the spirit of this disclosure is not limited to just autonomous self-driving cars. For example, the lead vehicle 220 may be a completely autonomous vehicle, semi-autonomous vehicle, or a manual human-driven vehicle. Depending on the capabilities of the lead vehicle 220, the vehicle may be equipped with the appropriate sensors 221 and other electronic components to enable the automation/semi-automation, as is already known in the relevant art of autonomous/semi-autonomous vehicles. Similarly, a following (drone) vehicle 210 may be equipped with the appropriate hardware and software to operate as an autonomous vehicle, semi-autonomous vehicle, or a manually-driven vehicle. In contrast, however, in some examples, the following drone vehicle 210 may be equipped with less hardware and/or software than a lead vehicle 220 because to some extent, the following drone vehicle 210 may rely upon the lead vehicle 220 to provide guidance and commands for controlling the speed, acceleration, braking, cornering, route, and other operation of the following vehicle 210. For example, a following drone vehicle 210 may transmit data to the lead vehicle 220 using its short-range wireless communications system 212, and rely upon long-range wireless communication capabilities 222 of the lead vehicle to forward the data to the appropriate final destination. At least one benefit of such an arrangement is that the cost/price of a following drone vehicle 210 may be less than that of other vehicles (e.g., lead vehicle 220) due to reduced complexity and reduce hardware and/or software requirements.


In addition, the integrity of collected vehicle driving data may be validated by comparing, e.g., by a driving analysis computer, the driving data (e.g., location, speed, direction) from one vehicle's sensors 211 with corresponding driving data from a nearby vehicle 220. Based on the relative locations, speeds, and directions of travel of vehicles 210 and 220, an autonomous droning relationship determinator 226B of a driving analysis computer 214 may determine that the vehicles are engaged in an autonomous driving relationship. In one example, driving data of the nearby vehicle can be collected by a data acquiring component of a following/drone vehicle 210 via, for example, vehicle V2V. In one example, the driving data of the nearby vehicle may be directly received from the nearby vehicle.


In accordance with various aspects of the disclosure, in one example, an autonomous droning relationship determinator 226 may receive (in step 502) driving data indicating that a lead vehicle 220 is in an autonomous droning relationship with at least one following vehicle 210. The autonomous droning relationship determinator 226A may be executing in a driving analysis computer 224 at the lead vehicle 220. Alternatively, the autonomous droning relationship determinator 226B may be executing in a driving analysis computer 214 at one of the following/drone vehicles 210. In yet another alternative, the autonomous droning relationship determinator 226 may be distributed across the lead vehicle 220 and one or more following vehicles 210 such that the determination of whether a vehicle is involved in an autonomous droning relationship involves coordination/communication between more than one autonomous droning relationship determinators 226A, 226B.


Continuing with the preceding example, the autonomous droning relationship determinator 226 may determine, in this particular example, that an autonomous droning relationship persists so long as the lead vehicle 220 transmits commands (e.g., vehicle control commands) to the at least one following vehicle 210 on at least a regular interval (e.g., every second, every half second, every couple seconds, or other interval of time). The commands, upon receipt at an interface to a vehicle control computer 217 at the following vehicle 210, may cause the following drone vehicle 210 to control/alter its driving characteristics pursuant the commands. Examples of changes in driving characteristics include, but are not limited to, changing the speed of the drone vehicle 210, changing the direction of travel of the drone vehicle 210, and/or turning on/off headlights of the drone vehicle 210. Moreover, some commands may be to cause the drone vehicle 210 to hold steady its speed; although such commands might not expressly alter the driving characteristics of the vehicle 210, the command does cause the vehicle control computer 217 to control the vehicle 210. A person having ordinary skill in the art, after review of the entirety disclosed herein, will appreciate that the disclosure contemplates other examples of commands that alter the driving characteristics of a drone vehicle.


The lead vehicle 220 may generate commands and transmit them to one or more following vehicles 210. In one example, the driving analysis computer 224 of the lead vehicle 220 may take driving data as input and generate commands based on that driving data. For example, if the OBD interface 221 of the lead vehicle 220 indicates that the lead vehicle 220 is traveling at a speed of 45 mph in a particular cardinal direction (e.g., North), that information may be used to generate commands that a following drone vehicle 210 may process to adjust its speed and/or direction. A person having ordinary skill in the art, after review of the entirety disclosed herein, will appreciate that systems for autonomous droning are known and that aspects of these systems are incorporated into the illustrative system shown in FIG. 2.


In the preceding example, the autonomous droning relationship determinator 226B of a following vehicle 210 monitors for incoming commands on a regular interval, and if valid incoming commands are detected, the autonomous droning relationship determinator 226B determines that the following vehicle 210 is in an autonomous droning relationship. Likewise, a lead vehicle 220, using its short-range communications system (see FIG. 2), may regularly send/broadcast commands to one or more following drone vehicles; and while it is continuing to send such commands, may notify the autonomous droning relationship determinator 226A that the lead vehicle 220 is in an autonomous droning relationship.


In another example, the autonomous droning relationship determinator 226 may determine that the vehicle 210, 220 has entered into an autonomous droning relationship when an electronic handshake between the vehicles is sent, acknowledged, and/or established. With the handshake protocol completed, an autonomous droning relationship is established. Meanwhile, the autonomous droning relationship determinator 226 may monitor communications between its vehicle and the nearby vehicle to detect/determine when the autonomous droning relationship has terminated. The driving data (and optionally other data) collected while in the autonomous droning relationship may be sent to an autonomous droning reward calculator 702 for calculation of a possible reward amount.


In step 504, at or after termination of the autonomous droning relationship, an autonomous droning reward calculator 702 may calculate a reward amount. Alternatively, the autonomous droning reward calculator 702 may maintain a running count of the reward amount during the autonomous droning relationship. The running count may allow the autonomous droning reward calculator 702 to display in real-time, e.g., on a display screen located in the vehicle, the pending reward amount.


Referring to FIG. 7, an autonomous droning reward calculator 702 may calculate the reward amount using at least driving data. The driving data may include, but is not limited to, at least one of: speed of the lead vehicle, location of the lead vehicle, amount of time spent in the autonomous droning relationship, amount of distance traveled in the autonomous droning relationship, and/or other characteristics. Other examples of driving data include the time of day when the autonomous droning relationship exists, the following distance between the lead vehicle and a following drone vehicle (or in the case of multiple following vehicles, such as illustrated in FIG. 6, the distance between a vehicle and the vehicle immediately behind it), the vehicle's characteristics (e.g., vehicle type—SUV, sports car, sedan, convertible, etc., vehicle's turning radius, vehicle's maximum speed, vehicle time to accelerate from 0-60 mph, and other characteristics tied to the specific vehicle), driving risk characteristics/profile of the driver/operator of the lead vehicle 220, and other characteristics. For example, assuming numerous systems exist for automated droning, “System A” may use hardware and/or software algorithms different from that of competing “System B.” As a result, each of the systems may react differently when used in the real world, and as such, will earn a driving risk characteristic/profile commensurate with the amount of risk associated with the particular system. Therefore, an autonomous droning reward calculator 702 may calculate a higher reward amount for “System A” than for “System B,” in some examples. In another example where the lead vehicle 220 is manually driven, either the entire time of a trip or just a portion of the trip, by a human, the driving risk characteristics may be tied to the driver's driving risk score, which may take into account factors such as number of accidents, speeding ticket violations, number of submitted insurance claims, and other factors known for use in calculating a driving risk score.


In addition, referring to FIG. 7, other information may also be inputted into the autonomous droning reward calculator 702 for consideration in calculating a reward amount. For example, the congestion level (e.g., traffic) on a roadway where the lead vehicle 220 is being followed by drone vehicles may be an input to determining a reward amount. Other examples include, but are not limited to, the weather conditions the roadway, historical occurrences of incidents (e.g., vehicular accidents) on the roadway, and other factors related to the environment/surroundings in which the vehicle is operated. For example, the autonomous droning reward calculator 702 may adjust the reward amount based on the congestion level on the roadway being high. In one example, the autonomous droning reward calculator 702 may increase reward amounts during rush hour traffic to encourage vehicles 210 equipped with a vehicle control computer 217 to drone/follow a lead vehicle, thus alleviate traffic congestion. Congestion levels may be divided, in one example, into categories of high, medium, and low based on the whether the travel time through a particular roadway falls into the upper ⅓, middle ⅓, or lower ⅓ of possible travels times historically logged on that roadway. Likewise, weather conditions may play a role in determining the reward amount. In a fog situation, reward amounts may be very high in order to encourage both drone/following vehicles 210 and appropriate lead vehicles 220 (e.g., those vehicles that have been rated safe for operation in fog conditions, for example, because they are equipped with fog lights or other fog-ready equipment) to enter into an autonomous droning relationship.


The aforementioned external information from the preceding example of FIG. 7 may be stored at and retrieved from various data sources, such as an external traffic databases containing traffic data (e.g., amounts of traffic, average driving speed, traffic speed distribution, and numbers and types of accidents, etc.) about various times and locations, external weather databases containing weather data (e.g., rain, snow, sleet, and hail amounts, temperatures, wind, road conditions, visibility, etc.) at various times and locations, and other external data sources containing driving hazard data (e.g., road hazards, traffic accidents, downed trees, power outages, road construction zones, school zones, and natural disasters, etc.), and insurance company databases containing insurance data (e.g., driver score, coverage amount, deductible amount, premium amount, insured status) for the vehicle, driver, and/or other nearby vehicles and drivers. The aforementioned external information may, in some examples, be wirelessly transmitted from a remote server and/or database to the vehicle 220 for consideration by the autonomous droning reward calculator 702. As explained earlier, drone vehicles 210 may leverage additional hardware and/or software capabilities of a lead vehicle 220 to gain access to the aforementioned information, when desired. For example, a lead vehicle 220 may receive, through its long-range communications circuitry 222 (or mobile phone 225), the information and forward it to drone vehicles 210 via their short-range communications 212 systems. As such, the vehicles 210, 220 may input the information into their autonomous droning reward calculator 702 for consideration.



FIG. 7 shows the autonomous droning reward calculator 702 receiving numerous inputs and outputting an autonomous droning reward amount. In some examples, the autonomous droning reward calculator 702 may be an application-specific integrated circuit (ASIC) designed to perform the functionality described herein. In other examples, autonomous droning reward calculator 702 may use a processing unit (e.g., comprising a computer processor, such as an Intel™ x86 microprocessor or other special-purpose processors) and computer-executable instructions stored in a memory to cause a driving analysis computer 214 to perform the steps described herein.


Referring to step 504, the autonomous droning reward calculator 702 may calculate a reward amount using one or more of the numerous driving data and other information described herein. In one example, the autonomous droning reward calculator 702 may increase the reward amount as the amount of distance traveled while in the autonomous droning relationship increases. In another example, the autonomous droning reward calculator 702 may increase the reward amount if the speed of the lead vehicle stays within a threshold speed range while in the autonomous droning relationship. In other words, if the vehicle speed of the caravan is kept within a safe range, a higher reward amount may be awarded. In yet another example, the autonomous droning reward calculator 702 may increase the reward amount if the location of the lead vehicle stays within a geographic region while in the autonomous droning relationship. In another example, the autonomous droning reward calculator 702 may increase the reward amount as the amount of time spent while in the autonomous droning relationship increases.



FIG. 8 show a graphic illustration of the preceding example in which the autonomous droning reward calculator 702 increases the reward amount as the amount of time spent in while in the autonomous droning relationship increases. Graph 800A shows that in some examples, the autonomous droning reward calculator 702 may be configured to provide a reward amount in a linear proportion to the amount of time. As such, the graph is a straight line. Meanwhile, graph 800B shows an example where the autonomous droning reward calculator 702 may be configured to provide a reward amount as a function of the amount of time, but with a predefined maximum reward amount. As such, the relationship is defined such that once a drone vehicle 210 has been in an autonomous droning relationship for an extended period of time, the portion of the reward amount, which may be due to a measure of the amount of time in an autonomous droning relationship, remains generally the same value. Along these same lines, the autonomous droning reward calculator 702 may be configured with varying formulas/functions that establish relationships (e.g., linear, exponential, logarithmic, etc.) between the various inputs (e.g., such as those illustrated in FIG. 7) and the reward amount.


While in some examples the reward amount is a dollar amount (e.g., $10, $5.75, $0.98, etc.), in other examples, the reward amount may be a non-dollar amount. For example, the autonomous droning reward amount may be a percentage value (e.g., a 1.2% discount, a 0.5% discount, a 2.0% upcharge) applied with respect to a vehicle insurance policy, either prospectively and/or retrospectively to the beginning of the current policy term. In yet another example, the reward amount may be a voucher or coupon for goods and/or services.


In step 506, with the autonomous droning reward amount having been calculated in step 504, the reward amount is credited to the appropriate party, such as an account associated with the lead vehicle or a drone vehicle. In some examples, the lead vehicle 220 may receive the entire reward amount. In other examples, the reward amount may be divided and credited to all of the vehicle involved in the autonomous droning relationship. For example, a reward amount of $10 may be calculated in step 504, and $7 of it may be credited (in step 506) to a bank checking account associated with the lead vehicle 220, $2 of it may be credited (in step 506) to a bank checking account associated with the drone vehicle 210, and the remaining $1 may be credited to one or more government or private entities involved in the autonomous droning reward system. Examples of such government and/or private entities might include an insurance company providing the vehicle insurance policy to the lead vehicle, an insurance company providing vehicle insurance policy to the drone vehicle 210, a State's department of transportation, or other entity.


With reference to FIG. 9, an account processor 902 executing an account module 254 may serve to credit and/or debit, as appropriate, a rewards amount amongst various entities. In one example, a vehicle insurance company 904 insuring the lead vehicle 220 may provide the full reward amount. In other examples, the vehicle insurance company(s) 906 corresponding to one or more drone vehicles 210 following the lead vehicle may be responsible for crediting the autonomous droning reward amount to the lead vehicle's account 256A. In some hybrid-approaches, multiple insurance companies, including that of the lead vehicle and drone vehicle(s), may share the responsibility for awarding the autonomous droning reward amount. In other examples, the autonomous droning reward amount awarded to the lead vehicle's account may be provided by a drone vehicle (e.g., the account 256B associated with the drone vehicle may be debited accordingly).


Numerous different types of accounts are contemplated in accordance with various aspects of the disclosure. In some examples, a user's checking account at a financial institution may be credited and debited based on the autonomous droning reward amount calculated using at least driving data. The financial institution may be in communication, either directly or indirectly, with the autonomous droning reward system. As a result, usage-based rewards may be implemented in an autonomous droning reward system. In other examples, a vehicle's insurance policy account at a vehicle insurance company may be credited and debited based on the autonomous droning reward amount calculated using at least driving data. The insurance company may be in communication, either directly or indirectly, with the autonomous droning reward system.


Aspects of the disclosure pertain to analyzing driving data to determine driving characteristics associated with vehicle drafting relationships, determine driver rewards based on such characteristics, and allocating the driving rewards. Vehicle drafting, also referred to as slipstreaming, as used herein pertains to when two or more vehicles move in alignment to reduce the overall effect of drag due to exploiting the lead vehicle's slipstream. As a vehicle moves, high pressure is generated in front and low pressure behind. The difference in pressure creates a force known as drag. Drag force can account for a large amount of fuel consumption, especially at high speeds. The concept of drafting or slipstreaming is to utilize the regions of reduced pressure behind moving vehicles to lessen the oncoming drag experienced by the following vehicle, thereby reducing fuel consumption of the following vehicle. Slipstreaming can also reduce the fuel consumption of the lead vehicle but to a lesser extent than the lead vehicle. However, in order to realize appreciable savings in fuel consumption by engaging in drafting, vehicles may have to follow at distances less than legal following distances for manual driving. For example, a University of California-San Diego study determined that drag reduction was negligible at a distance of 288 feet at 65 mph and, therefore, show no increase in fuel savings. See Duan et al., Effects of Highway Slipstreaming on California Gas Consumption, MAE 171B-Mechanical Lab II, (Jun. 13, 2007) (hereinafter referred to as the “Duan Report”), which is hereby incorporated by reference herein in its entirety and a copy of which is filed in the Information Disclosure Statement concurrently filed with this application. However, autonomous driving and driving in autonomous droning relationships can allow for following at close distances in a safe manner. Accordingly, fuel savings associated with drafting can be realized utilizing autonomous driving and autonomous droning.



FIG. 4 is a flow diagram illustrating an example method of determining a drafting property associated with a first vehicle engaged in a drafting relationship and allocating a drafting reward based on the drafting property. This example method may be performed by one or more computing devices in a driving analysis system, such as vehicle-based driving analysis computers 214 and 224, a driving analysis computer 251 of a driving analysis server 250, user mobile computing devices 215 and 225, and/or other computer systems.


The steps shown in FIG. 4 describe performing analysis of vehicle operational data to determine drafting characteristics of vehicles engaged in a drafting relationship, determine a drafting property based on the drafting characteristic, and allocate a drafting award based on the drafting property. In step 401, vehicle operational data may be received from a first vehicle 210. As described above, a driving analysis computer 214 may receive and store vehicle driving data from a vehicle data acquiring component, including but not limited to the vehicle's internal computer systems and any combination of the vehicle's sensors 211 and/or communication systems. The data received in step 401 may include, for example, an identifier that the vehicle is engaged in a drafting relationship. The data received in step 401 may also include, for example, the location, speed, and direction of the vehicle 210, object proximity data from the vehicle's external cameras and proximity sensors, and data from the vehicle's various systems used to determine if the vehicle 210 is braking, accelerating, or turning, etc., and to determine the status of the vehicle's user-operated controls (e.g., head lights, turn signals, hazard lights, radio, phone, etc.), vehicle type, along with any other data collected by vehicle sensors 211 or data received from a nearby vehicle.


In step 402, the vehicle operational data is analyzed to determine whether the vehicle is engaged in a drafting relationship with another vehicle. For example, the driving data may include an identifier which indicates that the vehicle is engaged in a drafting relationship. In addition, for example, a driving analysis computer 214 in a first vehicle 210 may compare the driving data (e.g., location, speed, vehicle type) from its own vehicle sensors 211 (received in step 401) with the corresponding driving data (e.g., location, speed, vehicle) from a nearby vehicle 220. Based on the relative locations, speeds, and vehicle types of vehicles 210 and 220, the driving analysis computer 214 may determine that the vehicles are engaged in a drafting relationship. In an embodiment, the driving data of the nearby vehicle can be collected by the data acquiring component of the first vehicle 210 via, for example, vehicle V2V. In an embodiment, the driving data of the nearby vehicle can be received from the nearby vehicle directly. In an embodiment, the driving data of the nearby vehicle can be determined from vehicle sensors 211 of the first vehicle.


In step 403, the vehicle driving data received in steps 401 may be analyzed, and driving characteristics of the drafting relationship may be determined for the vehicles 210 and 220 based on the driving data. A driving characteristic of the drafting relationship may include any number actions or events performed by the vehicles in the relationship or driving conditions or circumstances impacting the drafting relationship. For example, a characteristic of a drafting relationship can include identification of a lead vehicle, identification of the following vehicle, vehicle spacing, vehicle speeds, vehicle types, miles-per-gallon readings, a weather condition, whether the vehicles are engaged in autonomous driving, and/or a driver safety rating. The driving characteristic of the drafting relationship can be determined by, for example, identifying the pertinent information from the received data, including actions or events performed by the vehicle or nearby vehicles. For example, the driving data may include a data point that identifies a vehicle as a lead vehicle in a drafting relationship. The driving characteristic can be determined by calculations using selected driver data. For example, vehicle spacing can be calculated from vehicle location data by, for example, identifying the locations of the respective vehicles and performing an analysis to determine the distance between the vehicles. The driving characteristic can also be identified by comparing the driving data associated with a first vehicle with the corresponding driving data of the nearby vehicle. Based on information from, for example, the relative locations, speeds, and vehicle types of vehicles 210 and 220, the driving analysis computer 214 may determine a driving characteristic of the drafting relationship involving the two vehicles. As used herein, a driving characteristic of a drafting relationship may also be referred to as a drafting characteristic.


In step 404, a drafting property is determined based on the drafting characteristic. As used herein, a drafting property can include any measure of efficiency realized from the drafting relationship. For example, a drafting property can include a drafting fuel savings rate, a percentage increase in miles-per-gallon, and a drafting fuel savings amount. In various embodiments, the drafting property can be determined from algorithms or identified from predetermined correlations based on the drafting characteristics. For example, where the drafting property is a drafting fuel savings amount, the drafting fuel savings amount can be determined by subtracting from a miles-per-gallon reading taken while the vehicle is drafting a miles-per-gallon reading taken while the vehicle is not drafting to determine a drafting miles-per-gallon savings rate and dividing a distance traveled by the vehicle by the first drafting miles-per-gallon savings rate. In addition, for example, the drafting property can be determined from a correlation of vehicle spacing vs. an increase in miles-per-gallon for specific vehicle. Such correlations can be based on, for example, vehicle lead type, vehicle following type, vehicle spacing, and vehicle speed. Such correlations can be prepared by, for example, methods including modeling and or wind tunnel experiments. Example methods for preparing such correlations include those disclosed in the Duan Report. Correlations for use in determining drafting properties can be stored on a driving data database 252.


In step 405, the driving analysis computer can allocate a drafting reward based on the drafting property. For example, drafting reward (e.g., remunerative payment) can include at least one of a cash payment, a carbon credit, a fuel credit, a tax credit, a rebate, and at least a portion of a drafting fuel savings amount associated with the first vehicle. The drafting amount can be allocated to an associated with the vehicle or vehicle owner to which the allocation is made. In addition, the reward can be allocated at regular intervals while the vehicles are engaged in the drafting relationship. For example, a reward can be allocated every second, minute, or mile in which the vehicles are engaged in the drafting relationship. The drafting reward can be determined according to rules or agreements established between drivers and/or owners of the vehicles. The system can further be configured to cause the reward to be received by the party to which it is allocated. For example, if the reward is a payment, the system can cause the payment to be debited from a bank account associated with the following vehicle and credited to the account of the lead vehicle.


In various embodiments, the steps of 401-405 can be applied to a caravan of vehicles engaged in a drafting relationship. In an embodiment, a following vehicle allocates a reward to the vehicle it is following, which may in turn be following another vehicle. In such examples, rewards provided by a vehicle can be offset by reward received by the vehicle.


In various embodiments, the system can notify a vehicle engaged in a drafting relationship of a drafting characteristic, a drafting property, and/or drafting reward. In an embodiment, the notification can be sent to the owner of the vehicle or entity which is authorized to receive the notifications. The notifications can be made in real-time. In addition, the notifications can be sent to, for example, a mobile phone or other device associated with the vehicle, owner, or authorized party.


In some examples, certain functionality may be performed in vehicle-based driving analysis computers 214 and 224, while other functionality may be performed by the driving analysis computer 251 at the driving analysis server 250. For instance, vehicle-based driving analysis computers 214 and 224 may continuously receive and analyze driving data for their own vehicles 210 and 220 and nearby vehicles (via V2V communications), and may determine driving characteristics (e.g., autonomous driving characteristics, drafting characteristics, etc.) for their own vehicles 210 and 220 and/or the other nearby vehicles. After the vehicle-based driving analysis computers 214 and 224 have determined the driving characteristics, indications of these characteristics may be transmitted to the server 250 so that the driving analysis computer 251 can perform the insurance policy determinations and adjustments based on the driving characteristics. Thus, the driving analysis server 250 may be optional in certain embodiments, and some or all of the driving analyses may be performed within the vehicles themselves.


While the aspects described herein have been discussed with respect to specific examples including various modes of carrying out aspects of the disclosure, those skilled in the art will appreciate that there are numerous variations and permutations of the above described systems and techniques that fall within the spirit and scope of the invention. In addition, where reference has been made in this disclosure to items in the figures, in some instances the alphabetic suffix (e.g., “A” or “B”) of a reference number has been omitted when it is desired to generally reference (e.g., “226”) the item without specificity as to which of the plurality of items corresponding to the same base reference number.

Claims
  • 1. An autonomous droning reward system comprising: a vehicle operational data collector, wherein the vehicle operational data collector is configured to collect driving data from at least one of vehicle sensors, vehicle On-Board diagnostics systems (OBD) systems, and vehicle-to-vehicle communication systems; andan autonomous droning reward device, comprising a processing unit comprising at least one computer processor;a memory unit; anda wireless receiver;wherein the autonomous droning reward device is configured to: receive, via the wireless receiver, the driving data;determine, from the driving data, whether a lead vehicle is in an autonomous droning relationship with a following vehicle, wherein the lead vehicle is in the autonomous droning relationship with the following vehicle so long as the lead vehicle transmits commands to the following vehicle and the following vehicle uses the commands to follow the lead vehicle and automate driving based on the driving of the lead vehicle;calculate a reward amount based on at least an amount of distance traveled while the lead vehicle is in the autonomous droning relationship with the following vehicle;determine an account associated with the lead vehicle; andcredit the account associated with the lead vehicle by the reward amount.
  • 2. The autonomous droning reward system of claim 1, wherein the following vehicle is a drone vehicle comprising: a short-range wireless communications circuitry configured to receive commands generated by the lead vehicle; andan interface to a vehicle control computer of the drone vehicle, wherein the commands are sent to the vehicle control computer through the interface to cause the drone vehicle to alter its driving characteristics.
  • 3. The autonomous droning reward system of claim 1, wherein the autonomous droning reward device is further configured to: adjust the reward amount based on at least one of congestion level on a roadway, weather conditions on the roadway, and historical occurrence of incidents on the roadway.
  • 4. The autonomous droning reward system of claim 1, wherein the autonomous droning reward device is further configured to: adjust the reward amount based on at least one of time of day when the lead vehicle is operated, vehicle characteristics of the lead vehicle, and driving risk characteristics of the lead vehicle.
  • 5. The autonomous droning reward system of claim 1, wherein the driving data comprises at least one of: speed of the lead vehicle, location of the lead vehicle, and amount of time spent in the autonomous droning relationship, wherein the autonomous droning reward device is further configured to: increase the reward amount as the amount of time spent in the autonomous droning relationship increases;increase the reward amount if the speed of the lead vehicle stays within a threshold speed range while in the autonomous droning relationship; andincrease the reward amount if the location of the lead vehicle stays within a predetermined geographic region while in the autonomous droning relationship.
  • 6. The autonomous droning reward system of claim 1, wherein the reward amount comprises a percentage discount of an insurance policy premium associated with the lead vehicle.
  • 7. The autonomous droning reward system of claim 1, further comprising: a short-range wireless communications circuitry configured to communicate commands between the lead vehicle and the following vehicle; anda long-range wireless communications circuitry configured to communicate the driving data.
  • 8. The autonomous droning reward system of claim 1, wherein the account is an auto insurance policy account, and wherein the reward amount is a dollar amount at least partially credited against a premium of the auto insurance policy account.
  • 9. The autonomous droning reward system of claim 1, wherein the autonomous droning reward device is further configured to: determine an account associated with one of the following vehicle; anddebit the account associated with the following vehicle by the reward amount.
  • 10. The autonomous droning reward system of claim 1, wherein the autonomous droning reward device is further configured to: determine an account associated with the following vehicle; andcredit the account associated with the following vehicle by no more than the reward amount.
  • 11. The autonomous droning reward system of claim 1, wherein the autonomous droning reward device is further configured to: apply a percentage discount to an insurance policy account of the following vehicle.
  • 12. The autonomous droning reward system of claim 1, further comprising: a display screen configured to display at least the reward amount, wherein the display screen is in at least one of the lead vehicle, the following vehicle, and a mobile phone.
  • 13. The autonomous droning reward system of claim 1, further comprising a driving analysis computing device configured to: determine a drafting property associated with the following vehicle; andcalculate a drafting reward based on the drafting property associated with the following vehicle; and credit the account associated with the lead vehicle by the drafting reward.
  • 14. A method comprising: receiving, by a wireless receiver of an autonomous droning reward device, driving data collected by a vehicle operational data collector, wherein the vehicle operational data collector is configured to collect the driving data from at least one of vehicle sensors, vehicle On-Board diagnostics systems (OBD) systems, and vehicle-to-vehicle communication systems;determining, by a processing unit of the autonomous droning reward device and from the driving data, whether a lead vehicle is in an autonomous droning relationship with a following vehicle, wherein the lead vehicle is in the autonomous droning relationship with the following vehicle so long as the lead vehicle transmits commands to the following vehicle such that the following vehicle uses the commands to follow the lead vehicle and automate driving based on the driving of the lead vehicle;calculating, by a processing unit of the autonomous droning reward device, a reward amount based on at least an amount of distance traveled while the lead vehicle is in the autonomous droning relationship with the following vehicle; andcrediting the reward amount to one or more accounts associated with at least one of the lead vehicle and the following vehicle.
  • 15. The method of claim 14, wherein the one or more accounts includes an auto insurance policy account, and wherein the reward amount is a dollar amount at least partially credited against a premium of the auto insurance policy account.
  • 16. The method of claim 14, further comprising: adjusting, by the autonomous droning reward device, the reward amount based on at least one of congestion level on a roadway traveled by the lead vehicle, weather conditions on the roadway, historical occurrence of incidents on the roadway, time of day when the lead vehicle is operated, vehicle characteristics of the lead vehicle, and driving risk characteristics of the lead vehicle.
  • 17. The method of claim 14, wherein the one or more accounts includes an auto insurance policy account, and wherein the reward amount is a dollar amount at least partially credited against a premium due on the auto insurance policy account.
  • 18. The method of claim 14, wherein the driving data comprises at least one of: speed of the lead vehicle, location of the lead vehicle, and amount of time spent in the autonomous droning relationship, the method further comprising at least one of: increasing the reward amount as the amount of time spent in the autonomous droning relationship increases;increasing the reward amount if the speed of the lead vehicle stays within a threshold speed range while in the autonomous droning relationship; andincreasing the reward amount if the location of the lead vehicle stays within a predetermined geographic region while in the autonomous droning relationship.
  • 19. The method of claim 14, wherein the reward amount comprises a percentage discount of an insurance policy associated with at least one of the lead vehicle and the following vehicle.
  • 20. The method of claim 14, further comprising determining a drafting property associated with the following vehicle, calculating a drafting reward based on the drafting property, and credit the drafting reward to an account associated with the lead vehicle.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/163,741 now U.S. Pat. No. 9,355,423 entitled “Reward System Related to a Vehicle-to-Vehicle Communication System” and filed on Jan. 24, 2014, which is incorporated by reference herein in its entirety.

US Referenced Citations (394)
Number Name Date Kind
83960 Heator Nov 1868 A
4119166 Ayotte et al. Oct 1978 A
4622636 Tachibana Nov 1986 A
4706072 Ikeyama Nov 1987 A
4926336 Yamada May 1990 A
5053964 Mister et al. Oct 1991 A
5270708 Kamishima Dec 1993 A
5295551 Sukonick Mar 1994 A
5430432 Camhi et al. Jul 1995 A
5465079 Bouchard et al. Nov 1995 A
5475387 Matsumoto Dec 1995 A
5572449 Tang et al. Nov 1996 A
5680122 Mio Oct 1997 A
5710565 Shirai et al. Jan 1998 A
5797134 McMillan et al. Aug 1998 A
5848373 DeLorme et al. Dec 1998 A
6026345 Shah et al. Feb 2000 A
6060989 Gehlot May 2000 A
6064970 McMillan et al. May 2000 A
6116369 King et al. Sep 2000 A
6128559 Saitou et al. Oct 2000 A
6186793 Brubaker Feb 2001 B1
6188950 Tsutsumi et al. Feb 2001 B1
6265978 Atlas Jul 2001 B1
6301530 Tamura Oct 2001 B1
6366207 Murphy Apr 2002 B1
6389351 Egawa et al. May 2002 B1
6415226 Kozak Jul 2002 B1
6502020 Lang Dec 2002 B2
6502035 Levine Dec 2002 B2
6647328 Walker Nov 2003 B2
6675094 Russell et al. Jan 2004 B2
6707378 MacNeille et al. Mar 2004 B2
6732024 Wilhelm Rekow et al. May 2004 B2
6780077 Baumgartner et al. Aug 2004 B2
6868386 Henderson et al. Mar 2005 B1
6931309 Phelan et al. Aug 2005 B2
6982635 Obradovich Jan 2006 B2
7054831 Koenig May 2006 B2
7116248 Lu et al. Oct 2006 B2
7133771 Nesbitt Nov 2006 B1
7186199 Baxter, Jr. Mar 2007 B1
7242112 Wolf et al. Jul 2007 B2
7286825 Shishido et al. Oct 2007 B2
7304589 Kagawa Dec 2007 B2
7315239 Cheng et al. Jan 2008 B2
7339483 Farmer Mar 2008 B1
7353111 Takahashi et al. Apr 2008 B2
7356516 Richey et al. Apr 2008 B2
7366892 Spaur et al. Apr 2008 B2
7389198 Dimitriadis Jun 2008 B1
7546206 Miller et al. Jun 2009 B1
7610210 Helitzer et al. Oct 2009 B2
7650211 Wang et al. Jan 2010 B2
7657370 Nagase et al. Feb 2010 B2
7657441 Richey et al. Feb 2010 B2
7660725 Wahlbin Feb 2010 B2
7664589 Etori et al. Feb 2010 B2
7739087 Qiu Jun 2010 B2
7805321 Wahlbin Sep 2010 B2
7818187 Wahlbin Oct 2010 B2
7821421 Tamir et al. Oct 2010 B2
7822384 Anschutz et al. Oct 2010 B2
7937278 Cripe et al. May 2011 B1
7966118 Kade Jun 2011 B2
7991629 Gay et al. Aug 2011 B2
8031062 Smith Oct 2011 B2
8065169 Oldham et al. Nov 2011 B1
8078349 Prada Gomez et al. Dec 2011 B1
8078382 Sugano et al. Dec 2011 B2
8086523 Palmer Dec 2011 B1
8090598 Bauer et al. Jan 2012 B2
8108083 Kameyama Jan 2012 B2
8139109 Schmiedel et al. Mar 2012 B2
8145393 Foster et al. Mar 2012 B2
8152589 Bowen et al. Apr 2012 B2
8160809 Farwell et al. Apr 2012 B2
8180655 Hopkins, III May 2012 B1
8195394 Zhu et al. Jun 2012 B1
8229618 Tolstedt et al. Jul 2012 B2
8280308 Anschutz et al. Oct 2012 B2
8280752 Cripe et al. Oct 2012 B1
8290701 Mason et al. Oct 2012 B2
8314718 Muthaiah et al. Nov 2012 B2
8326473 Simpson et al. Dec 2012 B2
8335607 Gatten et al. Dec 2012 B2
8352112 Mudalige Jan 2013 B2
8407139 Palmer Mar 2013 B1
8457827 Ferguson et al. Jun 2013 B1
8457892 Aso et al. Jun 2013 B2
8538785 Coleman et al. Sep 2013 B2
8549318 White et al. Oct 2013 B2
8554468 Bullock Oct 2013 B1
8566126 Hopkins, III Oct 2013 B1
8577703 McClellan et al. Nov 2013 B2
8595037 Hyde et al. Nov 2013 B1
8606512 Bogovich et al. Dec 2013 B1
8620575 Vogt et al. Dec 2013 B2
8620693 Schumann, Jr. Dec 2013 B1
8639535 Kazenas Jan 2014 B1
8659436 Ngo Feb 2014 B2
8676466 Mudalige Mar 2014 B2
8686844 Wine Apr 2014 B1
8718861 Montemerlo et al. May 2014 B1
8725311 Breed May 2014 B1
8750306 Yousefi et al. Jun 2014 B2
8757309 Schmitt et al. Jun 2014 B2
8781669 Teller et al. Jul 2014 B1
8798841 Nickolaou et al. Aug 2014 B1
8799036 Christensen et al. Aug 2014 B1
8812330 Cripe et al. Aug 2014 B1
8930269 He et al. Jan 2015 B2
8949016 Ferguson et al. Feb 2015 B1
8954226 Binion et al. Feb 2015 B1
8996303 Bogovich et al. Mar 2015 B1
9020751 Bogovich et al. Apr 2015 B1
9046374 Ricci Jun 2015 B2
9063543 An et al. Jun 2015 B2
9079587 Rupp et al. Jul 2015 B1
9141582 Brinkmann et al. Sep 2015 B1
9188985 Hobbs et al. Nov 2015 B1
9216737 Zhu et al. Dec 2015 B1
9262787 Binion et al. Feb 2016 B2
9330571 Ferguson et al. May 2016 B2
9355423 Slusar May 2016 B1
9355546 Kim et al. May 2016 B2
9373149 Abhyanker Jun 2016 B2
9384148 Muttik et al. Jul 2016 B2
9390451 Slusar Jul 2016 B1
9433843 Morlock Sep 2016 B2
9457814 Kim et al. Oct 2016 B2
9495874 Zhu et al. Nov 2016 B1
9618359 Weast et al. Apr 2017 B2
9648107 Penilla May 2017 B1
9679487 Hayward Jun 2017 B1
9715711 Konrardy et al. Jul 2017 B1
9739627 Chintakindi Aug 2017 B1
9758039 Hannon Sep 2017 B2
9767516 Konrardy et al. Sep 2017 B1
9792656 Konrardy et al. Oct 2017 B1
9801580 Iizuka et al. Oct 2017 B2
9851214 Chintakindi Dec 2017 B1
9858621 Konrardy Jan 2018 B1
9865019 Bogovich et al. Jan 2018 B2
9870649 Fields Jan 2018 B1
9904900 Cao Feb 2018 B2
9922374 Vose Mar 2018 B1
9928432 Sathyanarayana et al. Mar 2018 B1
9931062 Cavallaro et al. Apr 2018 B2
9932033 Slusar et al. Apr 2018 B2
9946334 Pala et al. Apr 2018 B2
9953300 Connor Apr 2018 B2
9972054 Konrardy May 2018 B1
10012510 Denaro Jul 2018 B2
10037578 Bogovich et al. Jul 2018 B2
10037580 Bogovich et al. Jul 2018 B2
10046618 Kirsch et al. Aug 2018 B2
10078871 Sanchez et al. Sep 2018 B2
10096038 Ramirez et al. Oct 2018 B2
10127737 Manzella et al. Nov 2018 B1
20010020902 Tamura Sep 2001 A1
20010039509 Dar et al. Nov 2001 A1
20020022920 Straub Feb 2002 A1
20020024464 Kovell et al. Feb 2002 A1
20020095249 Lang Jul 2002 A1
20020111725 Burge Aug 2002 A1
20020111738 Iwami et al. Aug 2002 A1
20020120396 Boies et al. Aug 2002 A1
20020128882 Nakagawa et al. Sep 2002 A1
20020178033 Yoshioka et al. Nov 2002 A1
20030043045 Yasushi et al. Mar 2003 A1
20030128107 Wilkerson Jul 2003 A1
20030182165 Kato et al. Sep 2003 A1
20030187704 Hashiguchi et al. Oct 2003 A1
20040021583 Lau et al. Feb 2004 A1
20040036601 Obradovich Feb 2004 A1
20040054452 Bjorkman Mar 2004 A1
20040068555 Satou Apr 2004 A1
20040098464 Koch et al. May 2004 A1
20040103006 Wahlbin et al. May 2004 A1
20040103010 Wahlbin et al. May 2004 A1
20040128613 Sinisi Jul 2004 A1
20040142678 Krasner Jul 2004 A1
20040153362 Bauer et al. Aug 2004 A1
20040236476 Chowdhary Nov 2004 A1
20040254698 Hubbard et al. Dec 2004 A1
20040260579 Tremiti Dec 2004 A1
20050091175 Farmer Apr 2005 A9
20050107951 Brulle-Drews et al. May 2005 A1
20050137757 Phelan et al. Jun 2005 A1
20050174217 Basir et al. Aug 2005 A1
20050228622 Jacobi Oct 2005 A1
20050256638 Takahashi et al. Nov 2005 A1
20050264404 Franczyk et al. Dec 2005 A1
20050273263 Egami et al. Dec 2005 A1
20050283503 Hancock et al. Dec 2005 A1
20050288046 Zhao et al. Dec 2005 A1
20060006990 Obradovich Jan 2006 A1
20060053038 Warren et al. Mar 2006 A1
20060055565 Kawamata et al. Mar 2006 A1
20060095301 Gay May 2006 A1
20060129313 Becker et al. Jun 2006 A1
20060129445 McCallum Jun 2006 A1
20060161341 Haegebarth et al. Jul 2006 A1
20060184321 Kawakami et al. Aug 2006 A1
20060206623 Gipps et al. Sep 2006 A1
20060221328 Rouly Oct 2006 A1
20060247852 Kortge et al. Nov 2006 A1
20060253307 Warren et al. Nov 2006 A1
20070021910 Iwami et al. Jan 2007 A1
20070027583 Tamir et al. Feb 2007 A1
20070032929 Yoshioka et al. Feb 2007 A1
20070136107 Maguire et al. Jun 2007 A1
20070167147 Krasner et al. Jul 2007 A1
20070182532 Lengning et al. Aug 2007 A1
20070216521 Guensler et al. Sep 2007 A1
20070256499 Pelecanos et al. Nov 2007 A1
20070257815 Gunderson et al. Nov 2007 A1
20070282638 Surovy Dec 2007 A1
20080004802 Horvitz Jan 2008 A1
20080013789 Shima et al. Jan 2008 A1
20080033637 Kuhlman et al. Feb 2008 A1
20080059007 Whittaker et al. Mar 2008 A1
20080059351 Richey et al. Mar 2008 A1
20080091309 Walker Apr 2008 A1
20080091490 Abrahams et al. Apr 2008 A1
20080114542 Nambata et al. May 2008 A1
20080148409 Ampunan et al. Jun 2008 A1
20080161987 Breed Jul 2008 A1
20080167757 Kanevsky et al. Jul 2008 A1
20080243558 Gupte Oct 2008 A1
20080258890 Follmer et al. Oct 2008 A1
20080288406 Seguin et al. Nov 2008 A1
20080319602 McClellan et al. Dec 2008 A1
20090012703 Aso et al. Jan 2009 A1
20090024419 McClellan et al. Jan 2009 A1
20090063201 Nowotarski et al. Mar 2009 A1
20090079839 Fischer et al. Mar 2009 A1
20090115638 Shankwitz et al. May 2009 A1
20090140887 Breed et al. Jun 2009 A1
20090312945 Sakamoto et al. Dec 2009 A1
20100023183 Huang et al. Jan 2010 A1
20100030586 Taylor et al. Feb 2010 A1
20100042314 Vogt et al. Feb 2010 A1
20100131300 Collopy et al. May 2010 A1
20100131304 Collopy et al. May 2010 A1
20100131307 Collopy et al. May 2010 A1
20100138244 Basir Jun 2010 A1
20100211270 Chin et al. Aug 2010 A1
20100250087 Sauter Sep 2010 A1
20100256852 Mudalige Oct 2010 A1
20100280751 Breed Nov 2010 A1
20100302371 Abrams Dec 2010 A1
20100324775 Kermani et al. Dec 2010 A1
20100332131 Horvitz et al. Dec 2010 A1
20110029170 Hyde et al. Feb 2011 A1
20110043350 Ben David Feb 2011 A1
20110071718 Norris et al. Mar 2011 A1
20110077028 Wilkes, III et al. Mar 2011 A1
20110161119 Collins Jun 2011 A1
20110173015 Chapman et al. Jul 2011 A1
20110202305 Willis et al. Aug 2011 A1
20110210867 Benedikt Sep 2011 A1
20120034876 Nakamura et al. Feb 2012 A1
20120053808 Arai et al. Mar 2012 A1
20120072243 Collins et al. Mar 2012 A1
20120083960 Zhu et al. Apr 2012 A1
20120101660 Hattori Apr 2012 A1
20120109418 Lorber May 2012 A1
20120123641 Ferrin et al. May 2012 A1
20120123806 Schumann, Jr. et al. May 2012 A1
20120173290 Collins et al. Jul 2012 A1
20120197669 Kote et al. Aug 2012 A1
20120209505 Breed et al. Aug 2012 A1
20120290146 Dedes et al. Nov 2012 A1
20120295592 Peirce Nov 2012 A1
20130006469 Green et al. Jan 2013 A1
20130006674 Bowne et al. Jan 2013 A1
20130006675 Bowne et al. Jan 2013 A1
20130013179 Lection et al. Jan 2013 A1
20130018549 Kobana et al. Jan 2013 A1
20130030606 Mudalige et al. Jan 2013 A1
20130037650 Heppe Feb 2013 A1
20130046559 Coleman et al. Feb 2013 A1
20130052614 Mollicone et al. Feb 2013 A1
20130066511 Switkes et al. Mar 2013 A1
20130073321 Hofmann et al. Mar 2013 A1
20130090821 Abboud et al. Apr 2013 A1
20130116920 Cavalcante et al. May 2013 A1
20130131906 Green et al. May 2013 A1
20130144657 Ricci Jun 2013 A1
20130147638 Ricci Jun 2013 A1
20130166325 Ganapathy et al. Jun 2013 A1
20130179198 Bowne et al. Jul 2013 A1
20130198737 Ricci Aug 2013 A1
20130198802 Ricci Aug 2013 A1
20130200991 Ricci et al. Aug 2013 A1
20130203400 Ricci Aug 2013 A1
20130204645 Lehman et al. Aug 2013 A1
20130212659 Maher et al. Aug 2013 A1
20130218603 Hagelstein et al. Aug 2013 A1
20130218604 Hagelstein et al. Aug 2013 A1
20130226441 Horita Aug 2013 A1
20130250933 Yousefi et al. Sep 2013 A1
20130253809 Jones et al. Sep 2013 A1
20130261944 Koshizen Oct 2013 A1
20130297097 Fischer et al. Nov 2013 A1
20130304513 Hyde et al. Nov 2013 A1
20130304514 Hyde et al. Nov 2013 A1
20130311002 Isaac Nov 2013 A1
20140074512 Hare et al. Mar 2014 A1
20140080098 Price Mar 2014 A1
20140088855 Ferguson Mar 2014 A1
20140108058 Bourne et al. Apr 2014 A1
20140113619 Tibbitts et al. Apr 2014 A1
20140136414 Abhyanker May 2014 A1
20140139341 Green et al. May 2014 A1
20140156133 Cullinane et al. Jun 2014 A1
20140156134 Cullinane et al. Jun 2014 A1
20140172221 Solyom et al. Jun 2014 A1
20140172290 Prokhorov et al. Jun 2014 A1
20140180723 Cote et al. Jun 2014 A1
20140210644 Breed Jul 2014 A1
20140257871 Christensen et al. Sep 2014 A1
20140257873 Hayward et al. Sep 2014 A1
20140266795 Tseng et al. Sep 2014 A1
20140272810 Fields et al. Sep 2014 A1
20140278586 Sanchez et al. Sep 2014 A1
20140300458 Bennett Oct 2014 A1
20140300494 Tseng et al. Oct 2014 A1
20140303827 Dolgov et al. Oct 2014 A1
20140310075 Ricci Oct 2014 A1
20140333468 Zhu et al. Nov 2014 A1
20140350970 Schumann, Jr. et al. Nov 2014 A1
20140358413 Trombley et al. Dec 2014 A1
20140379384 Duncan et al. Dec 2014 A1
20140379385 Duncan et al. Dec 2014 A1
20140380264 Misra et al. Dec 2014 A1
20150019266 Stempora Jan 2015 A1
20150025917 Stempora Jan 2015 A1
20150057931 Pivonka Feb 2015 A1
20150081404 Basir Mar 2015 A1
20150088334 Bowers et al. Mar 2015 A1
20150088550 Bowers et al. Mar 2015 A1
20150112543 Binion et al. Apr 2015 A1
20150112730 Binion et al. Apr 2015 A1
20150112731 Binion et al. Apr 2015 A1
20150112733 Baker et al. Apr 2015 A1
20150120124 Bartels et al. Apr 2015 A1
20150134181 Ollis May 2015 A1
20150142244 You et al. May 2015 A1
20150149017 Allard et al. May 2015 A1
20150149019 Pilutti et al. May 2015 A1
20150158486 Healey et al. Jun 2015 A1
20150161738 Stempora Jun 2015 A1
20150166059 Ko Jun 2015 A1
20150166062 Johnson et al. Jun 2015 A1
20150166069 Engelman et al. Jun 2015 A1
20150170287 Tirone et al. Jun 2015 A1
20150175168 Hoye et al. Jun 2015 A1
20150179062 Ralston et al. Jun 2015 A1
20150187013 Adams et al. Jul 2015 A1
20150187014 Adams et al. Jul 2015 A1
20150187015 Adams et al. Jul 2015 A1
20150187019 Fernandes et al. Jul 2015 A1
20150194055 Maass Jul 2015 A1
20150217763 Reichel et al. Aug 2015 A1
20150242953 Suiter Aug 2015 A1
20150248131 Fairfield et al. Sep 2015 A1
20150254955 Fields et al. Sep 2015 A1
20150266455 Wilson Sep 2015 A1
20150294422 Carver et al. Oct 2015 A1
20160009291 Pallett et al. Jan 2016 A1
20160036558 Ibrahim et al. Feb 2016 A1
20160065116 Conger Mar 2016 A1
20160086285 Jordan Peters et al. Mar 2016 A1
20160086393 Collins et al. Mar 2016 A1
20160089954 Rojas Villanueva Mar 2016 A1
20160090097 Grube et al. Mar 2016 A1
20160096531 Hoye et al. Apr 2016 A1
20160163198 Dougherty Jun 2016 A1
20160167652 Slusar Jun 2016 A1
20160189303 Fuchs Jun 2016 A1
20170011465 Anastassov et al. Jan 2017 A1
20170021764 Adams et al. Jan 2017 A1
20170120929 Siddiqui et al. May 2017 A1
20170154636 Geiger et al. Jun 2017 A1
20170210288 Briggs et al. Jul 2017 A1
20170219364 Lathrop et al. Aug 2017 A1
20170221149 Hsu-Hoffman et al. Aug 2017 A1
20180202822 DeLizio Jul 2018 A1
20180251128 Penilla Sep 2018 A1
20180376357 Tavares Coutinho Dec 2018 A1
20190101649 Jensen Apr 2019 A1
Foreign Referenced Citations (20)
Number Date Country
101131588 Feb 2008 CN
102010001006 Jul 2011 DE
1296305 Mar 2003 EP
2293255 Mar 2011 EP
2471694 Jul 2012 EP
3303083 Apr 2018 EP
2001039090 May 2001 WO
2005108928 Nov 2005 WO
2007102405 Sep 2007 WO
2008067872 Jun 2008 WO
2008096376 Aug 2008 WO
2012014042 Feb 2012 WO
2012150591 Nov 2012 WO
2013012926 Jan 2013 WO
2013126582 Aug 2013 WO
2013160908 Oct 2013 WO
2014148975 Sep 2014 WO
2016028228 Feb 2016 WO
2016122881 Aug 2016 WO
2016200762 Dec 2016 WO
Non-Patent Literature Citations (159)
Entry
“Background on: Self-driving cars and insruance”, Auto Technology, Insurance Information Institute, Inc. (Year: 2018).
Autonomous Cars will Usher in Things We Never Saw Coming, Opnions, PC Magazine Digital Edition (Year: 2016).
Xu, Qing et al., “Vehicle-to-Vehicle Safety Messaging in DSRC”; 2004.
EE Herald webpage, “DSRC Packet Sniffer, a vehicle-to-vehicle communication technology is under demo”; www.eeherald.com/section/news/nw10000198.html; dated Nov. 22, 2008.
Bai, Fan et al., “Reliability Analysis of DSRC Wireless Communication for Vehicle Safety”; Sep. 2006.
Kotani, Kazuya et al., “Inter-Vehicle Communication Protocol for Cooperatively Capturing and Sharing” Intersection Video ; date unkown but believed to be before 2011.
BC Technology Webpage; “CarCom Intercom System”; www.bctechnologyltd.co.uk/clarson-intercom-system-brochure.htm; downloaded May 29, 2013.
Wolf Intercom webpage; “Wolf Intercom Systems”; http://wolfintercom.com/; downloaded May 29, 2013.
Telephonics Webpage; “Integrated Communication Systems Wired & Wireless Secure Intercommunications”; www.telephonics.com/netcom.asp; downloaded May 29, 2013.
Car-to-Car webpage; “Car-2-Car Communication”; www.car-to-car.org/index.php?id=8; downloaded May 29, 2013.
Cohda Wireless webpage; www.cohdawireless.com/default.html; downloaded May 29, 2013.
Eichler, Stephen et al., “Car-to-Car Communication” dated Oct. 2006.
Oki Webpage “OKI Develops World's First DSRC Inter-vehicle Communication Attachment for Mobile Phones to Help Pedestrian Safety” dated Jan. 8, 2009.
Zeng, X., Yin, K., and Ge, H., “Hazardous Driving Prediction System,” Submission to the Connected Vehicle Technology Challenge,Sep. 24, 2014, 20 pages.
Jan. 15, 2015—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/163,761.
Jan. 21, 2015—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/163,719.
Quad City Intersection Traffic Accident Study, Davenport-Rock Island-Moline Urbanized Area 1993 data, Bi-State Regional Commission, Mar. 1996; http://ntl.bts.gov/lib/000/300/338/00338.pdf; 78 pages; downloaded Apr. 8, 2008.
Geographic Information Systems Using CODES Linked Data (Crash Outcome Data Evaluation System), U.S. Department of Transportation National Highway Traffic Safety Administration, Apr. 2001; http://ntl.bts.gov/lib/11000/11100/11149/809-201.pdf; 44 pages; downloaded Apr. 8, 2008.
Final Report: What Value May Geographic Information Systems Add to the Art of Identifying Crash Countermeasures? John S. Miller, Senior Research Scientist, Virginia Transportation Research Council, Charlottesville, Virginia, Apr. 1999; http://www.virginiadot.org/vtrc/main/online_reports/pdf/99-r13.pdf; 44 pages; downloaded Apr. 8, 2008.
Mapping the Streets of the World, Hilmar Schmundt, Speigel Online, May 12, 2006 03:37 PM, High Technology; http://www.spiegel.de/international/spiegel/0,1518,druck-415848,00.html; 2 pages; downloaded Jun. 25, 2008.
Patents: At the forefront of technological innovation, Printed from the Teleatlas.com website, 2007; http://www.teleatlas.com/WhyTeleAtlas/Innovation/Patents/index.htm; 1 page; downloaded Jun. 25, 2008.
Digital Collection—Metadata View; Quad City Intersection Traffic Accident Study: 1993 Data; http://ntlsearch.bts.gov/tris/record/nt1/338.html; 2 pages; downloaded Jun. 25, 2008.
Advanced Tracking Technologies, Inc., Shadow Tracker Prov5 Track Detail Map, http://www.advantrack.com/map_pro_3.htm; 1 page; downloaded Jun. 25, 2008.
Advanced Tracking Technologies, Inc.; Track Playback; http://www.advantrack.com/Animated-Track-Playback.htm; 1 page; downloaded Jun. 25, 2008.
What is Geocoding?, http://www.trpc.org/programs/gis/geocode.htm; 5 pages; downloaded Jun. 25, 2008.
Logistics, Not Consumers, Best Early Market for Premium Traffic Information, Sep. 25, 2006; http://auto.ihs.com/news/2006/abi-premium-traffic.htm; 2 pages; downloaded Jun. 25, 2008.
Property/Casualty Insurance Gaining Position With Technology; Telematics, the use of Wireless communications and Global Positioning System (GPS) tracking, may soon change the way automobile insurance, both personal and commercial, is priced. Individual rating of a driver, to supplement class rating, now appears to be feasible.; http;//www.towersperrin.com/TILLINGHAST/publications/publications/emphasis/Emphasis_2005_3/Holderedge.pdf; 4 pages; downloaded Apr. 8, 2008.
IVOX's Driver Score; Personal Lines; Benefits to using IVOX DriverScore; http://www.ivosdata.com/personal_lines.html; 1 page; downloaded Jul. 25, 2008.
Group1 Software; Point-Level Geocoding Option Geocoding Enrichment Solution; http://www.g1.com/PDF/Product/PointLevelGeocode.pdf; 2 pages; downloaded Apr. 8, 2008.
Integrated Enterprise Geo-Spatial Technology—Insurance Risk Examples by Brady Foust, Ph.D., Howard Botts, Ph.D. and Margaret Miller, Ph.D., Jan. 27, 2006; http://www.directionsmag.com/printer.php?artcicle_id-2081; 2 pages; downloaded Jun. 25, 2008.
How the Discounts Work; www.SaveAsYouDrive.com; http://www.saveasyouddrive.com/page.asp?pageid=34&print=true; 2 pages; downloaded Jun. 25, 2008.
Jan. 29, 2016—(U.S.) Notice of Allowance and Fee(s) Due—U.S. Appl. No. 14/163,741.
Mar. 17, 2016—(U.S.) Notice of Allowance and Fee(s) Due—U.S. Appl. No. 14/163,761.
Apr. 7, 2016—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/163,719.
Mar. 18, 2016—(WO) International Search Report—App PCT/US2016/013204.
Mar. 18, 2016—(WO) Search Report and Written Opinion—App PCT/US2016/013204.
Light, Donald, “A Scenario: The End of Auto Insurance,” May 8, 2012, downloaded Nov. 11, 2013 from ww.celent.com/reports/scenario-end-auto-insurance, 2 pages.
“Driver Assistance Systems,” Robert Bosch GmbH, downloaded Oct. 27, 2013 from www.bosch-automotivetechnology.us/en_us/us/driving_comfort_1/driving_comfort_systems_for_passenger_cars_2/driver_assistance_systems_5/driver_assistan . . . 12 pages.
“Autonomous Car,” Wikipedia, the free encyclopedia, downloaded Nov. 11, 2013 from en.wikipedia.org/wiki/Autonomous_car#cite_ref-28, 20 pages.
“Schlaue Autos von A bis Z.” Encyclopedia, downloaded Oct. 27, 2013 from www.bester-beifahrer.de/startseite/lexikon/, 15 pages.
“Chassis Systems Control, Adaptive Cruise Control: More comfortable driving,” Robert Bosch GmbH, Brochure downloaded Oct. 26, 2013, 4 pages.
“Get Ready for Automated Cars,” Houston Chronicle, Sep. 11, 2012, downloaded Nov. 11, 2013, 1 page.
Levy, Steven, Salmon, Felix, Stokes, Jon, “Artificial Intelligence is Here. In Fact, It's All Around Us. But Its Nothing Like We Expected,” Jan. 2011, 14 pages.
“Driverless cars study: 1 in 5 would let computers do the driving,” Nov. 4, 2013, downloaded Dec. 19, 2013 from http://www.carinsurance.com/press/driverless-cars-survey-results.aspx, 2 pages.
Shladover, Steven E. “What if Cars Could Drive Themselves,” Access Magazine, University of California Transportation Center, UC Berkeley, Apr. 1, 2000, downloaded Dec. 19, 2013, 7 pages.
Kim, Mun Hyun, Dickerson, Julie, Kosko, Bart, “Fuzzy throttle and brake control for platoons of smart cars,” University of Southern California, Received May 1995, revised Aug. 1995, downloaded Dec. 19, 2013, 26 pages.
“A velocity control strategy for vehicular collision avoidance system,” Abstract downloaded on May 9, 2013 from ieeexplore.ieee.org/xpl/articleDetails.jsp?tp=arnumber=1626838contentType=Conference+Publications&queryText%3DA+velocity+control+strategy+for . . . , 1 page.
“Fuzzy system representation of car-following phenomena,” Abstract downloaded on May 9, 2013 from ieeexplore.ieee.org/xpl/articleDetails.jsp?tp=&arnumber=527798&contentType32 Conference+Publications&queryText%3DFuzzy+system+representation+of . . . , 1 page.
“Direct adaptive longitudinal control of vehicle platoons,” Abstract downloaded on May 9, 2013 from ieeexplore.ieee.org/xpl/articleDetails.jsp?tp=&arnumber=917908&contentType=Journals+%26+Magazines&queryText%3DDirect+adaptive+longitudinal+c . . . , 1 page.
Noguchi, Noboru, Will, Jeff, Reid, Joh, and Zhang, Qin, “Development of a master-slave robot system for farm operations,” Computers and Electronics in Agriculture 44 (2004), 19 pages.
“Project SARTRE (Safe Road Trains for the Environment),” Road Traffic Technology, downloaded on May 9, 2013 from www.roadtraffic-technology.com/projects/the-sartre-project/, 3 pages.
“A semi-autonomous tractor in an intelligent master-slave vehicle system,” Oct. 2010, vol. 3, Issue 4, pp. 263-269, downloaded Dec. 19, 2013 from http://link.springer.com/article/10.1007%2Fs11370-010-0071-6, 4 pages.
“Development of an intelligent master-slave system between agricultural vehicles,” Abstract downloaded on Dec. 19, 2013 from http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5548056&url=http%3A%2F%2Fi . . . ,1 page.
“A leader-follower formation flight control scheme for UAV helicopters,” Abstract downloaded on Dec. 19, 2013 from http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=4636116&url=http%3A%2F%2Fi . . . , 1 page.
Kurian, Bonny, “Auto-Insurance—Driving into the sunset?”, Tala Consultancy Services, 2013.
O'Brien, Christine, “Autonomous Vehicle Technology: Consideration for the Auto Insurance Industry”, University Transportation Resource Center (The 2nd Connected Vehicles Symposium, Rutgers University), Jun. 17, 2013.
Anderson, James M. et al., “Autonomous Vehicle Program: A Guide for Policymakers”, Rand Corporation: Transportation, Space, and Technology Program, 2014.
Marchant, Gary E. et al., “The Coming Collision Between Autonomous Vehicles and the Liability System”, Santa Clara Law Review (vol. 52: No. 4 (Article 6)), Dec. 17, 2012.
“The Munix Advantage”, AUMA, retrieved Apr. 8, 2014, <http://www.auma.ca/live/digitalAssets/71/71248_MUNIX_onepager.pdf>.
“The Use of Occupation and Education Factors in Automobile Insurance”, State of New Jersey: Department of Banking and Insurance, Apr. 2008.
“Preliminary Statement of Policy Concerning Automated Vehicles”, National Highway Traffic Safety Administration, retrieved Jun. 9, 2014.
Walker Smith, Bryant, “Summary of levels of Driving Automation for On-Road Vehicles”, Stanford Law School: The Center for Internet and Society, Dec. 18, 2013, <http://cyberlaw.stanford.edu/blog/2013/12/sae-levels-driving-automation>.
Auto Insurance Discounts, Liberty Mutual Insurance, downloaded from http://www.libertymutual.com/auto-insurance/auto-insurance-coverage/auto-insurance-discounts, Jan. 8, 2014, 2 pages.
Autonomous Vehicles Regulations, California Department of Motor Vehicles, 2011, downloaded from www.dmv.ca.gov/vr/autonomous/auto.htm, Jan. 2, 2014, 3 pages.
Sharma, Aroma, Autonomous Vehicle Conf Recap 2012: “Driving the Future: The Legal Implications of Autonomous Vehicles,” High Tech Law Institute, downloaded from law.scu.edu/hightech/autonomousvehicleconfrecap2012/, Jan. 2, 2014, 7 pages.
Strumpf, Dan, “Corporate News: Driverless Cars Face Issues of Liability”, Strumpf Dan, The Wall Street Journal Asia [Hong Kong ] Jan. 29, 2013: 19, downloaded from http://search.proquest.com.ezproxy, Jan. 8, 2014, 2 pages.
Lienert, Anita, Drivers Would Opt for Autonomous Cars to Save on Insurance, Study Finds: Published: Nov. 7, 2013, downloaded from www.edmunds.com/car-news/drivers-would-opt-for-autonomous-cars-to-save-on-insurance-study-finds.html on Jan. 2, 2014, 6 pages.
Neumann, Peter G. and Contributors, “Risks to the Public,” ACM SIGSOFT Software Engineering Notes, Jul. 2012 vol. 37 No. 4, pp. 20-29.
Kirkpatrick, Keith, “Legal issues with Robots,” Communications of the ACM, Nov. 2013, vol. 56 No. 11, pp. 17-19.
Ingolfo, Silvia, and Silva Souza, Vitor E. “Law and Adaptivity in Requirements Engineering,” SEAMS 2013, pp. 163-168.
O'Donnell, Anthony, “Prepare for Deep Auto Insurance Premium Drop Scenario, Celent Report Advises,” Insurance & Technology, May 8, 2012, downloaded from http://www.insurancetech.com/claims/prepare-for-deep-auto-insurance-premium/232901645?printer_friendly=this-page, Jan. 9, 2014, 3 pages.
Top issues: An annual report “The insurance industry in 2013; Strategy: Reshaping auto insurance”, vol. 5, 2013, 6 pages.
O'Donnell, Anthony, “Rapid Emergence of Driverless Cars Demands Creation of Legal Infrastructure, Stanford Scholar Says,” Insurance & Technology—Online, Jan. 3, 2013, downloaded from http: I I search.proquest.com .ezproxy.apollolibrary.com/ docview / 12 66 314 720 /fulltext/ 142 DA8916CC2 E861A14/ 11 ?accountid = 3 5812, Jan. 8, 2014, 2 pages.
“Self-driving cars: The next revolution” (kpmg.com | cargroup.org), 2012, 36 pages.
The autonomous car: The road to driverless driving, May 30, 2013, downloaded from analysis.telematicsupdate.com/v2x-safety/autonomous-car-road-driverless-driving on Jan. 2, 2014, 6 pages.
Ruquet, Mark E., “Who Insures a Driverless Car”? Property & Casualty 360, Oct. 1, 2012, downloaded from http:/ / search.proquest.com on Jan. 8, 2014, 2 pages.
Neil, Dan, “Who's Behind the Wheel? Nobody. The driverless car is coming. And we all should be glad it is,” Wall Street Journal (Online) [New York, N.Y] Sep. 24, 2012, downloaded from http:/ /search.proquest.com on Jan. 8, 2014, 4 pages.
Bylund, Anders, “Would You Buy a Self-Driving Car to Save 80% on Auto Insurance?” The Motley Fool, Nov. 27, 2013, http://www.dailyfinance.com/2013/11/27/would-you-buy-a-self-driving-car-to-save-80-on-car/, 2 pages.
Litman, Todd, “Autonomous Vehicle Implementation Predictions Implications for Transport Planning,” Victoria Transport Policy Institute, Dec. 23, 2013, 19 pages.
VentureBeat.com webpage; “Cisco and NXP encourage car communication to make driving safer” www.venturebeat.com/2013/01/04/cisco-and-nxp-encourage-car-communication-to-make-driving-safer/, Rebecca Grant dated Jan. 4, 2013.
Yang et al. “A vehicle-to-vehicle communication protocol for cooperative collision warning”; Aug. 2004.
Festag et al., “Vehicle-to-vehicle and road-side sensor communication for enhanced road safety”; Nov. 2008.
Zalstein, David, Car Advice. com webpage, “First large-scale vehicle-to-vehicle communication technology test unveiled” dated Aug. 22, 2012, www.caradvice.com.au/187379/first-large-scale-vehicle-to-vehicle-communication-technology-test-unveiled/basic-rgb-4/, 3 pages.
BMW.com webpage; “BMW Technology Guide: Car-to-car communication” www.bmw.com/com/en/insights/technology/technology_guide/articles/cartocar_communication.html; downloaded Apr. 5, 2013.
NEC.com webpage; “Car2Car Communication” www/nec.com/en/global.onlinetv/en/society/car_commu_I:html; downloaded Apr. 5, 2013.
Wardzinski, Dynamic risk assessment in autonomous vehicles motion planning, IEEE, 1st International Conference on Information Technology, Gdansk, May 18-21, 2008 [retrieved on Jul. 25, 2016], Retrieved from the Internet, <URL:http://kio.pg.gda.pl/lag/download/2008-IEEE%20ICIS-Dynamic%20Risk%20Assessment.pdf>, 4 pages.
U.S. Appl. No. 61/391,271, Oct. 8, 2010, Appendix to the Specification, “Appendix B”, (incorporated by reference in US 20120083960, Zhu, J. et al)) (Year: 2010).
U.S. Appl. No. 61/391,271, Oct. 8, 2010, Specification, “Google 3.8-292” (incorporated by reference in US 2012-0083960 (Zhu, J. et al)) (Year: 2010).
Apr. 19, 2018 (U.S.) Final Office Action—U.S. Appl. No. 14/458,764.
May 14, 2018 (U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,826.
May 15, 2018 (U.S.) Notice of Allowance—U.S. Appl. No. 14/163,719.
May 17, 2018 (U.S.) Notice of Allowance—U.S. Appl. No. 14/697,131.
May 18, 2018 (U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,744.
Jun. 6, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 15/015,623.
Jun. 14, 2018—(U.S.) Final Office Action—U.S. Appl. No. 14/458,796.
Jun. 5, 2018—(CA) Office Action—App 2,975,087.
Feb. 12, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/673,150.
Mar. 9, 2018—(U.S.) Non-Final Office Action—U.S. Appl. No. 15/013,523.
Mar. 14, 2018—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/607,433.
Mar. 13, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 15/206,521.
Mar. 29, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/697,141.
Apr. 2, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/697,153.
Mar. 30, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/816,299.
Apr. 2, 2018—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/816,336.
Aug. 14, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/607,433.
Sep. 4, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 14/816,336.
Oct. 11, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 15/206,521.
Sharma, Devansh, “Development of Leader-Follower Robot in IIT Bombay,” 4 pages, retrieved May 30, 2013, date unknown, but prior to the filing of this application.
Dec. 6, 2018—(U.S.) Notice of Allowance—U.S. Appl. No. 15/013,523.
Baronti, et al, “Distributed Sensor for Steering Wheel Grip Force Measurement in Driver Fatigue Detection,” Department of Engineering and Information, University of Pisa, Italy, pp. 1-4. (Year: 2009).
Ji, et al, “Real-Time Nonintrusive Monitoring and Prediction of Driver Fatigue,” IEEE Transactions on Vehicular Technology, vol. 53, No. 4, pp. 1-17 (Year: 2004).
Oct. 22, 2018 (CA) Office Action—App. 2,988,134.
Jan. 8, 2019—(U.S.) Notice of Allowance—U.S. Appl. No. 15/206,521.
Jan. 23, 2019—(U.S.) Final Office Action—U.S. Appl. No. 14/458,826.
U.S. Provisional Appl. No. 61/391,271, filed Oct. 8, 2010, Specification, “Google 3.8-292”, “Autonomous Vehicles”, Zhu et al., 56 pages (Year: 2010).
U.S. Provisional Appl. No. 61/391,271, filed Oct. 8, 2010, Appendix to the Specification, “Appendix B”, “User Interface for Displaying Internal State of Autonomous Driving System”, Zhu et al., 37 pages (Year: 2010).
Jul. 25, 2019—(US) Non-Final Office Action—U.S. Appl. No. 16/294,103.
Jan. 28, 2019—(U.S.) Final Office Action—U.S. Appl. No. 14/458,744.
“Your Questions Answered: Driverless Cars”, Stephen Harris, The Engine (Online), Feb 17, 2014; n/a. ProQuest. Web. Jan. 18, 2019 (Year 2014).
Feb. 11, 2019—(EP) Supplementary Search Report—EP16743839.9.
Mar. 21, 2019 (CA) Office Action—App. 2,975,087.
May 1, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,764.
May 3, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,796.
May 15, 2019 (EP) Extended European Search Report—App. 16808098.4.
Jun. 18, 2019 (U.S.) Notice of Allowance and Fees Due—U.S. Appl. No. 15/206,521.
Jun. 26, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,744.
Aug. 22, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 15/974,861.
Aug. 22, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 16/021,593.
Aug. 27, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 16/021,678.
Sep. 19, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 14/458,826.
Zhu et al., U.S. Provisional Appl. No. 61/391,271, filed Oct. 8, 2010, Specification “Google 3.8-392”, “Autonomous Vehicles”, 56 pages, Year 2010.
Zhu et al., U.S. Provisional Appl. No. 61/391,271, filed Oct. 8, 2010, Appendix to the Specification, “Appendix B”, “User Interface for Displaying Internal State of Autonomous Driving System”, 37 pages, Year 2010.
Feb. 26, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 15/206,521.
May 21, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,744.
Mar. 19, 2020—(U.S.) Non-Final Office Action—U.S. Appl. No. 16/294,103.
Apr. 16, 2020—(U.S.) Non-Final Office Action—U.S. Appl. No. 15/890,701.
Apr. 20, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 15/206,521.
Apr. 20, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 15/827,860.
Apr. 29, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 15/900,861.
Oct. 30, 2019—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,764.
Wu et al, “Petri Net Modeling of the Cooperation Behavior of a Driver and a Copilot in an Advanced Driving Assistance System”, IEEE Transactions on Intelligtent Transportation Systems, vol. 12, Issue 4, Dec. 1, 2011, pp. 977-989, Year 2011.
Oct. 31, 2019—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,744.
Nov. 4, 2019—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,796.
J.F. Coughlin, B. Reimer, B. Mehler, “Monitoring Managing and Motivating Driver Safety and Well-Being”, IEEE Pervasive Comput., vol. 10 No. 3, pp. 14-21, Year 2011.
Nov. 25, 2019—(In) Office Action—Application No. 201727043994.
Dec. 12, 2019—(U.S.) Non-Final Office Action—U.S. Appl. No. 15/827,860.
Dec. 11, 2019 (U.S.) Non-Final Office Action—U.S. Appl. No. 16/102,089.
May 15, 2019 (EP) European Extended Search Report—Application No. 16808098.4.
Harris, Stephen., “Your Questions Answered: Driverless Cars,” The Engineer (Online) Feb. 17, 2014: n/a. ProQuest. Web. Jan. 18, 2019 (Year: 2014).
Jan. 27, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,826.
“Driver Monitors: Improving Transportation Safety and Enhancing Performance Through Behavioral Change”, Ballard, T., Melton, A., and Sealy, I., Society of Petroleum Engineers, Jan. 1, 2004, Year 2004.
Jan. 31, 2020—(U.S.) Final Office Action—U.S. Appl. No. 16/021,593.
Feb. 3, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,744.
Feb. 4, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,764.
“Petri Net Modeling of the Cooperation Behavior of a Driver and a Copilot in an Advanced Driving Assistance System”, Wu et al., IEEEE Transportation on Intelligent Transportation Systems, vol. 12, Issue 4, Dec .1, 2011, pp. 377-989 (Year 2011).
Feb. 6, 2020—(U.S.) Final Office Action—U.S. Appl. No. 15/974,861.
Feb. 5, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,796.
Feb. 14, 2020—(U.S.) Final Office Action—U.S. Appl. No. 16/021,678.
Jun. 1, 2020—(U.S.) Notice of Allowance—U.S. Appl. No. 14/458,826.
Continuations (1)
Number Date Country
Parent 14163741 Jan 2014 US
Child 15166638 US