The technical field generally relates to the vehicles and, more specifically, to methods and systems for generating situation awareness graphs using cameras from different vehicles.
Many vehicles include various systems for improved operation of the vehicle, including use of cameras for detecting objects and other surroundings in proximity to the vehicle. However, in certain situations, a vehicle's cameras may be occluded or blocked by other obstacles on the road, and/or one or more other conditions may be present that may make it difficult for a camera of a single vehicle to ascertain situations awareness for the vehicle.
Accordingly, it is desirable to provide improved methods and systems to facilitate situation awareness for a vehicle having a camera. Furthermore, other desirable features and characteristics of the present invention will become apparent from the subsequent detailed description of the invention and the appended claims, taken in conjunction with the accompanying drawings and this background of the invention.
In one exemplary embodiment, a method is provided that includes: obtaining first camera images from a first camera onboard a first vehicle; generating, via one or more computer processors, a first situation awareness graph with respect to objects near the first vehicle, using the first camera images; obtaining second camera images from a second camera of a second device that is in proximity to the first vehicle; generating, via one or more computer processors, a second situation awareness graph with the respect to the objects, using the second camera images; and generating, via one or more computer processor, a global situation awareness graph with respect to the objects, by merging the first situation awareness graph with the second situation awareness graph, using respective first and second weights for the first and second situation awareness graphs.
Also in one embodiment, the step of generating the first situation awareness graph includes generating a first static situation awareness graph with respect to objects near the first vehicle at a particular time, using the first camera images; the step of generating the second situation awareness graph includes generating a second static situation awareness graph with respect to the objects at the particular time, using the second camera images; and the step of generating the global situation awareness graph includes generating a static global awareness graph with respect to the objects at the particular time, by merging the first static situation awareness graph with the second static situation awareness graph, using respective first and second weights for the first and second static situation awareness graphs.
Also in one embodiment, the step generating the static global situation awareness graph includes generating the static global situation awareness graph using the respective first and second weights for the first and second static situation awareness graphs based on a principal components analysis of respective Mahalanobis Distances from the first and second static situation awareness graphs.
Also in one embodiment, the step of generating the global situation awareness graph includes generating, via one or more computer processors, a time-evolving dynamic global situational awareness map with respect to the objects using the first situation awareness graph and the second situation awareness graph over multiple periods of time, using respective first and second weights for the first and second situation awareness graphs.
Also in one embodiment, the step of generating the global situation awareness graph includes generating, via one or more computer processors, a multi-layer dynamic bipartite graph for predicting the trajectory of detected objects, using the first situation awareness graph and the second situation awareness graph over multiple periods of time, using respective first and second weights for the first and second situation awareness graphs.
Also in one embodiment, the step of generating the global situation awareness graph includes generating, via one or more computer processors, a multi-layer dynamic probabilistic graph for predicting the trajectory of detected objects, using initial and posterior probabilities from the first situation awareness graph and the second situation awareness graph over multiple periods of time, utilizing a Bayesian particle filter.
Also in one embodiment, the second device includes a second vehicle that is in proximity to the first vehicle.
Also in one embodiment, the second device includes infrastructure that is in proximity to the vehicle.
Also in one embodiment, the method further includes: estimating, via one or more processors, one or more parameters as to one or more occluded vehicles of the detected objects, based on the global situation awareness graph; and controlling the first vehicle, the second vehicle, or both, in a manner to avoid contact with the one or more occluded vehicles, using the one or more parameters that were estimated based on the global situation awareness graph.
In another exemplary embodiment, a system is provided that includes: a first camera onboard a first vehicle, the first camera configured to generate first camera images from the first vehicle; and one or more computer processors configured to at least facilitate: generating a first situation awareness graph with respect to objects near the first vehicle, using the first camera images; and generating a global situation awareness graph with respect to the objects, by merging the first situation awareness graph with a second situation awareness graph that was generated using second camera images from a second camera of a second device that is in proximity to the first vehicle, using respective first and second weights for the first and second situation awareness graphs.
Also in embodiment, the system further includes a transceiver configured to receive the second camera images from the second device, the second situation awareness graph, or both.
Also in one embodiment, the second device includes a second vehicle that is in proximity to the first vehicle, and the transceiver is further configured to transmit, via instructions from the one or more computer processors, the first camera images, the first situation awareness graph, or both, for use by the second vehicle.
Also in one embodiment, the one or more processors are configured to at least facilitate: generating a first static situation awareness graph with respect to objects near the first vehicle at a particular time, using the first camera images; generating a second static situation awareness graph with respect to the objects at the particular time, using the second camera images; and generating a static global awareness graph with respect to the objects at the particular time, by merging the first static situation awareness graph with the second static situation awareness graph, using respective first and second weights for the first and second static situation awareness graphs.
Also in one embodiment, the one or more processors are configured to at least facilitate generating the static global situation awareness graph using the respective first and second weights for the first and second static situation awareness graphs based on a principal components analysis of respective Mahalanobis Distances from the first and second static situation awareness graphs.
Also in one embodiment, the one or more processors are configured to at least facilitate generating a time-evolving dynamic global situational awareness map with respect to the objects using the first situation awareness graph and the second situation awareness graph over multiple periods of time, using respective first and second weights for the first and second situation awareness graphs.
Also in one embodiment, the one or more processors are configured to at least facilitate generating a multi-layer dynamic bipartite graph for predicting the trajectory of detected objects, using the first situation awareness graph and the second situation awareness graph over multiple periods of time, using respective first and second weights for the first and second situation awareness graphs.
Also in one embodiment, the second device includes a second vehicle, and the one or more processors are configured to at least facilitate: estimating one or more parameters as to one or more occluded vehicles of the detected objects, based on the global situation awareness graph; and controlling the first vehicle, the second vehicle, or both, in a manner to avoid contact with the one or more occluded vehicles, using the one or more parameters that were estimated based on the global situation awareness graph.
In another exemplary embodiment, a vehicle is provided that includes a first camera and one or more computer processors. The first camera is onboard the vehicle, the first camera configured to generate first camera images from the vehicle; and one or more computer processors configured to at least facilitate: generating a first situation awareness graph with respect to objects near the vehicle, using the first camera images; and generating a global situation awareness graph with respect to the objects, by merging the first situation awareness graph with a second situation awareness graph that was generated using second camera images from a second camera of a second device that is in proximity to the vehicle, using respective first and second weights for the first and second situation the first camera images.
Also in one embodiment, the vehicle further includes a transceiver configured to receive the second camera images from the second device, the second situation awareness graph, or both, and to transmit, via instructions from the one or more computer processors, the first camera images, the first situation awareness graph, or both, for use by the second device.
The present disclosure will hereinafter be described in conjunction with the following drawing figures, wherein like numerals denote like elements, and wherein:
The following detailed description is merely exemplary in nature and is not intended to limit the disclosure or the application and uses thereof. Furthermore, there is no intention to be bound by any theory presented in the preceding background or the following detailed description.
In certain embodiments, the cameras 102 are controlled via a control system 104, as depicted in
In various embodiments, the vehicle 100 preferably comprises an automobile. The vehicle 100 may be any one of a number of different types of automobiles, such as, for example, a sedan, a wagon, a truck, or a sport utility vehicle (SUV), and may be two-wheel drive (2WD) (i.e., rear-wheel drive or front-wheel drive), four-wheel drive (4WD) or all-wheel drive (AWD), and/or various other types of vehicles in certain embodiments. In certain embodiments, the vehicle 100 may also comprise a motorcycle or other vehicle, and/or one or more other types of mobile platforms (e.g., a robot, a ship, and so on) and/or other systems, for example having a camera image with a fixed referenced point.
The vehicle 100 includes the above-referenced body 110 that is arranged on a chassis 112. The body 110 substantially encloses other components of the vehicle 100. The body 110 and the chassis 112 may jointly form a frame. The vehicle 100 also includes a plurality of wheels 114. The wheels 114 are each rotationally coupled to the chassis 112 near a respective corner of the body 110 to facilitate movement of the vehicle 100. In one embodiment, the vehicle 100 includes four wheels 114, although this may vary in other embodiments (for example for trucks and certain other vehicles).
A drive system 116 is mounted on the chassis 112, and drives the wheels 114. The drive system 116 preferably comprises a propulsion system. In certain exemplary embodiments, the drive system 116 comprises an internal combustion engine and/or an electric motor/generator, coupled with a transmission thereof. In certain embodiments, the drive system 116 may vary, and/or two or more drive systems 116 may be used. By way of example, the vehicle 100 may also incorporate any one of, or combination of, a number of different types of propulsion systems, such as, for example, a gasoline or diesel fueled combustion engine, a “flex fuel vehicle” (FFV) engine (i.e., using a mixture of gasoline and alcohol), a gaseous compound (e.g., hydrogen and/or natural gas) fueled engine, a combustion/electric motor hybrid engine, and an electric motor.
As depicted in
The cameras 102 provide images for viewing on one or more displays 108 that are disposed inside the vehicle 100 (i.e. inside the body 110 of the vehicle 100). In addition, in various embodiments, the camera images are also transmitted to one or more other vehicles. In addition, in certain embodiments, the camera images are also provided to a vehicle control system for use in vehicle control, and so on.
In various embodiments, the navigation system 106 provides location information for the vehicle 100. For example, in various embodiments, the navigation system 106 comprises a satellite-based system, such as a global positioning system (GPS) and/or other satellite-based system, and provides location information regarding a current position of the vehicle 100. In certain embodiments, the navigation system 106, and/or one or more components thereof, may be disposed within and/or be part of the control system 104. In other embodiments, the navigation system 106 may be coupled to the control system 104.
In various embodiments, the display 108 displays images, such as from the cameras 102 of the vehicle 100, and in certain embodiments also from respective cameras from other vehicles. In one embodiment, the display 108 is located on a center console of the vehicle 100. However, this may vary in other embodiments. In various other embodiments, the display 108 may be part of a radio display, a navigation display, and/or other display, for example as part of or in proximity to the center console. In certain other embodiments, the display 108 may be part of one or more other vehicle components, such as a rear view mirror. In one exemplary embodiment the display 108 comprises a liquid crystal display (LCD) screen or a light emitting diode (LED) screen. However, this may vary in other embodiments.
The control system 104 controls operation of the cameras 102, and generates situation awareness graphs and instructions for control of the vehicle 100 based on data from the cameras 102 of the vehicle 100 as well as data from other cameras of other nearby vehicles. In various embodiments, the control system 104 provides these and other functions in accordance with the steps of the process 300 discussed further below in connection with the implementation of
In various embodiments, the control system 104 is disposed within the body 110 of the vehicle 100. In one embodiment, the control system 104 is mounted on the chassis 112. In certain embodiments, the control system 104 and/or one or more components thereof may be disposed outside the body 110, for example on a remote server, in the cloud, or in a remote smart phone or other device where image processing is performed remotely. In addition, in certain embodiments, the control system 104 may be disposed within and/or as part of the cameras 102, navigation system 106, and/or display 108, and/or within and/or or as part of one or more other vehicle systems.
Also, as depicted in
As depicted in
The sensor array 120 generates sensor data, and provides the sensor data to the controller 124 for processing. As depicted in
The transceiver 122 transmits messages to, and receives messages from, other vehicles. Specifically, in various embodiments, the transceiver 122 transmits (via instructions provided by the controller 124) camera images, under appropriate circumstances, to other vehicles and/or infrastructure. Also in various embodiments, the transceiver 122 also receives images and other information, under appropriate circumstances, from other vehicles and/or infrastructure. It will be appreciated that in certain embodiments the transceiver 122 may comprise separate transmitters and/or receivers, or the like.
The controller 124 controls operation of the control system 104, and facilitates the control of situation awareness for the vehicle 100, including use of camera images and sharing of camera images and other information between vehicles and/or infrastructure on the roadway, and the generating of static and dynamic situation awareness graphs for the vehicle 100 with respect to detected objects, using the camera images from the cameras of the vehicle 100 along with camera images from cameras of other nearby vehicles and/or infrastructure. In certain embodiments, the controller 124 also controls various functionality of the vehicle 100 (e.g., steering and braking), for example to avoid obstacles, using the camera images and data. In various embodiments, the controller 124 provides these and other functions in accordance with the steps of the process 300 discussed further below in connection with the implementation of
In one embodiment, the controller 124 is coupled to the cameras 102, the navigation system 106, the sensor array 120, the transceiver 122, and the display 108. Also in one embodiment, the controller 124 is disposed within the control system 104, within the vehicle 100. In certain embodiments, the controller 124 (and/or components thereof, such as the processor 132 and/or other components) may be part of and/or disposed within the cameras 102, the navigation system 106, the display 108, and/or one or more other vehicle components. Also in certain embodiments, the controller 124 may be disposed in one or more other locations of the vehicle 100. In addition, in certain embodiments, multiple controllers 124 may be utilized (e.g. one controller 124 within the vehicle 100 and another controller within the cameras 102, the navigation system 106, and/or the display 108), among other possible variations. In addition, in certain embodiments, the controller can be placed outside the vehicle, such as in a remote server, in the cloud or on a remote smart device.
As depicted in
In the depicted embodiment, the computer system of the controller 124 includes a processor 132, a memory 134, an interface 136, a storage device 138, and a bus 140. The processor 132 performs the computation and control functions of the controller 124, and may comprise any type of processor or multiple processors, single integrated circuits such as a microprocessor, or any suitable number of integrated circuit devices and/or circuit boards working in cooperation to accomplish the functions of a processing unit. During operation, the processor 132 executes one or more programs 142 contained within the memory 134 and, as such, controls the general operation of the controller 124 and the computer system of the controller 124, generally in executing the processes described herein, such as the process 300 described further below in connection with
The memory 134 can be any type of suitable memory. For example, the memory 134 may include various types of dynamic random access memory (DRAM) such as SDRAM, the various types of static RAM (SRAM), and the various types of non-volatile memory (PROM, EPROM, and flash). In certain examples, the memory 134 is located on and/or co-located on the same computer chip as the processor 132. In the depicted embodiment, the memory 134 stores the above-referenced program 142 along with one or more stored values 144.
The bus 140 serves to transmit programs, data, status and other information or signals between the various components of the computer system of the controller 124. The interface 136 allows communication to the computer system of the controller 124, for example from a system driver and/or another computer system, and can be implemented using any suitable method and apparatus. In one embodiment, the interface 136 obtains the various data from the cameras 102, the navigation system 106, the transceiver 122, and/or the sensor array 120. The interface 136 can include one or more network interfaces to communicate with other systems or components. The interface 136 may also include one or more network interfaces to communicate with technicians, and/or one or more storage interfaces to connect to storage apparatuses, such as the storage device 138.
The storage device 138 can be any suitable type of storage apparatus, including direct access storage devices such as hard disk drives, flash systems, floppy disk drives and optical disk drives. In one exemplary embodiment, the storage device 138 comprises a program product from which memory 134 can receive a program 142 that executes one or more embodiments of one or more processes of the present disclosure, such as the steps of the process 300 (and any sub-processes thereof) described further below in connection with
The bus 140 can be any suitable physical or logical means of connecting computer systems and components. This includes, but is not limited to, direct hard-wired connections, fiber optics, infrared and wireless bus technologies. During operation, the program 142 is stored in the memory 134 and executed by the processor 132.
It will be appreciated that while this exemplary embodiment is described in the context of a fully functioning computer system, those skilled in the art will recognize that the mechanisms of the present disclosure are capable of being distributed as a program product with one or more types of non-transitory computer-readable signal bearing media used to store the program and the instructions thereof and carry out the distribution thereof, such as a non-transitory computer readable medium bearing the program and containing computer instructions stored therein for causing a computer processor (such as the processor 132) to perform and execute the program. Such a program product may take a variety of forms, and the present disclosure applies equally regardless of the particular type of computer-readable signal bearing media used to carry out the distribution. Examples of signal bearing media include: recordable media such as floppy disks, hard drives, memory cards and optical disks, and transmission media such as digital and analog communication links. It will be appreciated that cloud-based storage and/or other techniques may also be utilized in certain embodiments. It will similarly be appreciated that the computer system of the controller 124 may also otherwise differ from the embodiment depicted in
As will be explained in greater detail in connection with the process 300 discussed further below in connection with
As depicted in
As depicted in
Camera data is obtained for the vehicles (step 304). In various embodiments, camera images are obtained from one or more of the cameras 102 of the vehicle 100 of
Other sensor data is obtained (step 306). In various embodiments, sensor data is obtained from one or more other sensors 128 of the vehicle 100 of
The camera data is transmitted (step 308). In various embodiments, the different vehicles 100, 202 prepare camera feeds (e.g., still and/or video images, and/or associated information) for transmission to one another, and/or to other vehicles and/or infrastructure, and/or to the remote server 218, via the wireless networks 216. In various embodiments, the other sensor data is also transmitted as part of step 308. Also in various embodiments, the data is transmitted via the transceiver 122 of
In various embodiments, the camera data is aggregated (step 310). In certain embodiments, each vehicle aggregates camera data from itself and from other nearby vehicles (e.g., via the processor 132 of the vehicle 100 of
In certain embodiments, a global static graph is generated and utilized during a first sub-sequence 311 of steps, as described below. In various embodiments, the global static graph comprises a graphical representation for situation awareness for the vehicle(s) based on the camera data from the vehicles 100, 202 and/or infrastructure 203 at a particular point in time, as set forth below in accordance with sub-sequence 311 (corresponding to steps 312-317, described below).
First, during step 312, individual vehicle static graphs are generated for the various vehicles (e.g., the vehicles 100, 202) and/or infrastructure (e.g., infrastructure 203). Specifically, with reference to
Next during step 313, in various embodiments, principal components analysis is performed with respect to Mahalanobis Distances from the individual static graphs of step 312 (e.g., the first vehicle static graph 400 of the first vehicle 100 and the second vehicle static graphs 410 of the second vehicle 202 and/or infrastructure 203). For example, in one exemplary embodiments the Mahalanobis Distance is calculated with respect to identified objects of corresponding vectors for each of the individual static graphs, in accordance with the following equation:
in which “MahalD” represents the Mahalanobis Distance, “VB” represents the first vehicle 100, “WB” represents the second vehicle (or infrastructure), and “S” represents the covariance of these two vectors VB and WB. Also in various embodiments, respective different individual vehicle static graphs are determined to refer to the same detected object if the Mahalanobis Distance is less than a predetermined threshold distance. In certain embodiments, the predetermined threshold distance is a post-calibrated parameter that is generated by experts (e.g., from manufacturer of the vehicle 100). In various embodiments, one or more different calculations may be utilized, for example using one or more different equations. In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Weights are determined for the different individual static graphs (step 314). In certain embodiments, the weights comprise a weighted average of the static graphs (e.g., a weighted average of each of the respective points and distances from the first vehicle static graph 400 and the second static graph(s) 410 of
In various embodiments, the different individual static graphs are merged together in accordance with their respective weights (step 315). As a result, the global static graph is generated (step 316). Specifically, the global static graph represents a combined view of the surroundings of the vehicles 100, 202 and/or infrastructure 203 (including the objects 204, 206 in proximity thereto), leveraging the camera data from both of the vehicles 100, 202 and/or infrastructure 203 combined together. In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Specifically, with reference to
In various embodiments, the global static graph 420 (and associated viewpoint 422) includes a broader and/or more comprehensive view of nearby objects, from incorporating the respective static graphs 400, 410 (and respective associated viewpoints 402, 412) from the different cameras of the respective vehicles 100, 202 and/or infrastructure 203. For example, as depicted in
Also in various embodiments, the global static graph 420 includes the list of observed objects (e.g., objects 204, 206, which may include other vehicles and/or other objects), along with six degrees of freedom for the vehicle 100 generating the static graph. Also in various embodiments, the six degrees of freedom pose estimates of the observed objects, including a distance and an angle to the observed objects.
Once the global static graph is generated, the situation awareness of the vehicles is updated in accordance with the global static graph (step 316). In various embodiments, estimated parameters for the detected objects 204, 206 (e.g., estimated locations, distances, and angles from the respective vehicles 100, 202) are updated using the global static graph 420. In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
With reference back to step 310, also in various embodiments, a dynamic time-evolving situation awareness graph is generated and utilized during a second sub-sequence 321 of steps, as described below. In various embodiments, the dynamic time-evolving situation awareness graph comprises a graphical representation for situation awareness for the vehicle(s) and/or infrastructure over time based on the camera data from the vehicles 100, 202 and/or infrastructure 203 at different points in time, as set forth below in accordance with sub-sequence 321 (corresponding to steps 322-329, described below).
First, during step 322, a global static situational awareness graph is generated. In various embodiments, the global static situational awareness graph is generated by one or more processors (e.g., via the processor 132 of the vehicle 100 of
In addition, in various embodiments, a correlation function is generated for the first vehicle (step 323). In various embodiments, the correlation function pertains to a comparison of the values of the individual vehicle graphs for the vehicle 100 at different points in time in comparison with respective values of the global graphs at the points in time (e.g., as a measure of accuracy and/or precision of the values for the cameras 102 of the first vehicle 100). Also in various embodiments, the correlation function for the first vehicle 100 includes a weight factor that includes a factor comparing a position (e.g., a position of a detected object) in frame 1 (e.g., at a first time t0) to position in frame 2 (e.g., at a second time t1), and a factor comparing visual words) defining a connection from a detected object (e.g., object1) in one frame to the same detected object (e.g., object1) in next frame based on data from first vehicle 100, and so on. In various embodiments, similar connections are utilized for subsequent frames (e.g., from time t1 to subsequent time t2, and so on). In various embodiments, these actions are taken by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Also in various embodiments, similar correlation functions are generated for other nearby vehicles and/or infrastructure based on their respective cameras (step 324). For example, with respect to the second vehicle 202, in various embodiments, the correlation function pertains to a comparison of the values of the individual vehicle graphs for the second vehicle 202 at different points in time in comparison with respective values of the global graphs at the points in time (e.g., as a measure of accuracy and/or precision of the values for the cameras of the second vehicle 202). Also in various embodiments, the correlation function for the second vehicle 202 includes a weight factor that includes a factor comparing a position (e.g., a position of a detected object) in frame 1 (e.g., at a first time t0) to a position in frame 2 (e.g., at a second time t1), and a factor comparing visual words) defining a connection from a detected object (e.g., object1) in one frame to the same detected object (e.g., object1) in next frame based on data from second vehicle 202, and so on. In various embodiments, similar connections are utilized for subsequent frames (e.g., from time t1 to subsequent time t2, and so on). In various embodiments, these actions are taken by one or more processors (e.g., via the processor 132 of the vehicle 100 of
A cost function is determined (step 325). In various embodiments, the results of the various correlations of steps 323-324 of the various vehicles (e.g., including the vehicle 100 and the second vehicle 202) are summed together to determine a cost function defining a connection from object1 (i.e., a first detected object) in frame 1 (e.g., at time t0) to object1 in frame 2 (e.g., at time t1). In addition, in various embodiments, similar cost functions are determined for each of the detected objects (e.g., for each of the objects, such as object2, object3, and so on detected by the first vehicle 100 and/or the second vehicle 202) from frame 1 (e.g., at time t0) to frame 2 (e.g., at time t1). Also in various embodiments, for each of the objects (e.g., object1, object2, object3, and so on), cost functions are similarly determined with respect to each vehicle (for each object) and/or with respect to each infrastructure between subsequent frames as well, such as between frame 2 (e.g., at time t1) and frame 3 (e.g., at time t2), and so on. Also in various embodiments, the determination of the cost function is made by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Also in various embodiments, a highest cost object is determined (step 326). In certain embodiments, the highest cost object is determined for each object (e.g., object1) for each frame sequence (e.g., between frame 1 and frame 2, and so on), based on an evaluation of the cost functions of step 325 in determining the highest cost object to object connection corresponding to matched objects from frame to frame. Also in various embodiments, these actions are made by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Weights are determined for the various vehicles and/or infrastructure (step 327). In various embodiments, weights are determined for each of the vehicles and, if applicable, infrastructure (e.g., the first vehicle 100, the second vehicle 202, and the infrastructure 203) as a measure of accuracy and/or precision for the values generated from the camera images for object detection by the first vehicle 100, the second vehicle 202, and the infrastructure 203 (and for any other participating vehicles and infrastructure). In various embodiments, the weight for the first vehicle 100 is determined by comparing the first vehicle 100's position for each detected object (e.g., as estimated using the images from the cameras 102 of the first vehicle 100) versus that represented in the global graph of step 322. Similarly, in various embodiments, the weight for the second vehicle 202 is determined by comparing the second vehicle 202's position for each detected object (e.g., as estimated using the images from the cameras of the second vehicle 202) versus that represented in the global graph of step 322 (in various embodiments, the weight for the infrastructure 203 would be similarly determined based on the images from the cameras of the infrastructure 203, and so on). Also in certain embodiments, the weights are determined also in part based on the correlation functions of steps 323-324, the cost functions of step 325, and the highest cost object of step 326. In certain embodiments, these actions are made by one or more processors (e.g., via the processor 132 of the vehicle 100 of
in which “W” represents the respective calculated weights, “V” represents the respective velocities, “θ” represents the respective angles with respect to detected objects, “fcor” and “gcor” represent respective correlations, and “KF” represents the extended Kalman Filter. In other embodiments, one or more different equations may be utilized.
A dynamic time-evolving situational awareness graph is generated (step 328). In various embodiments, the dynamic time-evolving situational awareness graph comprises a global dynamic time-evolving bipartite situational awareness graph over the various points of time, by updating the global graph of step 322 with the updated data from the different vehicles 100, 202 (and in certain embodiments, the infrastructure 203), merged together with the weights of step 327 in accordance with the analysis of steps 323-327. In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
With reference to
Once the dynamic time-evolving bipartite graph is generated, the situation awareness of the vehicles is updated in accordance with the dynamic time-evolving situation awareness graph (step 329). In various embodiments, estimated parameters for the detected objects 204, 206 (e.g., estimated locations, distances, and angles from the respective vehicles 100, 202 and/or infrastructure 203) are updated using the dynamic time-evolving bipartite graph of step 328. Also in various embodiments, a trajectory of the detected objects is predicted using the dynamic time-evolving bipartite graph of step 328. In various embodiments, the process then proceeds to step 350 (discussed directly below), in which the vehicle 100 is controlled at least in part based on the dynamic time-evolving graph of step.
With reference back to step 310, also in various embodiments, a dynamic time-evolving probabilistic situation awareness graph is generated and utilized during a third sub-sequence 331 of steps, as described below. In various embodiments, the dynamic time-evolving particle filter graph comprises a graphical representation for situation awareness for the vehicle(s) over time based on the camera data from the vehicles 100, 202 (and, in certain embodiments, from the infrastructure 203) at different points in time, as set forth below in accordance with sub-sequence 331 (corresponding to steps 332-346, described below).
First, during step 332, the first vehicle 100 maintains its own prior probability distribution (e.g., xt-1) the location of one or more detected objects, and also generates a new distribution based on the first vehicle 100's observation of the detected object (e.g., ut) in order to obtain the first vehicle 100's new probability distribution (e.g., xt) for the location of the detected object. In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
Also in various embodiments, during step 334, additional vehicles (e.g., the second vehicle 202 of
The probability distributions for the various vehicles are reported (step 336). Specifically, in certain embodiments, the probability distributions from the first vehicle 100 and the second vehicle 202 and/or the infrastructure 203 are transmitted from the respective vehicles via respective transmitters via the wireless networks 216 of
Also in various embodiments, a global probability distribution is generated (step 338). In certain embodiments, a weight-dependent number of samples are drawn from the respective updated probability distributions from the different vehicles and/or infrastructure, and the global probability distribution is generated by merging the probability distributions of the individual vehicles and/or infrastructure based on the respective weights. In certain embodiments, a weight-dependent number of samples from the first vehicle 100 and the second vehicle 202 and/or the infrastructure 203 are taken by one or more processors (e.g., via the processor 132 of the vehicle 100 of
in which vi represents each reporting car's estimated location of targeted car and ωi is the weight of that reporting car's prediction (accordingly, the above equation represents the weighted average of the predicted vehicle location, in certain embodiments).
Also in various embodiments, weights of the various vehicles and/or infrastructure are updated (step 340). In certain embodiments, the weight of the first vehicle 100 is updated based on a comparison of the first vehicle 100's probability distribution (i.e., of step 332) and the merged probability distribution (i.e., of step 334). In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
In various embodiments, the dynamic time-evolving particle filter graph is generated (step 344). For example, in various embodiments, one or more processors (e.g., via the processor 132 of the vehicle 100 of
In certain embodiments, the revised global probability distributions are generated based on resampling of data and corresponding recalculation of the results using an iterative process from collaborative cameras.
For example, in certain embodiments, the resampling is performed in accordance with the following equation (Equation 4):
Q(t)p(xt|xt-1,ut)
in which Q(t) represents re-sampling at time (t), and “p” represents the conditional probability at point in time (t) given previous probability at time (t−1) and the new observations ut from the vehicle at time (t). Specifically, in certain embodiments, the Q(t) equation is used to update the new distribution of targeted vehicle xt at time t, based on its own previous distribution x(t-1) at time t−1, and new observation by reporting vehicle ut. In certain embodiments, all the reporting vehicles will iterate and obtain new distributions of detected objects (e.g., targeted vehicles) based on all these inputs. Also in certain embodiments, this new derived distribution will be re-sampled as the new distribution of the targeted vehicle's location.
Also in certain embodiments, the recalculating is performed using the following equation:
in which “wi” represents the recalculated weight, and “P” represents respective conditional probabilities at time “t” given each reporting vehicle's observation. Also in certain embodiments, for each vehicle “i” at time “t”, “m” represents the conditional probability distribution function of the target vehicle position “z”, conditioned on its observation towards all other “landmarks” (e.g., known objects) on the map. Also in certain embodiments, this recalculation is similar to the original calculation of the weights described above, exception this is the summation of all vehicles (i=1 . . . n). In certain embodiments, Equation 5 is utilized to recalculate the latest weight for each individual reporting vehicle's observation. Accordingly, in various embodiments, the resampling process occurs and a newer distribution is generated based individual's contribution. As a result, in various embodiments, in this iterative process, some reporting vehicles will become “winners”, gaining more weight because its prediction match reality better. Similarly, also in various embodiments, some reporting vehicles will become “losers”, due to inaccurate prediction of its distribution (and therefor losing weight). Accordingly, in various embodiments, this equation is used to provide the weight update.
In various embodiments, these actions are performed by one or more processors (e.g., via the processor 132 of the vehicle 100 of
With reference to
Second, with reference to the second illustration 650, a global probability distribution 651 (also of step 344) is shown in another format, including a graphical depiction 660 of one of the detected objects 610 (comprising another vehicle) in proximity to the roadway 200 of
Once the dynamic time-evolving particle filter graph is generated, the situation awareness of the vehicles is updated in accordance with the dynamic time-evolving situation awareness graph (step 346). In various embodiments, estimated parameters for the detected objects 204, 206 (e.g., estimated locations, distances, and angles from the respective vehicles 100, 202) are updated using the time-evolving particle filter graph. Also in various embodiments, a trajectory of the detected objects is predicted using the dynamic time-evolving situational awareness graph. In various embodiments, the process then proceeds to step 350 (discussed directly below), in which the vehicle 100 is controlled at least in part based on the dynamic time-evolving particle filter graph.
During step 350, the vehicle 100 is controlled at least in part based on one or more of the graphs of steps 317, 328, and/or 344. In certain embodiments, the vehicle 100 is controlled using the global static graph of step 317, the dynamic time-evolving situation awareness graph of step 328, and the dynamic time-evolving particle filter graph of step 344. In certain other embodiments, one or two of the dynamic time-evolving situation awareness graph of step 328, and the dynamic time-evolving particle filter graph of step 344 may be utilized. Also in certain embodiments, the estimates of steps 317, 329, and/or 346 are utilized for controlling the vehicle 100 in step 350. In certain embodiments, one or more vehicle 100 actions (e.g., automatic braking, automatic steering, and so on) are controlled using the above graphs and estimates (and associated data) to avoid objects (e.g., to avoid other vehicles and/or other types of objects). In certain embodiments, one or more warnings, depictions of situation surroundings (e.g., of the detected objects) and the like are provided, for example via the display 108 of
Accordingly, the systems, vehicles, and methods thus provide for potentially improved situation awareness and control of vehicles, for example when certain objects may be occluded or blocked with respect to a camera of the vehicle. For example, in various embodiments, camera data from different vehicles and/or infrastructure is used to generate one or more global static graphs, dynamic time-evolving situation awareness graphs, and/or dynamic time-evolving particle filter graph when a vehicle is travelling in proximity to one or more other vehicles along a roadway.
It will be appreciated that the systems, vehicles, and methods may vary from those depicted in the Figures and described herein. For example, the vehicle 100, the cameras 102, the control system 104, the navigation system 106, the display 108, and/or components thereof of
While at least one exemplary embodiment has been presented in the foregoing detailed description, it should be appreciated that a vast number of variations exist. It should also be appreciated that the exemplary embodiment or exemplary embodiments are only examples, and are not intended to limit the scope, applicability, or configuration of the disclosure in any way. Rather, the foregoing detailed description will provide those skilled in the art with a convenient road map for implementing the exemplary embodiment or exemplary embodiments. It should be understood that various changes can be made in the function and arrangement of elements without departing from the scope of the disclosure as set forth in the appended claims and the legal equivalents thereof
Number | Name | Date | Kind |
---|---|---|---|
20140195138 | Stelzig | Jul 2014 | A1 |
20180173229 | Huang | Jun 2018 | A1 |
Number | Date | Country | |
---|---|---|---|
20190318041 A1 | Oct 2019 | US |