Situation awareness processor

Information

  • Patent Grant
  • 6470272
  • Patent Number
    6,470,272
  • Date Filed
    Monday, June 11, 2001
    23 years ago
  • Date Issued
    Tuesday, October 22, 2002
    21 years ago
Abstract
A plurality of events representative of a situation in which the host vehicle is operated are selected, including at least one set of related events. Input data is provided to an inference engine from either a first set of data representative of a target in a field of view of the host vehicle, a second set of data representative of the position or motion of the host vehicle, or a third set of data is representative of an environment of said host vehicle. The inference engine operates in accordance with an inference method to generate an output representative of a probability of occurrence of at least one event of the set of events, responsive to the input data, and possibly to one or more outputs at a past time. A countermeasure may be invoked responsive to one or more outputs from one or more inference engines.
Description




The above-identified applications are incorporated herein by reference.











In the accompanying drawings:





FIG. 1

illustrates a target in view of a host vehicle in a local coordinate system of the host vehicle;





FIG. 2

illustrates a block diagram system incorporating a situation awareness processor;





FIG. 3

illustrates a block diagram of host vehicle with a system incorporating a situation awareness processor;





FIG. 4

illustrates a general inference process of a situation awareness processor; and





FIG. 5

illustrates an example of an inference process of a situation awareness processor.











There exists a need for an improved predictive collision sensing or collision avoidance system for automotive applications that can sense and identify an environment of a host vehicle with sufficient range and accuracy so that proper countermeasures can be selected and taken sufficiently early to either avoid a collision; or to mitigate injury therefrom either to occupants of the host vehicle, or to pedestrians outside thereof. As used herein, the term predictive collision sensing system will also refer to a collision avoidance system, so as to mean a system that can sense and track targets in the environment of the host vehicle, and then either suggest, or automatically invoke, countermeasures, that would improve safety. Generally, a predictive collision sensing system tracks the motion of the host vehicle relative to its environment, or vice versa, for example, using a radar system with an associated target tracker. The environment may include both stationary and moving targets. An automotive environment is distinguished from other target tracking environments—for example, that of air or sea vessels—in that automotive vehicles are primarily operated in an environment that is constrained by roadways. There are, of course, exceptions to this, for example, parking lots or off-road driving conditions, but these exceptions generally account for a relatively small percentage of vehicular operating time, or for a relatively small risk of collisions for which there would be benefit from a predictive collision sensing system.




Referring to

FIGS. 1-3

, a situation awareness processor


10


is illustrated in a radar processing system


12


that processes radar data from a radar system


14


incorporated in a host vehicle


16


. Generally, the host vehicle


16


incorporates a sensor for sensing targets


18


in the environment thereof, and incorporates a system for actuating and/or controlling associated countermeasures responsive to the relative motion of the host vehicle


16


with respect to one or more targets. The sensor for sensing targets


18


is, for example, a radar or lidar sensor system that senses and tracks the location of targets


18


relative to the host vehicle


16


, and predicts if a collision between the host vehicle


16


and a target


18


is likely to occur, for example, as disclosed in commonly owned U.S. Pat. No. 6,085,151 that is incorporated herein by reference.




Referring to

FIG. 1

, the host vehicle


16


is seen traveling in the X-direction at a host speed


20


in the environment of one or more targets


18


, each at an associated bearing


22


and distance


24


relative to the host vehicle


16


, each having an associated target speed


26


at an associated heading


28


. Generally, the targets


18


may be either stationary, or moving, relative to the environment


30


of the host vehicle


16


.





FIG. 2

illustrates a block diagram of the radar processing system


12


, wherein each block in the block diagram comprises associated software modules that receive, prepare and/or process data provided by a radar system


14


mounted in the host vehicle. Data from the radar system


14


is preprocessed by a preprocessor


32


so as to generate radar data, for example, range, range rate, and azimuth angle, suitable for target tracking. The radar data typically covers a relatively wide field of view forward of the host vehicle, at least +/−5 degrees from the host vehicle's longitudinal axis, and possibly extending to +/−180 degrees, or larger, depending upon the radar and associated antenna configuration. A present exemplary system has a field of view of +/−55 degrees.




A tracker


34


converts radar output data (range, range rate & azimuth angle) into target speed and X-Y coordinates specifying the location of a target. For example, the system of U.S. Pat. No. 6,085,151 discloses a system for tracking multiple targets, and for clustering associated radar data for a single target. An associator


36


relates older track data to that from the latest scan, compiling a track history of each target.




A track map generator


38


generates a track map comprising a grid of heading and quality information from the track data as a record of the evolution of target motion relative to the host vehicle. The track map is updated with track data from subsequent targets


18


that enter the field-of-view of the host vehicle


16


, and old data fades from the map with time. Accordingly, the track map provides a representation of the paths followed by targets


18


relative to the host vehicle


16


, which paths are normally constrained by the associated roadways in the environment


30


of the host vehicle


16


. For example, a track map generator


38


is disclosed in a separately filed U.S. application Ser. No. 09/877,493, entitled Track Map Generator, filed on Jun. 8, 2001.




A situation awareness processor


10


uses 1) the track map, 2) data acquired indicative of the position and/or motion of the host vehicle


16


, i.e. host vehicle information


40


, and possibly 3) environment data


42


, to determine the most likely or appropriate driving situation from a set of possible driving situations. For example, the environment data


42


can include navigation data from a navigation system


44


, digital maps, real-time wireless inputs of highway geometry and nearby vehicles, and data from real-time transponders such as electromagnetic or optical markers built into highways; and can be used—along with target tracking data from the tracker


34


/associator


36


—by a road curvature estimator


46


to provide an estimate of road curvature to the situation awareness processor


10


. For example, navigation data such as the location and direction of a vehicle can be measured by a GPS receiver; by a dead reckoning system using measurements of vehicle heading from a compass or directional gyroscope, and vehicle distance and heading from wheel speed or rotation measurements, in conjunction with a map matching algorithm; or a combination thereof.




The situation awareness processor


10


stores and interprets the track map from the track map generator


38


, and compares the progress over time of several target tracks. Evaluation of the relative positions and progress of the tracked targets permits identification of various driving situations, for example a location situation, a traffic situation, a driving maneuver situation, or the occurrence of sudden events. Several approaches can be used to identify the situation, for example set theoretic reasoning (using for example random set theory or evidential reasoning); Bayesian inference; or a neural network.




Examples of location situations include a divided or undivided highway, an intersection, a freeway entrance or exit, a parking lot or off-highway situation, and a stopped object on the left or right of the host vehicle. Examples of traffic situations include crowded traffic, loose traffic, or normal traffic. Examples of driving maneuver situations include target cut-in, host vehicle lane or speed changing, or target speed changing.




The situation estimated by the situation awareness processor


10


, together with collision and crash severity estimates from a collision estimator


48


and a crash severity estimator


50


respectively, are used as inputs to a response generator


52


to select an appropriate countermeasure


54


, for example, using a decision matrix. The decision of a particular response by the response generator


52


may be based on, for example, a rule-based system (an expert system), a neural network, or another decision means.




Examples of countermeasures


54


that can be activated include, a warning device to warn the driver to take corrective action, for example 3D audio warning (for example, as disclosed in commonly owned U.S. Pat. No. 5,979,586 that is incorporated by reference herein); various means for taking evasive action to avoid a collision, for example the engine throttle, the vehicle transmission, the vehicle braking system, or the vehicle steering system; and various means for mitigating injury to an occupant if a collision is unavoidable, for example a motorized safety belt pretensioner, or internal or external airbags. The particular one or more countermeasures


54


selected, and the manner by which that one or more countermeasures


54


are activated, actuated, or controlled, depends up the situation identified by the situation awareness processor


10


, and upon the collision and crash severity estimates. By way of example, one potential scenario is that the response to encroachment into the host's lane of travel would be different depending upon whether the target is coming from the opposite direction or going the same way as the host vehicle


16


, but cutting into the lane thereof. By considering the traffic situation giving rise to the threat, the countermeasures


54


can be better adapted to mitigate that threat. By using a radar system, or generally a predictive collision sensing system, to sense targets within range of the host vehicle, the countermeasures


54


may be implemented prior to an actual collision so as to either avoid a collision, or to mitigate occupant injury therefrom.




Referring to

FIG. 3

, the radar system


14


and radar processing system


12


are seen incorporated in host vehicle


16


, wherein the host vehicle information


40


is provided by associated sensors, for example, operatively connected to a vehicle processor


56


that provides the resulting host vehicle information


40


to the situation awareness processor


10


. For example, one or more wheel speed sensors


58


can provide wheel speed measurements from one or more associated wheels of the vehicle, the average of which, for laterally opposed wheels, provides a measure of vehicle speed; the difference of which provides a measure of yaw rate. Alternately, or additionally, the vehicle speed may be measured from the rotation of a transmission output shaft. A steering angle sensor


60


operatively connected to the steering shaft provides a measure of steering angle responsive to the turning of the steering wheel by the driver. The yaw rate of the host vehicle


16


may alternately, or additionally, be sensed by a gyro


62


, wherein the integration of yaw rate provides a measure of the heading of the host vehicle


16


.




The situational awareness processor


10


transforms the sensor information collected by various sensors into a coherent understanding of the driving environment, such that a more proper response can be invoked responsive to an emergency. Stated in another way, the situational awareness processor


10


interprets the meaning of sensor data. The inputs to the situational awareness processor


10


are transformed from a “state” domain into an “event” domain. Whereas variables in state domain are usually represented numerically, the variables in event domain are commonly represented as events and by a confidence of the occurrence thereof. The situational awareness processor


10


identifies an “event” corresponding to particular a “situation”. Tables 1-3 provide examples, and associated definitions, of various events.




Inputs for identifying relevant events are obtained by processing the original sensor reports including radar reports, host vehicle speed and yaw rate, GPS host location report and digital map report. These inputs include, but are not limited to, those listed in Table 4, the information of which, except for host absolute location, type of road and approaching road structure, is interrelated by the coordinate system of FIG.


1


. Host absolute location is defined in the coordinate system of the digital map, and the type of road and approach road structures are expressed in terms of events.













TABLE 1









Event







ID




Events Related to Host Vehicle Location

























A1




ivided Highway




Traffic on the opposite direction is separated








relatively far away






A2




ndivided




There is traffic on the opposite direction







ighway




nearby, with most traffic parallel and at








a relatively high speed






A3




ntersection




Traffics with different, intersecting directions






A4




reeway




Divided Highway with no Intersection, and a








relatively higher speed than for a Highway






A5




reeway Entrance




Connection into a freeway from another high-








way of any type






A6




reeway Exit




Connection out of a freeway into another high-








way of any type






A7




arking Lot




Place where there is a collection of stopped








vehicles, in some cases located irregularly,








and where the host vehicle is required








to perform tight maneuvers






A8




ff-Highway




Place where the surrounding environment is








irregular, and where the host vehicle may








perform almost any kind of








maneuver























TABLE 2









Event







ID




Events Related to Traffic

























B1




rowded Traffic




A higher volume of traffic for a particular type








of location






B2




oose Traffic




A lower volume of traffic for a particular type








of location






B3




ormal Traffic




A standard volume of traffic for a particular








type of location






B4




raffic on Left




Another moving object on the left of the host






B5




raffic on Right




Another moving object on the right of the host






B6




raffic in Lane




Another moving object within the lane of








the host






B7




topped Traffic




A stopped object on the left of the host







on Left






B8




topped Traffic




A stopped object on the right of the host







on Right






B9




topped Traffic




A stopped object within the lane of the host







in Lane























TABLE 3









Event







ID




Events Related to Relative Position Change

























C1




Left Traffic




An object is cutting-in the host's lane







Cut-in




from the left






C2




Right Traffic




An object is cutting-in the host's lane







Cut-in




from the right






C3




Host Lane




The host vehicle is moving into the left lane







Changing to







Left Lane






C4




Host Lane




The host vehicle is moving into the right lane







Changing to







Right Lane






C5




Left Traffic




An object to the left of the host increases







Accelerating




its speed






C6




Right Traffic




An object to the right of the host increases







Accelerating




its speed






C7




Lead Traffic




An object in the front of the host increases its







Accelerating




speed






C8




Left Traffic




An object to the left of the host decreases







Decelerating




its speed






C9




Right Traffic




An object to the right of the host decreases







Decelerating




its speed






C10




Lead Traffic




An object in the front of the host decreases







Decelerating




its speed






C11




Host Accelerating




The host speed increases






C12




Host Decelerating




The host speed decreases






C13




Traffic Turn Left




An object changes its heading to








the left relative to host






C14




Traffic Turn




An object changes its heading to the right







Right




relative to host






C15




Host Turn Left




The host changes its heading to the left






C16




Host Turn Right




The host changes its heading to the right






















TABLE 4











Inputs to Situation Awareness Processor













Type




Input




Source of Information









Target




Number of Targets




Obtained from processing forward






Related




Target Heading




looking sensor(s), such as radar,






Information




Target Speed




lidar, or camera, with tracking







Target Location




algorithm







Relative to Host






Host Related




Host Speed




Vehicle speed sensor in the






Information





transmission, or wheel speed








sensors, of host vehicle







Host Yaw Rate




Gyro, or differential wheel speed







Host Heading




GPS, compass, gyro and/or








differential wheel speed







Host Absolute




Host navigation system: GPS;







Location




or dead reckoning from








compass and wheel rotation;








with map matching






Environment




Track Map




A recording of previously obtained






Information





target trajectories into grids








covering the field-of-view, wherein








information carried by previous








tracks is saved in a








smoothed manner







Road Curvature




Obtained from digital map, or








processing of host yaw








rate and target information







Type of Road




Obtained from a digital map through







Approaching Road




matching with the







Structure




host absolute location














The output of the situational awareness processor


10


comprises a list of identified events, and associated confidence levels. The events collectively define the current driving situation. The confidence can be represented in several ways, such as probability and possibility, each generated by a different inference (reasoning) mechanism, as discussed further hereinbelow. The outputs are determined by an inference process, which is a recursive decision making process that correlates events with their associated and corresponding inputs. A decision is given in terms of a confidence for the occurrence of a particular event. A general formula for calculating the confidence of an event can is given as:








p




k




=f


(


p




k-1




,e




1




,e




2




,K,e




n


)  (1)






where p


k


is the confidence level, and e


i


are the inputs, at time k.




Various methods can be used in the inference engines that specify the recursive updating mechanism. Among them, the most widely used are 1) Dempster-Shafer Evidential Reasoning, 2) Fuzzy Logic Reasoning, 3) Bayesian Inference, and 4) Neural Network. Random set theory can also used in place of the evidential reasoning or fuzzy logic reasoning as a more general approach.




In preparation for using a particular inference engine, the above-described events are first grouped into several classes, each containing a number of mutually exclusive atomic events. Inference is made within each class, although the output of one inference engine can be fed back as an input to another inference engine. A decision on the nature of the current driving situation is made responsive to the collective output of all inference engines. For example, Table 5 lists a set of reorganized event classes based upon the events listed in Tables 1-3.




The decision process based on the reorganized events essentially forms a networked inference engine, which is illustrated generally in

FIG. 4

, and which is illustrated more specifically in FIG.


5


.




By way of example, the decision process is now illustrated by deriving a Bayesian inference of Approaching Road Structure. The Bayesian formula is given by:









p


(


A
i

|
B

)




=





p


(

B
|

A
i


)




p


(

A
i

)




p


(
B
)












=





p


(

B
|

A
i


)




p


(

A
i

)







j
=
1

n








p


(

B
|

A
j


)




p


(

A
j

)

















Denoting e


κ


, κ=a, b, Λ k, as all the information available to the inference engine, and {e


κ


} as the collection of all available inputs, e


a


collectively refers to all the information carried by the track map, which can be further decomposed into a set of G grids, i.e. e


a


{e


ah




ξ


,e


aq




ξ


,ξ=1, 2, Λ G}, where e


ah




ξ


and e


aq




ξ


are heading and quality information of grid ξ, respectively. For N targets, then e


e


=N. Superscripts can be used to distinguish targets, e.g. e


b




r


, e


c




r


and e


d




r


represent speed, heading, and location of target r, r=1, 2 . . . N, respectively. Of the target information, each e


d




r


contains two elements, defining a target's location in X and Y coordinates, i.e. e


d




r


=(x


r


, y


r


) Road type and approaching road structure are events in nature. Each event is from a class of possible events as defined earlier. Generally, a superscript can be used to differentiate an event from peers in its class, e.g. e


i




α


and e


j




β


specifies a road type and an approaching road structure, respectively.

















TABLE 5











Class






Event







ID




Class Name




Events in Class




ID(s)













CA




Approaching




Intersection




A3








Road




Freeway Entrance




A5








Structure




Freeway Exit




A6









Unknown







CB




Road Type




Divided Highway




A1









Undivided Highway




A2









Freeway




A4









Parking Lot




A7









Off-Highway




A8









Unknown







CC




Traffic




Crowded




B1








Volume




Loose




B2









Normal




B3









Unknown







CD




Traffic




Left




B4








Location




Right




B5









In Lane




B6









Unknown







CE




Stopped




Left




B7








Object




Right




B8








Location




In Lane




B9









Unknown







CF




Traffic




Cut-in From Left




C1








Maneuver




Cut-in From Right




C2









Turn Left




C13









Turn Right




C14









Straight Acceleration




C5, C6, C7









Straight Deceleration




C8, C9, C10









Nonmaneuver







CG




Host




Lane Changing to




C3








Maneuver




Left









Lane Changing to




C4









Right









Turn Left




C15









Turn Right




C16









Straight Acceleration




C11









Straight Deceleration




C12









Nonmaneuver















The output of the inference network comprises the seven events from the seven classes defined earlier, wherein event is denoted as S, subscripts are used to denote the class of an event, and superscript are used to denote the intended event of the class specified by the subscript. Accordingly, S


η




λ


is an output event for class η and event λ in the class. For example, if η=1, λ=2, then S


η




λ


means freeway entrance.




Given the above definitions, the Bayesian inference for Approaching Road Structure can be derived as follows. Under Bayesian framework, the outcome of an inference is the posterior probability of a given event. For approaching road structures, the probabilities to be evaluated are S


1




λ


, λ=1, 2, 3, 4, given the collection of evidence {e


κ


} and knowledge of a related event in class S


2


, are given by:







p


(



S
1
λ

|

S
2

λ
0



,

{

e
κ

}


)


=



p


(



{

e
κ

}

|

S
1
λ


,

S
2

λ
0



)




p


(


S
1
λ

|

S
2

λ
0



)








λ
1

=
1

4








p


(



{

e
κ

}

|

S
1

λ
1



,

S
2

λ
0



)




p


(


S
1

λ
1


|

S
2

λ
0



)















wherein, λ


0


=1, 2, Λ 6 is the index for events belonging to the road type class.




Using the previously obtained p(S


1




λ


|S


2




λ






0




, {e


κ


}) to approximate current p(S


1




λ


|S


2




λ






0




), provides the following recursive probability updating equation:








p
k



(



S
1
λ

|

S
2

λ
0



,

{

e
κ

}


)


=



p


(



{

e
κ

}

|

S
1
λ


,

S
2

λ
0



)





p

k
-
1




(



S
1
λ

|

S
2

λ
0



,

{

e
k

}


)








λ
1

=
1

4








p


(



{

e
κ

}

|

S
1

λ
1



,

S
2

λ
0



)





p

k
-
1




(



S
1

λ
1


|

S
2

λ
0



,

{

e
k

}


)















wherein p({e


κ


}|S


1




λ


, S


2




λ






0




) is the probability density of set {e


κ


} given the condition that events S


1




λ


and S


2




λ






0




are true. Considering the individual elements in {e


κ


} gives:







p


(



{

e
κ

}

|

S
1
λ


,

S
2

λ
0



)


=


p


(


e
a

,

{


e
b
r

,

e
c
r

,

e
d
r


}

,

e
f

,

e
i
α

,


e
j
β

|

S
1
λ


,

S
2

λ
0



)


=


p


(



e
a

|

S
1
λ


,

S
2

λ
0



)




p


(



e
f

|

S
1
λ


,

S
2

λ
0



)




p


(



e
i
α

|

S
1
λ


,

S
2

λ
0



)




p


(



e
j
β

|

S
1
λ


,

S
2

λ
0



)


×




r
=
1

N




p


(



e
b
r

|

S
1
λ


,

S
2

λ
0



)




p


(



e
c
r

|

S
1
λ


,

S
2

λ
0



)




p


(



e
d
r

|

S
1
λ


,

S
2

λ
0



)
















wherein individual conditional probabilities/probability densities can be found as:










p


(



e
a

|

S
1
λ


,

S
2

λ
0



)









p


(



e
a

|

S
1
λ


,

S
2

λ
0



)


=




ξ
=
1

G



[



p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
TRUE


)




p


(


ξ
=

TRUE




|

S
1
λ



,

S
2

λ
0



)



+


p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
FALSE


)




p


(


ξ
=

FALSE
|

S
1
λ



,

S
2

λ
0



)




]











As






p


(


ξ
=

TRUE




|

S
1
λ



,

S
2

λ
0



)




]


=



e
aq
ξ






and






p


(


ξ
=

FALSE
|

S
1
λ



,

S
2

λ
0



)




]


=

1
-

e
aq
ξ




,




then





(
1
)

















p


(



e
a

|

S
1
λ


,

S
2

λ
0



)


=




ξ
=
1

G




[



p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
TRUE


)




e
aq


+






p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
FALSE


)




(

1
-

e
aq
ξ


)



]

.












A suitable way to model p(e


ah




ξ


|S


1




λ


, S


2




λ






0




, ξ=TRUE) is to a use Guassian mixture, i.e. an approximation involving a summation of several Guassian distributions:







p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
TRUE


)


=


1

N
f







i
=
1


N
f









1

2


πσ
i









(


e
ah
ξ

-

μ
i


)

2


2


σ
i
2

















where N


ƒ


is the number of features selected for an approaching road structure. For example, the following parameters may be chosen for an intersection: N


ƒ


=3, μ


1


=0°, μ


2


=90°, μ


3


=90°, and σ


1





2





3


=10°.




The distribution p(e


ah




ξ


|S


1




λ


, S


2




λ






0




, ξ=FALSE) may be modeled as a uniform distribution, e.g.,







p


(



e
ah
ξ

|

S
1
λ


,

S
2

λ
0


,

ξ
=
FALSE


)


=


1

180

°


.











(2) p(e


b




r


|S


1




λ


, S


2




λ






0




), p(e


c




r


|S


1




λ


, S


2




λ






0




), p(e


d




r


|S


1




λ


, S


2




λ






0




) and p(e


ƒ


|S


1




λ


, S


2




λ






0




) can be handled similar to p(e


a


|S


1




λ


, S


2




λ






0




).




(3) p(e


i




α


|S


1




λ


, S


2




λ






0




):




Probability of event-type input e


i




α


can be decided as







p


(



e
i
α

|

S
1
λ


,

S
2

λ
0



)


=

{





p


(


e
i
α

=
TRUE

)


,





if






S
2

λ
0



=





e
i
α








1
-

p


(


e
i
α

=
TRUE

)



,





if






S
2

λ
0





e
i
α
















Where p(e


i




α


=TRUE) can be obtained by the quality measure of GPS/map matching.




(4) p(e


i




β


|S


1




λ


, S


2




λ






0




) can be handled similar to p(e


i




α


|S


1




λ


, S


2




λ






0




).




The actual forms and parameters of above probability terms would be obtained based on experiments, with simplification to reduce computational burden.




The above inference engine can be replaced by any of the other three widely used reasoning techniques, namely fuzzy logic reasoning, Dempster-Shafer evidential reasoning and neural networks. Particularly, a collection of atomic events can be specified based on the classes of events to be recognized, for example, as are illustrated in Table 6.
















TABLE 6











Set




Set




Events in Set




Event

















ID




Name




ID




Name




ID(s)




Definition









SA




Approaching




SA1




Intersection




3




Traffics at different







Road







directions meet each







Structure







other








SA2




Entrance




5




A road connected into











the host's road in a











near parallel manner








SA3




Exit




6




A road connected out











of the host's road in











a near parallel manner






SB




Road Type




SB1




Divided




1




Road with only one









Road





direction of traffic








SB2




Undivided




2




Traffic with opposite









Road





direction is close








SB3




Highway




1, A2




Road of any type with











higher speed and











traffics are parallel








SB4




Parking Lot




7




Same as A7








SB5




Off-Highway




8




Same as A8






SC




Traffic




SC1




Crowded




1




Same as B1







Volume




SC2




Loose




2




Same as B2








SC3




Normal




3




Same as B3






SD




Traffic




SD1




Left




4, B7




An object is on the left







Location







of the host








SD2




Right




5, B8




An object is on the











right of the host








SD3




In Lane




6, B9




An object is in the











same lane of the host






SE




Object




SE1




Moving





The location of an







Nature







object is changing











relative to the ground








SE2




Stopped





The location of an











object is not changing











relative to the ground






SF




Traffic




SF1




Turn Left




13




Heading of the







Maneuver







velocity of an object











changes to the left of











the host








SF2




Turn Right




14




Heading of the











velocity of an object











changes to the right











of the host








SF3




Acceleration




5, C6,




Object speed increases










C7








SF4




Deceleration




8, C9,




Object speed decreases










C10






SG




Host




SG1




Turn Left




15




Heading of the







Maneuver







velocity of host











changes to the left











of the host








SG2




Turn Right




16




Heading of the











velocity of host











changes to the right











of the host








SG3




Acceleration




11




Host speed increases








SG4




Deceleration




12




Host speed decreases








SG5




Lane Change




3, C4




Host moves to another











lane














Note that the defined atomic events are not exactly the same as the situations. Some are intersections of several events, some are unions of events, some are actually part of a situation, and some have not been defined. However, these atomic sets collectively can represent the situations defined earlier.




Given the defined sets of atomic events, the output events can be obtained by joining the atomic events, as illustrated in Table 7.
















TABLE 7













A1 = (SB1, SB3)




B1 = (SC1)




C1 = (B4, SF1)







A2 = (SB2, SB3)




B2 = (SC2)




C2 = (B5, SF2)







A3 = (SA1)




B3 = (SC3)




C3 = (SG1, SG5)







A4 = (SB1, SB2, SA1)




B4 = (SD1, SE1)




C4 = (SG2, SG5)







A5 = (A4, SA2)




B5 = (SD2, SE1)




C5 = (SF3, B4)







A6 = (A4, SA3)




B6 = (SD3, SE1)




C6 = (SF3, B5)







A7 = (SB4)




B7 = (SD1, SE2)




C7 = (SF3, B6)







A8 = (SB5)




B8 = (SD2, SE2)




C8 = (SF4, B4)








B9 = (SD3, SE2)




C9 = (SF4, B5)









C10 = (SF4, B6)









C11 = (SG3)









C12 = (SG4)









C13 = (SF1)









C14 = (SF2)









C15 = (SG1)









C16 = (SG2)















The Dempster-Schafer mass functions assigned to the events can be obtained based on experiments. Mass updating can also obtained based on evidential reasoning theory.




While specific embodiments have been described in detail in the foregoing detailed description and illustrated in the accompanying drawings, those with ordinary skill in the art will appreciate that various modifications and alternatives to those details could be developed in light of the overall teachings of the disclosure. Accordingly, the particular arrangements disclosed are meant to be illustrative only and not limiting as to the scope of the invention, which is to be given the full breadth of the appended claims and any and all equivalents thereof.



Claims
  • 1. A method of identifying a situation in which a host vehicle is operated, comprising:a. selecting a plurality of events, wherein each event of said plurality of events is representative of a situation in which the host vehicle is operated, and at least two of said events are different from one another; b. selecting at least one set of said events that are related to a situation in which the host vehicle is operated; c. providing for a set of data from at least one of a first set of data, a second set of data and a third set of data, wherein said first set of data is representative of a target in a field of view of the host vehicle, said second set of data is representative of the position or motion of the host vehicle, and said third set of data is representative of an environment of said host vehicle; d. selecting an inference method for a first inference engine, wherein said inference engine comprises at least one input and an output; e. selecting as a first input to said inference engine at least one element of said first set of data, said second set of data and said third set of data; and f. generating a first output from said first inference engine responsive to said first input, wherein said first output is representative of a probability of occurrence of at least one event of said set of said events.
  • 2. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein a plurality of said events are related to a location of the host vehicle.
  • 3. A method of identifying a situation in which a host vehicle is operated, as recited in claim 2, wherein each of said plurality of said events related to said location of the host vehicle is a situation selected from an operation of the host vehicle in a vicinity of a divided or undivided highway; an operation of the host vehicle in a vicinity of an intersection, an operation of the host vehicle in a vicinity of a freeway, an operation of the host vehicle in a vicinity of an entrance or exit to a freeway entrance; an operation of the host vehicle in a vicinity of a parking lot; and an operation of the host vehicle in an off-highway environment.
  • 4. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein a plurality of said events are related to a traffic situation.
  • 5. A method of identifying a situation in which a host vehicle is operated, as recited in claim 4, wherein each of said plurality of said events related to a traffic situation comprises operation of the host vehicle in an environment selected from crowded, loose or normal traffic; a moving or stationary object to the left or right of the host vehicle; and a moving or stationary object in a lane of the host vehicle.
  • 6. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein a plurality of said events are related to a relative position change of a target with respect to the host vehicle.
  • 7. A method of identifying a situation in which a host vehicle is operated, as recited in claim 6, wherein each of said plurality of said events related to a relative position change of a target with respect to the host vehicle is selected from an object cutting in a lane of the host vehicle from the left or right of the host vehicle; the host vehicle changing lanes to the left or right; a change in speed of traffic to the left or right of the host vehicle; a change in speed of traffic in the same lane as the host vehicle; a change in speed of the host vehicle; a change in direction of traffic to the left or right; and a change in direction of the host vehicle to the left or right.
  • 8. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein at least two events of said set of events are mutually exclusive.
  • 9. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein said first set of data is selected from a count of targets in a field of view of the host vehicle, a target heading, a target speed, and a target location relative to the host vehicle, and at least a portion of said first set of data is generated by a tracker operatively coupled to a radar system.
  • 10. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein said second set of data is selected from a speed of the host vehicle, a yaw rate of the host vehicle, a heading of the host vehicle, and a location of the host vehicle.
  • 11. A method of identifying a situation in which a host vehicle is operated, as recited in claim 10, further comprising providing for reading said location of the host vehicle from a navigation system.
  • 12. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein said third set of data is selected from a track map representative of a composite path of targets in an environment of the host vehicle, a curvature or type of a road upon which the host vehicle is operated, data representative of a fixed structure in the field of view of the host vehicle, and a digital map of a road surface, wherein said track map is generated by a track map generator from data is generated by a tracker operatively coupled to a radar system.
  • 13. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, wherein said inference method is selected from a Dempster-Shafer evidential reasoning method, a fuzzy logic reasoning method, a Bayesian inference method, a neural network and a reasoning method based on random set theory.
  • 14. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, further comprising:a. selecting as a second input to said inference engine an output from an inference engine, wherein said output is at a past time relative to said first input; and b. generating an output from said first inference engine responsive to said first and second inputs.
  • 15. A method of identifying a situation in which a host vehicle is operated, as recited in claim 1, further comprising selecting a countermeasure responsive to at least one said output of at least one said inference engine.
Parent Case Info

The instant application claims the benefit of U.S. Provisional Application Serial No. 60/210,878 filed on Jun. 9, 2000 (5701-00266). The instant application is related to U.S. application Ser. No. 09/877,493, entitled Track Map Generator, filed on Jun. 8, 2001 (5701-01265),

US Referenced Citations (32)
Number Name Date Kind
2709804 Chance et al. May 1955 A
3177485 Taylor, Jr. Apr 1965 A
3603994 Williams et al. Sep 1971 A
3699573 Andrews et al. Oct 1972 A
3725918 Fleischer et al. Apr 1973 A
3869601 Metcalf Mar 1975 A
3971018 Isbister et al. Jul 1976 A
4623966 O'Sullivan Nov 1986 A
5051751 Gray Sep 1991 A
5138321 Hammer Aug 1992 A
5170440 Cox Dec 1992 A
5202691 Hicks Apr 1993 A
5307289 Harris Apr 1994 A
5406289 Barker et al. Apr 1995 A
5471214 Faibish et al. Nov 1995 A
5537119 Poore, Jr. Jul 1996 A
5594414 Namngani Jan 1997 A
5598164 Reppas et al. Jan 1997 A
5627768 Uhlmann et al. May 1997 A
5631639 Hibino et al. May 1997 A
5633642 Hoss et al. May 1997 A
5684473 Hibino et al. Nov 1997 A
5689264 Ishikawa et al. Nov 1997 A
5703593 Campbell et al. Dec 1997 A
5948043 Mathis Sep 1999 A
5959552 Cho Sep 1999 A
5959574 Poore, Jr. Sep 1999 A
6085151 Farmer et al. Jul 2000 A
6161071 Shuman et al. Dec 2000 A
6226389 Lemelson et al. May 2001 B1
6275231 Obradovich Aug 2001 B1
6275773 Lemelson et al. Aug 2001 B1
Non-Patent Literature Citations (2)
Entry
Goebel, Greg, “An Introduction to Fuzzy Control Systems V1.0,” Internet Document, source unknown, Dec. 23, 1995.
Goebel, Greg, “An Introduction to Fuzzy Control Systems V1.2,” Internet Document, http://vectorsite.tripod.com/ttfuzzy.html, Oct. 1, 2000; downloaded on Sep. 14, 2001.
Provisional Applications (1)
Number Date Country
60/210878 Jun 2000 US