This invention relates to tracking a body by nonlinear and non-Gaussian parametric filtering.
Tracking a body, for example a robot or other physical agent, as it moves in a physical space can be based on measurements that depend on the movement of the body in the space. For example, sensor measurements may provide partial or noisy information regarding the position of the body. Over time, such measurements can be accumulated and combined with information related to the anticipated movement of the body to track the motion of the body. In some cases, the tracking is probabilistic in that the knowledge that is gleaned from the measurements is represented as a probability distribution representing the uncertainty in the position. Generally, the process of accumulating the measurements and combining them with the anticipated movement is referred to as a type of filtering. The characteristics of the body, for example, the position, velocity, pose, etc. are together referred to as the “state” of the body.
Filtering can be applied to problems in which a state of a body evolves according to system dynamics and deterministic and/or random inputs to the system, and the system is observered via measurements (“observations”) that depend on the state. A goal of filtering can include recovering an estimate of the evolution of the system state based on measurements. Such filtering of the measurements to recover the system state can be a challenging problem, for example, because the measurements are noisy (i.e., have high degree of uncertainty) or because there may not be a direct way to derive the system state from the measurements even if they were perfectly known. One example where such filtering is useful is in navigation in which the system state to be recovered includes the position and velocity of a vehicle and the measurements include reference based (e.g., GPS or beacon based) and/or interial (e.g., accelerometer or gyroscope based) sensor measurements of the vehicle's position. In the discussion below, the variables y1, y2, . . . refer to the state of the system at ordered discrete times t1, t2, . . . , and the variables z1, z2, . . . refer to the measurements at those times.
One approach to dealing with the uncertainty in the state evolution or the relationship between the system state and sensor measurements is to use probabilistic models, such as a model for the distribution p(yk|yk−1) that characterizes the evolution of the state from time tk−1 to time tk, and a distribution p(zk|yk) that characterizes the measurement distribution at time tk given a particular value yk of the system state at time tk. Further assumptions on the probability distributions can be made, including a Markov assumption on the system evolution p(yk|yk−1, . . . , y1)=p(yk|yk−1) and a conditional independence assumption on the observations such that p(z1, . . . , zk|y1, . . . , yk)=πip(zi|yi).
The Markov and conditional independence assumptions used in a number of filtering techniques in which the distribution of the state at time tk−1 based on all measurements up to that time, denoted p(yk−1|z1:k−1) where z1:k−1 refers to the sequence z1, . . . , zk−1, is updated based on zk to yield the distribution p(yk|z1:k). In particular, Bayes rule is used as the basis of two steps:
The prediction step may be implemented in the case of a continuous state space based on the equation
b) Update equation:
which may be broken up into two steps:
When the initial distribution of the system state is Gaussian as well as both dynamic and measurement models are linear-Gaussian, the prediction and update steps retain the Gaussian nature of the distributions, leading to simple closed-form solutions. The recursion given by such analytic solution is the well-known Kalman Filter. However, linear and Gaussian models are not adequate for navigation systems deployed in harsh environments, causing Kalman filters to suffer severe performance degradations. Other well-know approaches to implementing the prediction and update steps include use of a discrete state space in a Hidden Markov Model.
One approach to dealing with non-linear form of the system or observation model is to use linearize approximations of the models. Generally, some such approaches are referred to as Extended Kalman Filters.
Nonparametric techniques based on Monte Carlo sampling have also been used to handle such complex scenarios but with a computational complexity beyond the reach of many navigation devices, especially those that are small and low-cost. One such approach is refered to as a Particle Filter. Generally, such approaches represent the posterior distribution of the positional state as a set of random samples with associated weights (i.e., a probability density composed of a set of weighted impulses), each representing a “particle.” The prediction and update steps generally involve moving the samples (or possibly removing and adding samples) and determining new weights for the samples.
Another technique called Rao-Blackwellized particle filter takes advantage of the specific structure by marginalizing out some of the variables analytically and solving the remaining variables using particle filters. However, since this method needs to partition the state-space, it is only useful in some specific cases.
The dimension of a filter is the number of real numbers that the filter has to store in each step, where this dimension is a measure of the complexity of the filter. For instance, Kalman-like filters have to store a mean and a covariance matrix of a state vector of size d , therefore their dimension is d(1+(d+1)/2). On the other hand, particle filters have a dimension of M(d+1), where M is the number of particles which usually has to be several orders of magnitude larger than d to achieve acceptable performances.
Each navigation problem has an underlying structure that is determined by the dynamic and measurements models. Generally, traditional techniques for navigation such as the Kalman filter do not capture the nonlinear and non-Gaussian behavior of measurements in wireless localization and navigation systems deployed in harsh environments. Nonparametric filters as particle filters can cope with the measurements behavior in such environments at the expense of a computational complexity beyond the reach of low-cost navigation devices.
In essence, suboptimal filtering techniques choose a family of distributions and perform the prediction, update, and normalization steps in such a way that the estimated posterior always lies within the same family. For example, approaches such as EKF or UKF choose the family of Gaussian distributions and make the approximations by Taylor series expansions or deterministic sample points, respectively. Particle filters use mixtures of deltas as the family to approximate the distribution. The performance of each technique in terms of accuracy and complexity depends on the suitability of the family and accuracy/complexity of the approximation process.
In one aspect, in general, a new parametric filtering technique is applicable to nonlinear and/or non-Gaussian navigation problems. In some aspects, the methodology decomposes the problem in such a way that a linear-Gaussian part can still be solved in closed form, while the remaining computations are solved by a new technique for condensing a high dimensional distribution to a low dimensional one, which is referred to herein as Belief Condensation (BC). This condensation involves obtaining the distribution, which belonging to a certain family, closest (e.g., measured by Kullback-Leibler (KL) divergence) to the resulting distribution after the filtering steps. The BC filter (BCF) has a performance similar to particle filter, but with a much smaller dimension.
In another aspect, in general, a moving body having state y is tracked according to state-dependent measurements z from one or more sensors. For time each of a series of times tk, a data representation characterizing a distribution (e.g., a probability distribution) of the state of the body at that time in the series is formed based on measurements up to that time. This distribution for that time tk belongs to a family of distributions corresponding to that time. Forming the data representation characterizing the probability distribution of the state of the body at time tk includes accepting a data representation ξk−1 characterizing a distribution of the state yk−1 of the system at a previous time tk−1 without consideration of a measurement zk that depends on the state yk for the time tk in the series, for example, based only on measurements up to that previous time. This distribution belongs to a parametric family −1. A data representation of the measurement zk that depends on the state yk for the time tk in the series is also accepted. A first data representation characterizing a first distribution of the state yk of the body at the time tk in the series based on the measurements up to the previous time and the measurement for the time in the series is then determined. This distribution does not belong to the family corresponding to that time tk. A second data representation ξk characterizing a second distribution of the state yk of the body at the time tk in the series based on the measurements up to the previous time and the measurement for the time in the series is then determined. This second distribution does belong to the family corresponding to that time tk and forms an approximation of the first distribution of the state yk.
Aspects can include one or more of the following features.
Each family corresponding to each time tk comprises a family of probability distributions.
Each family corresponding to each time tk is a family of mixture probability distributions.
Each family corresponding to each time tk is a family of mixtures of exponential probability distributions.
The family of distributions for a plurality of times is a same family.
The first data representation characterizing a first probability distribution of the state yk of the system at the time tk comprises parameters defining a parametric distribution.
Forming the first data representation characterizing a first probability distribution of the state yk of the system at the time tk includes forming a data representation characterizing a predicted probability distribution of the state yk of the system at the time tk in the series based on the measurements up to the previous time tk−1, and forming the data representation characterizing the first probability distribution from using the measurement zk for the time in the series and the predicted probability distribution.
Determining the second data representation 4 characterizing a second distribution of the state yk of the system at the time tk includes iteratively forming a series of data representations μk[n] representing corresponding successive approximations of first probability distribution of the state of the body .
Iteratively forming the series of data representations includes, at each iteration, determining expected values of statistics of a distribution of the state of the body based on the measurement zk for the time in the series and the first probability distribution of the state of the body.
Accepting the data representation of the measurement zk includes accepting a data representation of a measurement from a sensor belonging to the group consisting of location-reference based sensor and an inertial sensor.
The moving body is a mobile device, a vehicle, and/or a pedestrian.
In another aspect, in general, a tracking system is adapted for tracking a moving body having state y , according to state-dependent measurements z from one or more sensors. The tracking system comprising a tracking filter coupled to a sensor system. The tracking filter comprises a predictor configured to accepting a data representation ξk−1 characterizing a distribution of the state yk−1 of the system at a each time tk−1 in a series of times based on measurements up to that previous time, the distribution belonging to a parametric family −1. The tracking filter further comprises an updater configured to accept a data representation of a measurement zk that depends on the state yk for each time tk in the series of times, form a first data representation characterizing a first distribution of the state yk of the body at the time tk in the series based on the measurements up to the previous time and the measurement for the time in the series, the distribution not belonging the family corresponding to that time tk, and determining a second data representation μk characterizing a second distribution of the state yk of the body at the time tk in the series based on the measurements up to the previous time and the measurement for the time in the series. The second distribution belongs to the family corresponding to that time tk and forms an approximation of the first distribution of the state yk.
In another aspect, in general, a tracking method includes maintaining a data storage for data including data characterizing a distribution of a location of a mobile device as a mixture of a plurality of distribution components. The following steps are repeated at each of a series of times. A data representation of a distribution of a predicted location of the device formed based on the distribution of the location of the device and data representing motion characteristics of the device. Positioning system measurements associated with a location of the device at a sensing time are accepted. A data representation characterizing a first distribution of an updated location of the device at the sensing time is formed based on the distribution of the predicted location and the sensor measurements, including identifying a plurality of locations and corresponding weights. The data characterizing the distribution of the current location of the device is iteratively updated. The distribution of the current location forms an approximation of the first distribution of the updated location according to the identified locations and weights. In some examples, the positioning system measurements include pseudorange measurements between the device and a plurality of reference locations, which may be subject to non-linear distortions (e.g., multipath distortion and a non-line-of-sight distortion) of true range measurements between the device and the plurality of reference locations.
In another aspect, in general, software stored on a computer-readable medium comprises instructions for causing a data processing system to perform all the steps of any of the methods presented above.
In another aspect, in general, a software implemented tracking system includes a data processing system configured to perform all the steps of any of the methods presented above.
Aspects can have one or more of the following advantages.
Accurate tracking of a body is achieved with less computation than can be achieved using other available techniques, particularly in situation in which sensor measurements represent substantially non-linear functions of the system state being tracked.
Other features and advantages of the invention are apparent from the following description, and from the claims.
Referring to
In the discussion below, we assume that the state yk of dynamic system 110 is represented as a vector of real-valued quantities. Furthermore, in at least some embodiments, the system state is assumed to evolve according to a known probabilistic model, for example, a linear Gaussian model yk+1=Hkyk+nk in which nk is a Gaussian random variable. In other emboduments, the state evolution is assumed to be according a more generally non-linear non-Gaussian model p(yk|yk−1). Similarly, the sensor system is assumed to provide an observation according to a model p(zk|yk).
As introduced above, exact filtering is not possible using available techniques for continuous-state models in general cases where dynamic or measurements models are nonlinear and/or non-Gaussian. In those cases the exact posterior distribution in each time step cannot be determined in closed-form based on the prediction and update steps, and hence one approach is to track the sequence of posterior distributions through approximations. Referring to
Generally, the BCF approximates the complex distributions emerging in the filtering steps by a parametric distribution that is suitable for inference and propagation in next time instants. In some examples, the approximate distributions computed by the BCF belong to a family of distributions that belongs to exponential families or mixtures of exponential families. As an example, the BCF 150 calculates distributions {circumflex over (p)}(yk|z1:k) that are mixtures of multivariate Gaussian distributions.
At least conceptually, and in some implementations using modules and/or software procedures corresponding to the steps, the BFC 150 can be considered to implement the following steps:
Referring to
The predictor 160 determines and outputs a representation of the predicted distribution {circumflex over (p)}(yk|z1k−1). For example, in a case that the system dynamics are linear-Gaussian, (i.e., determined by the system matrix Hk−1 and mean and covariance of the input nk−1), the representation of the distribution is again a Gaussian mixture representable by the set of mixture weights, means, and covariances of the mixture elements.
A condensed updater 170 accepts the representation of {circumflex over (p)}(yk|z1:k−1) and the sensor measurement zk, and outputs a representation of the updated distribution {circumflex over (p)}(yk|z1:k). For example, in the case of the family of distributions being m -element mixtures of Gaussian distributions, the output of the condensed updater ζis again a set of mixture weights, means, and covariances of the mixture elements.
In some implementations, the condensed updater 170 is implemented by an updater 172 followed by a condenser 174. The updater 172 accepts the representation of {circumflex over (p)}(yk|z1:k−1) and the sensor measurement zk. The updater is configured with or accepts a representation of the measurement distribution p(zk|yk). In general, this distribution does not have a Gaussian form. For example, the distribution of zk may depend in significantly non-linear fashion on the system state yk. In some examples, the updater 172 outputs a representation of the updated distribution {circumflex over (p)}(yk|z1:k). Such a representation may comprise parameter values for a parametric distribution. In some examples, the representation may comprised parameter values for configuring a numeric procedure for pointwise evaluation of the function. In some examples, rather than providing the representation of {tilde over (p)}(yk|z1:k), which might allow evaluation of the distribution value at specific points yk in the state space, the updater may be configured to accept specific points yk in the state space and evaluate (or approximate) {tilde over (p)}(yk|z1:k) at those specific points.
The condenser 174 then forms the approximation {circumflex over (p)}(yk|z1:k) corresponding to the distribution {tilde over (p)}(yk|z1:k) produced by the updater 172. The distribution {circumflex over (p)}(yk|z1:k) belongs to the family of distributions , and for the discussion below, is assumed to be parameterized by a set of parameters ξk.
In some examples, it is not possible to determine ξk in closed form based on the representation of {tilde over (p)}(yk|z1:k) provided by the updater. In some examples, an iterative procedure is used to determine a sequence of parameters ξk8n], for n=1, 2, . . . , which approaches or converges to the best approximation of {tilde over (p)}(yk|z1:k) in the family of distributions . For example, an initial value of ξk[1] may be the value of ξk−1, which characterizes {circumflex over (p)}(yk−1|z1:k−1).
In some examples, the quality of the approximation is measured according to a Kullback-Leibler divergence, defined as
for two distributions, g(y) and h(y) is used as quantity that determine the degree to which {circumflex over (p)}(yk−1|z1:k−1){circumflex over (p)}(yk|z1:k) approximates {tilde over (p)}(yk|z1:k). Specifically, the sequence of parameters μk[n], for n=1,2, . . . , is chosen to successively reduce (or approximated with the goal of reducing) the quantity
d
k
[n]
=D
KL({tilde over (p)}(yk|zhd 1:k)∥{circumflex over (p)}[n](yk|z1:k)
where {circumflex over (p)}[n](yk|z1:k) is parameterized by ξk[n], at each iteration for n=1,2, . . . .
Notice that for any constant C>0
and therefore it is only necessary to know the distribution {tilde over (p)}(yk|z1:k) to condense up to a constant.
In some examples, the procedure for determining the successive parameter values ξk[n] follows a procedure that is similar to the Expectation-Maximization (EM) Algorithm (Dempster, A. P.; Laird, N. M.; Rubin, D. B. (1977). “Maximum Likelihood from Incomplete Data via the EM Algorithm”. Journal of the Royal Statistical Society. Series B (Methodological) 39 (1): 1-38.)
In the case of a mixture distribution {circumflex over (p)}(yk|z1:k)=Σjαjqi(yk; θi), for instance when each distribution qi(yk;θi)=(yk;μi, Σi) is Gaussian, the iteration can be expressed as follows. (For simplicity, the subscript k and the conditional dependence on z1:k are omitted in certain equations below, and k(yk zl:k) is represented by q(y; ξ[n])=Σjαjqi(y; θi[n]).)
for i=1, . . . , m do
Note that it is not required that each of the element distributions qi(y; θi) is Gaussian. Other distributions in an exponential family can be used, and rather computing weighted expectations of statistics y and y yT, as shown above in the Gaussian example, appropriate other statistics are used in the analogous manner that other statistics are used in parameter estimation using the EM Algorithm.
The main complexity of procedure outlined above is evaluating expectations of the form q(·){{tilde over (p)}(·)f(·)}, is the distribution to condense, q(·) is a distribution in an exponential family, an f(·) is an elementary function. Since the expectations are taken with respect to a member of an exponential family, the required integrations can be efficiently computed.
Numerical integration methods require several evaluations of the integrand. In the following we call M the number of evaluations required by a numerical integration method. Monte Carlo techniques for numerical integration evaluate the integrand in randomly chosen points while quadrature rules obtain deterministically such points. Monte Carlo techniques can approximate very general integrals at the expenses of needing high values of M . In our case we need to compute expectations with respect to tractable distributions, this fact can be used to improve the numerical integration. For example, if q(·) is a Gaussian distribution, very efficient quadrature rules are known, where for state vectors of dimension n , only M=2n or M=n2+1 point-wise evaluations are needed to obtain cubature formulae of degree 3 or 5. These quadrature rules need values of M several orders of magnitude smaller than Monte Carlo methods, and they scale linearly or quadratically with the dimension of the state vector. The remaining complexity lies in the evaluation of pe) in the M points used in the cubature formulae. Notice that {tilde over (p)}(·) can be evaluated in the same points for all the expectations in each iteration (i.e., for each value of n).
The point evaluation of {tilde over (p)}(·) in can be efficiently performed because the distribution {circumflex over (p)}(yk−1|zk−1) is a mixture of distributions belonging to exponential families. In particular, let
Therefore, the complexity of the point evaluation of {tilde over (p)}(·) lies only in the m expectations over the distribution qi(yk−1) of the function p(yk|yk−1) which is easy to evaluate. Moreover, if the dynamic model is linear and Gaussian, and the distribution {circumflex over (p)}(yk−1|z1:k−1) is a mixture of Gaussians, evaluation of {tilde over (p)}(·) is trivial since the expectations involved can be easily obtained in closed-form.
In summary, the computational complexity lies in the computation of integrals in each iteration of the BC process, and the complexity of these computations depends on the number of evaluations of elemental functions needed. The number of the point-wise evaluations needed is M or M2 depending on whether the prediction step can be computed analytically or not. In addition, the tractability of the exponential families used in the BC process can lead to efficient quadrature rules which need small values of M.
We now describe the implementation for the specific case where the dynamic model is Gaussian and linear, and the family is mixtures of Gaussians. Pseudocode for implementation case is shown in
Due to the easiness of the prediction step, this particular case leads to a relatively simple BCF implementation.
At time tk (k ∈ ), if
Notice that in this case the distribution to condense is easy to evaluate point-wise. Therefore, each expectation in the iterations of the filter needs M evaluations of elemental functions, where M can be small since efficient quadrature rules for expectations over Gaussian distributions can be applied.
The pseudocode of
The approaches described above can be applied to a navigation task in which x(t) and o(t) denote the position and orientation of the agent node at time t, where o(t) can be represented by a rotation vector Both x(t) and o(t) can be modeled as analytic functions of time. Hence, at each time tk, they can be approximated by truncated Taylor expansions. For example, for the position at time tk we have
where the error of this approximation is
and t* is some point in the interval [tk,tk+1]
The dynamic model can be written as
y
k+1
=H
k
y
k
+n
k
where the matrix Hk is obtained from the Taylor series expansions, and the approximation error nk can be modeled as a random variable. Commonly this error is modeled as a zero-mean Gaussian variable (i.e. a discrete Wiener process). Thus, the dynamic model for the state vector in navigation can be considered as linear and Gaussian with a wide generality.
Navigation systems can use a variety of measurements obtained by multiple sensors such as radio frequency (RF) receivers, the global positioning system (GPS) receivers, inertial measurements units (IMU), Doppler sensors, compasses, etc. The set of measurements obtained by the agent in each time instant tk form the vector zk. The relationship between the positional state vector and zk can be described by the likelihood model p(zk|yk) as introduced above.
In one example of a navigation system, we focus on the case in which the agents obtains IMU and GPS measurements in harsh environments. An IMU device takes two kinds of measurements; the angular velocity about the body frame, zω, and the force zf measured by gyros and accelerometers, respectively.
The gyro measurement for angular velocity at each time instant is given by
z
ω
=w+b
in
ω
+S
in
ω
w+n
ω
where w ∈3 is the true angular velocity, binω∈3 and Sinω∈ are in-run bias and scale factor (both following linear-Gaussian dynamics), and nω ∈3 is a Gaussian vector with zero mean and covariance matrix Kω. Hence, p(zkω|yki) is a Gaussian distribution with mean w+binω+Sinωw and covariance matrix Kω.
Similarly, the measurement for the force at each time instant is given by
where a, f ∈3 are the true acceleration and force, respectively, g ∈3 is the gravity, binf ∈3 and Sinf ∈ are in-run bias and scale factor (both following linear-Gaussian dynamics), and of nf ∈ is a Gaussian vector with zero mean and covariance matrix Kf ∈3×3. Moreover, C(o) ∈3×3 can be written, by the Rodrigues' rotation formula, as
where [o]x is the skew-symmetric form of the rotation vector
Hence, p(zkf|yk) is a Gaussian distribution with mean (1+Sinf)·C†(o)·(a−g)+binf and covariance matrix Kf. Note that the relationship between the force measurement and the state vector is nonlinear.
The measurement of a pseudorange from satellite i with known position xiG is given by
z
i
G
=∥x
i
G
−x∥+c·b
t
+c·∥x
i
G
−x∥S
t
+B
i
+b
i
+n
i
where c is the propagation speed, bt
Note that the relationship between the pseudorange measurements and the state vector is nonlinear. Moreover, the distribution p(zkG∥xk, bt
Below we show the performance of proposed BCF by simulations with measurements emulating sensors' behavior in harsh propagation environments. Consider a scenario where one agent obtains both GPS and IMU measurements. We simulated GPS measurements from 4 satellites in NLOS conditions. The white thermal noise of such measurements was modeled as a Gaussian random variable with zero mean and standard deviation of 2 m, while the positive bias introduced by the NLOS propagation was modeled as an exponential random variable with mean 6 m, 8 m, 10 m, and 14 m for each satellite. The error in the IMU force measurements was modeled as a zero-mean Gaussian random variable with 0.07 N standard deviation, and the error in the angular velocity measurements as (0,0.02 rad/sec). The motion of the agent was simulated as shown in
In such scenario we filter the positional state of the agent by using the EKF, the GMF, the sampling importance resampling (SIR) particle filter, and the proposed filter. We simulate 100 positions, where the number of components for the filters using mixtures of Gaussians is N=10 (
Implementations of the approaches described can be implemented in software, in hardware, or in a combination of software and hardware. The software can include instructions stored on a computer-readable medium for causing a data processing system to perform the functions and procedures described above. Hardware can include special-purpose hardware, such as application-specific integrated circuits (ASIC). In some examples, the hardware is represented in a stored software representation using a hardware description language (HDL), such as Verilog. In some implementations, the approach is distributed with some modules being implemented in separate system, for example, in the case of a navigation system, with some modules being implemented on the vehicle being tracked and some modules being implemented remotely from the vehicle.
It is to be understood that the foregoing description is intended to illustrate and not to limit the scope of the invention, which is defined by the scope of the appended claims. Other embodiments are within the scope of the following claims.
This invention was made with government support under contract ECCS-0901034 awarded by the National Science Foundation. The government has certain rights in the invention.