The present invention relates to statistical signal processing and more particularly to sequential Bayesian filtering using density morphing and mode propagation.
Some areas of computational research such as real-time machine vision, speech recognition, real-time object tracking, etc., provide a complex challenge to an application designer. For example, in an object-tracking application the complexity of the system increases due to the random nature of the captured image data. For example, using machine vision to track a live person in a given environment would require complex processing of captured image data concerning the person and the environment around such person. Hence, researchers have applied probabilistic and numerical techniques to deal with complex systems with stochastic nature in the field of system dynamics and measurements.
Dynamic system tracking applications typically use estimations for determining changes in the dynamic system states over time. Dynamic states are generally modeled using discrete time approach where measurements are made available at discrete time intervals. Typically, two models—a system model, which evaluates the evolution of the system states over time and a measurement model that links noisy measurements to states—are required.
It is possible to apply several state estimation techniques using the dynamic and measurements model. One such estimation technique uses a Bayesian approach. Typical statistical methods include Kalman-Bucy filtering and Particle filtering. Kalman-Bucy filtering provides minimum-mean-squared-error state estimations only for linear stochastic systems with Gaussian initial condition on the state and Gaussian noise, For nonlinear stochastic systems or stochastic systems involving non-Gaussian random variables, particle filtering provides a numerical solution to the state estimation over time. Particle filtering is described next.
Using importance sampling, particle filtering tracks the evolution of state variable with respect to time by probability density functions (PDF) with a set of weighted samples. Several variables may be of interest in a given application. For example, in an object tracking machine vision application, the variables of interest could be the object's position, the temperature, velocity, shape, etc. For each variable, samples in the state spaces are considered to be particles, where each of them has a weight that signifies its probability. Particle filtering is used to estimate the states of such variables by recursively applying predictions and updates to the particle set.
While conventional particle filtering has become a popular technique for dynamic tracking applications, it has several drawbacks. Real-world applications typically involve nonlinear stochastic systems with a high-dimensional state space. However, particle filtering technique becomes inefficient and impractical when the dimension of the state space is high, where an inhibitively large number of samples need to be tracked. Efficient sampling in high-dimensional space is very difficult. Sample degeneracy and sample impoverishment are common issues associated with particle filtering. Hence, there is a need for a technique that provides better filtering performance by avoiding problems associated with particle filtering.
A system and method for analyzing dynamic systems with linear or nonlinear dynamic models by using Bayesian filtering. The system performs prediction over a state model of the dynamic system to generate a prediction density that has one or multiple modes. The system approximates a function of a measurement probability and prediction density. An update module updates the conditional density function using the measurement probability (i.e. likelihood function) and prediction density. A mode finding and mixture reduction process reduces the number of kernels to achieve mode propagation. The system maintains, over a time span, analytic representations in the form of mixtures of kernels for prior density functions, prediction density functions, measurement likelihood function and posterior density functions.
Density morphing can be performed to track the evolution of the density function of the state distribution over time. A prediction module can be configured to apply an unscented transformation on a prior probability density function of a prior state model of the dynamic system. The update module can then generate a continuous function to approximate measurement likelihood. A density morphing module tracks density evolution over the time and reconstructs the continuous function from sparse samples. The density morphing module propagates continuous density functions over time and maintains analytic representations in the form of mixtures of kernels for representing the density functions.
The number of mixtures generated during the update as a part of the recursive prediction-update iterations can be reduced by approximating densities near the modes by using Gaussian components with predetermined bandwidths.
Further, incremental sampling and interpolation can be used to approximate the measurement likelihood and the posterior PDF. Bandwidth of the kernels can be data-driven, i.e., determined by the local density of the sample distribution. Posterior probability density function can be generated by a product of the reconstructed continuous likelihood functions and predicted density. Throughout the recursive process of filtering, the intermediate density functions can be maintained in the form of mixtures of kernels.
Preferred embodiments of the invention are described with reference to the accompanying drawings, of which:
The preferred embodiments of the present invention will be described with reference to the appended drawings.
Bayesian filtering has been used for a number of applications including computer vision. In vision applications, the object or event of interest can be modeled by a dynamic system. For example, target tracking and belief propagation are applications where Bayesian filtering has been applied extensively. In the framework of Bayesian filtering, the state variable {xt;t=0, 1, 2, . . . } is characterized by its probability density function (PDF) estimated from the sequence of measurements {z1, z2, . . . zt}. The state evolution and measurement function are given by
xt=ƒt(xt−1)+vt−1 (1)
zt=ht(xt,nt) (2)
where {vt;t=0, 1, . . . } is an Independent and Identically Distributed (“i.i.d”) process noise sequence and {nt;t=1, 2, . . . } is an i.i.d. measurement noise sequence. A convenient Bayesian solution to the dynamic system is given by a recursive Bayesian filter, where the conditional density of the state variable, given that the measurements are propagated through the stages of prediction and update. This process is typically performed in a recursive manner, where the measurements are processed sequentially and the posterior probability density function is constructed recursively as shown by the equations below:
Prediction: p(xt|z1:t−1)=∫p(xt|xt−1)p(xt−1|z1:t−1)dxt−1 (3)
Update: p(xt|z1:t)=κt·p(zt|xt)p(xt|z1:t−1) (4)
where z1:t={z1, . . . , zt}, κt=[∫p(zt|xt)p(xt|z1:t−1)dxt]−1 is a constant independent of xt. In equation (3) above, p(xt|z1:t−1) is the prediction distribution, p(xt|xt−1) is the system dynamic model, also called motion model in computer vision applications, and p(xt−1|z1:t−1) is the prior distribution.
When the state equation (1) is linear and the posterior density of xt is Gaussian, the Kalman filter provides the optimal solution in terms of achieving minimum-mean-squared-error estimation. However, for many practical situations characterized by non-linear and/or non-Gaussian distributions, the solutions to optimal algorithm is not tractable due to the lack of a closed form expression for the integration result in prediction. For non-linear and/or non-Gaussian Bayesian filtering, various particle filtering methods have been developed, where the posterior density function is represented by a number of random samples (i.e. particles) {xti;i=1, 2 . . . } and their associated weights {wti;i=1, 2 . . . }.
In the particle filtering technique, the conditional density is propagated through sampling, re-sampling and weight modification.
A mixture density function of the conditional density is described as:
where K(x,xtk;Htk) is a kernel function centered at xtk, with bandwidth Htk. The Bayesian filtering through density morphing and mode propagation is described next.
In the prediction step 12, the system applies an unscented transformation to transform the prior PDF described by a mixture density function
into the prediction PDF described by a mixtured density function
According equation (3), the prediction PDF is written as follows
The computation of the prediction PDF in a nonlinear system is not straightforward due to lack of analytic solution for the integration in (8). To maintain a Gaussian mixture representation for the prediction PDF, we approximate each summation term ∫p(xt|xt−1)K(xt−1,xt−1k;Ht−1k)dxt−1 in (8) by a Gaussian kernel K(xt,
Where κ∈R, (√{square root over ((N+κ)Ht−1k))}i is the i-th row or column of the matrix square root of (N+κ)Ht−1k, and Wi is the weight associated with point xt−1,ik. Denote y=ƒt(xt−1) as the resulting random variable of propagating xt−1 through a nonlinear transformation ƒt.
The mean μy and covariance Hy of y are calculated from the transformed sigma points as follows
The mean and covariance of y obtained by the unscented transformation are precise to the second order of the Taylor series expansion of any nonlinear function ƒt. The choice of parameter κ affects the error introduced by higher order moments. Applying the unscented transformation to the Gaussian components of K(xt−1, xt−1k;Ht−1k) in (8) according to (3), the approximation of the prediction PDF (7) is derived as follows:
where Qt is the variance of process noise vt. It should be noted that here random sampling strategy can also be employed.
At step 14, the predicted density (7) is used to draw a set of samples {{tilde over (x)}{tilde over ( )}tj}. The likelihood of the measurement is p(zt|{tilde over (x)}{tilde over ( )}tj) calculated at these sample points.
At step 16, the likelihood function p(zt|xt) of the measurement is approximated as kernel mixtures. The approximation is obtained by interpolating the likelihood function from samples {{tilde over (x)}{tilde over ( )}tj}, assuming that the measurement likelihood is reasonably smooth.
An outline of the approximation process can be stated as:
1. Analytic sampling; place samples in sparsely sampled area and area of significance (high measurement probability); discovery and refinement.
2. Interpolation; data-responsive variable bandwidth; and weight adjustment.
The density morphing and mode reduction algorithm progressively improves the approximation of the likelihood through incremental sampling. In contrast to various improved particle filters, a continuous approximation of the likelihood function is interpolated from all the samples available from each sampling step.
Sampling and distribution update is repeated over multiple steps. At each step, new samples are drawn from the approximate likelihood function recovered in the previous step. New samples are then added into the sample set which is used to obtain a new approximation of the distribution.
Given a set of samples, the likelihood or posterior distribution is approximated by a mixture function composed of weighted Gaussian kernels whose centers are set as sample locations and whose bandwidths are determined by the distance to their k-th nearest neighbors (kNN), which reflects the local density condition of the actual sample distribution.
The kernel bandwidths are data driven and lead from a smooth to fine interpolation of the underlying distribution function. In the initial step, with a small number of samples that are sparsely distributed in the state space, kernel bandwidths are set large and the interpolation algorithm gives a smooth approximation of the distribution.
As the sample size increases, samples are distributed densely, which leads to small kernel bandwidths, and the interpolation result reveals fine details of the landscape of the underlying distribution function. Since the analytic representation provides a global view of the landscape of the likelihood function, efficient sampling is possible, More samples should be placed in the region where the likelihood has significant values. The algorithm is summarized as follows with prediction distribution used as the initial sampling function.
It is well known that closer the sampling function is to the distribution function under estimation, the better density estimate can be obtained from samples. It has been shown that, to recover a distribution function of interest, the best choice of the sampling function, which minimizes the variance of distribution estimate, is determined by the unknown distribution itself. Therefore, optimal sampling is not available in many situations.
Previous approaches choose to use the prediction PDF as the sampling function to estimate the measurement likelihood, which could lead to severe sample degeneracy when the likelihood is very “peaky” as compared to the sampling function. In the density morphing and mode propagation approach, sampling and updating of the sampling function simultaneously is done simultaneously through incremental sampling. The advantage is to bring the sampling function close to the distribution under estimation, so that the resulting density estimate is less sensitive to the initial sampling function.
At step 18, an update on the posterior PDF is performed. With the prediction density (7) and the measurement likelihood (12), the conditional density function is updated as
where κt=└∫p(zt|xt)p(xt|z1:t−1)dxt┘ is a constant independent of xt. Due to the mixture representation of (7) and (12), the conditional density p(xt|z1:t) can be represented by a mixture density function with increased number of kernels.
If the prediction PDF p(xt|z1:t−1) and the measurement likelihood p(zt|xt) are each represented by a Gaussian mixture with N1 and N2 components respectively, the resulting posterior PDF p(xt|z1:t) has N=N1N2 components. This number could explode in recursive filtering. Therefore, an algorithm or a technique is needed to reduce the mixture size. The need for approximating the sum or the product of Gaussian mixtures also arises in several other contexts, including nonparametric belief propagation and training mixture models.
A given application may bypass the approximating measurement likelihood, and use the incremental sampling and interpolation method to directly reconstruct a mixture representation for the product of the prediction PDF and the measurement likelihood. By normalizing the product, one obtains a kernel mixture representation of the posterior PDF.
At step 20, a mode finding and mixture reduction procedure is performed. To maintain a reasonable number of kernels in the posterior PDF, a mode finding and mixture reduction procedure through variable-bandwidth mean-shift is applied to reduce the number of mixtures in (17). The main idea is to locate the main modes of the posterior PDF and derive a more concise representation with fewer kernels using the modes to approximate the original density function. Mixture reduction is now described in detail. Given a Gaussian mixture resulting from the sum or the product of several Gaussian mixtures,
one can treat ƒ(x) as a density estimate from sample population {xi} with variable bandwidth {Hi}. A new algorithm called mode reduction produces an approximation of the original density, with a reduced number of Gaussian components. The algorithm is founded by the Variable Bandwidth MeanShift theory and performs a mode finding procedure on kernel density estimates.
From (18), one can compute the gradient as well as the Hessian at x
Starting from x, the variable bandwidth MeanShift produces a convergent sequence:
At the convergence location
ym=(Σi=1Nwi(ym)Hi−1)Σi=1Nwi(ym)Hi31 1xi, we have ∇ƒ(x)|x=ym=0.
To verify that this is indeed a mode, one needs to verify that the Hessian is negative definite (∇∇T)ƒ(x)|x=ym<0. By starting from each xi and running the Variable bandwidth MeanShift procedure, we can locate all the modes of ƒ(x). In addition, we divide all the samples {xi} into disjoint sets {xj,i} such that samples in the same set {xj,1,xj,2, . . . } converge to the same mode mj. The idea of our algorithm is to approximate the density near the modes using Gaussian components with proper bandwidth. For a Gaussian component k(x;μj,Σj), the gradient and the Hessian computed at the mean μj, also being the mode, are:
To approximate the density near the mode, we equate the location, gradient and Hessian of the Gaussian component with the parameters computed at a mode mj.
μj=mj
∇K(x;μj,Σj)|x=μj=∇ƒ(x)|x=mj
(∇∇T)K(x;μj,Σj)|x=μj=(∇∇T)ƒ(x)|x=mj
Following (21), the Gaussian kernel is decided by:
given that the Hessian P is negative definite. For those samples xj whose convergence point is not a mode of f(x), we maintain the Gaussian component k (x;xj,Hj) in the reduced mixture approximation. The resulting approximation is expressed as:
(mj,Σj)=(mj,(∇∇T)ƒ(x)|x=mj) or (xj
∥{tilde over (ƒ)}(x)−ƒ(x)∥2=cTQc+qTb+F (25)
where c=[c1, . . . , cM]T is the weight vector, Q is a M×M matrix,
and q is a M×1 vector, q=−2[<K(x;m1,ƒ(x)>, . . . , <K(x;mM,ƒ(x)ƒ(x)>]T.
The notation <ƒ(x), g(x)>=∫ƒ(x)g(x)dx is used.
The weights of the new components are obtained by solving the following convex quadratic programming problem:
In general, the means {mj} and the bandwidths {Σj} of the new components are up to further optimization over the cost function (25).
After mode finding and mixture reduction, the conditional density is estimated by a reduced number of kernels.
Density morphing and Mode propagation have several potential advantages over the conventional particle filtering. Through the Variable-bandwidth Mean-shift based mode finding algorithm, the number of modes in the posterior is determined automatically and naturally. This information helps to decide the number of samples to be drawn. With the conventional particle filtering methods, it is difficult to determine how to efficiently sample the state space, especially when the dimensions are high.
The optimal bandwidth can also be determined by imposing certain smoothness of the measurement probability and the posterior. Thus, optimal sampling given the kernel density and estimation constraints are achieved. Optimal sampling is especially critical when the dimension of the state space is high, which is often encountered in real applications. In mode propagation, kernel estimation is used to replace particles, and with variable bandwidth it is able to reduce the effects of degeneracy and sample impoverishment.
Experimental results for the above are presented below. The nonlinear process model and the non-stationary observation model used in the experiments are:
xt+1=1+sin(wπt)+φxt+vt
zt=φxt2+nt
where vt is i.i.d., Gaussian noise with zero mean and variance nv2, vt˜N(0,nv2), w=4e−2 and φ=0.5 are scalar parameters. The observation noise nt is drawn from a Gaussian distribution N(0,12). In the initial step, a Gaussian mixture distribution describes the state variable.
To approximate the measurement distribution, a system based on at least one embodiment of the present invention samples the mean and a number of sigma points from each Gaussian in the predicted state distribution. Assuming that the predicted state distribution has a mixture of N Gaussians denoted as N(mi,σi2)i=1, . . . , n. m samples are drawn from the prediction distribution. Sample weights are proportional to the measurement probability. Then the measurement distribution is composed of m Gaussian mixture, and—nm Gaussian kernels are used to estimate the posterior distribution in the final step. The mode-fusion procedure based on the Variable-bandwidth Mean-shift algorithm is applied to simplify the state description with a reduced number of Gaussians.
In the experiments described next, nine points are sampled from each Gaussian, and we set k=√{square root over (2)}/3. The
It is to be understood that the present invention may be implemented in various forms of hardware, software, firmware, special purpose processors, or a combination thereof. In at least one embodiment, the present invention may be implemented in software as an application program tangibly embodied on a program storage device. The application program may be uploaded to, and executed by, a machine capable of computing and having any suitable architecture.
Referring to
The computer platform 101 also includes an operating system and micro instruction code. The various processes and functions described herein may either be part of the micro instruction code or part of the application program (or a combination thereof) which is executed via the operating system. In addition, various other peripheral devices may be connected to the computer platform such as an additional data storage device and a printing device.
It is to be further understood that, because some of the constituent system components and method steps depicted in the accompanying figures may be implemented in software, the actual connections between the system components (or the process steps) may differ depending upon the manner in which the present invention is programmed. Given the teachings of the present invention provided herein, one of ordinary skill in the related art will be able to contemplate these and similar implementations or configurations of the present invention.
The Bayesian filtering method based on mode propagation can be generally applied to any purpose object tracking. For example, to 3D human tracking, vehicle tracking, and tracking problems from a large range of applications and domains. Software modules can be built for prediction, sampling, approximation and update as described above in detail. Additionally, input and output devices or other modules can be included depending upon a given application. Those skilled in the art will appreciate that these are only examples of applications where Bayesian filtering based on mode propagation can be applied.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present invention as defined by the appended claims.
This application claims the benefit of U.S. Provisional Application, “Density Morphing And Mode Propagation For Bayseian Filtering”, Ser. No. 60/490,326, filed on Jul. 25, 2003, the contents of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
6981040 | Konig et al. | Dec 2005 | B1 |
7128167 | Dunlop et al. | Oct 2006 | B2 |
20030065409 | Raeth et al. | Apr 2003 | A1 |
20030068082 | Comaniciu et al. | Apr 2003 | A1 |
20030233197 | Padilla et al. | Dec 2003 | A1 |
20040081359 | Bascle et al. | Apr 2004 | A1 |
20050114079 | Rui et al. | May 2005 | A1 |
20050286764 | Mittal et al. | Dec 2005 | A1 |
20060093188 | Blake et al. | May 2006 | A1 |
Number | Date | Country | |
---|---|---|---|
20050038638 A1 | Feb 2005 | US |
Number | Date | Country | |
---|---|---|---|
60490326 | Jul 2003 | US |