4 - Sampling methods: particle filter

In the previous tutorials we encountered some shortcomings in describing distributions as Gaussians, albeit with considerable flexibility in coping with the non-linear transforms.

Sampling methods offer an attractive alternative to such parametric methods in that there is no need for complicated though approximate covariance calculations. In this tutorial we look at a class of sequential Monte Carlo sampling methods, and in particular, the particle filter.

Colloquially we can think of a particle filter as a series of point samples being recursed through the predict-update stages of a Bayesian filter. The diversity of samples compensates for the lack of a covariance estimate, though often at the expense of increased computation requirements.

Background

In more detail, we seek to approximate the posterior state estimate as a sum of samples, or particles,

\[p(\textbf{x}_{k}|\textbf{z}_{1:k}) \approx \sum_{i} w_{k}^i \delta (\textbf{x}_{k} - \textbf{x}_{k}^i)\]

where \(w_{k}^i\) are weights such that \(\sum\limits_{i} w_{k}^i = 1\). This posterior can be calculated, and subsequently maintained, by successive applications of the Chapman-Kolmogorov equation and Bayes rule in an analogous manner to the Kalman family of filters of previous tutorials. There is considerable flexibility in how to sample from these various distributions and the interested reader can refer to [1] for more detail.

The present tutorial focuses on a so-called sequential importance resampling filter. This is facilitated by a number of Stone Soup classes. The weight-update equation is,

\[w^i_k = w^i_{k-1} \frac{p(\mathbf{z}_k|\mathbf{x}^i_k) p(\mathbf{x}^i_k|\mathbf{x}^1_{k-1})} {q(\mathbf{x}^i_k|\mathbf{x}^1_{k-1},\mathbf{z}^i_{1:k})}\]

where \(p(\mathbf{z}_k | \mathbf{x}^i_k)\) is the likelihood distribution (as defined by the MeasurementModel) and \(p(\mathbf{x}^i_k|\mathbf{x}^1_{k-1})\) is the transition probability distribution (TransitionModel). The \(q(\cdot)\) distribution – the importance density – should approximate the posterior distribution, while still being easy to sample from.

A common occurrence in such methods is that of sample impoverishment. After a few iterations, all but a small number of the particles will have negligible weight. This affects accuracy and wastes computation on particles with little effect on the estimate. Many resampling schemes exist and are designed to redistribute particles to areas where the posterior probability is higher. In Stone Soup such resampling is accomplished by a Resampler. More detail is provided in the example below.

Nearly-constant velocity example

We continue in the same vein as the previous tutorials.

Ground truth

Import the necessary libraries

import numpy as np

from datetime import datetime
from datetime import timedelta
start_time = datetime.now().replace(microsecond=0)

Initialise Stone Soup ground-truth and transition models.

from stonesoup.models.transition.linear import CombinedLinearGaussianTransitionModel, \
    ConstantVelocity
from stonesoup.types.groundtruth import GroundTruthPath, GroundTruthState

transition_model = CombinedLinearGaussianTransitionModel([ConstantVelocity(0.05),
                                                          ConstantVelocity(0.05)])
timesteps = [start_time]
truth = GroundTruthPath([GroundTruthState([0, 1, 0, 1], timestamp=start_time)])

Create the truth path

for k in range(1, 21):
    timesteps.append(start_time+timedelta(seconds=k))
    truth.append(GroundTruthState(
        transition_model.function(truth[k-1], noise=True, time_interval=timedelta(seconds=1)),
        timestamp=timesteps[k]))

Plot the ground truth.

from stonesoup.plotter import AnimatedPlotterly
plotter = AnimatedPlotterly(timesteps, tail_length=0.3)
plotter.plot_ground_truths(truth, [0, 2])
plotter.fig


Initialise the bearing, range sensor using the appropriate measurement model.

from stonesoup.models.measurement.nonlinear import CartesianToBearingRange
from stonesoup.types.detection import Detection

sensor_x = 50
sensor_y = 0

measurement_model = CartesianToBearingRange(
    ndim_state=4,
    mapping=(0, 2),
    noise_covar=np.diag([np.radians(0.2), 1]),
    translation_offset=np.array([[sensor_x], [sensor_y]])
)

Populate the measurement array

Plot those measurements



Set up the particle filter

Analogously to the Kalman family, we create a ParticlePredictor and a ParticleUpdater which take responsibility for the predict and update steps respectively. These require a TransitionModel and MeasurementModel as before. To cope with sample sparsity we also include a resampler, in this instance SystematicResampler, which is passed to the updater. It should be noted that there are many resampling schemes, and almost as many choices as to when to undertake resampling. The systematic resampler is described in [2], and in what follows below resampling is undertaken at each time-step. More resamplers that are included in Stone Soup are covered in the Resampler Tutorial

Use of Effective Sample Size resampler (ESS)

Resampling removes particles with a low weight and duplicates particles with a high weight. A side effect of this is that additional variance is added. Use of SystematicResampler at each time-step means that additional variance is being introduced when it may not necessarily be required. To reduce the additional variance, it may be optimal to resample less frequently.

The Effective Sample Size resampler (ESSResampler) compares the variance of the unnormalised weights of the particles to a pre-specified threshold, and only resamples when the variance is greater than this threshold. This threshold is often calculated by the ESS criterion (at time n) given by:

\[ESS = \left(\sum_{i=1}^{N} (W_{n}^i)^2\right)^{-1}\]
from stonesoup.predictor.particle import ParticlePredictor
predictor = ParticlePredictor(transition_model)
from stonesoup.resampler.particle import ESSResampler
resampler = ESSResampler()
from stonesoup.updater.particle import ParticleUpdater
updater = ParticleUpdater(measurement_model, resampler)

Initialise a prior

To start we create a prior estimate. This is a ParticleState which describes the state as a distribution of particles using StateVectors and weights. This is sampled from the Gaussian distribution (using the same parameters we had in the previous examples).

from scipy.stats import multivariate_normal

from stonesoup.types.numeric import Probability  # Similar to a float type
from stonesoup.types.state import ParticleState
from stonesoup.types.array import StateVectors

number_particles = 1000

# Sample from the prior Gaussian distribution
samples = multivariate_normal.rvs(np.array([0, 1, 0, 1]),
                                  np.diag([1.5, 0.5, 1.5, 0.5]),
                                  size=number_particles)

# Create prior particle state.
prior = ParticleState(state_vector=StateVectors(samples.T),
                      weight=np.array([Probability(1/number_particles)]*number_particles),
                      timestamp=start_time)

Run the tracker

We now run the predict and update steps, propagating the collection of particles and resampling when told to (at every step).

Plot the resulting track with the sample points at each iteration. Can also change ‘plot_history’ to True if wanted.

plotter.plot_tracks(track, [0, 2], particle=True, plot_history=False)
plotter.fig


Key points

  1. Sampling methods offer an attractive alternative to Kalman-based filtering for recursive state estimation.

  2. The particle filter trades off a more subtle quantification of a non-Gaussian estimate against increased computational effort.

  3. Very often particle filters encounter sample impoverishment and require a resampling step.

References

Total running time of the script: (0 minutes 1.157 seconds)

Gallery generated by Sphinx-Gallery