Kernel methods: the adaptive kernel Kalman filter

The AKKF is a model-based Bayesian filter designed for non-linear and non-Gaussian systems. It can avoid problematic resampling in most particle filters and reduce computational complexity. We aim to realise adaptive updates of the empirical kernel mean embeddings (KMEs) for posterior probability density functions (PDFs) using the AKKF, which is executed in the state space, measurement space, and reproducing kernel Hilbert space (RKHS).

  • In the state space, we generate the proposal state particles and propagate them through the motion model to obtain the state particles.

  • In the measurement space, the measurement particles are achieved by propagating the state particles through the measurement model.

  • All these particles are mapped into RKHSs as feature mappings and linearly predict and update the corresponding kernel weight mean vector and covariance matrix to approximate the empirical KMEs of the posterior PDFs in the RKHS.

Background and notation

Kernel mean embedding (KME)

The KME approach represents a PDF \(p(\mathbf{x})\) as an element in the RKHS using the following equation:

\[\mu_X := \mathbb{E}_X \left[\phi_\mathbf{x}(X)\right] = \int_{\mathcal{X}} \phi_\mathbf{x}(\mathbf{x}) p(\mathbf{x})d \mathbf{x}.\]

For joint PDFs of two or more variables, e.g., \(p(\mathbf{x}, \mathbf{y})\), they can be embedded into a tensor product feature space \(\mathcal{H}_\mathbf{x} \otimes \mathcal{H}_\mathbf{y}\) as an uncentred covariance operator:

\[\mathcal{C}_{XY} := \mathbb{E}_{XY} \left[\phi_\mathbf{x}(X)\otimes \phi_\mathbf{y}(Y)\right] = \int_{\mathcal{X}\times \mathcal{Y}} \phi_\mathbf{x}(\mathbf{x}) \otimes \phi_\mathbf{y}(\mathbf{y})p(\mathbf{x},\mathbf{y}) d \mathbf{x} d\mathbf{y}.\]

Similar to the KME of \(p(\mathbf{x})\), the KME of a conditional PDF \(p(X|\mathbf{y})\) is defined as

\[\mu_{X|\mathbf{y}} := \mathbb{E}_{X|\mathbf{y}} \left[\phi_\mathbf{x}(X)\right] = \int_{\mathcal{X}} \phi_\mathbf{x}(\mathbf{x}) p(\mathbf{x}|\mathbf{y}) d\mathbf{x}.\]

The difference between the KME \(\mu_{X|\mathbf{y}}\) and \(\mu_{X}\) is that \(\mu_{X}\) is a single element in the RKHS, while \(\mu_{X|\mathbf{y}}\) is a family of points, each indexed by fixing \(Y\) to a particular value \(\mathbf{y}\). A conditional operator \(\mathcal{C}_{X|Y}\) is defined under certain conditions as the linear operator, which takes the feature mapping of a fixed value \(\mathbf{y}\) as the input and outputs the corresponding conditional KME:

\[\mu_{X|\mathbf{y}} = \mathcal{C}_{X|Y} \phi_\mathbf{y}(\mathbf{y}) = \mathcal{C}_{XY}\left(\mathcal{C}_{YY} + {\kappa} I\right)^{-1} \phi_\mathbf{y}(\mathbf{y}).\]

Here, \(\kappa\) is a regularisation parameter to ensure that the inverse is well-defined, and \(I\) is the identity operator matrix.

Empirical KME of conditional distribution

As it is rare to have access to the actual underlying PDF mentioned above, we can alternatively estimate the KMEs using a finite number of samples/particles drawn from the corresponding PDF. The \(M\) sample pairs \(\mathcal{D}_{XY} = \{ (\mathbf{x}^{\{1\}}, \mathbf{y}^{\{1\}}), \dots, (\mathbf{x}^{\{M\}},\mathbf{y}^{\{M\}})\}\) are drawn independently and identically distributed (i.i.d.) from \(p(\mathbf{x},\mathbf{y})\) with the feature mappings \(\Phi := \left[\phi_{\mathbf{x}}(\mathbf{x}^{\{1\}}),\dots, \phi_{\mathbf{x}}(\mathbf{x}^{\{M\}})\right]\) and \(\Upsilon := \left[\phi_{\mathbf{y}}(\mathbf{y}^{\{1\}}),\dots, \phi_{\mathbf{y}}(\mathbf{y}^{\{M\}})\right]\). The estimate of the conditional embedding operator \(\hat{\mathcal{C}}_{X|Y}\) is obtained as a linear regression in the RKHS, as illustrated in figure 1. Then, the empirical KME of the conditional distribution, i.e., \(p(\mathbf{x}\mid\mathbf{y})\xrightarrow{\text{KME}} \hat{\mu}_{X|\mathbf{y}}\), is calculated by a linear algebra operation as:

\[p(\mathbf{x}\mid\mathbf{y}) \xrightarrow{\text{KME}}\hat{\mu}_{X|\mathbf{y}} = \hat{\mathcal{C}}_{X|Y}\phi_\mathbf{y}(\mathbf{y})= \Phi\left(G_{\mathbf{y}\mathbf{y}} + \kappa I\right)^{-1} \Upsilon^{\rm{T}}\phi_{\mathbf{y}}(\mathbf{y}) {\equiv} \Phi\mathbf{w}.\]

Here, \(\phi_{\mathbf{y}}(\mathbf{y})\) represents the feature mapping of the observation. The estimated empirical conditional operator \(\hat{\mathcal{C}}_{X|Y}\) is regarded as a linear regression in the RKHS. \(G_{\mathbf{y}\mathbf{y}} = \Upsilon^{\rm{T}} \Upsilon\) is the Gram matrix for the samples from the observed variable \(Y\).

The empirical KME \(\hat{\mu}_{X|\mathbf{y}}\) is a weighted sum of the feature mappings of the training samples. The weight vector includes \(M\) non-uniform weights, i.e., \({\mathbf{w}} = \left[w^{\{1\}}, \dots, w^{\{M\}} \right]^{\rm{T}}\), and is calculated as:

\[\mathbf{w} = \left(G_{\mathbf{y}\mathbf{y}} + \kappa I\right)^{-1} G_{:,\mathbf{y}}.\]

Here, the vector of kernel functions \(G_{:,\mathbf{y}} = \left[ k_{\mathbf{y}}(\mathbf{y}^{\{1\}}, \mathbf{y}), \dots, k_{\mathbf{y}}(\mathbf{y}^{\{M\}}, \mathbf{y}) \right]^{\rm{T}}\). The kernel weight vector \(\mathbf{w}\) is the solution to a set of linear equations in the RKHS. Unlike the particle filter, there are no non-negativity or normalisation constraints.

Illustration of Kernel Mean Embedding from data space to kernel feature space

Figure 1: This figure represents the KME of the conditional distribution \(p(X|\mathbf{Y})\) is embedded as a point in kernel feature space as \(\mu_{X|y} = \int_{\mathcal{X}}\phi_x(x) d P(x|y)\). Given the training data sampled from \(P(X, Y)\), the empirical KME of \(P(X|y)\) is approximated as a linear operation in RKHS, i.e., \(\hat{\mu}_{X|y}=\hat{C}_{X|Y}\phi_y(y)=\Phi\mathbf{w}\).

AKKF based on KME

The AKKF draws inspiration from the concepts of KME and Kalman Filter to address tracking problems in nonlinear systems. Unlike the particle filter, the AKKF approximates the posterior PDF using particles and weights, but not in the state space. Instead, it employs an empirical KME in an RKHS representation. The AKKF aims to estimate the hidden state \(\mathbf{x}_k\) at the \(k\)-th time slot given the corresponding observation \(\mathbf{y}_k\). This is achieved by embedding the posterior PDF into the RKHS as an empirical KME:

\[p(\mathbf{x}_k \mid \mathbf{y}_{1:k}) \rightarrow \hat{\mu}_{\mathbf{x}_k}^{+} = \Phi_k {\mathbf{w}}^{+}_{k}.\]

The feature mappings of particles \(\mathbf{x}_k^{\{1:M\}}\) are represented as \(\Phi_k\), i.e., \(\Phi_k = \left[\phi_{\mathbf{x}}(\mathbf{x}_k^{\{1\}}),\dots, \phi_{\mathbf{x}}(\mathbf{x}_k^{\{M\}})\right]\), and the weight vector \({\mathbf{w}}^{+}_{k}\) includes \(M\) non-uniform weights. The KME \(\hat{\mu}_{\mathbf{x}_k}^{+}\) is an element in the RKHS that captures the feature value of the distribution.

To obtain the empirical KME of the hidden state’s posterior PDF, the AKKF requires a set of generated particles’ feature mappings and their corresponding kernel weights. The particles are updated and propagated in the data space using a parametric dynamical state space model to capture the diversity of nonlinearities. The kernel weight mean vector \(\mathbf{w}_k^{\pm}\) and covariance matrix \(S_k^{\pm}\) are predicted and updated by matching (or approximating in a least squares manner in the feature space) with the state KME. This process ensures that the AKKF can effectively estimate the hidden state in nonlinear and non-Gaussian systems, leading to improved tracking performance.

Implement the AKKF

The AKKF consists of three modules, as depicted in figure 2: a predictor that utilises both prior and proposal information at time \(k-1\) to update the prior state particles and predict the kernel weight mean and covariance at time \(k\), an updater that employs the predicted values to update the kernel weight and covariance, and an updater that generates the proposal state particles.

Predictor takes prior and proposal

\({\color{blue}\blacksquare}\) In the state space, at time \(k\), the prior state particles are generated by passing the proposal particles at time \(k-1\), i.e., \(\tilde{\mathbf{x}}_{k-1}^{\{i=1:M\}}\), through the motion model as

\[\mathbf{x}_k^{\{i\}} = \mathtt{f}\left(\tilde{\mathbf{x}}_{k-1}^{\{i\}}, \mathbf{u}_{k}^{\{i\}} \right).\]

\({\color{orange}\blacksquare}\) In the RKHS, \({\mathbf{x}}_{k}^{\{i=1:M_{\text{A}}\}}\) are mapped as feature mappings \(\Phi_k\). Then, the predictive kernel weight vector \(\mathbf{w}^{-}_{k}\), and covariance matrix \({S}_{k}^{-}\), are calculated as

\[\begin{split}\mathbf{w}^{-}_{k} &= \Gamma_{k} \mathbf{w}^{+}_{k-1}\\ {S}_{k}^{-} &= \Gamma_{k} {S}^{+}_{k-1} \Gamma_{k} ^{\mathrm{T}} +V_{k}.\end{split}\]

Here, \(\mathbf{w}^{+}_{k-1}\) and \({S}_{k-1}^{+}\) are the posterior kernel weight mean vector and covariance matrix at time \(k-1\), respectively. The transition matrix \(\Gamma_{k}\) represents the change of sample representation, and \({V}_{k}\) represents the finite matrix representation of the transition residual matrix.

Updater uses prediction

\({\color{green}\blacksquare}\) In the measurement space, the measurement particles are generated according to the measurement model as:

\[\mathbf{y}_k^{\{i\}} = \mathtt{h}\left({\mathbf{x}}_{k}^{\{i\}}, \mathbf{v}_{k}^{\{i\}} \right).\]

\({\color{orange}\blacksquare}\) In the RKHS, \({\mathbf{y}}_{k}^{\{i=1:M_{\text{A}}\}}\) are mapped as feature mappings \(\Upsilon_k\). The posterior kernel weight vector and covariance matrix are updated as:

\[\begin{split}\mathbf{w}^{+}_{k} &= \mathbf{w}^{-}_{k} + Q_k\left(\mathbf{g}_{\mathbf{yy}_k} - G_\mathbf{yy}\mathbf{w}^{-}_{k}\right)\\ S_k^{+} &= S_k^{-} - Q_k G_\mathbf{yy}S_k^{-}\end{split}\]

Here, \(G_\mathbf{yy} = \Upsilon^\mathrm{T}_k\Upsilon_k\), \(\mathbf{g}_{\mathbf{yy}_k}=\Upsilon_k^\mathrm{T}\phi_\mathbf{y}(\mathbf{y}_k)\) represents the kernel vector of the measurement at time \(k\), \(Q_k\) is the kernel Kalman gain operator.

Proposal generated in updater

\({\color{blue}\blacksquare}\) The AKKF replaces \(X_{k} = \mathbf{x}_k^{\{i=1:M\}}\) with new weighted proposal particles \(\tilde{X}_{k} = \tilde{\mathbf{x}}_k^{\{i=1:M\}}\), where

\[\begin{split}\tilde{\mathbf{x}}_k^{\{i=1:M_{\text{A}}\}} &\sim \mathcal{N}\left(\mathbb{E}\left(X_{k}\right), \mathrm{Cov}\left(X_{k}\right)\right)\\ \mathbb{E}\left(X_{k}\right) &= X_{k}\mathbf{w}^{+}_{k}\\ \mathrm{Cov}\left(X_{k}\right) &= X_{k}S^{+}_{k} X_{k}^{\rm{T}}.\end{split}\]
Flow diagram of the AKKF

Figure 2: Flow diagram of the AKKF

Nearly-constant velocity & Bearing-only Tracking (BOT) example

We consider the BOT problem with one object moving in a 2-D space. The moving trajectory follows a nearly-constant velocity motion mode, as in the previous tutorials. The measurement is the actual bearing with an additional Gaussian error term that is non-linear.

Ground truth

import numpy as np
from datetime import datetime, timedelta

np.random.seed(50)

Initialise Stone Soup ground truth and transition models.

from stonesoup.types.groundtruth import GroundTruthPath, GroundTruthState
from stonesoup.models.transition.linear import CombinedLinearGaussianTransitionModel, \
                                               ConstantVelocity

start_time = datetime.now().replace(microsecond=0)

q_x = 1 * 10 ** (-4)
q_y = 1 * 10 ** (-4)
transition_model = CombinedLinearGaussianTransitionModel([ConstantVelocity(q_x),
                                                          ConstantVelocity(q_y)])

Create the ground truth path.

truth = GroundTruthPath([GroundTruthState([-0.3, 0.001, 0.7, -0.055], timestamp=start_time)])

num_steps = 30
for k in range(1, num_steps):
    truth.append(GroundTruthState(
        transition_model.function(truth[k - 1], noise=True, time_interval=timedelta(seconds=1)),
        timestamp=start_time + timedelta(seconds=k)))

Plot the ground truth trajectory.

from stonesoup.plotter import Plotter
plotter = Plotter()
plotter.plot_ground_truths(truth, [0, 2])
plotter.fig
AKKF

Initialise the bearing using the appropriate measurement model.

from stonesoup.models.measurement.nonlinear import Cartesian2DToBearing
measurement_model = Cartesian2DToBearing(
    ndim_state=4,
    mapping=(0, 2),
    noise_covar=np.array([[0.005 ** 2]])
    )

Generate the set of bearing only measurements

Set up the AKKF

We have designed the key new components required in Stone Soup for the AKKF to run, including defining different component types, such as the KernelParticleState, Kernel, AdaptiveKernelKalmanPredictor and AdaptiveKernelKalmanUpdater. Stone Soup provides the design choice of the Kernel class to enable the use of different kernels and will permit the AKKF to be used for a wide variety of dynamic and measurement models, as well as future extensions for joint tracking and parameter estimation problems.

KernelParticleState

The KernelParticleState inherits the functionality of the ParticleState and adds the kernel weight covariance matrix property, i.e., kernel_covar().

Kernel

The Kernel class provides a transformation from state space into the RKHS. The kernel can be either a polynomial or a Gaussian kernel. The polynomial kernels, QuadraticKernel and QuarticKernel, have the following properties:

  • c: a parameter that trades off the influence of higher - order versus lower - order terms in the polynomial.

  • ialpha: the inverse of \(\alpha\) and is the slope parameter that controls the influence of the dot product on the kernel value.

The Gaussian kernel, GaussianKernel has the following property

  • variance: the variance parameter of the Gaussian kernel.

AdaptiveKernelKalmanPredictor

The AdaptiveKernelKalmanPredictor is a subclass of KalmanPredictor and inherits the methods and properties of the KalmanPredictor. The AdaptiveKernelKalmanPredictor includes the following properties:

  • lambda_predictor: \(\lambda_{\tilde{K}}\), is a regularisation parameter to stabilise the inverse of the Gram matrix of state particles’ feature mappings. It typically falls within the proper range of \(\left[10^{-4}, 10^{-2}\right]\).

  • kernel: the Kernel class which is chosen to be used to map the state space into the RKHS.

AdaptiveKernelKalmanUpdater

The AdaptiveKernelKalmanUpdater is a subclass of KalmanUpdater and inherits the methods and properties of the KalmanUpdater. The AdaptiveKernelKalmanUpdater includes the following new properties:

  • lambda_updater: \(\kappa\) is a regularisation parameter to ensure the inverse of \(G_{\mathbf{y}\mathbf{y}} S_k ^ {-}\) is well - defined.

  • kernel: the Kernel class which is chosen to be used to map the measurement space into the kernel space.

Initialise a prior

To begin, we initialise a prior estimate for the tracking problem. This is achieved by KernelParticleState that describes the state as a distribution of particles, each associated with StateVectors and corresponding weights. The prior state is sampled from a Gaussian distribution.

from scipy.stats import multivariate_normal

from stonesoup.types.state import KernelParticleState
from stonesoup.types.array import StateVectors

number_particles = 30

# Sample from the prior Gaussian distribution
np.random.seed(50)
samples = multivariate_normal.rvs(np.squeeze(truth[0].state_vector),
                                  np.diag([0.1, 0.005, 0.1, 0.01])**2,
                                  size=number_particles)

# Create prior particle state.
prior = KernelParticleState(state_vector=StateVectors(samples.T),
                            weight=np.array([1/number_particles]*number_particles),
                            timestamp=start_time,
                            )

from stonesoup.kernel import QuadraticKernel
from stonesoup.predictor.kernel import AdaptiveKernelKalmanPredictor
predictor = AdaptiveKernelKalmanPredictor(transition_model=transition_model,
                                          kernel=QuadraticKernel(c=1, ialpha=1))

from stonesoup.updater.kernel import AdaptiveKernelKalmanUpdater
updater = AdaptiveKernelKalmanUpdater(measurement_model=measurement_model,
                                      kernel=QuadraticKernel(c=1, ialpha=1))

Run the tracker

Now, we execute the predict and update steps of the tracker. The Predictor module takes both the prior and proposal information to propagate the state particles, and it calculates the predictive kernel weight vector and covariance matrix.

Next, the Updater module uses the predictions to generate the measurement particles and updates the posterior kernel weight vector and covariance matrix accordingly. Additionally, the Updater generates new proposal particles at every step to refine the state estimate.

Plot the resulting track with the sample points at each iteration.

plotter = Plotter()
plotter.plot_ground_truths(truth, [0, 2], linewidth=3.0, color='black')
plotter.plot_tracks(track, [0, 2], track_label='AKKF - quadratic', color='royalblue')
plotter.fig
AKKF
plotter.plot_tracks(track, [0, 2], particle=True)
plotter.fig
AKKF

Total running time of the script: (0 minutes 0.742 seconds)

Gallery generated by Sphinx-Gallery