Kalman Filter algorithms (EKF,IEKF,UKF) are centralized in one single virtual class, mrpt::bayes::CKalmanFilterCapable. This class contains the system state vector and the system covariance matrix, as well as a generic method to execute one complete iteration of the selected algorithm. In practice, solving a specific problem requires deriving a new class from this virtual class and implementing a few methods such as transforming the state vector through the transition model, or computing the Jacobian of the observation model linearized at some given value of the state space.

## 1. Kalman Filters in the MRPT

Kalman Filter algorithms (EKF,IEKF,…) are centralized in one single virtual class, mrpt::bayes::CKalmanFilterCapable. This class contains the system state vector and the system covariance matrix, as well as a generic method to execute one complete iteration of the selected algorithm. In practice, solving a specific problem requires deriving a new class from this virtual class and implementing a few methods such as transforming the state vector through the transition model, or computing the Jacobian of the observation model linearized at some given value of the state space.

A set of parameters that are problem-independent can be changed in the member KF_options of this class, where the most important parameter is the selection of the KF algorithm. Currently it can be chosen from the following ones:

- Naive EKF: The basic EKF algorithm.
- Iterative Kalman Filter (IKF): This method re-linearizes the Jacobians around increasingly more accurate values of the state vector.

This class has been used to implement an efficient solution to 6D-SLAM.

Another implementation of Bayesian filtering in the MRPT C++ library are Particle Filters.

## 2. Writing a KF class for a specific problem

### 2.1. Deriving a new class

First a new class must be derived from mrpt::bayes::CKalmanFilterCapable. It should be declared a public method as an entry point for the user, which saves the input data (observations,actions,…) and calls the protected method runOneKalmanIteration() of the parent class.

There are two fundamental types of systems the user can build by deriving a new class:

- A normal problem: The size of the state vector remains unchanged over time. This size must be returned by the virtual method “get_vehicle_size()”.
- A mapping-like problem: The size of the state vector grows as new map elements are incorporated over time. In this case the first “get_vehicle_size()” elements of the

state vector correspond to the state of the vehicle/robot/camera/… and the rest of the state vector is a whole number of “get_feature_size()” sub-vectors, each describing one map element (landmark, feature,…).

### 2.2. The internal flow of the algorithm

The method mrpt::bayes::CKalmanFilterCapable::runOneKalmanIteration() will sequentially call each of the virtual methods, according to the following order:

- During the KF prediction stage:
- OnGetAction.
- OnTransitionModel.
- OnTransitionJacobian.
- OnTransitionNoise.

- OnNormalizeStateVector: This can be optionally implemented if required for the concrete problem.
- OnGetObservations: This is the ideal place for generating observations in those applications where it requires an estimate of the current state (e.g. in MonoSLAM, to predict where each landmark will be found onto the image). At this point the internal state vector and covariance contain the “prior distribution”, i.e. updated through the transition model. This is also the place where data association must be solved (in mapping problems).
- During the KF update stage:
- OnObservationModelAndJacobians: This method will be called just once in naive EKF, or several times in the EKF a la Davison.

- OnNormalizeStateVector.
- If the system implements a mapping problem, and the data association returned by “OnGetObservations” indicates the existence of unmapped observations, then the next method will be invoked for each of these new features:
- OnInverseObservationModel.

- OnPostIteration: A placeholder for any code the user wants to execute after each iteration (e.g. logging, visualization,…).

## 3. An example

An example of a KF implementation can be found in the samples directory (

samples/bayesianTracking) for the problem of tracking a vehicle from noisy range and bearing data. Here it will be derived the required equations to be implemented, as well as how they are actually implemented in this example. Note that this problem is also implemented as a Particle Filter in the same example in order to visualize side-to-side the performance of both approaches.

### 3.1. Problem Statement

The problem of range-bearing tracking is that of estimating the vehicle state:

where x and y are the Cartesian coordinates of the vehicle, and vx and vy are the linear velocities. Thus, we\’ll use a simple constant velocity model, where the transition function is:

We will consider the time interval between steps ?t as the action u of the system. The observation vector $z=(z_b ~ z_r)$latex consists of the bearing and range of the vehicle from some point (arbitrarily set to the origin):

Then, it is straightforward to obtain the required Jacobian of the transition function:

latexand the Jacobian of the observation model:

### 3.2 Implementation

The most important implemented methods are detailed below. For further details refer to the source code of the “bayesianTracking” example.

#### 3.2.1 The transition model

The constant-velocities model is implemented simply as:

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
/** Implements the transition model $latex \hat{x}_{k|k-1} = f( \hat{x}_{k-1|k-1}, u_k ) $ * \param in_u The vector returned by OnGetAction. * \param inout_x At input has $latex \hat{x}_{k-1|k-1} $, at output must have $latex \hat{x}_{k|k-1} $. * \param out_skip Set this to true if for some reason you want to skip the prediction step (to do not modify either the vector or the covariance). Default:false */ void CRangeBearing::OnTransitionModel( const KFArray_ACT &in_u, KFArray_VEH &inout_x, bool &out_skipPrediction ) const { // in_u[0] : Delta time // in_out_x: [0]:x [1]:y [2]:vx [3]: vy inout_x[0] += in_u[0] * inout_x[2]; inout_x[1] += in_u[0] * inout_x[3]; } |

#### 3.2.2 The transition model Jacobian

1 2 3 4 5 6 7 8 9 10 11 |
/** Implements the transition Jacobian $latex \frac{\partial f}{\partial x} $ * \param out_F Must return the Jacobian. * The returned matrix must be $N \times N$latex with N being either the size of the whole state vector or get_vehicle_size(). */ void CRangeBearing::OnTransitionJacobian(KFMatrix_VxV &F) const { F.unit(); F(0,2) = m_deltaTime; F(1,3) = m_deltaTime; } |

#### 3.2.3 The observations and the observation model

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
void CRangeBearing::OnGetObservationsAndDataAssociation( std::vector<KFArray_OBS> &out_z, vector_int &out_data_association, const vector<KFArray_OBS> &in_all_predictions, const KFMatrix &in_S, const vector_size_t &in_lm_indices_in_S, const KFMatrix_OxO &in_R ) { out_z.resize(1); out_z[0][0] = m_obsBearing; out_z[0][1] = m_obsRange; out_data_association.clear(); // Not used } /** Implements the observation prediction $latex h_i(x) $. * \param idx_landmark_to_predict The indices of the landmarks in the map whose predictions are expected as output. For non SLAM-like problems, this input value is undefined and the application should just generate one observation for the given problem. * \param out_predictions The predicted observations. */ void CRangeBearing::OnObservationModel( const vector_size_t &idx_landmarks_to_predict, std::vector<KFArray_OBS> &out_predictions ) const { // predicted bearing: kftype x = m_xkk[0]; kftype y = m_xkk[1]; kftype h_bear = atan2(y,x); kftype h_range = sqrt(square(x)+square(y)); // idx_landmarks_to_predict is ignored in NON-SLAM problems out_predictions.resize(1); out_predictions[0][0] = h_bear; out_predictions[0][1] = h_range; } |

### 3.3 Results

The following video shows the example samples/bayesianTracking running: