Auxiliary particle filter


The auxiliary particle filter is a particle filtering algorithm introduced by Pitt and Shephard in 1999 to improve some deficiencies of the sequential importance resampling algorithm when dealing with tailed observation densities.

Motivation

Particle filters approximate continuous random variable by particles with discrete probability mass, say for uniform distribution. The random sampled particles can be used to approximate the probability density function of the continuous random variable if the value.
The empirical prediction density is produced as the weighted summation of these particles:
, and we can view it as the "prior" density. Note that the particles are assumed to have the same weight.
Combining the prior density and the likelihood, the empirical filtering density can be produced as:
, where.
On the other hand, the true filtering density which we want to estimate is
The prior density can be used to approximate the true filtering density :
The weakness of the particle filters includes:
Therefore, the auxiliary particle filter is proposed to solve this problem.

Auxiliary particle filter

Auxiliary variable

Comparing with the empirical filtering density which has,
we now define, where.
Being aware that is formed by the summation of particles, the auxiliary variable represents one specific particle. With the aid of, we can form a set of samples which has the distribution. Then, we draw from these sample set instead of directly from. In other words, the samples are drawn from with different probability. The samples are ultimately utilized to approximate.
Take the SIR method for example:
The original particle filters draw samples from the prior density, while the auxiliary filters draw from the joint distribution of the prior density and the likelihood. In other words, the auxiliary particle filters avoid the circumstance which the particles are generated in the regions with low likelihood. As a result, the samples can approximate more precisely.

Selection of the auxiliary variable

The selection of the auxiliary variable affects and controls the distribution of the samples. A possible selection of can be:
, where and is the mean.
We sample from to approximate by the following procedure:
Following the procedure, we draw the samples from. Since is closely related to the mean, it has high conditional likelihood. As a result, the sampling procedure is more efficient and the value can be reduced.

Other point of view

Assume that the filtered posterior is described by the following M weighted samples:
Then, each step in the algorithm consists of first drawing a sample of the particle index which will be propagated from into the new step. These indexes are auxiliary variables only used as an intermediary step, hence the name of the algorithm. The indexes are drawn according to the likelihood of some reference point which in some way is related to the transition model :
This is repeated for, and using these indexes we can now draw the conditional samples:
Finally, the weights are updated to account for the mismatch between the likelihood at the actual sample and the predicted point :