HOME

TheInfoList



OR:

Particle filters, or sequential Monte Carlo methods, are a set of
Monte Carlo Monte Carlo (; ; french: Monte-Carlo , or colloquially ''Monte-Carl'' ; lij, Munte Carlu ; ) is officially an administrative area of the Principality of Monaco, specifically the ward of Monte Carlo/Spélugues, where the Monte Carlo Casino is ...
algorithms used to solve filtering problems arising in
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
and Bayesian statistical inference. The
filtering problem In the theory of stochastic processes, filtering describes the problem of determining the state of a system from an incomplete and potentially noisy set of observations. While originally motivated by problems in engineering, filtering found appli ...
consists of estimating the internal states in
dynamical systems In mathematics, a dynamical system is a system in which a function describes the time dependence of a point in an ambient space. Examples include the mathematical models that describe the swinging of a clock pendulum, the flow of water in a p ...
when partial observations are made and random perturbations are present in the sensors as well as in the dynamical system. The objective is to compute the posterior distributions of the states of a
Markov process A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happe ...
, given the noisy and partial observations. The term "particle filters" was first coined in 1996 by Del Moral about mean-field interacting particle methods used in
fluid mechanics Fluid mechanics is the branch of physics concerned with the mechanics of fluids ( liquids, gases, and plasmas) and the forces on them. It has applications in a wide range of disciplines, including mechanical, aerospace, civil, chemical and bio ...
since the beginning of the 1960s. The term "Sequential Monte Carlo" was coined by Liu and Chen in 1998. Particle filtering uses a set of particles (also called samples) to represent the
posterior distribution The posterior probability is a type of conditional probability that results from updating the prior probability with information summarized by the likelihood via an application of Bayes' rule. From an epistemological perspective, the posterior p ...
of a
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appea ...
given the noisy and/or partial observations. The state-space model can be nonlinear and the initial state and noise distributions can take any form required. Particle filter techniques provide a well-established methodology for generating samples from the required distribution without requiring assumptions about the state-space model or the state distributions. However, these methods do not perform well when applied to very high-dimensional systems. Particle filters update their prediction in an approximate (statistical) manner. The samples from the distribution are represented by a set of particles; each particle has a likelihood weight assigned to it that represents the
probability Probability is the branch of mathematics concerning numerical descriptions of how likely an Event (probability theory), event is to occur, or how likely it is that a proposition is true. The probability of an event is a number between 0 and ...
of that particle being sampled from the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
. Weight disparity leading to weight collapse is a common issue encountered in these filtering algorithms, however, it can be mitigated by including a resampling step before the weights become uneven. Several adaptive resampling criteria can be used including the
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers ...
of the weights and the relative
entropy Entropy is a scientific concept, as well as a measurable physical property, that is most commonly associated with a state of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynam ...
concerning the uniform distribution. In the resampling step, the particles with negligible weights are replaced by the new particles in the proximity of the particles with higher weights. From the statistical and probabilistic point of view, particle filters may be interpreted as mean-field particle interpretations of Feynman-Kac probability measures. These particle integration techniques were developed in
molecular chemistry Chemistry is the scientific study of the properties and behavior of matter. It is a natural science that covers the elements that make up matter to the compounds made of atoms, molecules and ions: their composition, structure, properties, ...
and
computational physics Computational physics is the study and implementation of numerical analysis to solve problems in physics for which a quantitative theory already exists. Historically, computational physics was the first application of modern computers in science, ...
by Theodore E. Harris and
Herman Kahn Herman Kahn (February 15, 1922 – July 7, 1983) was a founder of the Hudson Institute and one of the preeminent futurists of the latter part of the twentieth century. He originally came to prominence as a military strategist and systems theori ...
in 1951, Marshall N. Rosenbluth and
Arianna W. Rosenbluth Arianna Wright Rosenbluth (September 15, 1927 – December 28, 2020) was an American physicist who contributed to the development of the Metropolis–Hastings algorithm. She wrote the first full implementation of the Markov chain Monte Carlo meth ...
in 1955, and more recently by Jack H. Hetherington in 1984. In computational physics, these Feynman-Kac type path particle integration methods are also used in
Quantum Monte Carlo Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
, and more specifically Diffusion Monte Carlo methods. Feynman-Kac interacting particle methods are also strongly related to mutation-selection genetic algorithms currently used in
evolutionary computing In computer science, evolutionary computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they ...
to solve complex optimization problems. The particle filter methodology is used to solve
Hidden Markov Model A hidden Markov model (HMM) is a statistical Markov model in which the system being modeled is assumed to be a Markov process — call it X — with unobservable ("''hidden''") states. As part of the definition, HMM requires that there be an ob ...
(HMM) and
nonlinear filter In signal processing, a nonlinear (or non-linear) filter is a filter whose output is not a linear function of its input. That is, if the filter outputs signals ''R'' and ''S'' for two input signals ''r'' and ''s'' separately, but does not always o ...
ing problems. With the notable exception of linear-Gaussian signal-observation models (
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estimat ...
) or wider classes of models (Benes filter), Mireille Chaleyat-Maurel and Dominique Michel proved in 1984 that the sequence of posterior distributions of the random states of a signal, given the observations (a.k.a. optimal filter), has no finite recursion. Various other numerical methods based on fixed grid approximations,
Markov Chain Monte Carlo In statistics, Markov chain Monte Carlo (MCMC) methods comprise a class of algorithms for sampling from a probability distribution. By constructing a Markov chain that has the desired distribution as its equilibrium distribution, one can obtain ...
techniques, conventional linearization,
extended Kalman filter In estimation theory, the extended Kalman filter (EKF) is the nonlinear version of the Kalman filter which linearizes about an estimate of the current mean and covariance. In the case of well defined transition models, the EKF has been considered t ...
s, or determining the best linear system (in the expected cost-error sense) are unable to cope with large-scale systems, unstable processes, or when the nonlinearities are not sufficiently smooth. Particle filters and Feynman-Kac particle methodologies find application in signal and image processing,
Bayesian inference Bayesian inference is a method of statistical inference in which Bayes' theorem is used to update the probability for a hypothesis as more evidence or information becomes available. Bayesian inference is an important technique in statistics, a ...
,
machine learning Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine ...
, risk analysis and rare event sampling,
engineering Engineering is the use of scientific method, scientific principles to design and build machines, structures, and other items, including bridges, tunnels, roads, vehicles, and buildings. The discipline of engineering encompasses a broad rang ...
and robotics,
artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech re ...
,
bioinformatics Bioinformatics () is an interdisciplinary field that develops methods and software tools for understanding biological data, in particular when the data sets are large and complex. As an interdisciplinary field of science, bioinformatics combi ...
,
phylogenetics In biology, phylogenetics (; from Greek language, Greek wikt:φυλή, φυλή/wikt:φῦλον, φῦλον [] "tribe, clan, race", and wikt:γενετικός, γενετικός [] "origin, source, birth") is the study of the evolutionary his ...
, computational science, economics [
and or AND may refer to: Logic, grammar, and computing * Conjunction (grammar), connecting two words, phrases, or clauses * Logical conjunction in mathematical logic, notated as "∧", "⋅", "&", or simple juxtaposition * Bitwise AND, a boole ...
mathematical finance Mathematical finance, also known as quantitative finance and financial mathematics, is a field of applied mathematics, concerned with mathematical modeling of financial markets. In general, there exist two separate branches of finance that require ...
,
molecular chemistry Chemistry is the scientific study of the properties and behavior of matter. It is a natural science that covers the elements that make up matter to the compounds made of atoms, molecules and ions: their composition, structure, properties, ...
,
computational physics Computational physics is the study and implementation of numerical analysis to solve problems in physics for which a quantitative theory already exists. Historically, computational physics was the first application of modern computers in science, ...
,
pharmacokinetics Pharmacokinetics (from Ancient Greek ''pharmakon'' "drug" and ''kinetikos'' "moving, putting in motion"; see chemical kinetics), sometimes abbreviated as PK, is a branch of pharmacology dedicated to determining the fate of substances administered ...
, and other fields.


History


Heuristic-like algorithms

From a statistical and probabilistic viewpoint, particle filters belong to the class of branching/ genetic type algorithms, and mean-field type interacting particle methodologies. The interpretation of these particle methods depends on the scientific discipline. In
Evolutionary Computing In computer science, evolutionary computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they ...
, mean-field genetic type particle methodologies are often used as a heuristic and natural search algorithms (a.k.a.
Metaheuristic In computer science and mathematical optimization, a metaheuristic is a higher-level procedure or heuristic designed to find, generate, or select a heuristic (partial search algorithm) that may provide a sufficiently good solution to an optimizati ...
). In
computational physics Computational physics is the study and implementation of numerical analysis to solve problems in physics for which a quantitative theory already exists. Historically, computational physics was the first application of modern computers in science, ...
and
molecular chemistry Chemistry is the scientific study of the properties and behavior of matter. It is a natural science that covers the elements that make up matter to the compounds made of atoms, molecules and ions: their composition, structure, properties, ...
, they are used to solve Feynman-Kac path integration problems or to compute Boltzmann-Gibbs measures, top eigenvalues and ground states of Schrödinger operators. In
Biology Biology is the scientific study of life. It is a natural science with a broad scope but has several unifying themes that tie it together as a single, coherent field. For instance, all organisms are made up of cells that process hereditary i ...
and
Genetics Genetics is the study of genes, genetic variation, and heredity in organisms.Hartl D, Jones E (2005) It is an important branch in biology because heredity is vital to organisms' evolution. Gregor Mendel, a Moravian Augustinian friar wor ...
, they represent the evolution of a population of individuals or genes in some environment. The origins of mean-field type evolutionary computational techniques can be traced back to 1950 and 1954 with Alan Turing's work on genetic type mutation-selection learning machines and the articles by
Nils Aall Barricelli Nils Aall Barricelli (24 January 1912 – 27 January 1993) was a Norwegian-Italian mathematician. Barricelli's early computer-assisted experiments in symbiogenesis and evolution are considered pioneering in artificial life research. Barricel ...
at the
Institute for Advanced Study The Institute for Advanced Study (IAS), located in Princeton, New Jersey, in the United States, is an independent center for theoretical research and intellectual inquiry. It has served as the academic home of internationally preeminent scholar ...
in
Princeton, New Jersey Princeton is a municipality with a borough form of government in Mercer County, in the U.S. state of New Jersey. It was established on January 1, 2013, through the consolidation of the Borough of Princeton and Princeton Township, both of whi ...
. The first trace of particle filters in
statistical methodology Statistics (from German: ''Statistik'', "description of a state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a scientific, industria ...
dates back to the mid-1950s; the 'Poor Man's Monte Carlo', that was proposed by Hammersley et al., in 1954, contained hints of the genetic type particle filtering methods used today. In 1963,
Nils Aall Barricelli Nils Aall Barricelli (24 January 1912 – 27 January 1993) was a Norwegian-Italian mathematician. Barricelli's early computer-assisted experiments in symbiogenesis and evolution are considered pioneering in artificial life research. Barricel ...
simulated a genetic type algorithm to mimic the ability of individuals to play a simple game. In
evolutionary computing In computer science, evolutionary computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they ...
literature, genetic type mutation-selection algorithms became popular through the seminal work of John Holland in the early 1970s, and particularly his book published in 1975. In Biology and
Genetics Genetics is the study of genes, genetic variation, and heredity in organisms.Hartl D, Jones E (2005) It is an important branch in biology because heredity is vital to organisms' evolution. Gregor Mendel, a Moravian Augustinian friar wor ...
, the Australian geneticist Alex Fraser also published in 1957 a series of papers on the genetic type simulation of
artificial selection Selective breeding (also called artificial selection) is the process by which humans use animal breeding and plant breeding to selectively develop particular phenotypic traits (characteristics) by choosing which typically animal or plant m ...
of organisms. The computer simulation of the evolution by biologists became more common in the early 1960s, and the methods were described in books by Fraser and Burnell (1970) and Crosby (1973). Fraser's simulations included all of the essential elements of modern mutation-selection genetic particle algorithms. From the mathematical viewpoint, the conditional distribution of the random states of a signal given some partial and noisy observations is described by a Feynman-Kac probability on the random trajectories of the signal weighted by a sequence of likelihood potential functions.
Quantum Monte Carlo Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
, and more specifically Diffusion Monte Carlo methods can also be interpreted as a mean-field genetic type particle approximation of Feynman-Kac path integrals. The origins of Quantum Monte Carlo methods are often attributed to Enrico Fermi and Robert Richtmyer who developed in 1948 a mean-field particle interpretation of neutron-chain reactions, but the first heuristic-like and genetic type particle algorithm (a.k.a. Resampled or Reconfiguration Monte Carlo methods) for estimating ground state energies of quantum systems (in reduced matrix models) is due to Jack H. Hetherington in 1984. One can also quote the earlier seminal works of Theodore E. Harris and
Herman Kahn Herman Kahn (February 15, 1922 – July 7, 1983) was a founder of the Hudson Institute and one of the preeminent futurists of the latter part of the twentieth century. He originally came to prominence as a military strategist and systems theori ...
in particle physics, published in 1951, using mean-field but heuristic-like genetic methods for estimating particle transmission energies. In molecular chemistry, the use of genetic heuristic-like particle methodologies (a.k.a. pruning and enrichment strategies) can be traced back to 1955 with the seminal work of Marshall. N. Rosenbluth and Arianna. W. Rosenbluth. The use of genetic particle algorithms in advanced
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
and
Bayesian inference Bayesian inference is a method of statistical inference in which Bayes' theorem is used to update the probability for a hypothesis as more evidence or information becomes available. Bayesian inference is an important technique in statistics, a ...
is more recent. In January 1993, Genshiro Kitagawa developed a "Monte Carlo filter", a slightly modified version of this article appeared in 1996. In April 1993, Gordon et al., published in their seminal work an application of genetic type algorithm in Bayesian statistical inference. The authors named their algorithm 'the bootstrap filter', and demonstrated that compared to other filtering methods, their bootstrap algorithm does not require any assumption about that state space or the noise of the system. Independently, the ones by Pierre Del Moral and Himilcon Carvalho, Pierre Del Moral, André Monin, and Gérard Salut on particle filters published in the mid-1990s. Particle filters were also developed in signal processing in early 1989-1992 by P. Del Moral, J.C. Noyer, G. Rigal, and G. Salut in the LAAS-CNRS in a series of restricted and classified research reports with STCAN (Service Technique des Constructions et Armes Navales), the IT company DIGILOG, and th
LAAS-CNRS
(the Laboratory for Analysis and Architecture of Systems) on RADAR/SONAR and GPS signal processing problems.


Mathematical foundations

From 1950 to 1996, all the publications on particle filters, and genetic algorithms, including the pruning and resample Monte Carlo methods introduced in computational physics and molecular chemistry, present natural and heuristic-like algorithms applied to different situations without a single proof of their consistency, nor a discussion on the bias of the estimates and genealogical and ancestral tree-based algorithms. The mathematical foundations and the first rigorous analysis of these particle algorithms are due to Pierre Del Moral in 1996. The article also contains proof of the unbiased properties of a particle approximation of likelihood functions and unnormalized
conditional probability In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) has already occurred. This particular method relies on event B occur ...
measures. The unbiased particle estimator of the likelihood functions presented in this article is used today in Bayesian statistical inference. Branching type particle methodologies with varying population sizes were also developed toward the end of the 1990s by Dan Crisan, Jessica Gaines and Terry Lyons, and by Dan Crisan, Pierre Del Moral and Terry Lyons. Further developments in this field were made in 2000 by P. Del Moral, A. Guionnet and L. Miclo. The first central limit theorems are due to Pierre Del Moral and Alice Guionnet in 1999 and Pierre Del Moral and Laurent Miclo in 2000. The first uniform convergence results with respect to the time parameter for particle filters were developed in the end of the 1990s by Pierre Del Moral and Alice Guionnet. The first rigorous analysis of genealogical tree based particle filter smoothers is due to P. Del Moral and L. Miclo in 2001 The theory on Feynman-Kac particle methodologies and related particle filter algorithms was developed in 2000 and 2004 in the books. These abstract probabilistic models encapsulate genetic type algorithms, particle and bootstrap filters, interacting Kalman filters (a.k.a. Rao–Blackwellized particle filter ), importance sampling and resampling style particle filter techniques, including genealogical tree-based and particle backward methodologies for solving filtering and smoothing problems. Other classes of particle filtering methodologies include genealogical tree-based models, backward Markov particle models, adaptive mean-field particle models, island-type particle models, and particle Markov chain Monte Carlo methodologies.


The filtering problem


Objective

The objective of a particle filter is to estimate the posterior density of the state variables given the observation variables. The particle filter is designed for a
hidden Markov Model A hidden Markov model (HMM) is a statistical Markov model in which the system being modeled is assumed to be a Markov process — call it X — with unobservable ("''hidden''") states. As part of the definition, HMM requires that there be an ob ...
, where the system consists of both hidden and observable variables. The observable variables (observation process) are related to the hidden variables (state-process) by some functional form that is known. Similarly the dynamical system describing the evolution of the state variables is also known probabilistically. A generic particle filter estimates the posterior distribution of the hidden states using the observation measurement process. With respect to a state-space such as the one below: :\begin X_0&\to &X_1&\to &X_2&\to&X_3&\to &\cdots&\text\\ \downarrow&&\downarrow&&\downarrow&&\downarrow&&\cdots&\\ Y_0&&Y_1&&Y_2&&Y_3&&\cdots&\text \end the filtering problem is to estimate sequentially the values of the hidden states X_k, given the values of the observation process Y_0,\cdots,Y_k, at any time step ''k''. All Bayesian estimates of X_k follow from the posterior density p(x_k, y_0,y_1,...,y_k). The particle filter methodology provides an approximation of these conditional probabilities using the empirical measure associated with a genetic type particle algorithm. In contrast, the Markov Chain Monte Carlo or
importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally att ...
approach would model the full posterior p(x_0,x_1,...,x_k, y_0,y_1,...,y_k).


The Signal-Observation model

Particle methods often assume X_k and the observations Y_k can be modeled in this form: *X_0, X_1, \cdots is a
Markov process A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happe ...
on \mathbb R^ (for some d_x\geqslant 1) that evolves according to the transition probability density p(x_k, x_). This model is also often written in a synthetic way as *:X_k, X_=x_k \sim p(x_k, x_) :with an initial probability density p(x_0). *The observations Y_0, Y_1, \cdots take values in some state space on \mathbb^ (for some d_y\geqslant 1) and are conditionally independent provided that X_0, X_1, \cdots are known. In other words, each Y_k only depends on X_k. In addition, we assume conditional distribution for Y_k given X_k=x_k are absolutely continuous, and in a synthetic way we have *:Y_k, X_k=y_k \sim p(y_k, x_k) An example of system with these properties is: :X_k = g(X_) + W_ :Y_k = h(X_k) + V_k where both W_k and V_k are mutually independent sequences with known
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
s and ''g'' and ''h'' are known functions. These two equations can be viewed as
state space A state space is the set of all possible configurations of a system. It is a useful abstraction for reasoning about the behavior of a given system and is widely used in the fields of artificial intelligence and game theory. For instance, the toy ...
equations and look similar to the state space equations for the Kalman filter. If the functions ''g'' and ''h'' in the above example are linear, and if both W_k and V_k are
Gaussian Carl Friedrich Gauss (1777–1855) is the eponym of all of the topics listed below. There are over 100 topics all named after this German mathematician and scientist, all in the fields of mathematics, physics, and astronomy. The English eponymo ...
, the Kalman filter finds the exact Bayesian filtering distribution. If not, Kalman filter-based methods are a first-order approximation ( EKF) or a second-order approximation ( UKF in general, but if the probability distribution is Gaussian a third-order approximation is possible). The assumption that the initial distribution and the transitions of the Markov chain are continuous for the
Lebesgue measure In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of ''n''-dimensional Euclidean space. For ''n'' = 1, 2, or 3, it coincides wit ...
can be relaxed. To design a particle filter we simply need to assume that we can sample the transitions X_ \to X_k of the Markov chain X_k, and to compute the likelihood function x_k\mapsto p(y_k, x_k) (see for instance the genetic selection mutation description of the particle filter given below). The continuous assumption on the Markov transitions of X_k is only used to derive in an informal (and rather abusive) way different formulae between posterior distributions using the Bayes' rule for conditional densities.


Approximate Bayesian computation models

In certain problems, the conditional distribution of observations, given the random states of the signal, may fail to have a density; the latter may be impossible or too complex to compute. In this situation, an additional level of approximation is necessitated. One strategy is to replace the signal X_k by the Markov chain \mathcal X_k=\left(X_k,Y_k\right) and to introduce a virtual observation of the form :\mathcal Y_k=Y_k+\epsilon \mathcal V_k\quad\mbox\quad\epsilon\in ,1/math> for some sequence of independent random variables \mathcal V_k with known
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
s. The central idea is to observe that :\text\left(X_k, \mathcal Y_0=y_0,\cdots, \mathcal Y_k=y_k\right)\approx_ \text\left(X_k, Y_0=y_0,\cdots, Y_k=y_k\right) The particle filter associated with the Markov process \mathcal X_k=\left(X_k,Y_k\right) given the partial observations \mathcal Y_0=y_0,\cdots, \mathcal Y_k=y_k, is defined in terms of particles evolving in \mathbb R^ with a likelihood function given with some obvious abusive notation by p(\mathcal Y_k, \mathcal X_k). These probabilistic techniques are closely related to
Approximate Bayesian Computation Approximate Bayesian computation (ABC) constitutes a class of computational methods rooted in Bayesian statistics that can be used to estimate the posterior distributions of model parameters. In all model-based statistical inference, the like ...
(ABC). In the context of particle filters, these ABC particle filtering techniques were introduced in 1998 by P. Del Moral, J. Jacod and P. Protter. They were further developed by P. Del Moral, A. Doucet and A. Jasra.


The nonlinear filtering equation

Bayes' rule In probability theory and statistics, Bayes' theorem (alternatively Bayes' law or Bayes' rule), named after Thomas Bayes, describes the probability of an event, based on prior knowledge of conditions that might be related to the event. For examp ...
for conditional probability gives: :p(x_0, \cdots, x_k, y_0,\cdots,y_k) =\frac where :\begin p(y_0,\cdots,y_k) &=\int p(y_0,\cdots,y_k, x_0,\cdots, x_k) p(x_0,\cdots,x_k) dx_0\cdots dx_k \\ p(y_0,\cdots, y_k, x_0,\cdots ,x_k) &=\prod_^ p(y_l, x_l) \\ p(x_0,\cdots, x_k) &=p_0(x_0)\prod_^ p(x_l, x_) \end Particle filters are also an approximation, but with enough particles they can be much more accurate. The nonlinear filtering equation is given by the recursion with the convention p(x_0, y_0,\cdots,y_)=p(x_0) for ''k'' = 0. The nonlinear filtering problem consists in computing these conditional distributions sequentially.


Feynman-Kac formulation

We fix a time horizon n and a sequence of observations Y_0=y_0,\cdots,Y_n=y_n, and for each ''k'' = 0, ..., ''n'' we set: :G_k(x_k)=p(y_k, x_k). In this notation, for any bounded function ''F'' on the set of trajectories of X_k from the origin ''k'' = 0 up to time ''k'' = ''n'', we have the Feynman-Kac formula :\begin \int F(x_0,\cdots,x_n) p(x_0,\cdots,x_n, y_0,\cdots,y_n) dx_0\cdots dx_n &= \frac\\ &=\frac \end Feynman-Kac path integration models arise in a variety of scientific disciplines, including in computational physics, biology, information theory and computer sciences. Their interpretations are dependent on the application domain. For instance, if we choose the indicator function G_n(x_n)=1_A(x_n) of some subset of the state space, they represent the conditional distribution of a Markov chain given it stays in a given tube; that is, we have: :E\left(F(X_0,\cdots,X_n) , X_0\in A, \cdots, X_n\in A\right) =\frac and :P\left(X_0\in A,\cdots, X_n\in A\right)=E\left(\prod\limits_^ G_k(X_k)\right) as soon as the normalizing constant is strictly positive.


Particle filters


A Genetic type particle algorithm

Initially, such an algorithm starts with ''N'' independent random variables \left(\xi^i_0\right)_ with common probability density p(x_0). The genetic algorithm selection-mutation transitions :\xi_k:=\left(\xi^i_\right)_\stackrel \widehat_k:=\left(\widehat^i_\right)_\stackrel \xi_:=\left(\xi^i_\right)_ mimic/approximate the updating-prediction transitions of the optimal filter evolution (): * During the selection-updating transition we sample ''N'' (conditionally) independent random variables \widehat_k:=\left(\widehat^i_\right)_ with common (conditional) distribution ::\sum_^N \frac \delta_(dx_k) where \delta_a stands for the
Dirac measure In mathematics, a Dirac measure assigns a size to a set based solely on whether it contains a fixed element ''x'' or not. It is one way of formalizing the idea of the Dirac delta function, an important tool in physics and other technical fields. ...
at a given state a. * During the mutation-prediction transition, from each selected particle \widehat^i_k we sample independently a transition ::\widehat^i_k \longrightarrow\xi^i_ \sim p(x_, \widehat^i_k), \qquad i=1,\cdots,N. In the above displayed formulae p(y_k, \xi^i_k) stands for the likelihood function x_k\mapsto p(y_k, x_k) evaluated at x_k=\xi^i_k, and p(x_, \widehat^i_k) stands for the conditional density p(x_, x_k) evaluated at x_k=\widehat^i_k. At each time ''k'', we have the particle approximations :\widehat(dx_k, y_0,\cdots,y_k):=\frac \sum_^N \delta_ (dx_k) \approx_ p(dx_k, y_0,\cdots,y_k) \approx_ \sum_^N \frac \delta_(dx_k) and :\widehat(dx_k, y_0,\cdots,y_):=\frac\sum_^N \delta_(dx_k) \approx_ p(dx_k, y_0,\cdots,y_) In Genetic algorithms and
Evolutionary computing In computer science, evolutionary computation is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms, they ...
community, the mutation-selection Markov chain described above is often called the genetic algorithm with proportional selection. Several branching variants, including with random population sizes have also been proposed in the articles.


Monte Carlo principles

Particle methods, like all sampling-based approaches (e.g.,
Markov Chain Monte Carlo In statistics, Markov chain Monte Carlo (MCMC) methods comprise a class of algorithms for sampling from a probability distribution. By constructing a Markov chain that has the desired distribution as its equilibrium distribution, one can obtain ...
), generate a set of samples that approximate the filtering density :p(x_k, y_0, \cdots, y_k). For example, we may have ''N'' samples from the approximate posterior distribution of X_k, where the samples are labeled with superscripts as: :\widehat_k^1, \cdots, \widehat_k^. Then, expectations with respect to the filtering distribution are approximated by with :\widehat(dx_k, y_0,\cdots,y_k)=\frac\sum_^N \delta_(dx_k) where \delta_a stands for the
Dirac measure In mathematics, a Dirac measure assigns a size to a set based solely on whether it contains a fixed element ''x'' or not. It is one way of formalizing the idea of the Dirac delta function, an important tool in physics and other technical fields. ...
at a given state a. The function ''f'', in the usual way for Monte Carlo, can give all the moments etc. of the distribution up to some approximation error. When the approximation equation () is satisfied for any bounded function ''f'' we write :p(dx_k, y_0,\cdots,y_k):=p(x_k, y_0,\cdots,y_k) dx_k \approx_ \widehat(dx_k, y_0,\cdots,y_k)=\frac\sum_^N \delta_(dx_k) Particle filters can be interpreted as a genetic type particle algorithm evolving with mutation and selection transitions. We can keep track of the ancestral lines :\left(\widehat^_, \widehat^_,\cdots,\widehat^_,\widehat^i_\right) of the particles i=1,\cdots,N. The random states \widehat^_, with the lower indices l=0,...,k, stands for the ancestor of the individual \widehat^_=\widehat^i_k at level l=0,...,k. In this situation, we have the approximation formula with the
empirical measure In probability theory, an empirical measure is a random measure arising from a particular realization of a (usually finite) sequence of random variables. The precise definition is found below. Empirical measures are relevant to mathematical stat ...
:\widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_k):=\frac\sum_^N \delta_(d(x_0,\cdots,x_k)) Here ''F'' stands for any founded function on the path space of the signal. In a more synthetic form () is equivalent to :\begin p(d(x_0,\cdots,x_k), y_0,\cdots,y_k)&:=p(x_0,\cdots,x_k, y_0,\cdots,y_k) \, dx_0\cdots dx_k \\ &\approx_ \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_k) \\ &:=\frac\sum_^N \delta_(d(x_0,\cdots,x_k)) \end Particle filters can be interpreted in many different ways. From the probabilistic point of view they coincide with a mean-field particle interpretation of the nonlinear filtering equation. The updating-prediction transitions of the optimal filter evolution can also be interpreted as the classical genetic type selection-mutation transitions of individuals. The sequential importance resampling technique provides another interpretation of the filtering transitions coupling importance sampling with the bootstrap resampling step. Last, but not least, particle filters can be seen as an acceptance-rejection methodology equipped with a recycling mechanism.


Mean-field particle simulation


The general probabilistic principle

The nonlinear filtering evolution can be interpreted as a dynamical system in the set of probability measures of the following form \eta_=\Phi_\left(\eta_\right) where \Phi_ stands for some mapping from the set of probability distribution into itself. For instance, the evolution of the one-step optimal predictor \eta_n(dx_n) =p(x_n, y_0,\cdots,y_)dx_n satisfies a nonlinear evolution starting with the probability distribution \eta_0(dx_0)=p(x_0)dx_0. One of the simplest ways to approximate these probability measures is to start with ''N'' independent random variables \left(\xi^i_0\right)_ with common probability distribution \eta_0(dx_0)=p(x_0)dx_0 . Suppose we have defined a sequence of ''N'' random variables \left(\xi^i_n\right)_ such that :\frac\sum_^N \delta_(dx_n) \approx_ \eta_n(dx_n) At the next step we sample ''N'' (conditionally) independent random variables \xi_:=\left(\xi^i_\right)_ with common law . :\Phi_\left(\frac\sum_^N \delta_\right) \approx_ \Phi_\left(\eta_\right)=\eta_


A particle interpretation of the filtering equation

We illustrate this mean-field particle principle in the context of the evolution of the one step optimal predictors For ''k'' = 0 we use the convention p(x_0, y_0,\cdots,y_):=p(x_0). By the law of large numbers, we have :\widehat(dx_0)=\frac\sum_^N \delta_(dx_0)\approx_ p(x_0)dx_0 in the sense that :\int f(x_0)\widehat(dx_0)=\frac\sum_^N f(\xi^i_0)\approx_ \int f(x_0)p(dx_0)dx_0 for any bounded function f. We further assume that we have constructed a sequence of particles \left(\xi^i_k\right)_ at some rank ''k'' such that :\widehat(dx_k, y_0,\cdots,y_):=\frac\sum_^N \delta_(dx_k)\approx_~p(x_k~, ~y_0,\cdots,y_)dx_k in the sense that for any bounded function f we have :\int f(x_k)\widehat(dx_k, y_0,\cdots,y_)=\frac\sum_^N f(\xi^i_k)\approx_ \int f(x_k)p(dx_k, y_0,\cdots,y_)dx_k In this situation, replacing p(x_k, y_0,\cdots,y_) dx_k by the
empirical measure In probability theory, an empirical measure is a random measure arising from a particular realization of a (usually finite) sequence of random variables. The precise definition is found below. Empirical measures are relevant to mathematical stat ...
\widehat(dx_k, y_0,\cdots,y_) in the evolution equation of the one-step optimal filter stated in () we find that :p(x_, y_0,\cdots,y_k)\approx_ \int p(x_, x'_) \frac Notice that the right hand side in the above formula is a weighted probability mixture :\int p(x_, x'_) \frac=\sum_^N \frac p(x_, \xi^i_k)=:\widehat(x_, y_0,\cdots,y_k) where p(y_k, \xi^i_k) stands for the density p(y_k, x_k) evaluated at x_k=\xi^i_k, and p(x_, \xi^i_k) stands for the density p(x_, x_k) evaluated at x_k=\xi^i_k for i=1,\cdots,N. Then, we sample ''N'' independent random variable \left(\xi^i_\right)_ with common probability density \widehat(x_, y_0,\cdots,y_k) so that :\widehat(dx_, y_0,\cdots,y_):=\frac\sum_^N \delta_(dx_)\approx_ \widehat(x_, y_0,\cdots,y_) dx_ \approx_ p(x_, y_0,\cdots,y_)dx_ Iterating this procedure, we design a Markov chain such that :\widehat(dx_k, y_0,\cdots,y_):=\frac\sum_^N \delta_(dx_k) \approx_ p(dx_k, y_0,\cdots,y_):=p(x_k, y_0,\cdots,y_) dx_k Notice that the optimal filter is approximated at each time step k using the Bayes' formulae :p(dx_, y_0,\cdots,y_) \approx_ \frac=\sum_^N \frac~\delta_(dx_k) The terminology "mean-field approximation" comes from the fact that we replace at each time step the probability measure p(dx_k, y_0,\cdots,y_) by the empirical approximation \widehat(dx_k, y_0,\cdots,y_). The mean-field particle approximation of the filtering problem is far from being unique. Several strategies are developed in the books.


Some convergence results

The analysis of the convergence of particle filters was started in 1996 and in 2000 in the book and the series of articles. More recent developments can be found in the books, When the filtering equation is stable (in the sense that it corrects any erroneous initial condition), the bias and the variance of the particle particle estimates :I_k(f):=\int f(x_k) p(dx_k, y_0,\cdots,y_) \approx_ \widehat_k(f):=\int f(x_k) \widehat(dx_k, y_0,\cdots,y_) are controlled by the non asymptotic uniform estimates :\sup_\left\vert E\left(\widehat_k(f)\right)-I_k(f)\right\vert\leqslant \frac :\sup_E\left(\left widehat_k(f)-I_k(f)\right2\right)\leqslant \frac for any function ''f'' bounded by 1, and for some finite constants c_1,c_2. In addition, for any x\geqslant 0: :\mathbf \left ( \left, \widehat_k(f)-I_k(f)\\leqslant c_1 \frac+c_2 \sqrt\land \sup_\left, \widehat_k(f)-I_k(f)\\leqslant c \sqrt \right ) > 1-e^ for some finite constants c_1, c_2 related to the asymptotic bias and variance of the particle estimate, and some finite constant ''c''. The same results are satisfied if we replace the one step optimal predictor by the optimal filter approximation.


Genealogical trees and Unbiasedness properties


Genealogical tree based particle smoothing

Tracing back in time the ancestral lines :\left(\widehat^i_,\widehat^i_,\cdots,\widehat^i_,\widehat^i_\right), \quad \left(\xi^i_,\xi^i_,\cdots,\xi^i_,\xi_\right) of the individuals \widehat^i_\left(=\widehat^i_\right) and \xi^i_\left(=^i_\right) at every time step ''k'', we also have the particle approximations :\begin \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_k) &:=\frac\sum_^N \delta_(d(x_0,\cdots,x_k)) \\ &\approx_ p(d(x_0,\cdots,x_k), y_0,\cdots,y_k) \\ &\approx_ \sum_^N \frac \delta_(d(x_0,\cdots,x_k)) \\ & \ \\ \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_) &:=\frac\sum_^N \delta_(d(x_0,\cdots,x_k)) \\ &\approx_ p(d(x_0,\cdots,x_k), y_0,\cdots,y_) \\ &:=p(x_0,\cdots,x_k, y_0,\cdots,y_) dx_0,\cdots,dx_k \end These empirical approximations are equivalent to the particle integral approximations :\begin \int F(x_0,\cdots,x_n) \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_k) &:=\frac\sum_^N F\left(\widehat^i_,\cdots,\widehat^i_\right) \\ &\approx_ \int F(x_0,\cdots,x_n) p(d(x_0,\cdots,x_k), y_0,\cdots,y_k) \\ &\approx_ \sum_^N \frac F\left(\xi^i_, \cdots,\xi^i_ \right) \\ & \ \\ \int F(x_0,\cdots,x_n) \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_) &:=\frac \sum_^N F\left(\xi^i_,\cdots,\xi^i_\right) \\ &\approx_ \int F(x_0,\cdots,x_n) p(d(x_0,\cdots,x_k), y_0,\cdots,y_) \end for any bounded function ''F'' on the random trajectories of the signal. As shown in the evolution of the genealogical tree coincides with a mean-field particle interpretation of the evolution equations associated with the posterior densities of the signal trajectories. For more details on these path space models, we refer to the books.


Unbiased particle estimates of likelihood functions

We use the product formula :p(y_0,\cdots,y_n)=\prod_^n p(y_k, y_0,\cdots,y_) with :p(y_k, y_0,\cdots,y_)=\int p(y_k, x_k) p(dx_k, y_0,\cdots,y_) and the conventions p(y_0, y_0,\cdots,y_)=p(y_0) and p(x_0, y_0,\cdots,y_)=p(x_0), for ''k'' = 0. Replacing p(x_k, y_0,\cdots,y_)dx_k by the
empirical Empirical evidence for a proposition is evidence, i.e. what supports or counters this proposition, that is constituted by or accessible to sense experience or experimental procedure. Empirical evidence is of central importance to the sciences and ...
approximation :\widehat(dx_k, y_0,\cdots,y_):=\frac\sum_^N \delta_(dx_k) \approx_ p(dx_k, y_0,\cdots,y_) in the above displayed formula, we design the following unbiased particle approximation of the likelihood function :p(y_0,\cdots,y_n) \approx_ \widehat(y_0,\cdots,y_n)=\prod_^n \widehat(y_k, y_0,\cdots,y_) with :\widehat(y_k, y_0,\cdots,y_)=\int p(y_k, x_k) \widehat(dx_k, y_0,\cdots,y_)=\frac\sum_^N p(y_k, \xi^i_k) where p(y_k, \xi^i_k) stands for the density p(y_k, x_k) evaluated at x_k=\xi^i_k. The design of this particle estimate and the unbiasedness property has been proved in 1996 in the article. Refined variance estimates can be found in and.


Backward particle smoothers

Using Bayes' rule, we have the formula :p(x_0,\cdots,x_n, y_0,\cdots,y_) = p(x_n , y_0,\cdots,y_) p(x_, x_n, y_0,\cdots,y_ ) \cdots p(x_1, x_2,y_0,y_1) p(x_0, x_1,y_0) Notice that : \begin p(x_, x_,(y_0,\cdots,y_)) &\propto p(x_, x_)p(x_, (y_0,\cdots,y_)) \\ p(x_, (y_0,\cdots,y_) &\propto p(y_, x_)p(x_, (y_0,\cdots,y_) \end This implies that :p(x_, x_k, (y_0,\cdots,y_))=\frac Replacing the one-step optimal predictors p(x_, (y_0,\cdots,y_))dx_ by the particle
empirical measure In probability theory, an empirical measure is a random measure arising from a particular realization of a (usually finite) sequence of random variables. The precise definition is found below. Empirical measures are relevant to mathematical stat ...
s :\widehat(dx_, (y_0,\cdots,y_))=\frac\sum_^N \delta_(dx_) \left(\approx_ p(dx_, (y_0,\cdots,y_)):=(x_, (y_0,\cdots,y_)) dx_\right) we find that :\begin p(dx_, x_,(y_0,\cdots,y_)) &\approx_ \widehat(dx_, x_,(y_0,\cdots,y_)) \\ &:= \frac\\ &= \sum_^ \frac \delta_(dx_) \end We conclude that :p(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) \approx_ \widehat_(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) with the backward particle approximation :\begin \widehat_ (d(x_0,\cdots,x_n), (y_0,\cdots,y_)) = \widehat(dx_n, (y_0,\cdots,y_)) \widehat(dx_, x_n,(y_0,\cdots,y_)) \cdots \widehat(dx_1, x_2,(y_0,y_1)) \widehat(dx_0, x_1,y_0) \end The probability measure :\widehat_(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) is the probability of the random paths of a Markov chain \left(\mathbb X^_\right)_running backward in time from time k=n to time k=0, and evolving at each time step k in the state space associated with the population of particles \xi^i_k, i=1,\cdots,N. * Initially (at time k=n) the chain \mathbb X^_ chooses randomly a state with the distribution ::\widehat(dx_, (y_0,\cdots,y_))=\frac\sum_^N \delta_(dx_) * From time k to the time (k-1), the chain starting at some state \mathbb X^_=\xi^i_k for some i=1,\cdots,N at time k moves at time (k-1) to a random state \mathbb^_ chosen with the discrete weighted probability :\widehat(dx_, \xi^i_,(y_0,\cdots,y_))= \sum_^N\frac~\delta_(dx_) In the above displayed formula, \widehat(dx_, \xi^i_,(y_0,\cdots,y_)) stands for the conditional distribution \widehat(dx_, x_k, (y_0,\cdots,y_)) evaluated at x_k=\xi^i_. In the same vein, p(y_, \xi^j_) and p(\xi^i_k, \xi^j_) stand for the conditional densities p(y_, x_) and p(x_k, x_) evaluated at x_k=\xi^i_ and x_=\xi^j_. These models allows to reduce integration with respect to the densities p((x_0,\cdots,x_n), (y_0,\cdots,y_)) in terms of matrix operations with respect to the Markov transitions of the chain described above. For instance, for any function f_k we have the particle estimates :\begin \int p(d(x_0,\cdots,x_n)&, (y_0,\cdots,y_))f_k(x_k) \\ &\approx_ \int \widehat_(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) f_k(x_k) \\ &=\int \widehat(dx_n, (y_0,\cdots,y_)) \widehat(dx_, x_n,(y_0,\cdots,y_)) \cdots \widehat(dx_k, x_,(y_0,\cdots,y_k)) f_k(x_k) \\ &=\underbrace_\mathbb_ \cdots\mathbb M_ \begin f_k(\xi^1_k)\\ \vdots\\ f_k(\xi^N_k) \end \end where :\mathbb M_k= (\mathbb M_k(i,j))_: \qquad \mathbb M_k(i,j)=\frac This also shows that if :\overline(x_0,\cdots,x_n):=\frac\sum_^n f_k(x_k) then :\begin \int \overline(x_0,\cdots,x_n) p(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) &\approx_ \int \overline(x_0,\cdots,x_n) \widehat_(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) \\ &=\frac \sum_^n \underbrace_\mathbb M_\mathbb M_\cdots\mathbb_k \begin f_k(\xi^1_k)\\ \vdots\\ f_k(\xi^N_k) \end \end


Some convergence results

We shall assume that filtering equation is stable, in the sense that it corrects any erroneous initial condition. In this situation, the particle approximations of the likelihood functions are unbiased and the relative variance is controlled by :E\left(\widehat(y_0,\cdots,y_n)\right)= p(y_0,\cdots,y_n), \qquad E\left(\left frac-1\right2\right)\leqslant \frac, for some finite constant ''c''. In addition, for any x\geqslant 0: :\mathbf \left ( \left\vert \frac\log-\frac\log\right\vert \leqslant c_1 \frac+c_2 \sqrt \right ) > 1-e^ for some finite constants c_1, c_2 related to the asymptotic bias and variance of the particle estimate, and for some finite constant ''c''. The bias and the variance of the particle particle estimates based on the ancestral lines of the genealogical trees :\begin I^_k(F) &:=\int F(x_0,\cdots,x_k) p(d(x_0,\cdots,x_k), y_0,\cdots,y_) \\ &\approx_ \widehat^_k(F) \\ &:=\int F(x_0,\cdots,x_k) \widehat(d(x_0,\cdots,x_k), y_0,\cdots,y_) \\ &=\frac\sum_^N F\left(\xi^i_,\cdots,\xi^i_\right) \end are controlled by the non asymptotic uniform estimates :\left, E\left(\widehat^_k(F)\right)-I_k^(F)\\leqslant \frac, \qquad E\left(\left widehat^_k(F)-I_k^(F)\right2\right)\leqslant \frac, for any function ''F'' bounded by 1, and for some finite constants c_1, c_2. In addition, for any x\geqslant 0: :\mathbf \left ( \left, \widehat^_k(F)-I_k^(F)\right , \leqslant c_1 \frac+c_2 \sqrt \land \sup_\left, \widehat_k^(F)-I^_k(F)\ \leqslant c \sqrt \right ) > 1-e^ for some finite constants c_1, c_2 related to the asymptotic bias and variance of the particle estimate, and for some finite constant ''c''. The same type of bias and variance estimates hold for the backward particle smoothers. For additive functionals of the form :\overline(x_0,\cdots,x_n):=\frac\sum_f_k(x_k) with :I^_n(\overline) \approx_ I^_n(\overline):=\int \overline(x_0,\cdots,x_n) \widehat_(d(x_0,\cdots,x_n), (y_0,\cdots,y_)) with functions f_k bounded by 1, we have :\sup_ \leqslant \frac and :E\left(\left widehat^_n(F)-I_n^(F)\right2\right)\leqslant \frac+ \frac for some finite constants c_1,c_2,c_3. More refined estimates including exponentially small probability of errors are developed in.


Sequential Importance Resampling (SIR)


Monte Carlo filter and bootstrap filter

''Sequential importance Resampling (SIR)'', Monte Carlo filtering (Kitagawa 1993) and the bootstrap filtering algorithm (Gordon et al. 1993), are also commonly applied filtering algorithms, which approximate the filtering probability density p(x_k, y_0,\cdots,y_k) by a weighted set of ''N'' samples : \left \. The ''importance weights'' w^_k are approximations to the relative posterior probabilities (or densities) of the samples such that :\sum_^N w^_k = 1. Sequential importance sampling (SIS) is a sequential (i.e., recursive) version of
importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally att ...
. As in importance sampling, the expectation of a function ''f'' can be approximated as a weighted average : \int f(x_k) p(x_k, y_0,\dots,y_k) dx_k \approx \sum_^N w_k^ f(x_k^). For a finite set of samples, the algorithm performance is dependent on the choice of the ''proposal distribution'' : \pi(x_k, x_,y_)\, . The "''optimal" proposal distribution'' is given as the ''target distribution'' : \pi(x_k, x_,y_) = p(x_k, x_,y_)=\frac~p(x_k, x_). This particular choice of proposal transition has been proposed by P. Del Moral in 1996 and 1998. When it is difficult to sample transitions according to the distribution p(x_k, x_,y_) one natural strategy is to use the following particle approximation :\begin \frac p(x_k, x_)dx_k &\simeq_ \frac \widehat(dx_k, x_) \\ &= \sum_^N \frac \delta_(dx_k) \end with the empirical approximation : \widehat(dx_k, x_)= \frac\sum_^ \delta_(dx_k)~\simeq_ p(x_k, x_)dx_k associated with ''N'' (or any other large number of samples) independent random samples X^i_k(x_), i=1,\cdots,N with the conditional distribution of the random state X_k given X_=x_. The consistency of the resulting particle filter of this approximation and other extensions are developed in. In the above display \delta_a stands for the
Dirac measure In mathematics, a Dirac measure assigns a size to a set based solely on whether it contains a fixed element ''x'' or not. It is one way of formalizing the idea of the Dirac delta function, an important tool in physics and other technical fields. ...
at a given state a. However, the transition prior probability distribution is often used as importance function, since it is easier to draw particles (or samples) and perform subsequent importance weight calculations: : \pi(x_k, x_,y_) = p(x_k, x_). ''Sequential Importance Resampling'' (SIR) filters with transition prior probability distribution as importance function are commonly known as bootstrap filter and
condensation algorithm The condensation algorithm (Conditional Density Propagation) is a computer vision algorithm. The principal application is to detect and track the contour of objects moving in a cluttered environment. Object tracking is one of the more basic and diff ...
. ''Resampling'' is used to avoid the problem of degeneracy of the algorithm, that is, avoiding the situation that all but one of the importance weights are close to zero. The performance of the algorithm can be also affected by proper choice of resampling method. The ''
stratified sampling In statistics, stratified sampling is a method of sampling from a population which can be partitioned into subpopulations. In statistical surveys, when subpopulations within an overall population vary, it could be advantageous to sample each s ...
'' proposed by Kitagawa (1993) is optimal in terms of variance. A single step of sequential importance resampling is as follows: :1) For i=1,\cdots,N draw samples from the ''proposal distribution'' :: x^_k \sim \pi(x_k, x^_,y_) :2) For i=1,\cdots,N update the importance weights up to a normalizing constant: ::\hat^_k = w^_ \frac . : Note that when we use the transition prior probability distribution as the importance function, :: \pi(x_k^, x^_,y_) = p(x^_k, x^_), :this simplifies to the following : :: \hat^_k = w^_ p(y_k, x^_k), :3) For i=1,\cdots,N compute the normalized importance weights: :: w^_k = \frac :4) Compute an estimate of the effective number of particles as :: \hat_\mathit = \frac :This criterion reflects the variance of the weights. Other criteria can be found in the article, including their rigorous analysis and central limit theorems. :5) If the effective number of particles is less than a given threshold \hat_\mathit < N_, then perform resampling: ::a) Draw ''N'' particles from the current particle set with probabilities proportional to their weights. Replace the current particle set with this new one. ::b) For i=1,\cdots,N set w^_k = 1/N. The term "Sampling Importance Resampling" is also sometimes used when referring to SIR filters, but the term ''Importance Resampling'' is more accurate because the word "resampling" implies that the initial sampling has already been done.


Sequential importance sampling (SIS)

* Is the same as sequential importance resampling, but without the resampling stage.


"Direct version" algorithm

The "direct version" algorithm is rather simple (compared to other particle filtering algorithms) and it uses composition and rejection. To generate a single sample ''x'' at ''k'' from p_(x, y_): #Set ''n'' ''= 0'' (This will count the number of particles generated so far) #
Uniformly Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
choose an index ''i'' from the range \ #Generate a test \hat from the distribution p(x_k, x_) with x_=x_^ #Generate the probability of \hat using \hat from p(y_k, x_k),~\mbox~x_k=\hat where y_k is the measured value #Generate another
uniform A uniform is a variety of clothing worn by members of an organization while participating in that organization's activity. Modern uniforms are most often worn by armed forces and paramilitary organizations such as police, emergency services, se ...
u from , m_k/math> where m_k = \sup_ p(y_k, x_k) :1) Set ''n'' ''= 0'' (This will count the number of particles generated so far) :2)
Uniformly Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
choose an index i from the range \ :3) Generate a test \hat from the distribution p(x_k, x_) with x_=x_^ :4) Generate the probability of \hat using \hat from p(y_k, x_k),~\mbox~x_k=\hat where y_k is the measured value :5) Generate another
uniform A uniform is a variety of clothing worn by members of an organization while participating in that organization's activity. Modern uniforms are most often worn by armed forces and paramilitary organizations such as police, emergency services, se ...
u from , m_k/math> where m_k = \sup_ p(y_k, x_k) :6) Compare u and p\left(\hat\right) ::6a) If u is larger then repeat from step 2 ::6b) If u is smaller then save \hat as x_^ and increment n :7) If ''n

N'' then quit The goal is to generate P "particles" at ''k'' using only the particles from k-1. This requires that a Markov equation can be written (and computed) to generate a x_k based only upon x_. This algorithm uses the composition of the P particles from k-1 to generate a particle at ''k'' and repeats (steps 2–6) until P particles are generated at ''k''. This can be more easily visualized if ''x'' is viewed as a two-dimensional array. One dimension is ''k'' and the other dimension is the particle number. For example, x(k,i) would be the ith particle at k and can also be written x_k^ (as done above in the algorithm). Step 3 generates a ''potential'' x_k based on a randomly chosen particle (x_^) at time k-1 and rejects or accepts it in step 6. In other words, the x_k values are generated using the previously generated x_.


Applications

Particle filters and Feynman-Kac particle methodologies find application in several contexts, as an effective mean for tackling noisy observations or strong nonlinearities, such as: *
Bayesian inference Bayesian inference is a method of statistical inference in which Bayes' theorem is used to update the probability for a hypothesis as more evidence or information becomes available. Bayesian inference is an important technique in statistics, a ...
,
machine learning Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine ...
, risk analysis and rare event sampling *
Bioinformatics Bioinformatics () is an interdisciplinary field that develops methods and software tools for understanding biological data, in particular when the data sets are large and complex. As an interdisciplinary field of science, bioinformatics combi ...
*
Computational science Computational science, also known as scientific computing or scientific computation (SC), is a field in mathematics that uses advanced computing capabilities to understand and solve complex problems. It is an area of science that spans many disc ...
*
Economics Economics () is the social science that studies the Production (economics), production, distribution (economics), distribution, and Consumption (economics), consumption of goods and services. Economics focuses on the behaviour and intera ...
,
financial mathematics Mathematical finance, also known as quantitative finance and financial mathematics, is a field of applied mathematics, concerned with mathematical modeling of financial markets. In general, there exist two separate branches of finance that require ...
and
mathematical finance Mathematical finance, also known as quantitative finance and financial mathematics, is a field of applied mathematics, concerned with mathematical modeling of financial markets. In general, there exist two separate branches of finance that require ...
: particle filters can perform simulations which are needed to compute the high-dimensional and/or complex integrals related to problems such as dynamic stochastic general equilibrium models in macro-economics and option pricing *
Engineering Engineering is the use of scientific method, scientific principles to design and build machines, structures, and other items, including bridges, tunnels, roads, vehicles, and buildings. The discipline of engineering encompasses a broad rang ...
*
Fault detection and isolation Fault detection, isolation, and recovery (FDIR) is a subfield of control engineering which concerns itself with monitoring a system, identifying when a fault has occurred, and pinpointing the type of fault and its location. Two approaches can be ...
: in observer-based schemas a particle filter can forecast expected sensors output enabling fault isolation *
Molecular chemistry Chemistry is the scientific study of the properties and behavior of matter. It is a natural science that covers the elements that make up matter to the compounds made of atoms, molecules and ions: their composition, structure, properties, ...
and
computational physics Computational physics is the study and implementation of numerical analysis to solve problems in physics for which a quantitative theory already exists. Historically, computational physics was the first application of modern computers in science, ...
*
Pharmacokinetics Pharmacokinetics (from Ancient Greek ''pharmakon'' "drug" and ''kinetikos'' "moving, putting in motion"; see chemical kinetics), sometimes abbreviated as PK, is a branch of pharmacology dedicated to determining the fate of substances administered ...
*
Phylogenetics In biology, phylogenetics (; from Greek language, Greek wikt:φυλή, φυλή/wikt:φῦλον, φῦλον [] "tribe, clan, race", and wikt:γενετικός, γενετικός [] "origin, source, birth") is the study of the evolutionary his ...
*Robotics,
artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech re ...
:
Monte Carlo localization Monte Carlo localization (MCL), also known as particle filter localization,Ioannis M. Rekleitis.A Particle Filter Tutorial for Mobile Robot Localization" ''Centre for Intelligent Machines, McGill University, Tech. Rep. TR-CIM-04-02'' (2004). is an a ...
is a de facto standard in mobile robot localization Dieter Fox, Wolfram Burgard, Frank Dellaert, and Sebastian Thrun,
Monte Carlo Localization: Efficient Position Estimation for Mobile Robots
" ''Proc. of the Sixteenth National Conference on Artificial Intelligence'' John Wiley & Sons Ltd, 1999.
Sebastian Thrun, Wolfram Burgard, Dieter Fox
''Probabilistic Robotics''
MIT Press, 2005. Ch. 8.3 .
Sebastian Thrun, Dieter Fox, Wolfram Burgard, Frank Dellaert.

" ''Artificial Intelligence'' 128.1 (2001): 99–141.
* Signal and image processing: visual localization, tracking,
feature Feature may refer to: Computing * Feature (CAD), could be a hole, pocket, or notch * Feature (computer vision), could be an edge, corner or blob * Feature (software design) is an intentional distinguishing characteristic of a software item ...
recognition


Other particle filters

*
Auxiliary particle filter The auxiliary particle filter is a particle filtering algorithm introduced by Pitt and Shephard in 1999 to improve some deficiencies of the particle filter#Sequential importance sampling (SIS), sequential importance resampling (SIR) algorithm when ...
* Cost Reference particle filter * Exponential Natural Particle Filter * Feynman-Kac and mean-field particle methodologies * Gaussian particle filter * Gauss–Hermite particle filter * Hierarchical/Scalable particle filter * Nudged particle filter * Particle Markov-Chain Monte-Carlo, see e.g. pseudo-marginal Metropolis–Hastings algorithm. * Rao–Blackwellized particle filter * Regularized auxiliary particle filter * Rejection-sampling based optimal particle filter * Unscented particle filter


See also

*
Ensemble Kalman filter The ensemble Kalman filter (EnKF) is a recursive filter suitable for problems with a large number of variables, such as discretizations of partial differential equations in geophysical models. The EnKF originated as a version of the Kalman filter fo ...
*
Generalized filtering Generalized filtering is a generic Bayesian filtering scheme for nonlinear state-space models. It is based on a variational principle of least action, formulated in generalized coordinates of motion. Note that "generalized coordinates of motion" a ...
*
Genetic algorithm In computer science and operations research, a genetic algorithm (GA) is a metaheuristic inspired by the process of natural selection that belongs to the larger class of evolutionary algorithms (EA). Genetic algorithms are commonly used to gene ...
*
Mean-field particle methods Mean-field particle methods are a broad class of ''interacting type'' Monte Carlo algorithms for simulating from a sequence of probability distributions satisfying a nonlinear evolution equation. These flows of probability measures can always be int ...
*
Monte Carlo localization Monte Carlo localization (MCL), also known as particle filter localization,Ioannis M. Rekleitis.A Particle Filter Tutorial for Mobile Robot Localization" ''Centre for Intelligent Machines, McGill University, Tech. Rep. TR-CIM-04-02'' (2004). is an a ...
*
Moving horizon estimation Moving horizon estimation (MHE) is an optimization approach that uses a series of measurements observed over time, containing noise (random variations) and other inaccuracies, and produces estimates of unknown variables or parameters. Unlike determi ...
*
Recursive Bayesian estimation In probability theory, statistics, and machine learning, recursive Bayesian estimation, also known as a Bayes filter, is a general probabilistic approach for estimating an unknown probability density function ( PDF) recursively over time using inc ...


References


Bibliography

* * Del Moral, Pierre (2004).
Feynman-Kac formulae. Genealogical and interacting particle approximations
'. Springer. p. 575. "Series: Probability and Applications" * Del Moral, Pierre (2013).
Mean field simulation for Monte Carlo integration.
Chapman & Hall/CRC Press. p. 626. "Monographs on Statistics & Applied Probability" * * * * * * * * * * * * * * * * *


External links



Theoretical aspects and a list of application domains of particle filters
Sequential Monte Carlo Methods (Particle Filtering)
homepage on University of Cambridge
Dieter Fox's MCL Animations

Rob Hess' free software

SMCTC: A Template Class for Implementing SMC algorithms in C++

Java applet on particle filtering

vSMC : Vectorized Sequential Monte Carlo

Particle filter explained in the context of self driving car
{{DEFAULTSORT:Particle Filter Monte Carlo methods Computational statistics * Nonlinear filters Robot control Statistical mechanics Sampling techniques Stochastic simulation