Particle filters, also known as sequential Monte Carlo methods, are a set of
Monte Carlo
Monte Carlo ( ; ; or colloquially ; , ; ) is an official administrative area of Monaco, specifically the Ward (country subdivision), ward of Monte Carlo/Spélugues, where the Monte Carlo Casino is located. Informally, the name also refers to ...
algorithms used to find approximate solutions for
filtering problems for nonlinear state-space systems, such as
signal processing
Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, Scalar potential, potential fields, Seismic tomograph ...
and
Bayesian statistical inference.
The
filtering problem consists of estimating the internal states in
dynamical systems
In mathematics, a dynamical system is a system in which a Function (mathematics), function describes the time dependence of a Point (geometry), point in an ambient space, such as in a parametric curve. Examples include the mathematical models ...
when partial observations are made and random perturbations are present in the sensors as well as in the dynamical system. The objective is to compute the
posterior distributions of the states of a
Markov process
In probability theory and statistics, a Markov chain or Markov process is a stochastic process describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, ...
, given the noisy and partial observations. The term "particle filters" was first coined in 1996 by Pierre Del Moral about
mean-field interacting particle methods used in
fluid mechanics
Fluid mechanics is the branch of physics concerned with the mechanics of fluids (liquids, gases, and plasma (physics), plasmas) and the forces on them.
Originally applied to water (hydromechanics), it found applications in a wide range of discipl ...
since the beginning of the 1960s.
The term "Sequential Monte Carlo" was coined by
Jun S. Liu and Rong Chen in 1998.
Particle filtering uses a set of particles (also called samples) to represent the
posterior distribution
The posterior probability is a type of conditional probability that results from updating the prior probability with information summarized by the likelihood via an application of Bayes' rule. From an epistemological perspective, the posterior ...
of a
stochastic process
In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
given the noisy and/or partial observations. The state-space model can be nonlinear and the initial state and noise distributions can take any form required. Particle filter techniques provide a well-established methodology
for generating samples from the required distribution without requiring assumptions about the state-space model or the state distributions. However, these methods do not perform well when applied to very high-dimensional systems.
Particle filters update their prediction in an approximate (statistical) manner. The samples from the distribution are represented by a set of particles; each particle has a likelihood weight assigned to it that represents the
probability
Probability is a branch of mathematics and statistics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a number between 0 and 1; the larger the probability, the more likely an e ...
of that particle being sampled from the
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
. Weight disparity leading to weight collapse is a common issue encountered in these filtering algorithms. However, it can be mitigated by including a resampling step before the weights become uneven. Several adaptive resampling criteria can be used including the
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
of the weights and the relative
entropy
Entropy is a scientific concept, most commonly associated with states of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the micros ...
concerning the uniform distribution.
In the resampling step, the particles with negligible weights are replaced by new particles in the proximity of the particles with higher weights.
From the statistical and probabilistic point of view, particle filters may be interpreted as
mean-field particle interpretations of
Feynman-Kac probability measures.
These particle integration techniques were developed in
molecular chemistry and
computational physics
Computational physics is the study and implementation of numerical analysis to solve problems in physics. Historically, computational physics was the first application of modern computers in science, and is now a subset of computational science ...
by
Theodore E. Harris and
Herman Kahn
Herman Kahn (February 15, 1922 – July 7, 1983) was an American physicist and a founding member of the Hudson Institute, regarded as one of the preeminent futurists of the latter part of the twentieth century. He originally came to prominence ...
in 1951,
Marshall N. Rosenbluth and
Arianna W. Rosenbluth in 1955,
and more recently by Jack H. Hetherington in 1984.
In computational physics, these Feynman-Kac type path particle integration methods are also used in
Quantum Monte Carlo
Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
, and more specifically
Diffusion Monte Carlo methods.
Feynman-Kac interacting particle methods are also strongly related to
mutation-selection genetic algorithms currently used in
evolutionary computation
Evolutionary computation from computer science is a family of algorithms for global optimization inspired by biological evolution, and the subfield of artificial intelligence and soft computing studying these algorithms. In technical terms ...
to solve complex optimization problems.
The particle filter methodology is used to solve
Hidden Markov Model
A hidden Markov model (HMM) is a Markov model in which the observations are dependent on a latent (or ''hidden'') Markov process (referred to as X). An HMM requires that there be an observable process Y whose outcomes depend on the outcomes of X ...
(HMM) and
nonlinear filtering problems. With the notable exception of linear-Gaussian signal-observation models (
Kalman filter
In statistics and control theory, Kalman filtering (also known as linear quadratic estimation) is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, to produce estimates of unk ...
) or wider classes of models (Benes filter), Mireille Chaleyat-Maurel and Dominique Michel proved in 1984 that the sequence of posterior distributions of the random states of a signal, given the observations (a.k.a. optimal filter), has no finite recursion. Various other numerical methods based on fixed grid approximations,
Markov Chain Monte Carlo
In statistics, Markov chain Monte Carlo (MCMC) is a class of algorithms used to draw samples from a probability distribution. Given a probability distribution, one can construct a Markov chain whose elements' distribution approximates it – that ...
techniques, conventional linearization,
extended Kalman filter
In estimation theory, the extended Kalman filter (EKF) is the nonlinear version of the Kalman filter which linearizes about an estimate of the current mean and covariance. In the case of well defined transition models, the EKF has been considered t ...
s, or determining the best linear system (in the expected cost-error sense) are unable to cope with large-scale systems, unstable processes, or insufficiently smooth nonlinearities.
Particle filters and Feynman-Kac particle methodologies find application in
signal and image processing,
Bayesian inference
Bayesian inference ( or ) is a method of statistical inference in which Bayes' theorem is used to calculate a probability of a hypothesis, given prior evidence, and update it as more information becomes available. Fundamentally, Bayesian infer ...
,
machine learning
Machine learning (ML) is a field of study in artificial intelligence concerned with the development and study of Computational statistics, statistical algorithms that can learn from data and generalise to unseen data, and thus perform Task ( ...
,
risk analysis and rare event sampling,
engineering
Engineering is the practice of using natural science, mathematics, and the engineering design process to Problem solving#Engineering, solve problems within technology, increase efficiency and productivity, and improve Systems engineering, s ...
and robotics,
artificial intelligence
Artificial intelligence (AI) is the capability of computer, computational systems to perform tasks typically associated with human intelligence, such as learning, reasoning, problem-solving, perception, and decision-making. It is a field of re ...
,
bioinformatics
Bioinformatics () is an interdisciplinary field of science that develops methods and Bioinformatics software, software tools for understanding biological data, especially when the data sets are large and complex. Bioinformatics uses biology, ...
,
phylogenetics
In biology, phylogenetics () is the study of the evolutionary history of life using observable characteristics of organisms (or genes), which is known as phylogenetic inference. It infers the relationship among organisms based on empirical dat ...
,
computational science
Computational science, also known as scientific computing, technical computing or scientific computation (SC), is a division of science, and more specifically the Computer Sciences, which uses advanced computing capabilities to understand and s ...
,
economics
Economics () is a behavioral science that studies the Production (economics), production, distribution (economics), distribution, and Consumption (economics), consumption of goods and services.
Economics focuses on the behaviour and interac ...
and mathematical finance
Mathematical finance, also known as quantitative finance and financial mathematics, is a field of applied mathematics, concerned with mathematical modeling in the financial field.
In general, there exist two separate branches of finance that req ...
,
molecular chemistry,
computational physics
Computational physics is the study and implementation of numerical analysis to solve problems in physics. Historically, computational physics was the first application of modern computers in science, and is now a subset of computational science ...
,
pharmacokinetics
Pharmacokinetics (from Ancient Greek ''pharmakon'' "drug" and ''kinetikos'' "moving, putting in motion"; see chemical kinetics), sometimes abbreviated as PK, is a branch of pharmacology dedicated to describing how the body affects a specific su ...
, quantitative risk and insurance and other fields.
History
Heuristic-like algorithms
From a statistical and probabilistic viewpoint, particle filters belong to the class of
branching/
genetic type algorithms, and
mean-field type interacting particle methodologies. The interpretation of these particle methods depends on the scientific discipline. In
Evolutionary Computing,
mean-field genetic type particle methodologies are often used as heuristic and natural search algorithms (a.k.a.
Metaheuristic
In computer science and mathematical optimization, a metaheuristic is a higher-level procedure or heuristic designed to find, generate, tune, or select a heuristic (partial search algorithm) that may provide a sufficiently good solution to an op ...
). In
computational physics
Computational physics is the study and implementation of numerical analysis to solve problems in physics. Historically, computational physics was the first application of modern computers in science, and is now a subset of computational science ...
and
molecular chemistry, they are used to solve Feynman-Kac
path integration problems or to compute Boltzmann-Gibbs measures, top eigenvalues, and
ground states of
Schrödinger operators. In
Biology
Biology is the scientific study of life and living organisms. It is a broad natural science that encompasses a wide range of fields and unifying principles that explain the structure, function, growth, History of life, origin, evolution, and ...
and
Genetics
Genetics is the study of genes, genetic variation, and heredity in organisms.Hartl D, Jones E (2005) It is an important branch in biology because heredity is vital to organisms' evolution. Gregor Mendel, a Moravian Augustinians, Augustinian ...
, they represent the evolution of a population of individuals or genes in some environment.
The origins of mean-field type
evolutionary computational techniques can be traced back to 1950 and 1954 with
Alan Turing's work on genetic type mutation-selection learning machines and the articles by
Nils Aall Barricelli at the
Institute for Advanced Study
The Institute for Advanced Study (IAS) is an independent center for theoretical research and intellectual inquiry located in Princeton, New Jersey. It has served as the academic home of internationally preeminent scholars, including Albert Ein ...
in
Princeton, New Jersey
The Municipality of Princeton is a Borough (New Jersey), borough in Mercer County, New Jersey, United States. It was established on January 1, 2013, through the consolidation of the Borough of Princeton, New Jersey, Borough of Princeton and Pri ...
. The first trace of particle filters in
statistical methodology dates back to the mid-1950s; the 'Poor Man's Monte Carlo', that was proposed by
John Hammersley
John Michael Hammersley, (21 March 1920 – 2 May 2004) was a British mathematician best known for his foundational work in the theory of self-avoiding walks and percolation theory.
Early life and education
Hammersley was born in Helensburgh i ...
et al., in 1954, contained hints of the genetic type particle filtering methods used today. In 1963,
Nils Aall Barricelli simulated a genetic type algorithm to mimic the ability of individuals to play a simple game. In
evolutionary computing literature, genetic-type mutation-selection algorithms became popular through the seminal work of
John Holland in the early 1970s, particularly his book published in 1975.
In Biology and
Genetics
Genetics is the study of genes, genetic variation, and heredity in organisms.Hartl D, Jones E (2005) It is an important branch in biology because heredity is vital to organisms' evolution. Gregor Mendel, a Moravian Augustinians, Augustinian ...
, the Australian geneticist
Alex Fraser also published in 1957 a series of papers on the genetic type simulation of
artificial selection
Selective breeding (also called artificial selection) is the process by which humans use animal breeding and plant breeding to selectively develop particular phenotypic traits (characteristics) by choosing which typically animal or plant ...
of organisms. The computer simulation of the evolution by biologists became more common in the early 1960s, and the methods were described in books by Fraser and Burnell (1970) and Crosby (1973). Fraser's simulations included all of the essential elements of modern mutation-selection genetic particle algorithms.
From the mathematical viewpoint, the
conditional distribution
Conditional (if then) may refer to:
* Causal conditional, if X then Y, where X is a cause of Y
*Conditional probability, the probability of an event A given that another event B
* Conditional proof, in logic: a proof that asserts a conditional, ...
of the random states of a signal given some partial and noisy observations is described by a Feynman-Kac probability on the random trajectories of the signal weighted by a sequence of likelihood potential functions.
Quantum Monte Carlo
Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
, and more specifically
Diffusion Monte Carlo methods can also be interpreted as a mean-field genetic type particle approximation of Feynman-Kac path integrals.
The origins of
Quantum Monte Carlo
Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
methods are often attributed to
Enrico Fermi
Enrico Fermi (; 29 September 1901 – 28 November 1954) was an Italian and naturalized American physicist, renowned for being the creator of the world's first artificial nuclear reactor, the Chicago Pile-1, and a member of the Manhattan Project ...
and
Robert Richtmyer who developed in 1948 a mean-field particle interpretation of
neutron chain reactions, but the first heuristic-like and genetic type particle algorithm (a.k.a. Resampled or Reconfiguration Monte Carlo methods) for estimating ground state energies of quantum systems (in reduced matrix models) is due to Jack H. Hetherington in 1984.
One can also quote the earlier seminal works of
Theodore E. Harris and
Herman Kahn
Herman Kahn (February 15, 1922 – July 7, 1983) was an American physicist and a founding member of the Hudson Institute, regarded as one of the preeminent futurists of the latter part of the twentieth century. He originally came to prominence ...
in particle physics, published in 1951, using mean-field but heuristic-like genetic methods for estimating particle transmission energies. In molecular chemistry, the use of genetic heuristic-like particle methodologies (a.k.a. pruning and enrichment strategies) can be traced back to 1955 with the seminal work of
Marshall N. Rosenbluth and
Arianna W. Rosenbluth.
The use of
genetic particle algorithms in advanced
signal processing
Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, Scalar potential, potential fields, Seismic tomograph ...
and
Bayesian inference
Bayesian inference ( or ) is a method of statistical inference in which Bayes' theorem is used to calculate a probability of a hypothesis, given prior evidence, and update it as more information becomes available. Fundamentally, Bayesian infer ...
is more recent. In January 1993, Genshiro Kitagawa developed a "Monte Carlo filter",
a slightly modified version of this article appeared in 1996. In April 1993, Neil J. Gordon et al., published in their seminal work
an application of genetic type algorithm in Bayesian statistical inference. The authors named their algorithm 'the bootstrap filter', and demonstrated that compared to other filtering methods, their bootstrap algorithm does not require any assumption about that state space or the noise of the system. Independently, the ones by Pierre Del Moral
and Himilcon Carvalho, Pierre Del Moral,
André Monin, and Gérard Salut on particle filters published in the mid-1990s. Particle filters were also developed in signal processing in early 1989-1992 by P. Del Moral, J.C. Noyer, G. Rigal, and G. Salut in the
LAAS-CNRS in a series of restricted and classified research reports with STCAN (Service Technique des Constructions et Armes Navales), the IT company DIGILOG, and th
LAAS-CNRS(the Laboratory for Analysis and Architecture of Systems) on RADAR/SONAR and GPS signal processing problems.
Mathematical foundations
From 1950 to 1996, all the publications on particle filters, and genetic algorithms, including the pruning and resample Monte Carlo methods introduced in computational physics and molecular chemistry, present natural and heuristic-like algorithms applied to different situations without a single proof of their consistency, nor a discussion on the bias of the estimates and genealogical and ancestral tree-based algorithms.
The mathematical foundations and the first rigorous analysis of these particle algorithms are due to Pierre Del Moral
in 1996. The article
also contains proof of the unbiased properties of a particle approximation of likelihood functions and unnormalized
conditional probability
In probability theory, conditional probability is a measure of the probability of an Event (probability theory), event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This ...
measures. The unbiased particle estimator of the likelihood functions presented in this article is used today in Bayesian statistical inference.
Dan Crisan, Jessica Gaines, and
Terry Lyons,
as well as Pierre Del Moral, and Terry Lyons,
created branching-type particle techniques with various population sizes around the end of the 1990s. P. Del Moral, A. Guionnet, and L. Miclo
made more advances in this subject in 2000. Pierre Del Moral and Alice Guionnet
proved the first central limit theorems in 1999, and Pierre Del Moral and Laurent Miclo
proved them in 2000. The first uniform convergence results concerning the time parameter for particle filters were developed at the end of the 1990s by Pierre Del Moral and
Alice Guionnet.
The first rigorous analysis of genealogical tree-ased particle filter smoothers is due to P. Del Moral and L. Miclo in 2001
The theory on Feynman-Kac particle methodologies and related particle filter algorithms was developed in 2000 and 2004 in the books.
These abstract probabilistic models encapsulate genetic type algorithms, particle, and bootstrap filters, interacting Kalman filters (a.k.a. Rao–Blackwellized particle filter
[
]),
importance sampling
Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally at ...
and resampling style particle filter techniques, including genealogical tree-based and particle backward methodologies for solving filtering and smoothing problems. Other classes of particle filtering methodologies include genealogical tree-based models,
backward Markov particle models,
adaptive mean-field particle models,
island-type particle models, particle Markov chain Monte Carlo methodologies, Sequential Monte Carlo samplers and Sequential Monte Carlo Approximate Bayesian Computation methods and Sequential Monte Carlo ABC based Bayesian Bootstrap.
The filtering problem
Objective
A particle filter's goal is to estimate the posterior density of state variables given observation variables. The particle filter is intended for use with a
hidden Markov Model
A hidden Markov model (HMM) is a Markov model in which the observations are dependent on a latent (or ''hidden'') Markov process (referred to as X). An HMM requires that there be an observable process Y whose outcomes depend on the outcomes of X ...
, in which the system includes both hidden and observable variables. The observable variables (observation process) are linked to the hidden variables (state-process) via a known functional form. Similarly, the probabilistic description of the dynamical system defining the evolution of the state variables is known.
A generic particle filter estimates the posterior distribution of the hidden states using the observation measurement process. With respect to a state-space such as the one below:
:
the filtering problem is to estimate sequentially the values of the hidden states
, given the values of the observation process
at any time step ''k''.
All Bayesian estimates of
follow from the
posterior density . The particle filter methodology provides an approximation of these conditional probabilities using the empirical measure associated with a genetic type particle algorithm. In contrast, the Markov Chain Monte Carlo or
importance sampling
Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally at ...
approach would model the full posterior
.
The Signal-Observation model
Particle methods often assume
and the observations
can be modeled in this form:
*
is a
Markov process
In probability theory and statistics, a Markov chain or Markov process is a stochastic process describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, ...
on
(for some
) that evolves according to the transition probability density
. This model is also often written in a synthetic way as
*:
:with an initial probability density
.
*The observations
take values in some state space on
(for some
) and are conditionally independent provided that
are known. In other words, each
only depends on
. In addition, we assume conditional distribution for
given
are absolutely continuous, and in a synthetic way we have
*:
An example of system with these properties is:
:
:
where both
and
are mutually independent sequences with known
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
s and ''g'' and ''h'' are known functions. These two equations can be viewed as
state space
In computer science, a state space is a discrete space representing the set of all possible configurations of a system. It is a useful abstraction for reasoning about the behavior of a given system and is widely used in the fields of artificial ...
equations and look similar to the state space equations for the Kalman filter. If the functions ''g'' and ''h'' in the above example are linear, and if both
and
are
Gaussian, the Kalman filter finds the exact Bayesian filtering distribution. If not, Kalman filter-based methods are a first-order approximation (
EKF) or a second-order approximation (
UKF in general, but if the probability distribution is Gaussian a third-order approximation is possible).
The assumption that the initial distribution and the transitions of the Markov chain are continuous for the
Lebesgue measure
In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of higher dimensional Euclidean '-spaces. For lower dimensions or , it c ...
can be relaxed. To design a particle filter we simply need to assume that we can sample the transitions
of the Markov chain
and to compute the likelihood function
(see for instance the genetic selection mutation description of the particle filter given below). The continuous assumption on the Markov transitions of
is only used to derive in an informal (and rather abusive) way different formulae between posterior distributions using the Bayes' rule for conditional densities.
Approximate Bayesian computation models
In certain problems, the conditional distribution of observations, given the random states of the signal, may fail to have a density; the latter may be impossible or too complex to compute.
In this situation, an additional level of approximation is necessitated. One strategy is to replace the signal
by the Markov chain
and to introduce a virtual observation of the form
: