Monte Carlo Method In Statistical Physics
   HOME

TheInfoList



OR:

Monte Carlo in statistical physics refers to the application of the
Monte Carlo method Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that rely on repeated random sampling to obtain numerical results. The underlying concept is to use randomness to solve problems that might be determi ...
to problems in
statistical physics Statistical physics is a branch of physics that evolved from a foundation of statistical mechanics, which uses methods of probability theory and statistics, and particularly the Mathematics, mathematical tools for dealing with large populations ...
, or
statistical mechanics In physics, statistical mechanics is a mathematical framework that applies statistical methods and probability theory to large assemblies of microscopic entities. It does not assume or postulate any natural laws, but explains the macroscopic be ...
.


Overview

The general motivation to use the Monte Carlo method in statistical physics is to evaluate a multivariable integral. The typical problem begins with a system for which the Hamiltonian is known, it is at a given temperature and it follows the
Boltzmann statistics Ludwig Eduard Boltzmann (; 20 February 1844 – 5 September 1906) was an Austrian physicist and philosopher. His greatest achievements were the development of statistical mechanics, and the statistical explanation of the second law of thermod ...
. To obtain the mean value of some macroscopic variable, say A, the general approach is to compute, over all the
phase space In dynamical system theory, a phase space is a space in which all possible states of a system are represented, with each possible state corresponding to one unique point in the phase space. For mechanical systems, the phase space usually ...
, PS for simplicity, the mean value of A using the Boltzmann distribution: :\langle A\rangle=\int_ A_ \frac d\vec. where E(\vec)=E_ is the energy of the system for a given state defined by \vec - a vector with all the degrees of freedom (for instance, for a mechanical system, \vec = \left(\vec, \vec \right) ), \beta\equiv 1/k_bT and :Z= \int_ e^d\vec is the partition function. One possible approach to solve this multivariable integral is to exactly enumerate all possible configurations of the system, and calculate averages at will. This is done in exactly solvable systems, and in simulations of simple systems with few particles. In realistic systems, on the other hand, an exact enumeration can be difficult or impossible to implement. For those systems, the
Monte Carlo integration In mathematics, Monte Carlo integration is a technique for numerical integration using random numbers. It is a particular Monte Carlo method that numerically computes a definite integral. While other algorithms usually evaluate the integrand at ...
(and not to be confused with
Monte Carlo method Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that rely on repeated random sampling to obtain numerical results. The underlying concept is to use randomness to solve problems that might be determi ...
, which is used to simulate molecular chains) is generally employed. The main motivation for its use is the fact that, with the Monte Carlo integration, the error goes as 1/\sqrt, independently of the dimension of the integral. Another important concept related to the Monte Carlo integration is the
importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally att ...
, a technique that improves the computational time of the simulation. In the following sections, the general implementation of the Monte Carlo integration for solving this kind of problems is discussed.


Importance sampling

An estimation, under Monte Carlo integration, of an integral defined as :\langle A\rangle = \int_ A_ e^d\vec/Z is :\langle A\rangle \simeq \frac\sum_^N A_ e^/Z where \vec_i are uniformly obtained from all the phase space (PS) and N is the number of sampling points (or function evaluations). From all the phase space, some zones of it are generally more important to the mean of the variable A than others. In particular, those that have the value of e^ sufficiently high when compared to the rest of the energy spectra are the most relevant for the integral. Using this fact, the natural question to ask is: is it possible to choose, with more frequency, the states that are known to be more relevant to the integral? The answer is yes, using the
importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally att ...
technique. Lets assume p(\vec) is a distribution that chooses the states that are known to be more relevant to the integral. The mean value of A can be rewritten as :\langle A\rangle = \int_ p^(\vec) \frace^/Zd\vec = \int_ p^(\vec) A^_ e^/Zd\vec , where A^_ are the sampled values taking into account the importance probability p(\vec). This integral can be estimated by :\langle A\rangle \simeq \frac\sum_^N p^(\vec_i) A^_ e^/Z where \vec_i are now randomly generated using the p(\vec) distribution. Since most of the times it is not easy to find a way of generating states with a given distribution, the
Metropolis algorithm A metropolis () is a large city or conurbation which is a significant economic, political, and cultural center for a country or region, and an important hub for regional or international connections, commerce, and communications. A big c ...
must be used.


Canonical

Because it is known that the most likely states are those that maximize the Boltzmann distribution, a good distribution, p(\vec), to choose for the importance sampling is the Boltzmann distribution or canonic distribution. Let :p(\vec) = \frac be the distribution to use. Substituting on the previous sum, :\langle A\rangle \simeq \frac\sum_^N A^_. So, the procedure to obtain a mean value of a given variable, using metropolis algorithm, with the canonical distribution, is to use the Metropolis algorithm to generate states given by the distribution p(\vec) and perform means over A^_. One important issue must be considered when using the metropolis algorithm with the canonical distribution: when performing a given measure, i.e. realization of \vec_i, one must ensure that that realization is not correlated with the previous state of the system (otherwise the states are not being "randomly" generated). On systems with relevant energy gaps, this is the major drawback of the use of the canonical distribution because the time needed to the system de-correlate from the previous state can tend to infinity.


Multi-canonical

As stated before, micro-canonical approach has a major drawback, which becomes relevant in most of the systems that use Monte Carlo Integration. For those systems with "rough energy landscapes", the multicanonic approach can be used. The multicanonic approach uses a different choice for importance sampling: :p(\vec) = \frac where \Omega(E) is the
density of states In solid state physics and condensed matter physics, the density of states (DOS) of a system describes the number of modes per unit frequency range. The density of states is defined as D(E) = N(E)/V , where N(E)\delta E is the number of states i ...
of the system. The major advantage of this choice is that the energy histogram is flat, i.e. the generated states are equally distributed on energy. This means that, when using the Metropolis algorithm, the simulation doesn't see the "rough energy landscape", because every energy is treated equally. The major drawback of this choice is the fact that, on most systems, \Omega(E) is unknown. To overcome this, the
Wang and Landau algorithm The Wang and Landau algorithm, proposed by Fugao Wang and David P. Landau, is a Monte Carlo method designed to estimate the density of states of a system. The method performs a non-Markovian random walk to build the density of states by quickly vi ...
is normally used to obtain the DOS during the simulation. Note that after the DOS is known, the mean values of every variable can be calculated for every temperature, since the generation of states does not depend on \beta.


Implementation

On this section, the implementation will focus on the
Ising model The Ising model () (or Lenz-Ising model or Ising-Lenz model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent ...
. Lets consider a two-dimensional spin network, with L spins (lattice sites) on each side. There are naturally N = L^2 spins, and so, the phase space is discrete and is characterized by N spins, \vec = (\sigma_1,\sigma_2,...,\sigma_N) where \sigma_i\in \ is the spin of each lattice site. The system's energy is given by E(\vec) = \sum_^N\sum_ (1 - J_\sigma_i \sigma_j), where viz_i are the set of first neighborhood spins of i and J is the interaction matrix (for a ferromagnetic ising model, J is the identity matrix). The problem is stated. On this example, the objective is to obtain \langle M \rangle and \langle M^2 \rangle (for instance, to obtain the
magnetic susceptibility In electromagnetism, the magnetic susceptibility (Latin: , "receptive"; denoted ) is a measure of how much a material will become magnetized in an applied magnetic field. It is the ratio of magnetization (magnetic moment per unit volume) to the ap ...
of the system) since it is straightforward to generalize to other observables. According to the definition, M(\vec) = \sum_^N \sigma_i.


Canonical

First, the system must be initialized: let \beta=1/k_b T be the system's Boltzmann temperature and initialize the system with an initial state (which can be anything since the final result should not depend on it). With micro-canonic choice, the metropolis method must be employed. Because there is no right way of choosing which state is to be picked, one can particularize and choose to try to flip one spin at the time. This choice is usually called ''single spin flip''. The following steps are to be made to perform a single measurement. step 1: generate a state that follows the p(\vec) distribution: step 1.1: Perform TT times the following iteration: step 1.1.1: pick a lattice site at random (with probability 1/N), which will be called i, with spin \sigma_i. step 1.1.2: pick a random number \alpha \in ,1/math>. step 1.1.3: calculate the energy change of trying to flip the spin i: :\Delta E = 2\sigma_i \sum_\sigma_j and its magnetization change: \Delta M = -2\sigma_i step 1.1.4: if \alpha < \min(1, e^ ), flip the spin (\sigma_i = -\sigma_i ), otherwise, don't. step 1.1.5: update the several macroscopic variables in case the spin flipped: E = E + \Delta E, M = M + \Delta M after TT times, the system is considered to be not correlated from its previous state, which means that, at this moment, the probability of the system to be on a given state follows the Boltzmann distribution, which is the objective proposed by this method. step 2: perform the measurement: step 2.1: save, on a histogram, the values of M and M2. As a final note, one should note that TT is not easy to estimate because it is not easy to say when the system is de-correlated from the previous state. To surpass this point, one generally do not use a fixed TT, but TT as a ''tunneling time''. One tunneling time is defined as the number of steps 1. the system needs to make to go from the minimum of its energy to the maximum of its energy and return. A major drawback of this method with the ''single spin flip'' choice in systems like Ising model is that the tunneling time scales as a power law as N^ where z is greater than 0.5, phenomenon known as ''critical slowing down''.


Applicability

The method thus neglects dynamics, which can be a major drawback, or a great advantage. Indeed, the method can only be applied to static quantities, but the freedom to choose moves makes the method very flexible. An additional advantage is that some systems, such as the
Ising model The Ising model () (or Lenz-Ising model or Ising-Lenz model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent ...
, lack a dynamical description and are only defined by an energy prescription; for these the Monte Carlo approach is the only one feasible.


Generalizations

The great success of this method in statistical mechanics has led to various generalizations such as the method of
simulated annealing Simulated annealing (SA) is a probabilistic technique for approximating the global optimum of a given function. Specifically, it is a metaheuristic to approximate global optimization in a large search space for an optimization problem. It ...
for optimization, in which a fictitious temperature is introduced and then gradually lowered.


See also

*
Monte Carlo integration In mathematics, Monte Carlo integration is a technique for numerical integration using random numbers. It is a particular Monte Carlo method that numerically computes a definite integral. While other algorithms usually evaluate the integrand at ...
*
Metropolis algorithm A metropolis () is a large city or conurbation which is a significant economic, political, and cultural center for a country or region, and an important hub for regional or international connections, commerce, and communications. A big c ...
*
Importance sampling Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally att ...
*
Quantum Monte Carlo Quantum Monte Carlo encompasses a large family of computational methods whose common aim is the study of complex quantum systems. One of the major goals of these approaches is to provide a reliable solution (or an accurate approximation) of the ...
*
Monte Carlo molecular modeling Monte Carlo molecular modelling is the application of Monte Carlo methods to molecular problems. These problems can also be modelled by the molecular dynamics method. The difference is that this approach relies on equilibrium statistical mechanics ...


References

* * * * {{cite book , first1=Jerome , last1=Spanier , first2=Ely M. , last2=Gelbard , title=Monte Carlo Principles and Neutron Transport Problems , location= , publisher=Dover , year=2008 , isbn=978-0-486-46293-6 , chapter=Importance Sampling , pages=110–124 Computational chemistry Theoretical chemistry Computational physics