HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
mathematical physics Mathematical physics is the development of mathematics, mathematical methods for application to problems in physics. The ''Journal of Mathematical Physics'' defines the field as "the application of mathematics to problems in physics and the de ...
, a random matrix is a matrix-valued
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
—that is, a matrix in which some or all of its entries are sampled randomly from a
probability distribution In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
. Random matrix theory (RMT) is the study of properties of random matrices, often as they become large. RMT provides techniques like mean-field theory, diagrammatic methods, the cavity method, or the replica method to compute quantities like traces, spectral densities, or scalar products between eigenvectors. Many physical phenomena, such as the
spectrum A spectrum (: spectra or spectrums) is a set of related ideas, objects, or properties whose features overlap such that they blend to form a continuum. The word ''spectrum'' was first used scientifically in optics to describe the rainbow of co ...
of nuclei of heavy atoms, the
thermal conductivity The thermal conductivity of a material is a measure of its ability to heat conduction, conduct heat. It is commonly denoted by k, \lambda, or \kappa and is measured in W·m−1·K−1. Heat transfer occurs at a lower rate in materials of low ...
of a lattice, or the emergence of quantum chaos, can be modeled mathematically as problems concerning large, random matrices.


Applications


Physics

In
nuclear physics Nuclear physics is the field of physics that studies atomic nuclei and their constituents and interactions, in addition to the study of other forms of nuclear matter. Nuclear physics should not be confused with atomic physics, which studies th ...
, random matrices were introduced by
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
to model the nuclei of heavy atoms. Wigner postulated that the spacings between the lines in the spectrum of a heavy atom nucleus should resemble the spacings between the eigenvalues of a random matrix, and should depend only on the symmetry class of the underlying evolution. In
solid-state physics Solid-state physics is the study of rigid matter, or solids, through methods such as solid-state chemistry, quantum mechanics, crystallography, electromagnetism, and metallurgy. It is the largest branch of condensed matter physics. Solid-state phy ...
, random matrices model the behaviour of large disordered Hamiltonians in the mean-field approximation. In quantum chaos, the Bohigas–Giannoni–Schmit (BGS) conjecture asserts that the spectral statistics of quantum systems whose classical counterparts exhibit chaotic behaviour are described by random matrix theory. In
quantum optics Quantum optics is a branch of atomic, molecular, and optical physics and quantum chemistry that studies the behavior of photons (individual quanta of light). It includes the study of the particle-like properties of photons and their interaction ...
, transformations described by random unitary matrices are crucial for demonstrating the advantage of quantum over classical computation (see, e.g., the boson sampling model). Moreover, such random unitary transformations can be directly implemented in an optical circuit, by mapping their parameters to optical circuit components (that is
beam splitter A beam splitter or beamsplitter is an optical instrument, optical device that splits a beam of light into a transmitted and a reflected beam. It is a crucial part of many optical experimental and measurement systems, such as Interferometry, int ...
s and phase shifters). Random matrix theory has also found applications to the chiral Dirac operator in
quantum chromodynamics In theoretical physics, quantum chromodynamics (QCD) is the study of the strong interaction between quarks mediated by gluons. Quarks are fundamental particles that make up composite hadrons such as the proton, neutron and pion. QCD is a type of ...
,
quantum gravity Quantum gravity (QG) is a field of theoretical physics that seeks to describe gravity according to the principles of quantum mechanics. It deals with environments in which neither gravitational nor quantum effects can be ignored, such as in the v ...
in two dimensions, mesoscopic physics, spin-transfer torque, the fractional quantum Hall effect, Anderson localization, quantum dots, and
superconductors Superconductivity is a set of physical properties observed in superconductors: materials where electrical resistance vanishes and magnetic fields are expelled from the material. Unlike an ordinary metallic conductor, whose resistance decreases ...


Mathematical statistics and numerical analysis

In
multivariate statistics Multivariate statistics is a subdivision of statistics encompassing the simultaneous observation and analysis of more than one outcome variable, i.e., '' multivariate random variables''. Multivariate statistics concerns understanding the differ ...
, random matrices were introduced by John Wishart, who sought to estimate covariance matrices of large samples. Chernoff-, Bernstein-, and Hoeffding-type inequalities can typically be strengthened when applied to the maximal eigenvalue (i.e. the eigenvalue of largest magnitude) of a finite sum of random Hermitian matrices. Random matrix theory is used to study the spectral properties of random matrices—such as sample covariance matrices—which is of particular interest in high-dimensional statistics. Random matrix theory also saw applications in
neural networks A neural network is a group of interconnected units called neurons that send signals to one another. Neurons can be either Cell (biology), biological cells or signal pathways. While individual neurons are simple, many of them together in a netwo ...
and
deep learning Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression, and representation learning. The field takes inspiration from biological neuroscience a ...
, with recent work utilizing random matrices to show that hyper-parameter tunings can be cheaply transferred between large neural networks without the need for re-training. In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
, random matrices have been used since the work of
John von Neumann John von Neumann ( ; ; December 28, 1903 – February 8, 1957) was a Hungarian and American mathematician, physicist, computer scientist and engineer. Von Neumann had perhaps the widest coverage of any mathematician of his time, in ...
and Herman Goldstine to describe computation errors in operations such as
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
. Although random entries are traditional "generic" inputs to an algorithm, the concentration of measure associated with random matrix distributions implies that random matrices will not test large portions of an algorithm's input space.


Number theory

In
number theory Number theory is a branch of pure mathematics devoted primarily to the study of the integers and arithmetic functions. Number theorists study prime numbers as well as the properties of mathematical objects constructed from integers (for example ...
, the distribution of zeros of the Riemann zeta function (and other
L-function In mathematics, an ''L''-function is a meromorphic function on the complex plane, associated to one out of several categories of mathematical objects. An ''L''-series is a Dirichlet series, usually convergent on a half-plane, that may gi ...
s) is modeled by the distribution of eigenvalues of certain random matrices. The connection was first discovered by Hugh Montgomery and
Freeman Dyson Freeman John Dyson (15 December 1923 – 28 February 2020) was a British-American theoretical physics, theoretical physicist and mathematician known for his works in quantum field theory, astrophysics, random matrix, random matrices, math ...
. It is connected to the Hilbert–Pólya conjecture.


Free probability

The relation of free probability with random matrices is a key reason for the wide use of free probability in other subjects. Voiculescu introduced the concept of freeness around 1983 in an operator algebraic context; at the beginning there was no relation at all with random matrices. This connection was only revealed later in 1991 by Voiculescu; he was motivated by the fact that the limit distribution which he found in his free central limit theorem had appeared before in Wigner's semi-circle law in the random matrix context.


Computational neuroscience

In the field of computational neuroscience, random matrices are increasingly used to model the network of synaptic connections between neurons in the brain. Dynamical models of neuronal networks with random connectivity matrix were shown to exhibit a phase transition to chaos when the variance of the synaptic weights crosses a critical value, at the limit of infinite system size. Results on random matrices have also shown that the dynamics of random-matrix models are insensitive to mean connection strength. Instead, the stability of fluctuations depends on connection strength variation and time to synchrony depends on network topology. In the analysis of massive data such as fMRI, random matrix theory has been applied in order to perform dimension reduction. When applying an algorithm such as PCA, it is important to be able to select the number of significant components. The criteria for selecting components can be multiple (based on explained variance, Kaiser's method, eigenvalue, etc.). Random matrix theory in this content has its representative the Marchenko-Pastur distribution, which guarantees the theoretical high and low limits of the eigenvalues associated with a random variable covariance matrix. This matrix calculated in this way becomes the null hypothesis that allows one to find the eigenvalues (and their eigenvectors) that deviate from the theoretical random range. The components thus excluded become the reduced dimensional space (see examples in fMRI ).


Optimal control

In optimal control theory, the evolution of ''n'' state variables through time depends at any time on their own values and on the values of ''k'' control variables. With linear evolution, matrices of coefficients appear in the state equation (equation of evolution). In some problems the values of the parameters in these matrices are not known with certainty, in which case there are random matrices in the state equation and the problem is known as one of stochastic control. A key result in the case of linear-quadratic control with stochastic matrices is that the certainty equivalence principle does not apply: while in the absence of multiplier uncertainty (that is, with only additive uncertainty) the optimal policy with a quadratic loss function coincides with what would be decided if the uncertainty were ignored, the optimal policy may differ if the state equation contains random coefficients.


Computational mechanics

In computational mechanics, epistemic uncertainties underlying the lack of knowledge about the physics of the modeled system give rise to mathematical operators associated with the computational model, which are deficient in a certain sense. Such operators lack certain properties linked to unmodeled physics. When such operators are discretized to perform computational simulations, their accuracy is limited by the missing physics. To compensate for this deficiency of the mathematical operator, it is not enough to make the model parameters random, it is necessary to consider a mathematical operator that is random and can thus generate families of computational models in the hope that one of these captures the missing physics. Random matrices have been used in this sense, with applications in vibroacoustics, wave propagations, materials science, fluid mechanics, heat transfer, etc.


Engineering

Random matrix theory can be applied to the electrical and communications engineering research efforts to study, model and develop Massive Multiple-Input Multiple-Output ( MIMO) radio systems.


History

Random matrix theory first gained attention beyond mathematics literature in the context of nuclear physics. Experiments by Enrico Fermi and others demonstrated evidence that individual
nucleons In physics and chemistry, a nucleon is either a proton or a neutron, considered in its role as a component of an atomic nucleus. The number of nucleons in a nucleus defines the atom's mass number. Until the 1960s, nucleons were thought to be ele ...
cannot be approximated to move independently, leading
Niels Bohr Niels Henrik David Bohr (, ; ; 7 October 1885 – 18 November 1962) was a Danish theoretical physicist who made foundational contributions to understanding atomic structure and old quantum theory, quantum theory, for which he received the No ...
to formulate the idea of a compound nucleus. Because there was no knowledge of direct nucleon-nucleon interactions,
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
and Leonard Eisenbud approximated that the nuclear
Hamiltonian Hamiltonian may refer to: * Hamiltonian mechanics, a function that represents the total energy of a system * Hamiltonian (quantum mechanics), an operator corresponding to the total energy of that system ** Dyall Hamiltonian, a modified Hamiltonian ...
could be modeled as a random matrix. For larger atoms, the distribution of the energy eigenvalues of the Hamiltonian could be computed in order to approximate scattering cross sections by invoking the Wishart distribution.


Gaussian ensembles

The most-commonly studied random matrix distributions are the Gaussian ensembles: GOE, GUE and GSE. They are often denoted by their Dyson index, ''β'' = 1 for GOE, ''β'' = 2 for GUE, and ''β'' = 4 for GSE. This index counts the number of real components per matrix element.


Definitions

The Gaussian unitary ensemble \text(n) is described by the
Gaussian measure In mathematics, Gaussian measure is a Borel measure on finite-dimensional Euclidean space \mathbb^n, closely related to the normal distribution in statistics. There is also a generalization to infinite-dimensional spaces. Gaussian measures are na ...
with density \frac e^ on the space of n \times n Hermitian matrices H = (H_)_^n. Here Z_ = 2^ \left(\frac\right)^ is a normalization constant, chosen so that the integral of the density is equal to one. The term ''unitary'' refers to the fact that the distribution is invariant under unitary conjugation. The Gaussian unitary ensemble models Hamiltonians lacking time-reversal symmetry. The Gaussian orthogonal ensemble \text(n) is described by the Gaussian measure with density \frac e^ on the space of ''n'' × ''n'' real symmetric matrices ''H'' = (''H''''ij''). Its distribution is invariant under orthogonal conjugation, and it models Hamiltonians with time-reversal symmetry. Equivalently, it is generated by H = (G+G^T)/\sqrt, where G is an n\times n matrix with IID samples from the standard normal distribution. The Gaussian symplectic ensemble \text(n) is described by the Gaussian measure with density \frac e^ on the space of ''n'' × ''n'' Hermitian quaternionic matrices, e.g. symmetric square matrices composed of
quaternion In mathematics, the quaternion number system extends the complex numbers. Quaternions were first described by the Irish mathematician William Rowan Hamilton in 1843 and applied to mechanics in three-dimensional space. The algebra of quater ...
s, . Its distribution is invariant under conjugation by the symplectic group, and it models Hamiltonians with time-reversal symmetry but no rotational symmetry.


Point correlation functions

The ensembles as defined here have Gaussian distributed matrix elements with mean ⟨''H''''ij''⟩ = 0, and two-point correlations given by \langle H_ H_^* \rangle = \langle H_ H_ \rangle = \frac \delta_ \delta_ + \frac\delta_\delta_ , from which all higher correlations follow by Isserlis' theorem.


Moment generating functions

The moment generating function for the GOE isE ^= e^where \, \cdot \, _F is the
Frobenius norm In the field of mathematics, norms are defined for elements within a vector space. Specifically, when the vector space comprises matrices, such norms are referred to as matrix norms. Matrix norms differ from vector norms in that they must also ...
.


Spectral density

The joint probability density for the eigenvalues of GUE/GOE/GSE is given by where ''Z''''β'',''n'' is a normalization constant which can be explicitly computed, see Selberg integral. In the case of GUE (''β'' = 2), the formula (1) describes a determinantal point process. Eigenvalues repel as the joint probability density has a zero (of \betath order) for coinciding eigenvalues \lambda_j = \lambda_i, and Z_ = (2\pi)^ \prod_^n k!. More succinctly, \frac e^ , \Delta_n(\lambda), ^\betawhere \Delta_n is the Vandermonde determinant. The distribution of the largest eigenvalue for GOE, and GUE, are explicitly solvable. They converge to the Tracy–Widom distribution after shifting and scaling appropriately.


Convergence to Wigner semicircular distribution

The spectrum, divided by \sqrt, converges in distribution to the semicircular distribution on the interval 2, +2/math>: \rho(x) = \frac\sqrt. Here \sigma^2 is the variance of off-diagonal entries. The variance of the on-diagonal entries do not matter.


Distribution of level spacings

From the ordered sequence of eigenvalues \lambda_1 < \ldots < \lambda_n < \lambda_ < \ldots, one defines the normalized spacings s = (\lambda_ - \lambda_n)/\langle s \rangle, where \langle s \rangle =\langle \lambda_ - \lambda_n \rangle is the mean spacing. The probability distribution of spacings is approximately given by, p_1(s) = \fracs\, e^ for the orthogonal ensemble GOE \beta=1, p_2(s) = \fracs^2 \mathrm^ for the unitary ensemble GUE \beta=2, and p_4(s) = \fracs^4 e^ for the symplectic ensemble GSE \beta = 4. The numerical constants are such that p_\beta(s) is normalized: \int_0^\infty ds\,p_\beta(s) = 1 and the mean spacing is, \int_0^\infty ds\, s\, p_\beta(s) = 1, for \beta = 1,2,4 .


Generalizations

''Wigner matrices'' are random Hermitian matrices H_n = (H_n(i,j))_^n such that the entries \left\ above the main diagonal are independent random variables with zero mean and have identical second moments. The Gaussian ensembles can be extended for \beta \neq 1, 2, 4 using the Dumitriu-Edelman tridiagonal ensemble. ''Invariant matrix ensembles'' are random Hermitian matrices with density on the space of real symmetric/Hermitian/quaternionic Hermitian matrices, which is of the form \frac e^~, where the function is called the potential. The Gaussian ensembles are the only common special cases of these two classes of random matrices. This is a consequence of a theorem by Porter and Rosenzweig.


Spectral theory of random matrices

The spectral theory of random matrices studies the distribution of the eigenvalues as the size of the matrix goes to infinity.


Empirical spectral measure

The ''empirical spectral measure'' \mu_H of H is defined by \mu_(A) = \frac \, \# \left\ = N_, \quad A \subset \mathbb. or more succinctly, if \lambda_1,\ldots, \lambda_n are the eigenvalues of H \mu_(d\lambda) = \frac 1n \sum_i \delta_(d\lambda). Usually, the limit of \mu_ is a deterministic measure; this is a particular case of self-averaging. The
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
of the limiting measure is called the integrated density of states and is denoted ''N''(''λ''). If the integrated density of states is differentiable, its derivative is called the density of states and is denoted ''ρ''(''λ'').


Types of convergence

Given a matrix ensemble, we say that its spectral measures converge weakly to \rho iff for any measurable set A, the ensemble-average converges:\lim_ \mathbb E_H mu_H(A)= \rho(A)Convergence weakly almost surely: If we sample H_1, H_2, H_3, \dots independently from the ensemble, then with probability 1,\lim_ \mu_(A) = \rho(A)for any measurable set A. In another sense, weak almost sure convergence means that we sample H_1, H_2, H_3, \dots, not independently, but by "growing" (a
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
), then with probability 1, \lim_ \mu_(A) = \rho(A) for any measurable set A. For example, we can "grow" a sequence of matrices from the Gaussian ensemble as follows: * Sample an infinite doubly infinite sequence of standard random variables \_ . * Define each H_n = (G_n+G_n^T)/\sqrt where G_n is the matrix made of entries \_ . Note that generic matrix ensembles do not allow us to grow, but most of the common ones, such as the three Gaussian ensembles, do allow us to grow.


Global regime

In the ''global regime'', one is interested in the distribution of linear statistics of the form N_ = n^ \text f(H). The limit of the empirical spectral measure for Wigner matrices was described by
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
; see Wigner semicircle distribution and Wigner surmise. As far as sample covariance matrices are concerned, a theory was developed by Marčenko and Pastur.. The limit of the empirical spectral measure of invariant matrix ensembles is described by a certain integral equation which arises from
potential theory In mathematics and mathematical physics, potential theory is the study of harmonic functions. The term "potential theory" was coined in 19th-century physics when it was realized that the two fundamental forces of nature known at the time, namely g ...
.


Fluctuations

For the linear statistics , one is also interested in the fluctuations about ∫ ''f''(''λ'') ''dN''(''λ''). For many classes of random matrices, a central limit theorem of the form \frac \overset N(0, 1) is known.


The variational problem for the unitary ensembles

Consider the measure :\mathrm\mu_N(\mu)=\frace^\mathrm\lambda,\qquad H_N(\lambda)=-\sum\limits_\ln, \lambda_j-\lambda_k, +N\sum\limits_^N Q(\lambda_j), where Q(M) is the potential of the ensemble and let \nu be the empirical spectral measure. We can rewrite H_N(\lambda) with \nu as :H_N(\lambda)=N^2\left x-y, \mathrm\nu(x)\mathrm\nu(y)+\int Q(x)\mathrm\nu(x)\right the probability measure is now of the form :\mathrm\mu_N(\mu)=\frace^\mathrm\lambda, where I_Q(\nu) is the above functional inside the squared brackets. Let now :M_1(\mathbb)=\left\ be the space of one-dimensional probability measures and consider the minimizer :E_Q=\inf\limits_-\int\int_ \ln , x-y, \mathrm\nu(x)\mathrm\nu(y)+\int Q(x)\mathrm\nu(x). For E_Q there exists a unique equilibrium measure \nu_ through the Euler-Lagrange variational conditions for some real constant l :2\int_\mathbb\log , x-y, \mathrm\nu(y)-Q(x)=l,\quad x\in J :2\int_\mathbb\log , x-y, \mathrm\nu(y)-Q(x)\leq l,\quad x\in \mathbb\setminus J where J=\bigcup\limits_^q _j,b_j/math> is the support of the measure and define :q(x)=-\left(\frac\right)^2+\int \frac\mathrm\nu_(y). The equilibrium measure \nu_ has the following Radon–Nikodym density :\frac=\frac\sqrt.


Mesoscopic regime

The typical statement of the Wigner semicircular law is equivalent to the following statement: For each ''fixed'' interval lambda_0 - \Delta \lambda, \lambda_0 + \Delta \lambda/math> centered at a point \lambda_0, as N, the number of dimensions of the gaussian ensemble increases, the proportion of the eigenvalues falling within the interval converges to \int_ \rho(t) dt, where \rho(t) is the density of the semicircular distribution. If \Delta \lambda can be allowed to decrease as N increases, then we obtain strictly stronger theorems, named "local laws" or "mesoscopic regime". The mesoscopic regime is intermediate between the local and the global. In the ''mesoscopic regime'', one is interested in the limit distribution of eigenvalues in a set that shrinks to zero, but slow enough, such that the number of eigenvalues inside \to \infty . For example, the Ginibre ensemble has a mesoscopic law: For any sequence of shrinking disks with areas u inside the unite disk, if the disks have area A_n = O(n^) , the conditional distribution of the spectrum inside the disks also converges to a uniform distribution. That is, if we cut the shrinking disks along with the spectrum falling inside the disks, then scale the disks up to unit area, we would see the spectra converging to a flat distribution in the disks.


Local regime

In the ''local regime'', one is interested in the limit distribution of eigenvalues in a set that shrinks so fast that the number of eigenvalues remains O(1) . Typically this means the study of spacings between eigenvalues, and, more generally, in the joint distribution of eigenvalues in an interval of length of order 1/''n''. One distinguishes between ''bulk statistics'', pertaining to intervals inside the support of the limiting spectral measure, and ''edge statistics'', pertaining to intervals near the boundary of the support.


Bulk statistics

Formally, fix \lambda_0 in the interior of the support of N(\lambda). Then consider the point process \Xi(\lambda_0) = \sum_j \delta\Big( - n \rho(\lambda_0) (\lambda_j - \lambda_0) \Big)~, where \lambda_j are the eigenvalues of the random matrix. The point process \Xi(\lambda_0) captures the statistical properties of eigenvalues in the vicinity of \lambda_0. For the Gaussian ensembles, the limit of \Xi(\lambda_0) is known; thus, for GUE it is a determinantal point process with the kernel K(x, y) = \frac (the ''sine kernel''). The ''universality'' principle postulates that the limit of \Xi(\lambda_0) as n \to \infty should depend only on the symmetry class of the random matrix (and neither on the specific model of random matrices nor on \lambda_0). Rigorous proofs of universality are known for invariant matrix ensembles and Wigner matrices.


Edge statistics

One example of edge statistics is the Tracy–Widom distribution. As another example, consider the Ginibre ensemble. It can be real or complex. The real Ginibre ensemble has i.i.d. standard Gaussian entries \mathcal N(0, 1), and the complex Ginibre ensemble has i.i.d. standard complex Gaussian entries \mathcal N(0, 1/2) + i\mathcal N(0, 1/2). Now let G_n be sampled from the real or complex ensemble, and let \rho(G_n) be the absolute value of its maximal eigenvalue:\rho(G_n) := \max_j , \lambda_j, We have the following theorem for the edge statistics: This theorem refines the circular law of the Ginibre ensemble. In words, the circular law says that the spectrum of \frac G_n almost surely falls uniformly on the unit disc. and the edge statistics theorem states that the radius of the almost-unit-disk is about 1-\sqrt, and fluctuates on a scale of \frac, according to the Gumbel law.


Correlation functions

The joint probability density of the eigenvalues of n\times n random Hermitian matrices M \in \mathbf^ , with partition functions of the form Z_n = \int_ d\mu_0(M)e^ where V(x):=\sum_^\infty v_j x^j and d\mu_0(M) is the standard Lebesgue measure on the space \mathbf^ of Hermitian n \times n matrices, is given by p_(x_1,\dots, x_n) = \frac\prod_ (x_i-x_j)^2 e^. The k-point correlation functions (or ''marginal distributions'') are defined as R^_(x_1,\dots,x_k) = \frac \int_dx_ \cdots \int_ dx_ \, p_(x_1,x_2,\dots,x_n), which are skew symmetric functions of their variables. In particular, the one-point correlation function, or ''density of states'', is R^_(x_1) = n\int_dx_ \cdots \int_ dx_ \, p_(x_1,x_2,\dots,x_n). Its integral over a Borel set B \subset \mathbf gives the expected number of eigenvalues contained in B: \int_ R^_(x)dx = \mathbf\left(\#\\right). The following result expresses these correlation functions as determinants of the matrices formed from evaluating the appropriate integral kernel at the pairs (x_i, x_j) of points appearing within the correlator. Theorem yson-Mehta For any k, 1\leq k \leq n the k-point correlation function R^_ can be written as a determinant R^_(x_1,x_2,\dots,x_k) = \det_\left(K_(x_i,x_j)\right), where K_(x,y) is the nth Christoffel-Darboux kernel K_(x,y) := \sum_^\psi_k(x)\psi_k(y), associated to V, written in terms of the quasipolynomials \psi_k(x) = \, p_k(z)\, e^ , where \_ is a complete sequence of monic polynomials, of the degrees indicated, satisfying the orthogonilty conditions \int_ \psi_j(x) \psi_k(x) dx = \delta_.


Other classes of random matrices


Wishart matrices

''Wishart matrices'' are ''n'' × ''n'' random matrices of the form , where ''X'' is an ''n'' × ''m'' random matrix (''m'' ≥ ''n'') with independent entries, and ''X''* is its conjugate transpose. In the important special case considered by Wishart, the entries of ''X'' are identically distributed Gaussian random variables (either real or complex). The limit of the empirical spectral measure of Wishart matrices was found by Vladimir Marchenko and Leonid Pastur.


Random unitary matrices


Non-Hermitian random matrices


Selected bibliography


Books

* * * *


Survey articles

* * * * * *


Historic works

* * *


References


External links

* * {{DEFAULTSORT:Random Matrix Algebra of random variables Mathematical physics Probability theory