HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
mathematical physics Mathematical physics is the development of mathematics, mathematical methods for application to problems in physics. The ''Journal of Mathematical Physics'' defines the field as "the application of mathematics to problems in physics and the de ...
, a random matrix is a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
-valued
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
—that is, a matrix in which some or all of its entries are
sampled Sample or samples may refer to: * Sample (graphics), an intersection of a color channel and a pixel * Sample (material), a specimen or small quantity of something * Sample (signal), a digital discrete sample of a continuous analog signal * Sample ...
randomly from a
probability distribution In probability theory and statistics, a probability distribution is a Function (mathematics), function that gives the probabilities of occurrence of possible events for an Experiment (probability theory), experiment. It is a mathematical descri ...
. Random matrix theory (RMT) is the study of properties of random matrices, often as they become large. RMT provides techniques like
mean-field theory In physics and probability theory, Mean-field theory (MFT) or Self-consistent field theory studies the behavior of high-dimensional random (stochastic) models by studying a simpler model that approximates the original by averaging over Degrees of ...
, diagrammatic methods, the
cavity method The cavity method is a mathematical method presented by Marc Mézard, Giorgio Parisi and Miguel Angel Virasoro in 1987 to derive and solve some mean field-type models in statistical physics, specially adapted to disordered systems. The method ha ...
, or the replica method to compute quantities like traces, spectral densities, or scalar products between eigenvectors. Many physical phenomena, such as the
spectrum A spectrum (: spectra or spectrums) is a set of related ideas, objects, or properties whose features overlap such that they blend to form a continuum. The word ''spectrum'' was first used scientifically in optics to describe the rainbow of co ...
of nuclei of heavy atoms, the
thermal conductivity The thermal conductivity of a material is a measure of its ability to heat conduction, conduct heat. It is commonly denoted by k, \lambda, or \kappa and is measured in W·m−1·K−1. Heat transfer occurs at a lower rate in materials of low ...
of a lattice, or the emergence of quantum chaos, can be modeled mathematically as problems concerning large, random matrices.


Applications


Physics

In
nuclear physics Nuclear physics is the field of physics that studies atomic nuclei and their constituents and interactions, in addition to the study of other forms of nuclear matter. Nuclear physics should not be confused with atomic physics, which studies th ...
, random matrices were introduced by
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
to model the nuclei of heavy atoms. Wigner postulated that the spacings between the lines in the spectrum of a heavy atom nucleus should resemble the spacings between the
eigenvalues In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
of a random matrix, and should depend only on the symmetry class of the underlying evolution. In
solid-state physics Solid-state physics is the study of rigid matter, or solids, through methods such as solid-state chemistry, quantum mechanics, crystallography, electromagnetism, and metallurgy. It is the largest branch of condensed matter physics. Solid-state phy ...
, random matrices model the behaviour of large disordered Hamiltonians in the mean-field approximation. In quantum chaos, the Bohigas–Giannoni–Schmit (BGS) conjecture asserts that the spectral statistics of quantum systems whose classical counterparts exhibit chaotic behaviour are described by random matrix theory. In
quantum optics Quantum optics is a branch of atomic, molecular, and optical physics and quantum chemistry that studies the behavior of photons (individual quanta of light). It includes the study of the particle-like properties of photons and their interaction ...
, transformations described by random unitary matrices are crucial for demonstrating the advantage of quantum over classical computation (see, e.g., the boson sampling model). Moreover, such random unitary transformations can be directly implemented in an optical circuit, by mapping their parameters to optical circuit components (that is
beam splitter A beam splitter or beamsplitter is an optical instrument, optical device that splits a beam of light into a transmitted and a reflected beam. It is a crucial part of many optical experimental and measurement systems, such as Interferometry, int ...
s and phase shifters). Random matrix theory has also found applications to the chiral Dirac operator in
quantum chromodynamics In theoretical physics, quantum chromodynamics (QCD) is the study of the strong interaction between quarks mediated by gluons. Quarks are fundamental particles that make up composite hadrons such as the proton, neutron and pion. QCD is a type of ...
,
quantum gravity Quantum gravity (QG) is a field of theoretical physics that seeks to describe gravity according to the principles of quantum mechanics. It deals with environments in which neither gravitational nor quantum effects can be ignored, such as in the v ...
in two dimensions,
mesoscopic physics Mesoscopic physics is a subdiscipline of condensed matter physics that deals with materials of an intermediate size. These materials range in size between the nanoscale for a quantity of atoms (such as a molecule) and of materials measuring mic ...
,
spin-transfer torque Spin-transfer torque (STT) is an effect in which the orientation of a magnetic layer in a magnetic tunnel junction or spin valve can be modified using a spin-polarized current. Charge carriers (such as electrons) have a property known as spin ...
, the
fractional quantum Hall effect The fractional quantum Hall effect (fractional QHE or FQHE) is the observation of precisely quantized plateaus in the Hall conductance of 2-dimensional (2D) electrons at fractional values of e^2/h, where ''e'' is the electron charge and ''h'' i ...
,
Anderson localization In condensed matter physics, Anderson localization (also known as strong localization) is the absence of diffusion of waves in a ''disordered'' medium. This phenomenon is named after the American physicist P. W. Anderson, who was the first to su ...
,
quantum dots Quantum dots (QDs) or semiconductor nanocrystals are semiconductor particles a few nanometres in size with optical and electronic properties that differ from those of larger particles via quantum mechanical effects. They are a central topic i ...
, and
superconductors Superconductivity is a set of physical properties observed in superconductors: materials where electrical resistance vanishes and magnetic fields are expelled from the material. Unlike an ordinary metallic conductor, whose resistance decreases ...


Mathematical statistics and numerical analysis

In
multivariate statistics Multivariate statistics is a subdivision of statistics encompassing the simultaneous observation and analysis of more than one outcome variable, i.e., '' multivariate random variables''. Multivariate statistics concerns understanding the differ ...
, random matrices were introduced by John Wishart, who sought to estimate covariance matrices of large samples. Chernoff-, Bernstein-, and Hoeffding-type inequalities can typically be strengthened when applied to the maximal eigenvalue (i.e. the eigenvalue of largest magnitude) of a finite sum of random
Hermitian matrices In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -th ...
. Random matrix theory is used to study the spectral properties of random matrices—such as sample covariance matrices—which is of particular interest in
high-dimensional statistics In statistical theory, the field of high-dimensional statistics studies data whose dimension is larger (relative to the number of datapoints) than typically considered in classical multivariate analysis. The area arose owing to the emergence of man ...
. Random matrix theory also saw applications in
neural networks A neural network is a group of interconnected units called neurons that send signals to one another. Neurons can be either Cell (biology), biological cells or signal pathways. While individual neurons are simple, many of them together in a netwo ...
and
deep learning Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression, and representation learning. The field takes inspiration from biological neuroscience a ...
, with recent work utilizing random matrices to show that hyper-parameter tunings can be cheaply transferred between large neural networks without the need for re-training. In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
, random matrices have been used since the work of
John von Neumann John von Neumann ( ; ; December 28, 1903 – February 8, 1957) was a Hungarian and American mathematician, physicist, computer scientist and engineer. Von Neumann had perhaps the widest coverage of any mathematician of his time, in ...
and
Herman Goldstine Herman Heine Goldstine (September 13, 1913 – June 16, 2004) was a mathematician and computer scientist, who worked as the director of the IAS machine at the Institute for Advanced Study and helped to develop ENIAC, the first of the modern ...
to describe computation errors in operations such as
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
. Although random entries are traditional "generic" inputs to an algorithm, the
concentration of measure In mathematics, concentration of measure (about a median) is a principle that is applied in measure theory, probability and combinatorics, and has consequences for other fields such as Banach space theory. Informally, it states that "A random var ...
associated with random matrix distributions implies that random matrices will not test large portions of an algorithm's input space.


Number theory

In
number theory Number theory is a branch of pure mathematics devoted primarily to the study of the integers and arithmetic functions. Number theorists study prime numbers as well as the properties of mathematical objects constructed from integers (for example ...
, the distribution of zeros of the
Riemann zeta function The Riemann zeta function or Euler–Riemann zeta function, denoted by the Greek letter (zeta), is a mathematical function of a complex variable defined as \zeta(s) = \sum_^\infty \frac = \frac + \frac + \frac + \cdots for and its analytic c ...
(and other
L-function In mathematics, an ''L''-function is a meromorphic function on the complex plane, associated to one out of several categories of mathematical objects. An ''L''-series is a Dirichlet series, usually convergent on a half-plane, that may gi ...
s) is modeled by the distribution of eigenvalues of certain random matrices. The connection was first discovered by Hugh Montgomery and
Freeman Dyson Freeman John Dyson (15 December 1923 – 28 February 2020) was a British-American theoretical physics, theoretical physicist and mathematician known for his works in quantum field theory, astrophysics, random matrix, random matrices, math ...
. It is connected to the
Hilbert–Pólya conjecture In mathematics, the Hilbert–Pólya conjecture states that the non-trivial Zero of a function, zeros of the Riemann zeta function correspond to eigenvalues of a self-adjoint operator. It is a possible approach to the Riemann hypothesis, by means o ...
.


Free probability

The relation of
free probability Free probability is a mathematics, mathematical theory that studies non-commutative random variables. The "freeness" or free independence property is the analogue of the classical notion of statistical independence, independence, and it is connecte ...
with random matrices is a key reason for the wide use of free probability in other subjects. Voiculescu introduced the concept of freeness around 1983 in an operator algebraic context; at the beginning there was no relation at all with random matrices. This connection was only revealed later in 1991 by Voiculescu; he was motivated by the fact that the limit distribution which he found in his free central limit theorem had appeared before in Wigner's semi-circle law in the random matrix context.


Computational neuroscience

In the field of computational neuroscience, random matrices are increasingly used to model the network of synaptic connections between neurons in the brain. Dynamical models of neuronal networks with random connectivity matrix were shown to exhibit a phase transition to chaos when the variance of the synaptic weights crosses a critical value, at the limit of infinite system size. Results on random matrices have also shown that the dynamics of random-matrix models are insensitive to mean connection strength. Instead, the stability of fluctuations depends on connection strength variation and time to synchrony depends on network topology. In the analysis of massive data such as
fMRI Functional magnetic resonance imaging or functional MRI (fMRI) measures brain activity by detecting changes associated with blood flow. This technique relies on the fact that cerebral blood flow and neuronal activation are coupled. When an area o ...
, random matrix theory has been applied in order to perform dimension reduction. When applying an algorithm such as PCA, it is important to be able to select the number of significant components. The criteria for selecting components can be multiple (based on explained variance, Kaiser's method, eigenvalue, etc.). Random matrix theory in this content has its representative the Marchenko-Pastur distribution, which guarantees the theoretical high and low limits of the eigenvalues associated with a random variable covariance matrix. This matrix calculated in this way becomes the null hypothesis that allows one to find the eigenvalues (and their eigenvectors) that deviate from the theoretical random range. The components thus excluded become the reduced dimensional space (see examples in fMRI ).


Optimal control

In
optimal control Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations ...
theory, the evolution of ''n'' state variables through time depends at any time on their own values and on the values of ''k'' control variables. With linear evolution, matrices of coefficients appear in the state equation (equation of evolution). In some problems the values of the parameters in these matrices are not known with certainty, in which case there are random matrices in the state equation and the problem is known as one of
stochastic control Stochastic control or stochastic optimal control is a sub field of control theory that deals with the existence of uncertainty either in observations or in the noise that drives the evolution of the system. The system designer assumes, in a Bayesi ...
. A key result in the case of linear-quadratic control with stochastic matrices is that the certainty equivalence principle does not apply: while in the absence of multiplier uncertainty (that is, with only additive uncertainty) the optimal policy with a quadratic loss function coincides with what would be decided if the uncertainty were ignored, the optimal policy may differ if the state equation contains random coefficients.


Computational mechanics

In
computational mechanics Computational mechanics is the discipline concerned with the use of computational methods to study phenomena governed by the principles of mechanics. Before the emergence of computational science (also called scientific computing) as a "third wa ...
, epistemic uncertainties underlying the lack of knowledge about the physics of the modeled system give rise to mathematical operators associated with the computational model, which are deficient in a certain sense. Such operators lack certain properties linked to unmodeled physics. When such operators are discretized to perform computational simulations, their accuracy is limited by the missing physics. To compensate for this deficiency of the mathematical operator, it is not enough to make the model parameters random, it is necessary to consider a mathematical operator that is random and can thus generate families of computational models in the hope that one of these captures the missing physics. Random matrices have been used in this sense, with applications in vibroacoustics, wave propagations, materials science, fluid mechanics, heat transfer, etc.


Engineering

Random matrix theory can be applied to the electrical and communications engineering research efforts to study, model and develop Massive Multiple-Input Multiple-Output (
MIMO In radio, multiple-input and multiple-output (MIMO) () is a method for multiplying the capacity of a radio link using multiple transmission and receiving antennas to exploit multipath propagation. MIMO has become an essential element of wirel ...
) radio systems.


History

Random matrix theory first gained attention beyond mathematics literature in the context of nuclear physics. Experiments by
Enrico Fermi Enrico Fermi (; 29 September 1901 – 28 November 1954) was an Italian and naturalized American physicist, renowned for being the creator of the world's first artificial nuclear reactor, the Chicago Pile-1, and a member of the Manhattan Project ...
and others demonstrated evidence that individual
nucleons In physics and chemistry, a nucleon is either a proton or a neutron, considered in its role as a component of an atomic nucleus. The number of nucleons in a nucleus defines the atom's mass number. Until the 1960s, nucleons were thought to be ele ...
cannot be approximated to move independently, leading
Niels Bohr Niels Henrik David Bohr (, ; ; 7 October 1885 – 18 November 1962) was a Danish theoretical physicist who made foundational contributions to understanding atomic structure and old quantum theory, quantum theory, for which he received the No ...
to formulate the idea of a compound nucleus. Because there was no knowledge of direct nucleon-nucleon interactions,
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
and Leonard Eisenbud approximated that the nuclear
Hamiltonian Hamiltonian may refer to: * Hamiltonian mechanics, a function that represents the total energy of a system * Hamiltonian (quantum mechanics), an operator corresponding to the total energy of that system ** Dyall Hamiltonian, a modified Hamiltonian ...
could be modeled as a random matrix. For larger atoms, the distribution of the energy eigenvalues of the Hamiltonian could be computed in order to approximate scattering cross sections by invoking the
Wishart distribution In statistics, the Wishart distribution is a generalization of the gamma distribution to multiple dimensions. It is named in honor of John Wishart (statistician), John Wishart, who first formulated the distribution in 1928. Other names include Wi ...
.


Gaussian ensembles

The most-commonly studied random matrix distributions are the Gaussian ensembles: GOE, GUE and GSE. They are often denoted by their
Dyson Dyson may refer to: * Dyson (surname), people with the surname Dyson * Dyson (company), a Singaporean multinational home appliances company founded by James Dyson * Dyson (crater), a crater on the Moon * Dyson (operating system), a Unix general- ...
index, ''β'' = 1 for GOE, ''β'' = 2 for GUE, and ''β'' = 4 for GSE. This index counts the number of real components per matrix element.


Definitions

The Gaussian unitary ensemble \text(n) is described by the
Gaussian measure In mathematics, Gaussian measure is a Borel measure on finite-dimensional Euclidean space \mathbb^n, closely related to the normal distribution in statistics. There is also a generalization to infinite-dimensional spaces. Gaussian measures are na ...
with density \frac e^ on the space of n \times n
Hermitian matrices In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -th ...
H = (H_)_^n. Here Z_ = 2^ \left(\frac\right)^ is a normalization constant, chosen so that the integral of the density is equal to one. The term ''unitary'' refers to the fact that the distribution is invariant under unitary conjugation. The Gaussian unitary ensemble models Hamiltonians lacking time-reversal symmetry. The Gaussian orthogonal ensemble \text(n) is described by the Gaussian measure with density \frac e^ on the space of ''n'' × ''n'' real symmetric matrices ''H'' = (''H''''ij''). Its distribution is invariant under orthogonal conjugation, and it models Hamiltonians with time-reversal symmetry. Equivalently, it is generated by H = (G+G^T)/\sqrt, where G is an n\times n matrix with IID samples from the standard normal distribution. The Gaussian symplectic ensemble \text(n) is described by the Gaussian measure with density \frac e^ on the space of ''n'' × ''n'' Hermitian quaternionic matrices, e.g. symmetric square matrices composed of
quaternion In mathematics, the quaternion number system extends the complex numbers. Quaternions were first described by the Irish mathematician William Rowan Hamilton in 1843 and applied to mechanics in three-dimensional space. The algebra of quater ...
s, . Its distribution is invariant under conjugation by the
symplectic group In mathematics, the name symplectic group can refer to two different, but closely related, collections of mathematical groups, denoted and for positive integer ''n'' and field F (usually C or R). The latter is called the compact symplectic gr ...
, and it models Hamiltonians with time-reversal symmetry but no rotational symmetry.


Point correlation functions

The ensembles as defined here have Gaussian distributed matrix elements with mean ⟨''H''''ij''⟩ = 0, and two-point correlations given by \langle H_ H_^* \rangle = \langle H_ H_ \rangle = \frac \delta_ \delta_ + \frac\delta_\delta_ , from which all higher correlations follow by Isserlis' theorem.


Moment generating functions

The
moment generating function In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compare ...
for the GOE isE ^= e^where \, \cdot \, _F is the
Frobenius norm In the field of mathematics, norms are defined for elements within a vector space. Specifically, when the vector space comprises matrices, such norms are referred to as matrix norms. Matrix norms differ from vector norms in that they must also ...
.


Spectral density

The joint
probability density In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values ...
for the
eigenvalues In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
of GUE/GOE/GSE is given by where ''Z''''β'',''n'' is a normalization constant which can be explicitly computed, see
Selberg integral In mathematics, the Selberg integral is a generalization of Euler beta function to ''n'' dimensions introduced by Atle Selberg. Selberg's integral formula When Re(\alpha) > 0, Re(\beta) > 0, Re(\gamma) > -\min \left(\frac 1n , \frac, \frac\right) ...
. In the case of GUE (''β'' = 2), the formula (1) describes a determinantal point process. Eigenvalues repel as the joint probability density has a zero (of \betath order) for coinciding eigenvalues \lambda_j = \lambda_i, and Z_ = (2\pi)^ \prod_^n k!. More succinctly, \frac e^ , \Delta_n(\lambda), ^\betawhere \Delta_n is the
Vandermonde determinant In algebra, the Vandermonde polynomial of an ordered set of ''n'' variables X_1,\dots, X_n, named after Alexandre-Théophile Vandermonde, is the polynomial: :V_n = \prod_ (X_j-X_i). (Some sources use the opposite order (X_i-X_j), which changes the ...
. The distribution of the largest eigenvalue for GOE, and GUE, are explicitly solvable. They converge to the Tracy–Widom distribution after shifting and scaling appropriately.


Convergence to Wigner semicircular distribution

The spectrum, divided by \sqrt, converges in distribution to the semicircular distribution on the interval 2, +2/math>: \rho(x) = \frac\sqrt. Here \sigma^2 is the variance of off-diagonal entries. The variance of the on-diagonal entries do not matter.


Distribution of level spacings

From the ordered sequence of eigenvalues \lambda_1 < \ldots < \lambda_n < \lambda_ < \ldots, one defines the normalized spacings s = (\lambda_ - \lambda_n)/\langle s \rangle, where \langle s \rangle =\langle \lambda_ - \lambda_n \rangle is the mean spacing. The probability distribution of spacings is approximately given by, p_1(s) = \fracs\, e^ for the orthogonal ensemble GOE \beta=1, p_2(s) = \fracs^2 \mathrm^ for the unitary ensemble GUE \beta=2, and p_4(s) = \fracs^4 e^ for the symplectic ensemble GSE \beta = 4. The numerical constants are such that p_\beta(s) is normalized: \int_0^\infty ds\,p_\beta(s) = 1 and the mean spacing is, \int_0^\infty ds\, s\, p_\beta(s) = 1, for \beta = 1,2,4 .


Generalizations

''Wigner matrices'' are random Hermitian matrices H_n = (H_n(i,j))_^n such that the entries \left\ above the main diagonal are independent random variables with zero mean and have identical second moments. The Gaussian ensembles can be extended for \beta \neq 1, 2, 4 using the Dumitriu-Edelman tridiagonal ensemble. ''Invariant matrix ensembles'' are random Hermitian matrices with density on the space of real symmetric/Hermitian/quaternionic Hermitian matrices, which is of the form \frac e^~, where the function is called the potential. The Gaussian ensembles are the only common special cases of these two classes of random matrices. This is a consequence of a theorem by Porter and Rosenzweig.


Spectral theory of random matrices

The spectral theory of random matrices studies the distribution of the eigenvalues as the size of the matrix goes to infinity.


Empirical spectral measure

The ''empirical spectral measure'' \mu_H of H is defined by \mu_(A) = \frac \, \# \left\ = N_, \quad A \subset \mathbb. or more succinctly, if \lambda_1,\ldots, \lambda_n are the eigenvalues of H \mu_(d\lambda) = \frac 1n \sum_i \delta_(d\lambda). Usually, the limit of \mu_ is a deterministic measure; this is a particular case of self-averaging. The
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ever ...
of the limiting measure is called the integrated density of states and is denoted ''N''(''λ''). If the integrated density of states is differentiable, its derivative is called the
density of states In condensed matter physics, the density of states (DOS) of a system describes the number of allowed modes or quantum state, states per unit energy range. The density of states is defined as where N(E)\delta E is the number of states in the syste ...
and is denoted ''ρ''(''λ'').


Types of convergence

Given a matrix ensemble, we say that its spectral measures converge weakly to \rho iff for any measurable set A, the ensemble-average converges:\lim_ \mathbb E_H mu_H(A)= \rho(A)Convergence weakly almost surely: If we sample H_1, H_2, H_3, \dots independently from the ensemble, then with probability 1,\lim_ \mu_(A) = \rho(A)for any measurable set A. In another sense, weak almost sure convergence means that we sample H_1, H_2, H_3, \dots, not independently, but by "growing" (a
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
), then with probability 1, \lim_ \mu_(A) = \rho(A) for any measurable set A. For example, we can "grow" a sequence of matrices from the Gaussian ensemble as follows: * Sample an infinite doubly infinite sequence of standard random variables \_ . * Define each H_n = (G_n+G_n^T)/\sqrt where G_n is the matrix made of entries \_ . Note that generic matrix ensembles do not allow us to grow, but most of the common ones, such as the three Gaussian ensembles, do allow us to grow.


Global regime

In the ''global regime'', one is interested in the distribution of linear statistics of the form N_ = n^ \text f(H). The limit of the empirical spectral measure for Wigner matrices was described by
Eugene Wigner Eugene Paul Wigner (, ; November 17, 1902 – January 1, 1995) was a Hungarian-American theoretical physicist who also contributed to mathematical physics. He received the Nobel Prize in Physics in 1963 "for his contributions to the theory of th ...
; see
Wigner semicircle distribution The Wigner semicircle distribution, named after the physicist Eugene Wigner, is the probability distribution defined on the domain minus;''R'', ''R''whose probability density function ''f'' is a scaled semicircle, i.e. a semi-ellipse, centered at ...
and Wigner surmise. As far as sample covariance matrices are concerned, a theory was developed by Marčenko and Pastur.. The limit of the empirical spectral measure of invariant matrix ensembles is described by a certain integral equation which arises from
potential theory In mathematics and mathematical physics, potential theory is the study of harmonic functions. The term "potential theory" was coined in 19th-century physics when it was realized that the two fundamental forces of nature known at the time, namely g ...
.


Fluctuations

For the linear statistics , one is also interested in the fluctuations about ∫ ''f''(''λ'') ''dN''(''λ''). For many classes of random matrices, a central limit theorem of the form \frac \overset N(0, 1) is known.


The variational problem for the unitary ensembles

Consider the measure :\mathrm\mu_N(\mu)=\frace^\mathrm\lambda,\qquad H_N(\lambda)=-\sum\limits_\ln, \lambda_j-\lambda_k, +N\sum\limits_^N Q(\lambda_j), where Q(M) is the potential of the ensemble and let \nu be the empirical spectral measure. We can rewrite H_N(\lambda) with \nu as :H_N(\lambda)=N^2\left x-y, \mathrm\nu(x)\mathrm\nu(y)+\int Q(x)\mathrm\nu(x)\right the probability measure is now of the form :\mathrm\mu_N(\mu)=\frace^\mathrm\lambda, where I_Q(\nu) is the above functional inside the squared brackets. Let now :M_1(\mathbb)=\left\ be the space of one-dimensional probability measures and consider the minimizer :E_Q=\inf\limits_-\int\int_ \ln , x-y, \mathrm\nu(x)\mathrm\nu(y)+\int Q(x)\mathrm\nu(x). For E_Q there exists a unique equilibrium measure \nu_ through the Euler-Lagrange variational conditions for some real constant l :2\int_\mathbb\log , x-y, \mathrm\nu(y)-Q(x)=l,\quad x\in J :2\int_\mathbb\log , x-y, \mathrm\nu(y)-Q(x)\leq l,\quad x\in \mathbb\setminus J where J=\bigcup\limits_^q _j,b_j/math> is the support of the measure and define :q(x)=-\left(\frac\right)^2+\int \frac\mathrm\nu_(y). The equilibrium measure \nu_ has the following Radon–Nikodym density :\frac=\frac\sqrt.


Mesoscopic regime

The typical statement of the Wigner semicircular law is equivalent to the following statement: For each ''fixed'' interval lambda_0 - \Delta \lambda, \lambda_0 + \Delta \lambda/math> centered at a point \lambda_0, as N, the number of dimensions of the gaussian ensemble increases, the proportion of the eigenvalues falling within the interval converges to \int_ \rho(t) dt, where \rho(t) is the density of the semicircular distribution. If \Delta \lambda can be allowed to decrease as N increases, then we obtain strictly stronger theorems, named "local laws" or "mesoscopic regime". The mesoscopic regime is intermediate between the local and the global. In the ''mesoscopic regime'', one is interested in the limit distribution of eigenvalues in a set that shrinks to zero, but slow enough, such that the number of eigenvalues inside \to \infty . For example, the Ginibre ensemble has a mesoscopic law: For any sequence of shrinking disks with areas u inside the unite disk, if the disks have area A_n = O(n^) , the conditional distribution of the spectrum inside the disks also converges to a uniform distribution. That is, if we cut the shrinking disks along with the spectrum falling inside the disks, then scale the disks up to unit area, we would see the spectra converging to a flat distribution in the disks.


Local regime

In the ''local regime'', one is interested in the limit distribution of eigenvalues in a set that shrinks so fast that the number of eigenvalues remains O(1) . Typically this means the study of spacings between eigenvalues, and, more generally, in the joint distribution of eigenvalues in an interval of length of order 1/''n''. One distinguishes between ''bulk statistics'', pertaining to intervals inside the support of the limiting spectral measure, and ''edge statistics'', pertaining to intervals near the boundary of the support.


Bulk statistics

Formally, fix \lambda_0 in the interior of the
support Support may refer to: Arts, entertainment, and media * Supporting character * Support (art), a solid surface upon which a painting is executed Business and finance * Support (technical analysis) * Child support * Customer support * Income Su ...
of N(\lambda). Then consider the
point process In statistics and probability theory, a point process or point field is a set of a random number of mathematical points randomly located on a mathematical space such as the real line or Euclidean space. Kallenberg, O. (1986). ''Random Measures'', ...
\Xi(\lambda_0) = \sum_j \delta\Big( - n \rho(\lambda_0) (\lambda_j - \lambda_0) \Big)~, where \lambda_j are the eigenvalues of the random matrix. The point process \Xi(\lambda_0) captures the statistical properties of eigenvalues in the vicinity of \lambda_0. For the Gaussian ensembles, the limit of \Xi(\lambda_0) is known; thus, for GUE it is a determinantal point process with the kernel K(x, y) = \frac (the ''sine kernel''). The ''universality'' principle postulates that the limit of \Xi(\lambda_0) as n \to \infty should depend only on the symmetry class of the random matrix (and neither on the specific model of random matrices nor on \lambda_0). Rigorous proofs of universality are known for invariant matrix ensembles and Wigner matrices.


Edge statistics

One example of edge statistics is the Tracy–Widom distribution. As another example, consider the Ginibre ensemble. It can be real or complex. The real Ginibre ensemble has i.i.d. standard Gaussian entries \mathcal N(0, 1), and the complex Ginibre ensemble has i.i.d. standard complex Gaussian entries \mathcal N(0, 1/2) + i\mathcal N(0, 1/2). Now let G_n be sampled from the real or complex ensemble, and let \rho(G_n) be the absolute value of its maximal eigenvalue:\rho(G_n) := \max_j , \lambda_j, We have the following theorem for the edge statistics: This theorem refines the circular law of the Ginibre ensemble. In words, the circular law says that the spectrum of \frac G_n almost surely falls uniformly on the unit disc. and the edge statistics theorem states that the radius of the almost-unit-disk is about 1-\sqrt, and fluctuates on a scale of \frac, according to the Gumbel law.


Correlation functions

The joint probability density of the eigenvalues of n\times n random Hermitian matrices M \in \mathbf^ , with partition functions of the form Z_n = \int_ d\mu_0(M)e^ where V(x):=\sum_^\infty v_j x^j and d\mu_0(M) is the standard Lebesgue measure on the space \mathbf^ of Hermitian n \times n matrices, is given by p_(x_1,\dots, x_n) = \frac\prod_ (x_i-x_j)^2 e^. The k-point correlation functions (or ''marginal distributions'') are defined as R^_(x_1,\dots,x_k) = \frac \int_dx_ \cdots \int_ dx_ \, p_(x_1,x_2,\dots,x_n), which are skew symmetric functions of their variables. In particular, the one-point correlation function, or ''density of states'', is R^_(x_1) = n\int_dx_ \cdots \int_ dx_ \, p_(x_1,x_2,\dots,x_n). Its integral over a Borel set B \subset \mathbf gives the expected number of eigenvalues contained in B: \int_ R^_(x)dx = \mathbf\left(\#\\right). The following result expresses these correlation functions as determinants of the matrices formed from evaluating the appropriate integral kernel at the pairs (x_i, x_j) of points appearing within the correlator. Theorem yson-Mehta For any k, 1\leq k \leq n the k-point correlation function R^_ can be written as a determinant R^_(x_1,x_2,\dots,x_k) = \det_\left(K_(x_i,x_j)\right), where K_(x,y) is the nth Christoffel-Darboux kernel K_(x,y) := \sum_^\psi_k(x)\psi_k(y), associated to V, written in terms of the quasipolynomials \psi_k(x) = \, p_k(z)\, e^ , where \_ is a complete sequence of monic polynomials, of the degrees indicated, satisfying the orthogonilty conditions \int_ \psi_j(x) \psi_k(x) dx = \delta_.


Other classes of random matrices


Wishart matrices

''Wishart matrices'' are ''n'' × ''n'' random matrices of the form , where ''X'' is an ''n'' × ''m'' random matrix (''m'' ≥ ''n'') with independent entries, and ''X''* is its
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \mathbf is an n \times m matrix obtained by transposing \mathbf and applying complex conjugation to each entry (the complex conjugate ...
. In the important special case considered by Wishart, the entries of ''X'' are identically distributed Gaussian random variables (either real or complex). The limit of the empirical spectral measure of Wishart matrices was found by Vladimir Marchenko and Leonid Pastur.


Random unitary matrices


Non-Hermitian random matrices


Selected bibliography


Books

* * * *


Survey articles

* * * * * *


Historic works

* * *


References


External links

* * {{DEFAULTSORT:Random Matrix Algebra of random variables Mathematical physics Probability theory