HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
, a logit-normal distribution is a
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
of a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
whose
logit In statistics, the logit ( ) function is the quantile function associated with the standard logistic distribution. It has many uses in data analysis and machine learning, especially in data transformations. Mathematically, the logit is the ...
has a
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
. If ''Y'' is a random variable with a normal distribution, and ''t'' is the standard
logistic function A logistic function or logistic curve is a common S-shaped curve (sigmoid curve) with equation f(x) = \frac, where For values of x in the domain of real numbers from -\infty to +\infty, the S-curve shown on the right is obtained, with the ...
, then ''X'' = ''t''(''Y'') has a logit-normal distribution; likewise, if ''X'' is logit-normally distributed, then ''Y'' = 
logit In statistics, the logit ( ) function is the quantile function associated with the standard logistic distribution. It has many uses in data analysis and machine learning, especially in data transformations. Mathematically, the logit is the ...
(''X'')= log (''X''/(1-''X'')) is normally distributed. It is also known as the logistic normal distribution, which often refers to a multinomial logit version (e.g.). A variable might be modeled as logit-normal if it is a proportion, which is bounded by zero and one, and where values of zero and one never occur.


Characterization


Probability density function

The
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
(PDF) of a logit-normal distribution, for 0 < ''x'' < 1, is: : f_X(x;\mu,\sigma) = \frac\,\frac\, e^ where ''μ'' and ''σ'' are the
mean There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value (magnitude and sign) of a given data set. For a data set, the ''arithme ...
and
standard deviation In statistics, the standard deviation is a measure of the amount of variation or dispersion of a set of values. A low standard deviation indicates that the values tend to be close to the mean (also called the expected value) of the set, while ...
of the variable’s
logit In statistics, the logit ( ) function is the quantile function associated with the standard logistic distribution. It has many uses in data analysis and machine learning, especially in data transformations. Mathematically, the logit is the ...
(by definition, the variable’s logit is normally distributed). The density obtained by changing the sign of ''μ'' is symmetrical, in that it is equal to f(1-x;-''μ'',''σ''), shifting the mode to the other side of 0.5 (the midpoint of the (0,1) interval).


Moments

The moments of the logit-normal distribution have no analytic solution. The moments can be estimated by
numerical integration In analysis, numerical integration comprises a broad family of algorithms for calculating the numerical value of a definite integral, and by extension, the term is also sometimes used to describe the numerical solution of differential equations ...
, however numerical integration can be prohibitive when the values of \mu, \sigma^2are such that the density function diverges to infinity at the end points zero and one. An alternative is to use the observation that the logit-normal is a transformation of a normal random variable. This allows us to approximate the n-th moment via the following quasi Monte Carlo estimate E ^n\approx \frac \sum_^ \left( P\left(\Phi^_(i/K)\right) \right)^n, where P is the standard logistic function, and \Phi^_ is the inverse cumulative distribution function of a normal distribution with mean and variance \mu, \sigma^2.


Mode or modes

When the derivative of the density equals 0 then the location of the mode x satisfies the following equation: :\operatorname(x) = \sigma^2(2x-1)+\mu . For some values of the parameters there are two solutions, i.e. the distribution is
bimodal In statistics, a multimodal distribution is a probability distribution with more than one mode (statistics), mode. These appear as distinct peaks (local maxima) in the probability density function, as shown in Figures 1 and 2. Categorical, ...
.


Multivariate generalization

The logistic normal distribution is a generalization of the logit–normal distribution to D-dimensional probability vectors by taking a logistic transformation of a multivariate normal distribution.J. Atchison. "The Statistical Analysis of Compositional Data." Monographs on Statistics and Applied Probability, Chapman and Hall, 1986
Book
/ref>


Probability density function

The
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
is: : f_X( \mathbf; \boldsymbol , \boldsymbol ) = \frac \, \frac \, e^ \quad , \quad \mathbf \in \mathcal^D \;\; , where \mathbf_ denotes a vector of the first (D-1) components of \mathbf and \mathcal^D denotes the
simplex In geometry, a simplex (plural: simplexes or simplices) is a generalization of the notion of a triangle or tetrahedron to arbitrary dimensions. The simplex is so-named because it represents the simplest possible polytope in any given dimension. ...
of D-dimensional probability vectors. This follows from applying the additive logistic transformation to map a
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One d ...
random variable \mathbf \sim \mathcal \left( \boldsymbol , \boldsymbol \right) \; , \; \mathbf \in \mathbb^ to the simplex: : \mathbf = \left \frac , \dots , \frac , \frac \right\top The unique inverse mapping is given by: : \mathbf = \left \log \left( \frac \right) , \dots , \log \left( \frac \right) \right\top. This is the case of a vector x which components sum up to one. In the case of x with sigmoidal elements, that is, when : \mathbf = \left \log \left( \frac \right) , \dots , \log \left( \frac \right) \right\top we have : f_X( \mathbf; \boldsymbol , \boldsymbol ) = \frac \, \frac \, e^ where the log and the division in the argument are taken element-wise. This is because the Jacobian matrix of the transformation is diagonal with elements \frac.


Use in statistical analysis

The logistic normal distribution is a more flexible alternative to the
Dirichlet distribution In probability and statistics, the Dirichlet distribution (after Peter Gustav Lejeune Dirichlet), often denoted \operatorname(\boldsymbol\alpha), is a family of continuous multivariate probability distributions parameterized by a vector \boldsymb ...
in that it can capture correlations between components of probability vectors. It also has the potential to simplify statistical analyses of compositional data by allowing one to answer questions about log-ratios of the components of the data vectors. One is often interested in ratios rather than absolute component values. The probability simplex is a bounded space, making standard techniques that are typically applied to vectors in \mathbb^n less meaningful. Aitchison described the problem of spurious negative correlations when applying such methods directly to simplicial vectors. However, mapping compositional data in \mathcal^D through the inverse of the additive logistic transformation yields real-valued data in \mathbb^. Standard techniques can be applied to this representation of the data. This approach justifies use of the logistic normal distribution, which can thus be regarded as the "Gaussian of the simplex".


Relationship with the Dirichlet distribution

The Dirichlet and logistic normal distributions are never exactly equal for any choice of parameters. However, Aitchison described a method for approximating a Dirichlet with a logistic normal such that their Kullback–Leibler divergence (KL) is minimized: : K(p,q) = \int_ p \left( \mathbf \mid \boldsymbol \right) \log \left( \frac \right) \, d \mathbf This is minimized by: : \boldsymbol^* = \mathbf_p \left \log \left( \frac \right) \right\quad , \quad \boldsymbol^* = \textbf_p \left \log \left( \frac \right) \right/math> Using moment properties of the Dirichlet distribution, the solution can be written in terms of the
digamma Digamma or wau (uppercase: Ϝ, lowercase: ϝ, numeral: ϛ) is an archaic letter of the Greek alphabet. It originally stood for the sound but it has remained in use principally as a Greek numeral for 6. Whereas it was originally called ''waw' ...
\psi and trigamma \psi' functions: : \mu_i^* = \psi \left( \alpha_i \right) - \psi \left( \alpha_D \right) \quad , \quad i = 1 , \ldots , D-1 : \Sigma_^* = \psi' \left( \alpha_i \right) + \psi' \left( \alpha_D \right) \quad , \quad i = 1 , \ldots , D-1 : \Sigma_^* = \psi' \left( \alpha_D \right) \quad , \quad i \neq j This approximation is particularly accurate for large \boldsymbol. In fact, one can show that for \alpha_i \rightarrow \infty , i = 1 , \ldots , D, we have that p \left( \mathbf \mid \boldsymbol \right) \rightarrow q \left( \mathbf \mid \boldsymbol^* , \boldsymbol^* \right).


See also

*
Beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval , 1in terms of two positive parameters, denoted by ''alpha'' (''α'') and ''beta'' (''β''), that appear as ...
and
Kumaraswamy distribution In probability and statistics, the Kumaraswamy's double bounded distribution is a family of continuous probability distributions defined on the interval (0,1). It is similar to the Beta distribution, but much simpler to use especially in simulatio ...
, other two-parameter distributions on a bounded interval with similar shapes


References


Further reading

* Frederic, P. & Lad, F. (2008
Two Moments of the Logitnormal Distribution.
''Communications in Statistics-Simulation and Computation''. 37: 1263-1269 *


External links


logitnorm package
for R {{DEFAULTSORT:Logit-normal distribution Continuous distributions