Network Entropy
   HOME

TheInfoList



OR:

In
network science Network science is an academic field which studies complex networks such as telecommunication networks, computer networks, biological networks, Cognitive network, cognitive and semantic networks, and social networks, considering distinct eleme ...
, the network entropy is a disorder measure derived from
information theory Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
to describe the level of randomness and the amount of information encoded in a graph. It is a relevant metric to quantitatively characterize real
complex networks Complex Networks is an American media and entertainment company for youth culture, based in New York City. It was founded as a bi-monthly magazine, ''Complex'', by fashion designer Marc Eckō. Complex Networks reports on popular and emerging ...
and can also be used to quantify network complexity


Formulations

According to a 2018 publication by Zenil ''et al.'' there are several formulations by which to calculate network entropy and, as a rule, they all require a particular property of the graph to be focused, such as the adjacency matrix, degree sequence, degree distribution or number of bifurcations, what might lead to values of entropy that aren't invariant to the chosen network description.


Degree Distribution Shannon Entropy

The
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
can be measured for the network degree probability distribution as an average measurement of the heterogeneity of the network. \mathcal = - \sum_^ P(k) \ln This formulation has limited use with regards to complexity, information content, causation and temporal information. Be that as it may, algorithmic complexity has the ability to characterize any general or universal property of a graph or network and it is proven that graphs with low entropy have low algorithmic complexity because the statistical regularities found in a graph are useful for computer programs to recreate it. The same cannot be said for high entropy networks though, as these might have any value for algorithmic complexity.


Random Walker Shannon Entropy

Due to the limits of the previous formulation, it is possible to take a different approach while keeping the usage of the original Shannon Entropy equation. Consider a random walker that travels around the graph, going from a node i to any node j adjacent to i with equal probability. The probability distribution p_ that describes the behavior of this random walker would thus be p_ = \begin \frac, & \text A_ = 1 \\ 0, & \text A_ = 0 \\ \end, where (A_) is the graph adjacency matrix and k_i is the node i degree. From that, the
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
from each node \mathcal_i can be defined as \mathcal_i = - \sum_^ p_ \ln = \ln and, since max(k_i) = N - 1, the normalized node entropy \mathcal_i is calculated \mathcal_i = \frac = \frac = \frac This leads to a normalized network entropy \mathcal, calculated by averaging the normalized node entropy over the whole network: \mathcal = \frac \sum_^N \mathcal_i = \frac \sum_^N \ln The normalized network entropy is maximal \mathcal = 1 when the network is fully connected and decreases the sparser the network becomes \mathcal = 0. Notice that isolated nodes k_i = 0 do not have its probability p_ defined and, therefore, are not considered when measuring the network entropy. This formulation of network entropy has low sensitivity to hubs due to the logarithmic factor and is more meaningful for weighted networks., what ultimately makes it hard to differentiate scale-free networks using this measure alone.


Random Walker Kolmogorov–Sinai Entropy

The limitations of the random walker Shannon entropy can be overcome by adapting it to use a Kolmogorov–Sinai entropy. In this context, network entropy is the entropy of a stochastic matrix associated with the graph adjacency matrix (A_) and the random walker Shannon entropy is called the ''dynamic entropy'' of the network. From that, let \lambda be the dominant eigenvalue of (A_). It is proven that \ln \lambda satisfies a variational principal that is equivalent to the ''dynamic entropy'' for unweighted networks, i.e., the adjacency matrix consists exclusively of boolean values. Therefore, the
topological entropy In mathematics, the topological entropy of a topological dynamical system is a nonnegative extended real number that is a measure of the complexity of the system. Topological entropy was first introduced in 1965 by Adler, Konheim and McAndrew. Th ...
is defined as \mathcal = \ln \lambda This formulation is important to the study of network robustness, i.e., the capacity of the network to withstand random structural changes. Robustness is actually difficult to be measured numerically whereas the entropy can be easily calculated for any network, which is especially important in the context of non-stationary networks. The entropic fluctuation theorem shows that this entropy is positively correlated to robustness and hence a greater insensitivity of an observable to dynamic or structural perturbations of the network. Moreover, the eigenvalues are inherently related to the multiplicity of internal pathways, leading to a negative correlation between the
topological entropy In mathematics, the topological entropy of a topological dynamical system is a nonnegative extended real number that is a measure of the complexity of the system. Topological entropy was first introduced in 1965 by Adler, Konheim and McAndrew. Th ...
and the shortest average path length. Other than that, the Kolmogorov entropy is related to the
Ricci curvature In differential geometry, the Ricci curvature tensor, named after Gregorio Ricci-Curbastro, is a geometric object which is determined by a choice of Riemannian or pseudo-Riemannian metric on a manifold. It can be considered, broadly, as a measure ...
of the network, a metric that has been used to differentiate stages of cancer from gene co-expression networks, as well as to give hallmarks of financial crashes from stock correlation networks


Von Neumann entropy

Von Neumann entropy In physics, the von Neumann entropy, named after John von Neumann, is a measure of the statistical uncertainty within a description of a quantum system. It extends the concept of Gibbs entropy from classical statistical mechanics to quantum statis ...
is the extension of the classical Gibbs entropy in a quantum context. This entropy is constructed from a
density matrix In quantum mechanics, a density matrix (or density operator) is a matrix used in calculating the probabilities of the outcomes of measurements performed on physical systems. It is a generalization of the state vectors or wavefunctions: while th ...
\rho: historically, the first proposed candidate for such a density matrix has been an expression of the
Laplacian matrix In the mathematical field of graph theory, the Laplacian matrix, also called the graph Laplacian, admittance matrix, Kirchhoff matrix, or discrete Laplacian, is a matrix representation of a graph. Named after Pierre-Simon Laplace, the graph Lap ...
L associated with the network. The average von Neumann entropy of an ensemble is calculated as: _ = -\langle\mathrm\rho\log(\rho)\rangle For random network ensemble G(N,p), the relation between S_ and S is nonmonotonic when the average connectivity p(N-1) is varied. For canonical power-law network ensembles, the two entropies are linearly related. _ = \eta + \beta Networks with given expected degree sequences suggest that, heterogeneity in the expected degree distribution implies an equivalence between a quantum and a classical description of networks, which respectively corresponds to the von Neumann and the Shannon entropy. This definition of the Von Neumann entropy can also be extended to multilayer networks with tensorial approach and has been used successfully to reduce their dimensionality from a structural point of perspective. However, it has been shown that this definition of entropy does not satisfy the property of sub-additivity (see Von Neumann entropy's subadditivity), expected to hold theoretically. A more grounded definition, satisfying this fundamental property, has been introduced by
Manlio De Domenico Manlio De Domenico is an Italian physicist and complex systems scientist, currently Professor of Physics at the University of Padua and previously at the Fondazione Bruno Kessler in Trento (Italy). In 2014 he has co-founded the Mediterranean S ...
and Biamonte as a quantum-like Gibbs state \rho(\beta)=\frac where Z(\beta)=Tr ^/math> is a normalizing factor which plays the role of the partition function, and \beta is a tunable parameter which allows multi-resolution analysis. If \beta is interpreted as a temporal parameter, this density matrix is formally proportional to the propagator of a diffusive process on the top of the network. This feature has been used to build a statistical field theory of complex information dynamics, where the density matrix can be interpreted in terms of the super-position of streams operators whose action is to activate information flows among nodes. The framework has been successfully applied to analyze the protein-protein interaction networks of virus-human interactomes, including the
SARS-CoV-2 Severe acute respiratory syndrome coronavirus 2 (SARS‑CoV‑2) is a strain of coronavirus that causes COVID-19, the respiratory illness responsible for the COVID-19 pandemic. The virus previously had the Novel coronavirus, provisional nam ...
, to unravel the systemic features of infection of the latter at microscopic, mesoscopic and macroscopic scales, as well as to assess the importance of nodes for integrating information flows within the network and the role they play in network robustness. This approach has been generalized to deal with other types of dynamics, such as random walks, on the top of multilayer networks, providing an effective way to reduce the dimensionality of such systems without altering their structure. Using both classical and maximum-entropy random walks, the corresponding density matrices have been used to encode the network states of the human brain and to assess, at multiple scales, connectome’s information capacity at different stages of dementia.


Maximum Entropy Principle

The maximum entropy principle is a variational principal stating that the probability distribution best representing the current state of a system is the one which maximizes the Shannon entropy. This concept can be used to generate an ensemble of random graphs with given structural properties derived from the maximum entropy approach which, in its turn, describes the most probable network configuration: the maximum entropy principle allows for maximally unbiased information when lacking complete knowledge (microscopic configuration is not accessible, e.g.: we don't know the adjacency matrix). On the other hand, this ensemble serves as a null model when the actual microscopic configuration of the network is known, allowing to assess the significance of empirical patterns found in the network


Network Ensembles

It is possible to extend the network entropy formulations to instead measure the ensemble entropy. A set of networks that satisfies given structural characteristics can be treated as a network ensemble. Brought up by Ginestra Bianconi in 2007, the entropy of a network ensemble measures the level of the order or uncertainty of a network ensemble. The entropy is the logarithm of the number of graphs. Entropy can also be defined in one network. Basin entropy is the logarithm of the attractors in one
Boolean network A Boolean network consists of a discrete set of Boolean variables each of which has a Boolean function (possibly different for each variable) assigned to it which takes inputs from a subset of those variables and output that determines the sta ...
. Employing approaches from
statistical mechanics In physics, statistical mechanics is a mathematical framework that applies statistical methods and probability theory to large assemblies of microscopic entities. Sometimes called statistical physics or statistical thermodynamics, its applicati ...
, the complexity, uncertainty, and randomness of networks can be described by network ensembles with different types of constraints.


Gibbs and Shannon entropy

By analogy to statistical mechanics,
microcanonical ensemble In statistical mechanics, the microcanonical ensemble is a statistical ensemble that represents the possible states of a mechanical system whose total energy is exactly specified. The system is assumed to be isolated in the sense that it canno ...
s and
canonical ensemble In statistical mechanics, a canonical ensemble is the statistical ensemble that represents the possible states of a mechanical system in thermal equilibrium with a heat bath at a fixed temperature. The system can exchange energy with the hea ...
s of networks are introduced for the implementation. A partition function Z of an ensemble can be defined as: Z = \sum_ \delta \left vec(\mathbf)-\vec\right\exp\left(\sum_h_\Theta(a_) + r_a_\right) where \vec(\mathbf)=\vec is the constraint, and a_ (a_ \geq ) are the elements in the
adjacency matrix In graph theory and computer science, an adjacency matrix is a square matrix used to represent a finite graph (discrete mathematics), graph. The elements of the matrix (mathematics), matrix indicate whether pairs of Vertex (graph theory), vertices ...
, a_ > 0 if and only if there is a link between node i and node j. \Theta(a_) is a step function with \Theta(a_) = 1 if x > 0, and \Theta(a_) = 0 if x = 0. The auxiliary fields h_ and r_ have been introduced as analogy to the bath in classical mechanics. For simple undirected networks, the partition function can be simplified as Z = \sum_ \prod_\delta(\textrm_(\)) \exp\left(\sum_\sum_h_(\alpha)\delta_\right) where a_\in\alpha, \alpha is the index of the weight, and for a simple network \alpha=\. Microcanonical ensembles and canonical ensembles are demonstrated with simple undirected networks. For a microcanonical ensemble, the
Gibbs entropy The concept entropy was first developed by German physicist Rudolf Clausius in the mid-nineteenth century as a thermodynamic property that predicts that certain spontaneous processes are irreversible or impossible. In statistical mechanics, entrop ...
\Sigma is defined by: \begin \Sigma &= \frac \log\mathcal \\ &= \frac \log Z, _ \end where \mathcal indicates the cardinality of the ensemble, i.e., the total number of networks in the ensemble. The probability of having a link between nodes i and j, with weight \alpha is given by: \pi_(\alpha) = \frac For a canonical ensemble, the entropy is presented in the form of a
Shannon entropy Shannon may refer to: People * Shannon (given name) * Shannon (surname) * Shannon (American singer), stage name of singer Brenda Shannon Greene (born 1958) * Shannon (South Korean singer), British-South Korean singer and actress Shannon Arrum ...
: =-\sum_\sum_ \pi_(\alpha) \log \pi_(\alpha)


Relation between Gibbs and Shannon entropy

Network ensemble G(N,L) with given number of nodes N and links L, and its conjugate-canonical ensemble G(N,p) are characterized as microcanonical and canonical ensembles and they have Gibbs entropy \Sigma and the Shannon entropy S, respectively. The Gibbs entropy in the G(N,p) ensemble is given by: \Sigma = \log\left(\begin\cfrac\\L\end\right) For G(N,p) ensemble, _ = p = \cfrac Inserting p_ into the Shannon entropy: \Sigma = S/N+\cfrac\left log\left( \cfrac \right) - \log\left(\cfrac-L\right)\right/math> The relation indicates that the Gibbs entropy \Sigma and the Shannon entropy per node S/N of random graphs are equal in the
thermodynamic limit In statistical mechanics, the thermodynamic limit or macroscopic limit, of a system is the Limit (mathematics), limit for a large number of particles (e.g., atoms or molecules) where the volume is taken to grow in proportion with the number of ...
N\to\infty.


See also

*
Canonical ensemble In statistical mechanics, a canonical ensemble is the statistical ensemble that represents the possible states of a mechanical system in thermal equilibrium with a heat bath at a fixed temperature. The system can exchange energy with the hea ...
*
Microcanonical ensemble In statistical mechanics, the microcanonical ensemble is a statistical ensemble that represents the possible states of a mechanical system whose total energy is exactly specified. The system is assumed to be isolated in the sense that it canno ...
* Maximum-entropy random graph model * Graph entropy


References

{{reflist Network Entropy