Hopfield model
   HOME

TheInfoList



OR:

A Hopfield network (or Ising model of a neural network or Ising–Lenz–Little model) is a form of recurrent
artificial neural network Artificial neural networks (ANNs), usually simply called neural networks (NNs) or neural nets, are computing systems inspired by the biological neural networks that constitute animal brains. An ANN is based on a collection of connected unit ...
and a type of
spin glass In condensed matter physics, a spin glass is a magnetic state characterized by randomness, besides cooperative behavior in freezing of spins at a temperature called 'freezing temperature' ''Tf''. In ferromagnetic solids, component atoms' magne ...
system popularised by
John Hopfield John Joseph Hopfield (born July 15, 1933) is an American scientist most widely known for his invention of an associative neural network in 1982. It is now more commonly known as the Hopfield network. Biography Hopfield was born in 1933 to Po ...
in 1982 as described earlier by Little in 1974 based on
Ernst Ising Ernst Ising (; May 10, 1900 in Cologne, Germany – May 11, 1998 in Peoria, Illinois, USA) was a German physicist, who is best remembered for the development of the Ising model. He was a professor of physics at Bradley University until his r ...
's work with
Wilhelm Lenz Wilhelm Lenz (February 8, 1888 in Frankfurt am Main – April 30, 1957 in Hamburg) was a German physicist, most notable for his invention of the Ising model and for his application of the Laplace–Runge–Lenz vector to the old quantum mechanical ...
on the
Ising model The Ising model () (or Lenz-Ising model or Ising-Lenz model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent ...
. Hopfield networks serve as content-addressable ("associative") memory systems with
binary Binary may refer to: Science and technology Mathematics * Binary number, a representation of numbers using only two digits (0 and 1) * Binary function, a function that takes two arguments * Binary operation, a mathematical operation that ta ...
threshold nodes, or with continuous variables. Hopfield networks also provide a model for understanding human memory.


Origins

The
Ising model The Ising model () (or Lenz-Ising model or Ising-Lenz model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent ...
of a neural network as a memory model was first proposed by William A. Little in 1974, which was acknowledged by Hopfield in his 1982 paper. Networks with continuous dynamics were developed by Hopfield in his 1984 paper. A major advance in memory storage capacity was developed by Krotov and Hopfield in 2016 through a change in network dynamics and energy function. This idea was further extended by Demircigil and collaborators in 2017. The continuous dynamics of large memory capacity models was developed in a series of papers between 2016 and 2020. Large memory storage capacity Hopfield Networks are now called Dense Associative Memories or modern Hopfield networks.


Structure

The units in Hopfield nets are binary threshold units, i.e. the units only take on two different values for their states, and the value is determined by whether or not the unit's input exceeds its threshold U_i . Discrete Hopfield nets describe relationships between binary (firing or not-firing) neurons 1,2,\ldots,i,j,\ldots,N. At a certain time, the state of the neural net is described by a vector V , which records which neurons are firing in a binary word of N bits. The interactions w_ between neurons have units that usually take on values of 1 or −1, and this convention will be used throughout this article. However, other literature might use units that take values of 0 and 1. These interactions are "learned" via Hebb's law of association, such that, for a certain state V^s w_ = V_i^s V_j^s but w_ = 0 . (Note that the Hebbian learning rule takes the form w_ = (2V_i^s - 1)(2V_j^s -1) when the units assume values in \ .) Once the network is trained, w_ no longer evolve. If a new state of neurons V^ is introduced to the neural network, the net acts on neurons such that * V^_i \rightarrow 1 if \sum_j w_ V^_j > U_i * V^_i \rightarrow -1 if \sum_j w_ V^_j < U_i where U_i is the threshold value of the i'th neuron (often taken to be 0). In this way, Hopfield networks have the ability to "remember" states stored in the interaction matrix, because if a new state V^ is subjected to the interaction matrix, each neuron will change until it matches the original state V^ (see the Updates section below). The connections in a Hopfield net typically have the following restrictions: * w_=0, \forall i (no unit has a connection with itself) * w_ = w_, \forall i,j (connections are symmetric) The constraint that weights are symmetric guarantees that the energy function decreases monotonically while following the activation rules. A network with asymmetric weights may exhibit some periodic or chaotic behaviour; however, Hopfield found that this behavior is confined to relatively small parts of the phase space and does not impair the network's ability to act as a content-addressable associative memory system. Hopfield also modeled neural nets for continuous values, in which the electric output of each neuron is not binary but some value between 0 and 1. He found that this type of network was also able to store and reproduce memorized states. Notice that every pair of units ''i'' and ''j'' in a Hopfield network has a connection that is described by the connectivity weight w_ . In this sense, the Hopfield network can be formally described as a complete undirected graph G = \langle V, f\rangle , where V is a set of McCulloch–Pitts neurons and f:V^2 \rightarrow \mathbb R is a function that links pairs of units to a real value, the connectivity weight.


Updating

Updating one unit (node in the graph simulating the artificial neuron) in the Hopfield network is performed using the following rule: s_i \leftarrow \left\} with boundary conditions The main difference of these equations from the conventional feedforward networks is the presence of the second term, which is responsible for the feedback from higher layers. These top-down signals help neurons in lower layers to decide on their response to the presented stimuli. Following the general recipe it is convenient to introduce a Lagrangian function L^A(\) for the A-th hidden layer, which depends on the activities of all the neurons in that layer. The activation functions in that layer can be defined as partial derivatives of the Lagrangian With these definitions the energy (Lyapunov) function is given by \Big \sum\limits_^ x_i^A g_i^A - L^\Big- \sum\limits_^ \sum\limits_^ \sum\limits_^ g_i^ \xi^_ g_j^A, If the Lagrangian functions, or equivalently the activation functions, are chosen in such a way that the Hessians for each layer are positive semi-definite and the overall energy is bounded from below, this system is guaranteed to converge to a fixed point attractor state. The temporal derivative of this energy function is given by \tau_A \sum\limits_^ \frac \frac \frac \leq 0, Thus, the hierarchical layered network is indeed an attractor network with the global energy function. This network is described by a hierarchical set of synaptic weights that can be learned for each specific problem.


See also

* Associative memory (disambiguation) *
Autoassociative memory Autoassociative memory, also known as auto-association memory or an autoassociation network, is any type of memory that is able to retrieve a piece of data from only a tiny sample of itself. They are very effective in de-noising or removing interfer ...
*
Boltzmann machine A Boltzmann machine (also called Sherrington–Kirkpatrick model with external field or stochastic Ising–Lenz–Little model) is a stochastic spin-glass model with an external field, i.e., a Sherrington–Kirkpatrick model, that is a stochastic ...
– like a Hopfield net but uses annealed Gibbs sampling instead of gradient descent * Dynamical systems model of cognition *
Ising model The Ising model () (or Lenz-Ising model or Ising-Lenz model), named after the physicists Ernst Ising and Wilhelm Lenz, is a mathematical model of ferromagnetism in statistical mechanics. The model consists of discrete variables that represent ...
*
Hebbian theory Hebbian theory is a neuroscientific theory claiming that an increase in synaptic efficacy arises from a presynaptic cell's repeated and persistent stimulation of a postsynaptic cell. It is an attempt to explain synaptic plasticity, the adaptation ...


References

* * * * * *


External links

*
Hopfield Network Javascript
– Hopfield Neural Network JAVA Applet * *
Neural Lab Graphical Interface
– Hopfield Neural Network graphical interface (Python & gtk) {{Stochastic processes Neural network architectures