A type of Artificial neural network that has thresholded binary units (activation function is a Step function). It was a very simplified early model of Neural dynamics
See wiki
[11][1]Hopfield Nets (13 min) (Hinton)
They are related to Restricted Boltzmann machines
They are sometimes called recurrent cooperative networks, attractor neural networks, associative networks, Little and Hopfield networks, and spin glass networks. The random variant is a type of Random Boolean network.
If the connections are symmetric, there is a global Energy function (also called the "discontent" function), which is minimized by the network's dynamics. Often Hopfield network is used to refer explicitly to the network where the connections are symmetric. It is
where is the vector of neuron states (either +1 or -1), are the interaction energies (synaptic strengths), and , where is a constant. is the activation threshold of neuron .
The energy function corresponds to the Hamiltonian of a Spin glass
The Hopfield network model is defined as a threshold Boolean network. See Dynamics of Boolean networks.
In the network a neuron fires when its membrane potential is over the threshold. Mathematically,
Update rule types
there are four types of modes of updating the net- work's state: sequential, random, synchronous parallel, and asyn- chronous parallel. In sequential and random updating, time advances in 1/N of a synaptic delay, and the neurons have immediate effects on each other. In parallel updating, the effects appear only after a synaptic delay. In synchronous parallel updating, time advances in one synaptic delay, whereas in asynchronous parallel updating it advances in a fraction of a synaptic delay. The latter mode is the most physiological, but the least frequently used for recurrent networks.
When a Hopfield network (with symmetric connections) is updated in parallel, it may contain stable limit cycles. However, when it is updated sequentially, or randomly, it always falls into a single local minimum energy state. These stable states have associated Basin of attraction, which can depend on the update sequence.
Hopfield proposed in 1982 that Memoryes could be energy minima of symmetric Hopfield nets. In terms of dynamics, they are stable firing patterns/constellations.
The idea of memories as energy minima was proposed as I. A. RIchards in 1924 in "Principles of Literary Criticism". The idea of memories being distributed over wide neural networks, and not just in individual parts of the brain goes back to experiments by Lashley.
This can also be seen as kind of Classification
Topological characterization of a Spin Glass transition in a Random Boolean Hopfield Network
[11][3]Hopfield nets with hidden units (10 min)
Multilayer perceptrons, Hopfield’s associative memories, and restricted Boltzmann machines
See more at Neural Networks: An Introduction [2 ed.] by Muller et al., chapter 3, and at Content-addressable memory. Also in Corticonics book, section 5.5.
Need to learn weights so that each stored pattern corresponds to a stable configuration of the network and so that small deviations from it will be automatically corrected by the network dynamics.
The smaller the ratio , where is the number of patterns, and is the number of neurons, the greater the likelihood that the pattern v is a stable configuration of the network, and that the pattern is recovered by a damaged pattern in a single update step.
If is comparable to , the patterns can't be recovered. Using tools from Statistical physics, one can show that there is a phase transition between these two cases at .
See Statistical mechanics of neural networks
We note that Hebb originally spoke of strengthening synapses in which the presynaptic activity slightly preceded the postsynaptic activity. Thereby, Hebb's cell assemblies produced a sequence of activities at the recalled memory, not a group of neurons that coactivated each other in a stable fashion. For treatment of this more realistic kind of memory/learning, see Spike-timing-dependent plasticity, and "polychrony" in Spiking neural network.
When embedding multiple memories in a network, one faces the prob- lem that in addition to the main valleys associated with the embedded memories, one generates multiple shallow valleys in the discontent ter- rain. It is possible to prevent the network from being trapped in one of these shallow valleys by making the transition from the firing state to the nonfiring state gradual, as shown in Figure 5.5.5.
This can be interpreted as adding non-zero Temperature to the system, which allows the system to escape the shallow local minima. Gradual cooling (Simulated annealing) can also be used. Periodic changes in background excitation (fluctuations ~ temperature) are known to occur by way of the diffuse projection systems of the cortex (Section 1.3.3, of Corticonics book)
Superious minima also occur with linear combinations of the stored memories. The effects of these can be reduced by requiring the memories to be as orthogonal as possible.
According to Amit et al. [1985b], in a network of N neurons with sym- metric connectivity and uncorrelated prototypes, as long as the number of embedded prototypes (P) is smaller than 0.14N, one obtains recall with a low probability of errors. As soon as P exceeds 0.144N, the probability of error increases very rapidly, and the network becomes useless [Nadal et al., 1986]. Note that the critical value of 14 percent is valid for large recurrent networks with uncorrelated prototypes and with synaptic strengths that are programmed according to Hebb's rule. Pro- gramming rules that allow for embedding many more stable (but corre- lated) states have been developed [Kohonen, 1984; Palm, 1986a; Gard- ner, 1988].
See more at Neural Networks: An Introduction [2 ed.] by Muller et al., chapter 3, and at Content-addressable memory
Notes from corticonics book These are mostly on random Boolean thresholded networks (with connections which need not be symmetric).
A kind of Artificial neuron
Definition of point neuron (following Corticonics book)
https://en.m.wikipedia.org/wiki/Post-tetanic_potentiation
Model described here is a threshold network model. It is also a random network model. Analysis officer stability is also presented, different stable steady states are found depending on the threshold.
Netlets. A small random networks with sparse connections.
Dynamical system analysis is performed, including web diagrams, and bifurcation diagrams, when an external fixed activation is varied. This shows a typical hysteresis Curve. It also shows that as the background activation is increased, it is easier to take the Network to the active stable steady state, that is, the neurones are more willing to fire when the background level is increased just as it is found in real neurones.
It is hard to estabilize neuronal networks at low firing levels. Cortical networks appear optimized for fast inhibitory feedback.
Random vs actual networks. These comments apply to other models, like Random Boolean networks.
it appears that in each cortical area, a major portion of the excitation that drives the ongoing activity is derived from activity coming through the white matter. The inhibition is generated locally. This arrangement assures the local stability of the ongoing activity. When the entire cortex is viewed as one huge network, stability is as- sured by the short delay of the local inhibitory feedback, as compared with the conduction times of excitation through the white matter.
even completely random networks can generate stable firing patterns (modes of activity). Using models, one can create a whole spectrum of stable modes by a variety of assumptions regarding the connectivity, the distribution of synaptic strengths, and the internal prop- erties of each neuron. The question remains to what degree the modes of activity observed in the cortex are due to the internal structure of the cortex and to what degree these modes are due to pacemakerlike activity derived from deeper nuclei.
There are several attractive features, as well as weaknesses of the Hopfield network model, as a model of real Neurons, particularly in the Cortex. These are mostly discussed in the Corticonics book (pages 187-193).
Is there any recent research on the applicability of the model, and the idea of content-addressable memory to the Brain?