Hopfield network

cosmos 28th November 2016 at 7:32pm
Boolean network Energy-based model

A type of Artificial neural network that has thresholded binary units (activation function is a Step function). It was a very simplified early model of Neural dynamics

See wiki

[11][1]Hopfield Nets (13 min) (Hinton)

They are related to Restricted Boltzmann machines

They are sometimes called recurrent cooperative networks, attractor neural networks, associative networks, Little and Hopfield networks, and spin glass networks. The random variant is a type of Random Boolean network.

Hopfield network energy

If the connections are symmetric, there is a global Energy function (also called the "discontent" function), which is minimized by the network's dynamics. Often Hopfield network is used to refer explicitly to the network where the connections are symmetric. It is

E[S(t)]=12j,iSi(t)JijSj(t)+iSi(t)TiE[\mathbf{S}(t)] = -\frac{1}{2}\sum\limits_{j,i} S_i(t) J_{ij}S_j(t) + \sum\limits_i S_i (t) T_i

where S(t)\mathbf{S}(t) is the vector of neuron states (either +1 or -1), JijJ_{ij} are the interaction energies (synaptic strengths), and Ti=2θiijJijT_i = 2\theta_i - \sum\limits_{i\neq j} J_{ij}, where θi\theta_i is a constant. TiT_i is the activation threshold of neuron ii.

The energy function corresponds to the Hamiltonian of a Spin glass

Dynamics of Hopfield networks

The Hopfield network model is defined as a threshold Boolean network. See Dynamics of Boolean networks.

In the network a neuron fires when its membrane potential is over the threshold. Mathematically, jJijSj>Ti\sum\limits_j J_{ij} S_j > T_i

Update rule types

there are four types of modes of updating the net- work's state: sequential, random, synchronous parallel, and asyn- chronous parallel. In sequential and random updating, time advances in 1/N of a synaptic delay, and the neurons have immediate effects on each other. In parallel updating, the effects appear only after a synaptic delay. In synchronous parallel updating, time advances in one synaptic delay, whereas in asynchronous parallel updating it advances in a fraction of a synaptic delay. The latter mode is the most physiological, but the least frequently used for recurrent networks.

When a Hopfield network (with symmetric connections) is updated in parallel, it may contain stable limit cycles. However, when it is updated sequentially, or randomly, it always falls into a single local minimum energy state. These stable states have associated Basin of attraction, which can depend on the update sequence.

Hopfield associative memories

Hopfield proposed in 1982 that Memoryes could be energy minima of symmetric Hopfield nets. In terms of dynamics, they are stable firing patterns/constellations.

  • The binary threshold decision rule can then be used to "clean up" incomplete or corrupted memories/thoughts (but which still fall in the basin of attraction of the real memory)
  • This is a type of Content-addressable memory, i.e. you can access an item just by knowing part of its content

The idea of memories as energy minima was proposed as I. A. RIchards in 1924 in "Principles of Literary Criticism". The idea of memories being distributed over wide neural networks, and not just in individual parts of the brain goes back to experiments by Lashley.

This can also be seen as kind of Classification

Storage rule

Topological characterization of a Spin Glass transition in a Random Boolean Hopfield Network

[11][3]Hopfield nets with hidden units (10 min)

https://www.quora.com/Why-use-reduced-Boltzmann-machines-instead-of-Hopfield-networks-for-deep-belief-networks

Multilayer perceptrons, Hopfield’s associative memories, and restricted Boltzmann machines

See more at Neural Networks: An Introduction [2 ed.] by Muller et al., chapter 3, and at Content-addressable memory. Also in Corticonics book, section 5.5.

Learning by Hebb's rule

Need to learn weights so that each stored pattern corresponds to a stable configuration of the network and so that small deviations from it will be automatically corrected by the network dynamics.

The smaller the ratio p/Np/N, where pp is the number of patterns, and NN is the number of neurons, the greater the likelihood that the pattern v is a stable configuration of the network, and that the pattern is recovered by a damaged pattern in a single update step.

If pp is comparable to NN, the patterns can't be recovered. Using tools from Statistical physics, one can show that there is a phase transition between these two cases at p/N0.14p/N \approx 0.14.

See Statistical mechanics of neural networks

–>David MacKay lectures

We note that Hebb originally spoke of strengthening synapses in which the presynaptic activity slightly preceded the postsynaptic activity. Thereby, Hebb's cell assemblies produced a sequence of activities at the recalled memory, not a group of neurons that coactivated each other in a stable fashion. For treatment of this more realistic kind of memory/learning, see Spike-timing-dependent plasticity, and "polychrony" in Spiking neural network.

Spurious minima

When embedding multiple memories in a network, one faces the prob- lem that in addition to the main valleys associated with the embedded memories, one generates multiple shallow valleys in the discontent ter- rain. It is possible to prevent the network from being trapped in one of these shallow valleys by making the transition from the firing state to the nonfiring state gradual, as shown in Figure 5.5.5.

This can be interpreted as adding non-zero Temperature to the system, which allows the system to escape the shallow local minima. Gradual cooling (Simulated annealing) can also be used. Periodic changes in background excitation (fluctuations ~ temperature) are known to occur by way of the diffuse projection systems of the cortex (Section 1.3.3, of Corticonics book)

Superious minima also occur with linear combinations of the stored memories. The effects of these can be reduced by requiring the memories to be as orthogonal as possible.

Memory capacity of the network

According to Amit et al. [1985b], in a network of N neurons with sym- metric connectivity and uncorrelated prototypes, as long as the number of embedded prototypes (P) is smaller than 0.14N, one obtains recall with a low probability of errors. As soon as P exceeds 0.144N, the probability of error increases very rapidly, and the network becomes useless [Nadal et al., 1986]. Note that the critical value of 14 percent is valid for large recurrent networks with uncorrelated prototypes and with synaptic strengths that are programmed according to Hebb's rule. Pro- gramming rules that allow for embedding many more stable (but corre- lated) states have been developed [Kohonen, 1984; Palm, 1986a; Gard- ner, 1988].

See more at Neural Networks: An Introduction [2 ed.] by Muller et al., chapter 3, and at Content-addressable memory


Notes from corticonics book These are mostly on random Boolean thresholded networks (with connections which need not be symmetric).

Point neuron (threshold model neuron)

A kind of Artificial neuron

Definition of point neuron (following Corticonics book)

https://en.m.wikipedia.org/wiki/Post-tetanic_potentiation

Model described here is a threshold network model. It is also a random network model. Analysis officer stability is also presented, different stable steady states are found depending on the threshold.

Netlets. A small random networks with sparse connections.

Dynamical system analysis is performed, including web diagrams, and bifurcation diagrams, when an external fixed activation is varied. This shows a typical hysteresis Curve. It also shows that as the background activation is increased, it is easier to take the Network to the active stable steady state, that is, the neurones are more willing to fire when the background level is increased just as it is found in real neurones.

It is hard to estabilize neuronal networks at low firing levels. Cortical networks appear optimized for fast inhibitory feedback.

Random vs actual networks. These comments apply to other models, like Random Boolean networks.

Externally driven ongoing activity

it appears that in each cortical area, a major portion of the excitation that drives the ongoing activity is derived from activity coming through the white matter. The inhibition is generated locally. This arrangement assures the local stability of the ongoing activity. When the entire cortex is viewed as one huge network, stability is as- sured by the short delay of the local inhibitory feedback, as compared with the conduction times of excitation through the white matter.

Modes of activity in random networks

even completely random networks can generate stable firing patterns (modes of activity). Using models, one can create a whole spectrum of stable modes by a variety of assumptions regarding the connectivity, the distribution of synaptic strengths, and the internal prop- erties of each neuron. The question remains to what degree the modes of activity observed in the cortex are due to the internal structure of the cortex and to what degree these modes are due to pacemakerlike activity derived from deeper nuclei.

Relations to real cortical Neurons

There are several attractive features, as well as weaknesses of the Hopfield network model, as a model of real Neurons, particularly in the Cortex. These are mostly discussed in the Corticonics book (pages 187-193).

Is there any recent research on the applicability of the model, and the idea of content-addressable memory to the Brain?